SlideShare a Scribd company logo
1 of 71
Download to read offline
PNPG Algorithm Applications References References References Acronyms
Projected Nesterov’s Proximal-Gradient Algorithm
for Sparse Signal RecoveryŽ
Aleksandar Dogandžić
Electrical and Computer Engineering
Ž joint work with Renliang Gu, Ph.D. student
supported by
1 / 71
PNPG Algorithm Applications References References References Acronyms
Terminology and Notation
soft-thresholding operator for a D .ai /N
iD1 2 RN :
T .a/ i
D sign.ai / max jai j ; 0 I
“ ” is the elementwise version of “ ”;
proximal operator for function r.x/ scaled by :
prox r a D arg min
x
1
2 kx ak2
2 C r.x/:
"-subgradient (Rockafellar 1970, Sec. 23):
@"r.x/ ,
˚
g 2 Rp
j r.z/ r.x/ C .z x/T
g "; 8z 2 Rp
«
:
2 / 71
PNPG Algorithm Applications References References References Acronyms
Introduction I
For most natural signals x,
# significant coefficients of .x/ signal size p
where
.x/ W Rp
7! Rp0
is sparsifying transform.
3 / 71
PNPG Algorithm Applications References References References Acronyms
Sparsifying Transforms I
p pixels
linear
transform
$
DWT
0
100
200
300
400
500
600
700
800
900
1000
# significant coeffs p
.x/ D ‰T
x
where ‰ 2 Rp p0
is a known sparsifying dictionary matrix.
4 / 71
PNPG Algorithm Applications References References References Acronyms
Sparsifying Transforms II
p pixels
transform
$
gradient map
# significant coeffs p
Œ .x/
p0
iD1 ,
s X
j2Ni
.xi xj /2
Ni is the index set of neighbors of xi .
5 / 71
PNPG Algorithm Applications References References References Acronyms
Convex-Set Constraint
x 2 C
where C is a nonempty closed convex set.
Example: the nonnegative signal set
C D R
p
C
is of significant practical interest and applicable to X-ray CT,
SPECT, PET, and MRI.
6 / 71
PNPG Algorithm Applications References References References Acronyms
Goal
Sense the significant components of .x/ using a small number of
measurements.
Define the noiseless measurement vector .x/, where
. / W Rp
7! RN
(N Ä p).
Example: Linear model
.x/ D ˆx
where ˆ 2 RN p is a known sensing matrix.
7 / 71
PNPG Algorithm Applications References References References Acronyms
Penalized NLL
objective
function
f .x/ D L.x/ C u k .x/k1 C IC .x/
ž
r.x/
convex differentiable
negative log-likelihood (NLL)
convex penalty term
u > 0 is a scalar tuning constant
C Â cl dom L.x/
8 / 71
PNPG Algorithm Applications References References References Acronyms
Comment
Our objective function f .x/ is
convex
has a convex set as minimum (unique if L.x/ is strongly
convex),
not differentiable with respect to the signal x
cannot apply usual gradient- or Newton-type algorithms,
need proximal-gradient (PG) schemes.
9 / 71
PNPG Algorithm Applications References References References Acronyms
Goals
Develop a fast algorithm with
O k 2
convergence-rate and
iterate convergence guarantees
for minimizing f .x/ that
is general (for a diverse set of NLLs),
requires minimal tuning, and
is matrix-free , a must for solving large-scale problems.
involves only matrix-vector multiplications implementable using, e.g.,
function handle in Matlab
10 / 71
PNPG Algorithm Applications References References References Acronyms
Majorizing Function
Define the quadratic approximation of the NLL L.x/:
Qˇ x j xx D L.xx/ C .x xx/T
rL.xx/ C
1
2ˇ
kx xxk2
2
with ˇ chosen so that Qˇ x j xx majorizes L.x/ in the
neighborhood of x D xx.
¯x(j)¯x(i)
y
0 x
L(x)
Qβ(i) (x|¯x(i))
Qβ(j) (x|¯x(j))
11 / 71
PNPG Algorithm Applications References References References Acronyms
¯x(j)¯x(i)
y
0 x
L(x)
Qβ(i) (x|¯x(i))
Qβ(j) (x|¯x(j))
¯x(j)¯x(i)
y
0
L(x)
Qβ(i) (x|¯x(i))
Qβ(j) (x|¯x(j))
Q10β(j) (x|¯x(j))
Figure 1: Majorizing function: Impact of ˇ.
No need for strict majorization, sufficient to majorize in the
neighborhood of xx where we wish to move next!
12 / 71
PNPG Algorithm Applications References References References Acronyms
PNPG Method: Iteration i
B.i/
D ˇ.i 1/
=ˇ.i/
Â.i/
D
˚
1; i Ä 1
1
C
q
b C B.i/ Â.i 1/ 2
; i > 1
xx.i/
D PC
Â
x.i 1/
C
Â.i 1/ 1
Â.i/ x.i 1/
x.i 2/
Ã
accel. step
x.i/
D proxˇ.i/ur xx.i/
ˇ.i/
rL xx.i/
Á
PG step
where ˇ.i/
> 0 is an adaptive step size:
satisfies
L x.i/
Ä Qˇ.i/ x.i/
j xx.i/ majorization condition,
is as large as possible.
more
13 / 71
PNPG Algorithm Applications References References References Acronyms
PNPG Method: Iteration i
allows general dom L
B.i/
D ˇ.i 1/
=ˇ.i/
Â.i/
D
˚
i; i Ä 1
1
C
q
b C B.i/ Â.i 1/ 2
; i > 1
xx.i/
D PC
Â
x.i 1/
C
Â.i 1/ 1
Â.i/ x.i 1/
x.i 2/
Ã
accel. step
x.i/
D proxˇ.i/ur xx.i/
ˇ.i/
rL xx.i/
Á
PG step
where ˇ.i/
> 0 is an adaptive step size:
satisfies
L x.i/
Ä Qˇ.i/ x.i/
j xx.i/ majorization condition,
is as large as possible.
14 / 71
PNPG Algorithm Applications References References References Acronyms
PNPG Method: Iteration i
B.i/
D ˇ.i 1/
=ˇ.i/
Â.i/
D
˚
i; i Ä 1
1
C
q
b C B.i/ Â.i 1/ 2
; i > 1
xx.i/
D PC
Â
x.i 1/
C
Â.i 1/ 1
Â.i/ x.i 1/
x.i 2/
Ã
accel. step
x.i/
D proxˇ.i/ur xx.i/
ˇ.i/
rL xx.i/
Á
PG step
where ˇ.i/
> 0 is an adaptive step size:
satisfies
L x.i/
Ä Qˇ.i/ x.i/
j xx.i/ majorization condition,
needs to hold for x.i/
, not for all x!
L x — Qˇ.i/ x j xx.i/ in general, for an arbitrary x! more
15 / 71
PNPG Algorithm Applications References References References Acronyms
Algorithm 1: PNPG method
Input: x. 1/
, u, n, m, , Á, and threshold
Output: arg min
x
f .x/
Initialization: Â.0/
0, x.0/
0, i 0, Ä 0, ˇ.1/
by the BB method
repeat
i i C 1 and Ä Ä C 1
while true do // backtracking search
evaluate B.i/
, Â.i/
, and xx.i/
if xx.i/
… dom L then // domain restart
Â.i 1/
1 and continue
solve the PG step
if majorization condition holds then
break
else
if ˇ.i/
> ˇ.i 1/
then // increase n
n n C m
ˇ.i/
ˇ.i/
and Ä 0
if i > 1 and f x.i/
> f x.i 1/
then // function restart
Â.i 1/
1, i i 1, and continue
if convergence condition holds then
declare convergence
if Ä n then // adapt step size
Ä 0 and ˇ.iC1/
ˇ.i/
=
else
ˇ.iC1/
ˇ.i/
until convergence declared or maximum number of iterations exceeded
16 / 71
PNPG Algorithm Applications References References References Acronyms
10 7
10 6
10 5
10 4
0 50 100 150 200 250 300
StepSizeˇ.i/
# of iterations
PNPG .n D 4/
PNPG .n D 1/
n
Figure 2: Illustration of step-size selection for Poisson generalized linear
model (GLM) with identity link.
17 / 71
PNPG Algorithm Applications References References References Acronyms
Momentum Illustration
xx.i/
D PC
Â
yx
.i/
¥x.i 1/
C Â.i 1/ 1
Â.i/ x.i 1/
x.i 2/
 
momentum term
prevents zigzagging
Ã
x(i−2)
x(i−1)
x(i)
x(i)
−∇L(x(i)
)
C
18 / 71
PNPG Algorithm Applications References References References Acronyms
Comments on the extrapolation term Â.i/
I
Â.i/
D 1
C
q
b C B.i/ Â.i 1/ 2
; i 2
where
2; b 2 0; 1=4
are momentum tuning constants.
To establish O k 2
convergence of PNPG, need
Â.i/
Ä 1
2 C
q
1
4 C B.i/ Â.i 1/ 2
; i 2:
controls the rate of increase of Â.i/
.
19 / 71
PNPG Algorithm Applications References References References Acronyms
Comments on the extrapolation term Â.i/
II
Â.i/
x
? implies stronger momentum.
Effect of step size
In the “steady state” where ˇ.i 1/
D ˇ.i/
, Â.i/
x
? aproximately
linearly with i, with slope 1= .
Changes in the step size affect Â.i/
:
ˇ.i/
< ˇ.i 1/
step size decrease, faster increase of Â.i/
,
ˇ.i/
> ˇ.i 1/
step size increase, decrease or slower increase of Â.i/
than in the steady state.
20 / 71
PNPG Algorithm Applications References References References Acronyms
Proximal Mapping
To compute
prox r a D arg min
x
1
2 kx ak2
2 C r.x/
use
for `1-norm penalty with .x/ D ‰T
x, alternating direction
method of multipliers (ADMM)
o iterative
for total-variation (TV)-norm penalty with gradient map .x/,
an inner iteration with the TV-based denoising method in
(Beck and Teboulle 2009b).
o iterative
21 / 71
PNPG Algorithm Applications References References References Acronyms
Inexact PG Steps
B.i/
D ˇ.i 1/
=ˇ.i/
Â.i/
D
˚
i; i Ä 1
1
C
q
b C B.i/ Â.i 1/ 2
; i > 1
xx.i/
D PC
Â
x.i 1/
C
Â.i 1/ 1
Â.i/
x.i 1/
x.i 2/
Ã
accel. step
x.i/
Ñ".i/ proxˇ.i/ur xx.i/
ˇ.i/
rL xx.i/
Á
PG step
Because of their iterative nature, PG steps are inexact: ".i/
quantifies the precision of the PG step in Iteration i.
more
22 / 71
PNPG Algorithm Applications References References References Acronyms
Remark (Monotonicity)
The projected Nesterov’s proximal-gradient (PNPG) iteration with
restart is non-increasing:
f x.i/
Ä f x.i 1/
if the inexact PG steps are sufficiently accurate and satisfy
".i/
Ä
p
ı.i/
where
ı.i/
, x.i/
x.i 1/ 2
2
is the local variation of signal iterates.
23 / 71
PNPG Algorithm Applications References References References Acronyms
Convergence Criterion
p
ı.i/ < x.i/
2
where > 0 is the convergence threshold.
more
Restart
The goal of function and domain restarts is to ensure that
the PNPG iteration is monotonic and
xx.i/
and x.i/
remain within dom f .
more
24 / 71
PNPG Algorithm Applications References References References Acronyms
Summary of PNPG Approach
Combine
convex-set projection with
Nesterov acceleration.
Apply
adaptive step size,
restart.
more
25 / 71
PNPG Algorithm Applications References References References Acronyms
Why?
Thanks to step-size adaptation, no need for Lipschitz
continuity of the gradient of the NLL.
dom L does not have to be Rp
.
Extends the application of the Nesterov’s accelerationŽ
to more
general measurement models than those used previously.
Ž
Y. Nesterov, “A method of solving a convex programming problem with
convergence rate O.1=k2
/,” Sov. Math. Dokl., vol. 27, 1983, pp. 372–376.
26 / 71
PNPG Algorithm Applications References References References Acronyms
Theorem (Convergence of the Objective Function)
Assume
NLL L.x/ is convex and differentiable and r.x/ is convex,
C Â dom L: no need for domain restart.
Consider the PNPG iteration without restart.
27 / 71
PNPG Algorithm Applications References References References Acronyms
Theorem (Convergence of the Objective Function)
f x.k/
f x?
Ä 2
x.0/ x? 2
2
C E.k/
2
p
ˇ.1/ C
Pk
iD1
p
ˇ.i/
Á2
where
E.k/
,
kX
iD1
Â.i/
".i/ 2
error term, accounts for inexact PG steps
x?
, arg min
x
f .x/
28 / 71
PNPG Algorithm Applications References References References Acronyms
Comments
f x.k/
f x?
Ä 2
x.0/ x? 2
2
C E.k/
2
p
ˇ.1/ C
Pk
iD1
p
ˇ.i/
Á2
Step sizes ˇ.i/
x
?, convergence-rate upper bound
?
y.
better initialization, convergence-rate upper bound #.
smaller prox-step approx. error, convergence-rate bound #.
29 / 71
PNPG Algorithm Applications References References References Acronyms
Corollary
Under the condition of the Theorem,
f x.k/
f x?
Ä 2
x.0/ x? 2
2
C E.k/
2k2ˇminŸ
O k 2 if E.C1/ < C1
provided that
ˇmin ,
C1
min
kD1
ˇ.k/
> 0:
The assumption that the step-size sequence is lower-bounded by a
strictly positive quantity is weaker than Lipschitz continuity of
rL.x/ because it is guaranteed to have ˇmin > =L if rL.x/ has a
Lipschitz constant L.
30 / 71
PNPG Algorithm Applications References References References Acronyms
Theorem (Convergence of Iterates)
Assume that
1 NLL L.x/ is convex and differentiable and r.x/ is convex,
2 C Â dom L, hence no need for domain restart,
3 cumulative error term E.k/
converges: E.C1/
< C1,
4 momentum tuning constants satisfy > 2 and b 2 0; 1= 2
,
5 the step-size sequence ˇ.i/ C1
iD1
is bounded within the range
ˇmin; ˇmax ; .ˇmin > 0/.
R The sequence of PNPG iterates x.i/
without restart converges
weakly to a minimizer of f .x/. a minimizer of f .x/.
31 / 71
PNPG Algorithm Applications References References References Acronyms
Theorem (Convergence of Iterates)
strict inequality
Assume that
1 NLL L.x/ is convex and differentiable and r.x/ is convex,
2 C Â dom L, hence no need for domain restart,
3 cumulative error term E.k/
converges: E.C1/
< C1,
4 momentum tuning const. satisfy > 2 and b 2 0; 1= 2
,
5 the step-size sequence ˇ.i/ C1
iD1
is bounded within the range
ˇmin; ˇmax ; .ˇmin > 0/.
R The sequence of PNPG iterates x.i/
without restart converges
weakly to a minimizer of f .x/. a minimizer of f .x/.
narrower than 0; 1=4
32 / 71
PNPG Algorithm Applications References References References Acronyms
Idea of Proof.
Recall the inequality
Â.i/
D 1
C
q
b C B.i/ Â.i 1/ 2
Ä 1
2 C
q
1
4 C B.i/ Â.i 1/ 2
used to establish convergence of the objective function.
Assumption 4:
> 2; b 2 0; 1= 2
creates a sufficient “gap” in this inequality that allows us to
show faster convergence of the objective function than the
previous theorem and
establish the convergence of iterates.
R Inspired by (Chambolle and Dossal 2015).
33 / 71
PNPG Algorithm Applications References References References Acronyms
Introduction
Signal reconstruction from Poisson-distributed measurements with
affine model for the mean-signal intensity is important for
tomographic (Ollinger and Fessler 1997),
astronomic, optical, microscopic (Bertero et al. 2009),
hyperspectral (Willett et al. 2014)
imaging.
34 / 71
PNPG Algorithm Applications References References References Acronyms
PET: Coincidence detection due to positron decay and annihilation
(Prince and Links 2015).
35 / 71
PNPG Algorithm Applications References References References Acronyms
Measurement Model
N independent measurements y D .yn/N
nD1 follow the Poisson
distribution with means
Œˆx C bn
where
ˆ 2 R
N p
C ; b
are the known sensing matrix and intercept term
.

the intercept b models background radiation and scattering, obtained, e.g.,
by calibration before the measurements y have been collected
36 / 71
PNPG Algorithm Applications References References References Acronyms
Existing Work
The sparse Poisson-intensity reconstruction algorithm (SPIRAL)§
approximates the logarithm function
in the underlying NLL by adding a
small positive term to it and then
descends a regularized NLL
objective function with proximal
steps that employ Barzilai-Borwein
(BB) step size in each iteration,
followed by backtracking.
§
Z. T. Harmany et al., “This is SPIRAL-TAP: Sparse Poisson intensity
reconstruction algorithms—theory and practice,” IEEE Trans. Image Process.,
vol. 21, no. 3, pp. 1084–1096, Mar. 2012.
37 / 71
PNPG Algorithm Applications References References References Acronyms
PET Image Reconstruction
128 128 concentration map x.
Collect the photons from 90 equally
spaced directions over 180ı
, with 128
radial samples at each direction,
Background radiation, scattering effect,
and accidental coincidence combined to-
gether lead to a known intercept term b.
The elements of the intercept term
are set to a constant equal to 10 % of
the sample mean of ˆx: b D
1T
ˆx
10N
1.
The model, choices of parameters in the PET system setup, and
concentration map have been adopted from Image Reconstruction
Toolbox (IRT) (Fessler 2016).
38 / 71
PNPG Algorithm Applications References References References Acronyms
Numerical Example
Main metric for assessing the performance of the compared
algorithms is relative square error (RSE)
RSE D
kyx xtruek2
2
kxtruek2
2
where xtrue and yx are the true and reconstructed signal,
respectively.
All iterative methods use the convergence threshold
D 10 6
and have the maximum number of iterations limited to 104
.
Regularization constant u has the form
u D 10a
:
We vary a in the range Œ 6; 3 with a grid size of 0.5 and
search for the reconstructions with the best RSE performance.
39 / 71
PNPG Algorithm Applications References References References Acronyms
Compared Methods
Filtered backprojection (FBP) (Ollinger and Fessler 1997) and
PG methods that aim at minimizing f .x/ with nonnegative x:
C D R
p
C:
All iterative methods initialized by FBP reconstructions.
Matlab implementation available at
http://isucsp.github.io/imgRecSrc/npg.html.
40 / 71
PNPG Algorithm Applications References References References Acronyms
PG Methods
PNPG with . ; b/ D .2; 1=4/.
AT (Auslender and Teboulle 2006) implemented in the
templates for first-order conic solvers (TFOCS) package
(Becker et al. 2011) with a periodic restart every 200 iterations
(tuned for its best performance) and our proximal mapping.
SPIRAL, when possible.
41 / 71
PNPG Algorithm Applications References References References Acronyms
(a) radio-isotope concentration (b) attenuation map
Figure 3: (a) True emission image and (b) density map.
42 / 71
PNPG Algorithm Applications References References References Acronyms
RSE=3.09 %
(a) FBP
RSE=0.66 %
(b) `1
Figure 4: Reconstructions of the emission concentration map for expected
total annihilation photon count (SNR) equal to 108
.
43 / 71
PNPG Algorithm Applications References References References Acronyms
RSE=0.66 %
(a) `1
RSE=0.22 %
(b) TV
Figure 5: Comparison of the two sparsity regularizations.
44 / 71
PNPG Algorithm Applications References References References Acronyms
10 1
100
101
102
103
104
105
106
0 20 40 60 80 100 120 140
.i/
Time/s
AT
PNPG .n D 1/
PNPG .n D m D 0/
PNPG .n D m D 4/
(a) `1, 1T
E.y/ D 108
10 4
10 3
10 2
10 1
100
101
102
103
104
105
106
0 10 20 30 40 50 60
.i/
Time/s
SPIRAL
AT
PNPG .n D 1/
PNPG .n D m D 0/
PNPG .n D m D 4/
(b) TV, 1T
E.y/ D 107
Figure 6: Centered objectives as functions of CPU time.
45 / 71
PNPG Algorithm Applications References
10 1
100
101
102
103
104
105
106
0 20 40 60 80 100 120 140
.i/
Time/s
AT
PNPG .n D 1/
PNPG .n D m D 0/
PNPG .n D m D 4/
(a) `1, 1T
E.y/ D 108
10 4
10 3
10 2
10 1
100
101
102
103
104
105
106
0 10 20 30 40 50 60
.i/
Time/s
SPIRAL
AT
PNPG .n D 1/
PNPG .n D m D 0/
PNPG .n D m D 4/
(b) TV, 1T
E.y/ D 107
Figure 7: Centered objectives as functions of CPU time.
50 / 101
PNPG Algorithm Applications References References References Acronyms
Linear Model with Gaussian Noise
L.x/ D
1
2
ky ˆxk2
2
where
y 2 RN
is the measurement vector and
the elements of the sensing matrix ˆ are independent,
identically distributed (i.i.d.), drawn from the standard normal
distribution.
We select the `1-norm sparsifying signal penalty with linear map:
.x/ D ‰T
x:
46 / 71
PNPG Algorithm Applications References References References Acronyms
−0.5
0
0.5
1
1.5
2
2.5
0 200 400 600 800 1000
Figure 7: True signal.
47 / 71
PNPG Algorithm Applications References References References Acronyms
Comments I
More methods available for comparison:
sparse reconstruction by separable approximation (SpaRSA)
(Wright et al. 2009),
generalized forward-backward (GFB) (Raguet et al. 2013),
primal-dual splitting (PDS) (Condat 2013).
Select the regularization parameter u as
u D 10a
U; U , ‰T
rL.0/ 1
where a is an integer selected from the interval Œ 9; 1 and U
is an upper bound on u of interest.
Choose the nonnegativity convex set:
C D R
p
C:
48 / 71
PNPG Algorithm Applications References References References Acronyms
Comments II
If we remove the convex-set constraint by setting C D Rp
,
PNPG iteration reduces to the Nesterov’s proximal gradient
iteration with adaptive step size that imposes signal sparsity
only in the analysis form (termed NPGS).
49 / 71
PNPG Algorithm Applications References References References Acronyms
−0.5
0
0.5
1
1.5
2
2.5
0 200 400 600 800 1000
RSE=0.011 %
(a) PNPG
−0.5
0
0.5
1
1.5
2
2.5
0 200 400 600 800 1000
RSE=0.23 %
(b) NPGS
Figure 8: PNPG and NPGS reconstructions for N=p D 0:34.
50 / 71
PNPG Algorithm Applications References References References Acronyms
10−8
10−6
10−4
10−2
100
102
104
0 10 20 30 40 50 60 70 80 90
∆(i)
time/s
(a) a D 5; N=p D 0:34
10−8
10−6
10−4
10−2
100
102
104
0 10 20 30 40 50 60 70 80 90
∆(i)
time/s
SPIRAL
SpaRSA (cont.)
PNPG
PNPG (cont.)
GFB
PDS
AT
(b) a D 6; N=p D 0:49
Figure 9: Centered objectives as functions of CPU time.
51 / 71
PNPG Algorithm Applications References References References Acronyms
10−8
10−6
10−4
10−2
100
102
104
0 10 20 30 40 50 60 70 80
time/s
(a) a D 4; N=p D 0:34
10−8
10−6
10−4
10−2
100
102
104
0 10 20 30 40 50 60 70 80
time/s
SPIRAL
SpaRSA (cont.)
PNPG
PNPG (cont.)
GFB
PDS
AT
(b) a D 5; N=p D 0:49
Figure 10: Centered objectives as functions of CPU time.
52 / 71
PNPG Algorithm Applications References References References Acronyms
10−8
10−6
10−4
10−2
100
102
104
0 5 10 15 20 25 30
time/s
(a) a D 3; N=p D 0:34
10−8
10−6
10−4
10−2
100
102
104
0 5 10 15 20
time/s
SPIRAL
SpaRSA (cont.)
PNPG
PNPG (cont.)
GFB
PDS
AT
(b) a D 4; N=p D 0:49
Figure 11: Centered objectives as functions of CPU time.
53 / 71
PNPG Algorithm Applications References References References Acronyms
Publications
R. G. and A. D. (May 2016), Projected Nesterov’s
proximal-gradient algorithm for sparse signal reconstruction
with a convex constraint, version 5. arXiv: 1502.02613
[stat.CO].
R. G. and A. D., “Projected Nesterov’s proximal-gradient
signal recovery from compressive Poisson measurements,”
Proc. Asilomar Conf. Signals, Syst. Comput., Pacific Grove,
CA, Nov. 2015, pp. 1490–1495.
54 / 71
PNPG Algorithm Applications References References References Acronyms
Preliminary Work
R. G. and A. D., “A fast proximal gradient algorithm for
reconstructing nonnegative signals with sparse transform
coefficients,” Proc. Asilomar Conf. Signals, Syst. Comput.,
Pacific Grove, CA, Nov. 2014, pp. 1662–1667.
R. G. and A. D. (Mar. 2015), Reconstruction of nonnegative
sparse signals using accelerated proximal-gradient algorithms,
version 3. arXiv: 1502.02613v3 [stat.CO].
55 / 71
PNPG Algorithm Applications References References References Acronyms
Conclusion
Developed a fast framework for reconstructing signals that are
sparse in a transform domain and belong to a closed convex set
by employing a projected proximal-gradient scheme with
Nesterov’s acceleration, restart and adaptive step size.
Applied the proposed framework to construct the first
Nesterov-accelerated Poisson compressed-sensing
reconstruction algorithm.
Derived convergence-rate upper-bound that accounts for
inexactness of the proximal operator.
Proved convergence of iterates.
Our PNPG approach is computationally efficient compared with
the state-of-the-art.
56 / 71
PNPG Algorithm Applications References References References Acronyms
References I
A. Auslender and M. Teboulle, “Interior gradient and proximal methods for
convex and conic optimization,” SIAM J. Optim., vol. 16, no. 3, pp. 697–725,
2006.
A. Beck and M. Teboulle, “A fast iterative shrinkage-thresholding algorithm for
linear inverse problems,” SIAM J. Imag. Sci., vol. 2, no. 1, pp. 183–202, 2009.
A. Beck and M. Teboulle, “Fast gradient-based algorithms for constrained
total variation image denoising and deblurring problems,” IEEE Trans. Image
Process., vol. 18, no. 11, pp. 2419–2434, 2009.
S. R. Becker, E. J. Candès, and M. C. Grant, “Templates for convex cone
problems with applications to sparse signal recovery,” Math. Program. Comp.,
vol. 3, no. 3, pp. 165–218, 2011. [Online]. Available: http://cvxr.com/tfocs.
M. Bertero, P. Boccacci, G. Desiderà, and G. Vicidomini, “Image deblurring
with Poisson data: From cells to galaxies,” Inverse Prob., vol. 25, no. 12,
pp. 123006-1–123006-26, 2009.
S. Bonettini, I. Loris, F. Porta, and M. Prato, “Variable metric inexact
line-search-based methods for nonsmooth optimization,” SIAM J. Optim., vol.
26, no. 2, pp. 891–921, 2016.
57 / 71
PNPG Algorithm Applications References References References Acronyms
References II
S. Bonettini, F. Porta, and V. Ruggiero, “A variable metric forward-backward
method with extrapolation,” SIAM J. Sci. Comput., vol. 38, no. 4,
A2558–A2584, 2016.
A. Chambolle and C. Dossal, “On the convergence of the iterates of the ‘fast
iterative shrinkage/thresholding algorithm’,” J. Optim. Theory Appl., vol. 166,
no. 3, pp. 968–982, 2015.
L. Condat, “A primal–dual splitting method for convex optimization involving
Lipschitzian, proximable and linear composite terms,” J. Optim. Theory Appl.,
vol. 158, no. 2, pp. 460–479, 2013.
J. A. Fessler, Image reconstruction toolbox, [Online]. Available:
http://www.eecs.umich.edu/~fessler/code (visited on 08/23/2016).
Z. T. Harmany, R. F. Marcia, and R. M. Willett, “This is SPIRAL-TAP:
Sparse Poisson intensity reconstruction algorithms—theory and practice,”
IEEE Trans. Image Process., vol. 21, no. 3, pp. 1084–1096, Mar. 2012.
Y. Nesterov, “A method of solving a convex programming problem with
convergence rate O.1=k2
/,” Sov. Math. Dokl., vol. 27, 1983, pp. 372–376.
58 / 71
PNPG Algorithm Applications References References References Acronyms
References III
B. O‘Donoghue and E. Candès, “Adaptive restart for accelerated gradient
schemes,” Found. Comput. Math., vol. 15, no. 3, pp. 715–732, 2015.
J. M. Ollinger and J. A. Fessler, “Positron-emission tomography,” IEEE
Signal Process. Mag., vol. 14, no. 1, pp. 43–55, 1997.
B. T. Polyak, “Some methods of speeding up the convergence of iteration
methods,” USSR Comput. Math. Math. Phys., vol. 4, no. 5, pp. 1–17, 1964.
B. T. Polyak, Introduction to Optimization. New York: Optimization Software,
1987.
J. L. Prince and J. M. Links, Medical Imaging Signals and Systems, 2nd ed.
Upper Saddle River, NJ: Pearson, 2015.
H. Raguet, J. Fadili, and G. Peyré, “A generalized forward-backward splitting,”
SIAM J. Imag. Sci., vol. 6, no. 3, pp. 1199–1226, 2013.
R. T. Rockafellar, Convex Analysis. Princeton, NJ: Princeton Univ. Press,
1970.
59 / 71
PNPG Algorithm Applications References References References Acronyms
References IV
S. Salzo. (May 2016), The variable metric forward-backward splitting
algorithm under mild differentiability assumptions, arXiv: 1605.00952
[math.OC].
S. Villa, S. Salzo, L. Baldassarre, and A. Verri, “Accelerated and inexact
forward-backward algorithms,” SIAM J. Optim., vol. 23, no. 3,
pp. 1607–1633, 2013.
R. M. Willett, M. F. Duarte, M. A. Davenport, and R. G. Baraniuk, “Sparsity
and structure in hyperspectral imaging: Sensing, reconstruction, and target
detection,” IEEE Signal Process. Mag., vol. 31, no. 1, pp. 116–126, Jan. 2014.
S. J. Wright, R. D. Nowak, and M. A. T. Figueiredo, “Sparse reconstruction
by separable approximation,” IEEE Trans. Signal Process., vol. 57, no. 7,
pp. 2479–2493, 2009.
60 / 71
PNPG Algorithm Applications References References References Acronyms
Adaptive Step Size
1 if no step-size backtracking events or increase attempts for n
consecutive iterations, start with a larger step size
xˇ.i/
D
ˇ.i 1/
(increase attempt)
where 2 .0; 1/ is a step-size adaptation parameter;
otherwise start with
xˇ.i/
D ˇ.i 1/
I
2 (backtracking search) select
ˇ.i/
D ti xˇ.i/
(2)
where ti 0 is the smallest integer such that (2) satisfies the
majorization condition (2); backtracking event corresponds to
ti > 0.
3 if max ˇ.i/
; ˇ.i 1/
< xˇ.i/
, increase n by a nonnegative
integer m: n n C m: back
61 / 71
PNPG Algorithm Applications References References References Acronyms
Restart
Whenever f x.i/
> f x.i 1/
or xx.i/
2 C n dom L, we set
Â.i 1/
D 1 (restart)
and refer to this action as function restart (O‘Donoghue and Candès
2015) or domain restart respectively.
back
62 / 71
PNPG Algorithm Applications References References References Acronyms
Inner Convergence Criteria
TV: x.i;j/
x.i;j 1/
2
Ä Á
p
ı.i 1/ (3a)
`1: max s.i;j/
‰T
x.i;j/
2
; s.i;j/
s.i;j 1/
2
Á
Ä Á ‰T
x.i 1/
x.i 2/
2
(3b)
where j is the inner-iteration index,
x.i;j/
is the iterate of x in the jth inner iteration step within
the ith step of the (outer) PNPG iteration, and
Á 2 .0; 1/
is the convergence tuning constant chosen to trade off the
accuracy and speed of the inner iterations and provide
sufficiently accurate solutions to the proximal mapping.
back
63 / 71
PNPG Algorithm Applications References References References Acronyms
Definition (Inexact Proximal Operator (Villa et al. 2013))
We say that x is an approximation of proxur a with "-precision,
denoted by
x Ñ" proxur a
if
a x
u
2 @"2
2u
r.x/:
Note: This definition implies
kx proxur ak2
2 Ä "2
:
back
64 / 71
PNPG Algorithm Applications References References References Acronyms
Relationship with FISTA I
PNPG can be thought of as a generalized fast iterative
shrinkage-thresholding algorithm (FISTA) (Beck and Teboulle
2009a) that accommodates
convex constraints,
more general NLLs,‘
and (increasing) adaptive step size
thanks to this step-size adaptation, PNPG does not require
Lipschitz continuity of the gradient of the NLL.
back
‘
FISTA has been developed for the linear Gaussian model.
65 / 71
PNPG Algorithm Applications References References References Acronyms
Relationship with FISTA II
Need B.i/
to derive theoretical guarantee for convergence
speed of the PNPG iteration.
In contrast with PNPG, FISTA has a non-increasing step size
ˇ.i/
, which allows for setting
B.i/
D 1
for all i:
Â.i/
D
1
2
Ä
1 C
q
1 C 4 Â.i 1/ 2
:
A simpler version of FISTA is
Â.i/
D
1
2
C Â.i 1/
D
i C 1
2
for i 1, which corresponds to . ; b/ D .2; 0/.
back
Y. Nesterov, “A method of solving a convex programming problem with
convergence rate O.1=k2
/,” Sov. Math. Dokl., vol. 27, 1983, pp. 372–376.
66 / 71
PNPG Algorithm Applications References References References Acronyms
Relationship with AT
(Auslender and Teboulle 2006):
Â.i/
D
1
2
Ä
1 C
q
1 C 4 Â.i 1/ 2
xx.i/
D
Â
1
1
Â.i/
Ã
x.i 1/
C
1
Â.i/
zx.i 1/
zx.i/
D proxÂ.i/ˇ.i/ur zx.i 1/
Â.i/
ˇ.i/
rL xx.i/
Á
x.i/
D
Â
1
1
Â.i/
Ã
x.i 1/
C
1
Â.i/
zx.i/
Other variants with infinite memory are available at (Becker et al.
2011).
back
67 / 71
PNPG Algorithm Applications References References References Acronyms
Heavy-ball Method
(Polyak 1964; Polyak 1987):
x.i/
D proxˇ.i/ur x.i 1/
ˇ.i/
rL x.i 1/
Á
C ‚.i/
x.i 1/
x.i 2/
:
back
68 / 71
PNPG Algorithm Applications References References References Acronyms
Glossary I
ADMM alternating direction method of multipliers. 21
BB Barzilai-Borwein. 37
CPU central processing unit. 45
CT computed tomography. 6
FBP filtered backprojection. 40
FISTA fast iterative shrinkage-thresholding algorithm. 66, 67
GFB generalized forward-backward. 49
GLM generalized linear model. 17
i.i.d. independent, identically distributed. 47
IRT Image Reconstruction Toolbox. 38
MRI magnetic resonance imaging. 6
70 / 71
PNPG Algorithm Applications References References References Acronyms
Glossary II
NLL negative log-likelihood. 10, 11, 26, 27, 31, 32, 37, 66
NPGS Nesterov’s proximal-gradient sparse. 50, 51
PDS primal-dual splitting. 49
PET positron emission tomography. 6, 35, 38
PG proximal-gradient. 9, 16, 22, 23, 28, 40
PNPG projected Nesterov’s proximal-gradient. 23, 24, 27, 31,
32, 41, 51, 57, 64, 66, 67
RSE relative square error. 39
SpaRSA sparse reconstruction by separable approximation. 49
SPECT single photon emission computed tomography. 6
SPIRAL sparse Poisson-intensity reconstruction algorithm. 37,
41
TFOCS templates for first-order conic solvers. 41
TV total-variation. 21
71 / 71

More Related Content

What's hot

Design and realization of iir digital band stop filter using modified analog ...
Design and realization of iir digital band stop filter using modified analog ...Design and realization of iir digital band stop filter using modified analog ...
Design and realization of iir digital band stop filter using modified analog ...
Subhadeep Chakraborty
 
Fm G As
Fm G AsFm G As
Fm G As
ysemet
 
Development of Reliability Analysis and Multidisciplinary Design Optimization...
Development of Reliability Analysis and Multidisciplinary Design Optimization...Development of Reliability Analysis and Multidisciplinary Design Optimization...
Development of Reliability Analysis and Multidisciplinary Design Optimization...
Altair
 
Stack Hybridization: A Mechanism for Bridging Two Compilation Strategies in a...
Stack Hybridization: A Mechanism for Bridging Two Compilation Strategies in a...Stack Hybridization: A Mechanism for Bridging Two Compilation Strategies in a...
Stack Hybridization: A Mechanism for Bridging Two Compilation Strategies in a...
Yusuke Izawa
 

What's hot (20)

learning vhdl by examples
learning vhdl by exampleslearning vhdl by examples
learning vhdl by examples
 
GPU Computing
GPU ComputingGPU Computing
GPU Computing
 
International Journal of Engineering Research and Development
International Journal of Engineering Research and DevelopmentInternational Journal of Engineering Research and Development
International Journal of Engineering Research and Development
 
Linear Phase FIR Low Pass Filter Design Based on Firefly Algorithm
Linear Phase FIR Low Pass Filter Design Based on Firefly Algorithm Linear Phase FIR Low Pass Filter Design Based on Firefly Algorithm
Linear Phase FIR Low Pass Filter Design Based on Firefly Algorithm
 
Tools and Methods for Continuously Expanding Software Applications
Tools and Methods for Continuously Expanding Software ApplicationsTools and Methods for Continuously Expanding Software Applications
Tools and Methods for Continuously Expanding Software Applications
 
Design and realization of iir digital band stop filter using modified analog ...
Design and realization of iir digital band stop filter using modified analog ...Design and realization of iir digital band stop filter using modified analog ...
Design and realization of iir digital band stop filter using modified analog ...
 
PERFORMANCE ANALYSIS OF PARALLEL POLLARD’S RHO FACTORING ALGORITHM
PERFORMANCE ANALYSIS OF PARALLEL POLLARD’S  RHO FACTORING ALGORITHM PERFORMANCE ANALYSIS OF PARALLEL POLLARD’S  RHO FACTORING ALGORITHM
PERFORMANCE ANALYSIS OF PARALLEL POLLARD’S RHO FACTORING ALGORITHM
 
A new efficient fpga design of residue to-binary converter
A new efficient fpga design of residue to-binary converterA new efficient fpga design of residue to-binary converter
A new efficient fpga design of residue to-binary converter
 
Cg in Two Pages
Cg in Two PagesCg in Two Pages
Cg in Two Pages
 
Fm G As
Fm G AsFm G As
Fm G As
 
Optimized Reversible Vedic Multipliers for High Speed Low Power Operations
Optimized Reversible Vedic Multipliers for High Speed Low Power OperationsOptimized Reversible Vedic Multipliers for High Speed Low Power Operations
Optimized Reversible Vedic Multipliers for High Speed Low Power Operations
 
Microprocessor Week 7: Branch Instruction
Microprocessor Week 7: Branch InstructionMicroprocessor Week 7: Branch Instruction
Microprocessor Week 7: Branch Instruction
 
Lecture04
Lecture04Lecture04
Lecture04
 
Development of Reliability Analysis and Multidisciplinary Design Optimization...
Development of Reliability Analysis and Multidisciplinary Design Optimization...Development of Reliability Analysis and Multidisciplinary Design Optimization...
Development of Reliability Analysis and Multidisciplinary Design Optimization...
 
OPTIMIZED MULTIPLIER USING REVERSIBLE MULTICONTROL INPUT TOFFOLI GATES
OPTIMIZED MULTIPLIER USING REVERSIBLE MULTICONTROL INPUT TOFFOLI GATESOPTIMIZED MULTIPLIER USING REVERSIBLE MULTICONTROL INPUT TOFFOLI GATES
OPTIMIZED MULTIPLIER USING REVERSIBLE MULTICONTROL INPUT TOFFOLI GATES
 
A Follow-up Cg Runtime Tutorial for Readers of The Cg Tutorial
A Follow-up Cg Runtime Tutorial for Readers of The Cg TutorialA Follow-up Cg Runtime Tutorial for Readers of The Cg Tutorial
A Follow-up Cg Runtime Tutorial for Readers of The Cg Tutorial
 
Improved ant colony optimization for quantum cost reduction
Improved ant colony optimization for quantum cost reductionImproved ant colony optimization for quantum cost reduction
Improved ant colony optimization for quantum cost reduction
 
Wilmott Nyc Jul2012 Nag Talk John Holden
Wilmott Nyc Jul2012 Nag Talk John HoldenWilmott Nyc Jul2012 Nag Talk John Holden
Wilmott Nyc Jul2012 Nag Talk John Holden
 
slides
slidesslides
slides
 
Stack Hybridization: A Mechanism for Bridging Two Compilation Strategies in a...
Stack Hybridization: A Mechanism for Bridging Two Compilation Strategies in a...Stack Hybridization: A Mechanism for Bridging Two Compilation Strategies in a...
Stack Hybridization: A Mechanism for Bridging Two Compilation Strategies in a...
 

Similar to Projected Nesterov's Proximal-Gradient Algorithm for Sparse Signal Recovery

MarcoCeze_defense
MarcoCeze_defenseMarcoCeze_defense
MarcoCeze_defense
Marco Ceze
 

Similar to Projected Nesterov's Proximal-Gradient Algorithm for Sparse Signal Recovery (20)

15.sp.dictionary_draft.pdf
15.sp.dictionary_draft.pdf15.sp.dictionary_draft.pdf
15.sp.dictionary_draft.pdf
 
A Hybrid Formulation between Differential Evolution and Simulated Annealing A...
A Hybrid Formulation between Differential Evolution and Simulated Annealing A...A Hybrid Formulation between Differential Evolution and Simulated Annealing A...
A Hybrid Formulation between Differential Evolution and Simulated Annealing A...
 
236628934.pdf
236628934.pdf236628934.pdf
236628934.pdf
 
MUMS: Bayesian, Fiducial, and Frequentist Conference - Model Selection in the...
MUMS: Bayesian, Fiducial, and Frequentist Conference - Model Selection in the...MUMS: Bayesian, Fiducial, and Frequentist Conference - Model Selection in the...
MUMS: Bayesian, Fiducial, and Frequentist Conference - Model Selection in the...
 
B046050711
B046050711B046050711
B046050711
 
Synchrophasor Fundamentals: from Computation to Implementation
Synchrophasor Fundamentals: from Computation to ImplementationSynchrophasor Fundamentals: from Computation to Implementation
Synchrophasor Fundamentals: from Computation to Implementation
 
A COMPARATIVE STUDY OF BACKPROPAGATION ALGORITHMS IN FINANCIAL PREDICTION
A COMPARATIVE STUDY OF BACKPROPAGATION ALGORITHMS IN FINANCIAL PREDICTIONA COMPARATIVE STUDY OF BACKPROPAGATION ALGORITHMS IN FINANCIAL PREDICTION
A COMPARATIVE STUDY OF BACKPROPAGATION ALGORITHMS IN FINANCIAL PREDICTION
 
B010411016
B010411016B010411016
B010411016
 
Nested dynamic programming algorithm
Nested dynamic programming algorithmNested dynamic programming algorithm
Nested dynamic programming algorithm
 
Design of Low Power Reconfigurable IIR filter with Row Bypassing Multiplier
Design of Low Power Reconfigurable IIR filter with Row Bypassing MultiplierDesign of Low Power Reconfigurable IIR filter with Row Bypassing Multiplier
Design of Low Power Reconfigurable IIR filter with Row Bypassing Multiplier
 
Optimization of Unit Commitment Problem using Classical Soft Computing Techni...
Optimization of Unit Commitment Problem using Classical Soft Computing Techni...Optimization of Unit Commitment Problem using Classical Soft Computing Techni...
Optimization of Unit Commitment Problem using Classical Soft Computing Techni...
 
lachhab2013.pdf
lachhab2013.pdflachhab2013.pdf
lachhab2013.pdf
 
Prpagation of Error Bounds Across reduction interfaces
Prpagation of Error Bounds Across reduction interfacesPrpagation of Error Bounds Across reduction interfaces
Prpagation of Error Bounds Across reduction interfaces
 
AbdoSummerANS_mod3
AbdoSummerANS_mod3AbdoSummerANS_mod3
AbdoSummerANS_mod3
 
Gene's law
Gene's lawGene's law
Gene's law
 
ANSSummer2015
ANSSummer2015ANSSummer2015
ANSSummer2015
 
Development of Multi-Level ROM
Development of Multi-Level ROMDevelopment of Multi-Level ROM
Development of Multi-Level ROM
 
Conducting and reporting the results of a cfd simulation
Conducting and reporting the results of a cfd simulationConducting and reporting the results of a cfd simulation
Conducting and reporting the results of a cfd simulation
 
Paper id 27201430
Paper id 27201430Paper id 27201430
Paper id 27201430
 
MarcoCeze_defense
MarcoCeze_defenseMarcoCeze_defense
MarcoCeze_defense
 

Recently uploaded

VIP Call Girls Palanpur 7001035870 Whatsapp Number, 24/07 Booking
VIP Call Girls Palanpur 7001035870 Whatsapp Number, 24/07 BookingVIP Call Girls Palanpur 7001035870 Whatsapp Number, 24/07 Booking
VIP Call Girls Palanpur 7001035870 Whatsapp Number, 24/07 Booking
dharasingh5698
 
Call Now ≽ 9953056974 ≼🔝 Call Girls In New Ashok Nagar ≼🔝 Delhi door step de...
Call Now ≽ 9953056974 ≼🔝 Call Girls In New Ashok Nagar  ≼🔝 Delhi door step de...Call Now ≽ 9953056974 ≼🔝 Call Girls In New Ashok Nagar  ≼🔝 Delhi door step de...
Call Now ≽ 9953056974 ≼🔝 Call Girls In New Ashok Nagar ≼🔝 Delhi door step de...
9953056974 Low Rate Call Girls In Saket, Delhi NCR
 
notes on Evolution Of Analytic Scalability.ppt
notes on Evolution Of Analytic Scalability.pptnotes on Evolution Of Analytic Scalability.ppt
notes on Evolution Of Analytic Scalability.ppt
MsecMca
 
Top Rated Call Girls In chittoor 📱 {7001035870} VIP Escorts chittoor
Top Rated Call Girls In chittoor 📱 {7001035870} VIP Escorts chittoorTop Rated Call Girls In chittoor 📱 {7001035870} VIP Escorts chittoor
Top Rated Call Girls In chittoor 📱 {7001035870} VIP Escorts chittoor
dharasingh5698
 
Integrated Test Rig For HTFE-25 - Neometrix
Integrated Test Rig For HTFE-25 - NeometrixIntegrated Test Rig For HTFE-25 - Neometrix
Integrated Test Rig For HTFE-25 - Neometrix
Neometrix_Engineering_Pvt_Ltd
 
VIP Call Girls Ankleshwar 7001035870 Whatsapp Number, 24/07 Booking
VIP Call Girls Ankleshwar 7001035870 Whatsapp Number, 24/07 BookingVIP Call Girls Ankleshwar 7001035870 Whatsapp Number, 24/07 Booking
VIP Call Girls Ankleshwar 7001035870 Whatsapp Number, 24/07 Booking
dharasingh5698
 
Call Girls In Bangalore ☎ 7737669865 🥵 Book Your One night Stand
Call Girls In Bangalore ☎ 7737669865 🥵 Book Your One night StandCall Girls In Bangalore ☎ 7737669865 🥵 Book Your One night Stand
Call Girls In Bangalore ☎ 7737669865 🥵 Book Your One night Stand
amitlee9823
 
Standard vs Custom Battery Packs - Decoding the Power Play
Standard vs Custom Battery Packs - Decoding the Power PlayStandard vs Custom Battery Packs - Decoding the Power Play
Standard vs Custom Battery Packs - Decoding the Power Play
Epec Engineered Technologies
 
FULL ENJOY Call Girls In Mahipalpur Delhi Contact Us 8377877756
FULL ENJOY Call Girls In Mahipalpur Delhi Contact Us 8377877756FULL ENJOY Call Girls In Mahipalpur Delhi Contact Us 8377877756
FULL ENJOY Call Girls In Mahipalpur Delhi Contact Us 8377877756
dollysharma2066
 
Call Girls in Netaji Nagar, Delhi 💯 Call Us 🔝9953056974 🔝 Escort Service
Call Girls in Netaji Nagar, Delhi 💯 Call Us 🔝9953056974 🔝 Escort ServiceCall Girls in Netaji Nagar, Delhi 💯 Call Us 🔝9953056974 🔝 Escort Service
Call Girls in Netaji Nagar, Delhi 💯 Call Us 🔝9953056974 🔝 Escort Service
9953056974 Low Rate Call Girls In Saket, Delhi NCR
 

Recently uploaded (20)

VIP Call Girls Palanpur 7001035870 Whatsapp Number, 24/07 Booking
VIP Call Girls Palanpur 7001035870 Whatsapp Number, 24/07 BookingVIP Call Girls Palanpur 7001035870 Whatsapp Number, 24/07 Booking
VIP Call Girls Palanpur 7001035870 Whatsapp Number, 24/07 Booking
 
Block diagram reduction techniques in control systems.ppt
Block diagram reduction techniques in control systems.pptBlock diagram reduction techniques in control systems.ppt
Block diagram reduction techniques in control systems.ppt
 
data_management_and _data_science_cheat_sheet.pdf
data_management_and _data_science_cheat_sheet.pdfdata_management_and _data_science_cheat_sheet.pdf
data_management_and _data_science_cheat_sheet.pdf
 
A Study of Urban Area Plan for Pabna Municipality
A Study of Urban Area Plan for Pabna MunicipalityA Study of Urban Area Plan for Pabna Municipality
A Study of Urban Area Plan for Pabna Municipality
 
Call Now ≽ 9953056974 ≼🔝 Call Girls In New Ashok Nagar ≼🔝 Delhi door step de...
Call Now ≽ 9953056974 ≼🔝 Call Girls In New Ashok Nagar  ≼🔝 Delhi door step de...Call Now ≽ 9953056974 ≼🔝 Call Girls In New Ashok Nagar  ≼🔝 Delhi door step de...
Call Now ≽ 9953056974 ≼🔝 Call Girls In New Ashok Nagar ≼🔝 Delhi door step de...
 
notes on Evolution Of Analytic Scalability.ppt
notes on Evolution Of Analytic Scalability.pptnotes on Evolution Of Analytic Scalability.ppt
notes on Evolution Of Analytic Scalability.ppt
 
Top Rated Call Girls In chittoor 📱 {7001035870} VIP Escorts chittoor
Top Rated Call Girls In chittoor 📱 {7001035870} VIP Escorts chittoorTop Rated Call Girls In chittoor 📱 {7001035870} VIP Escorts chittoor
Top Rated Call Girls In chittoor 📱 {7001035870} VIP Escorts chittoor
 
Integrated Test Rig For HTFE-25 - Neometrix
Integrated Test Rig For HTFE-25 - NeometrixIntegrated Test Rig For HTFE-25 - Neometrix
Integrated Test Rig For HTFE-25 - Neometrix
 
Call Girls Pimpri Chinchwad Call Me 7737669865 Budget Friendly No Advance Boo...
Call Girls Pimpri Chinchwad Call Me 7737669865 Budget Friendly No Advance Boo...Call Girls Pimpri Chinchwad Call Me 7737669865 Budget Friendly No Advance Boo...
Call Girls Pimpri Chinchwad Call Me 7737669865 Budget Friendly No Advance Boo...
 
VIP Call Girls Ankleshwar 7001035870 Whatsapp Number, 24/07 Booking
VIP Call Girls Ankleshwar 7001035870 Whatsapp Number, 24/07 BookingVIP Call Girls Ankleshwar 7001035870 Whatsapp Number, 24/07 Booking
VIP Call Girls Ankleshwar 7001035870 Whatsapp Number, 24/07 Booking
 
Call Girls In Bangalore ☎ 7737669865 🥵 Book Your One night Stand
Call Girls In Bangalore ☎ 7737669865 🥵 Book Your One night StandCall Girls In Bangalore ☎ 7737669865 🥵 Book Your One night Stand
Call Girls In Bangalore ☎ 7737669865 🥵 Book Your One night Stand
 
chapter 5.pptx: drainage and irrigation engineering
chapter 5.pptx: drainage and irrigation engineeringchapter 5.pptx: drainage and irrigation engineering
chapter 5.pptx: drainage and irrigation engineering
 
UNIT - IV - Air Compressors and its Performance
UNIT - IV - Air Compressors and its PerformanceUNIT - IV - Air Compressors and its Performance
UNIT - IV - Air Compressors and its Performance
 
ONLINE FOOD ORDER SYSTEM PROJECT REPORT.pdf
ONLINE FOOD ORDER SYSTEM PROJECT REPORT.pdfONLINE FOOD ORDER SYSTEM PROJECT REPORT.pdf
ONLINE FOOD ORDER SYSTEM PROJECT REPORT.pdf
 
Minimum and Maximum Modes of microprocessor 8086
Minimum and Maximum Modes of microprocessor 8086Minimum and Maximum Modes of microprocessor 8086
Minimum and Maximum Modes of microprocessor 8086
 
University management System project report..pdf
University management System project report..pdfUniversity management System project report..pdf
University management System project report..pdf
 
Generative AI or GenAI technology based PPT
Generative AI or GenAI technology based PPTGenerative AI or GenAI technology based PPT
Generative AI or GenAI technology based PPT
 
Standard vs Custom Battery Packs - Decoding the Power Play
Standard vs Custom Battery Packs - Decoding the Power PlayStandard vs Custom Battery Packs - Decoding the Power Play
Standard vs Custom Battery Packs - Decoding the Power Play
 
FULL ENJOY Call Girls In Mahipalpur Delhi Contact Us 8377877756
FULL ENJOY Call Girls In Mahipalpur Delhi Contact Us 8377877756FULL ENJOY Call Girls In Mahipalpur Delhi Contact Us 8377877756
FULL ENJOY Call Girls In Mahipalpur Delhi Contact Us 8377877756
 
Call Girls in Netaji Nagar, Delhi 💯 Call Us 🔝9953056974 🔝 Escort Service
Call Girls in Netaji Nagar, Delhi 💯 Call Us 🔝9953056974 🔝 Escort ServiceCall Girls in Netaji Nagar, Delhi 💯 Call Us 🔝9953056974 🔝 Escort Service
Call Girls in Netaji Nagar, Delhi 💯 Call Us 🔝9953056974 🔝 Escort Service
 

Projected Nesterov's Proximal-Gradient Algorithm for Sparse Signal Recovery

  • 1. PNPG Algorithm Applications References References References Acronyms Projected Nesterov’s Proximal-Gradient Algorithm for Sparse Signal RecoveryŽ Aleksandar Dogandžić Electrical and Computer Engineering Ž joint work with Renliang Gu, Ph.D. student supported by 1 / 71
  • 2. PNPG Algorithm Applications References References References Acronyms Terminology and Notation soft-thresholding operator for a D .ai /N iD1 2 RN : T .a/ i D sign.ai / max jai j ; 0 I “ ” is the elementwise version of “ ”; proximal operator for function r.x/ scaled by : prox r a D arg min x 1 2 kx ak2 2 C r.x/: "-subgradient (Rockafellar 1970, Sec. 23): @"r.x/ , ˚ g 2 Rp j r.z/ r.x/ C .z x/T g "; 8z 2 Rp « : 2 / 71
  • 3. PNPG Algorithm Applications References References References Acronyms Introduction I For most natural signals x, # significant coefficients of .x/ signal size p where .x/ W Rp 7! Rp0 is sparsifying transform. 3 / 71
  • 4. PNPG Algorithm Applications References References References Acronyms Sparsifying Transforms I p pixels linear transform $ DWT 0 100 200 300 400 500 600 700 800 900 1000 # significant coeffs p .x/ D ‰T x where ‰ 2 Rp p0 is a known sparsifying dictionary matrix. 4 / 71
  • 5. PNPG Algorithm Applications References References References Acronyms Sparsifying Transforms II p pixels transform $ gradient map # significant coeffs p Œ .x/ p0 iD1 , s X j2Ni .xi xj /2 Ni is the index set of neighbors of xi . 5 / 71
  • 6. PNPG Algorithm Applications References References References Acronyms Convex-Set Constraint x 2 C where C is a nonempty closed convex set. Example: the nonnegative signal set C D R p C is of significant practical interest and applicable to X-ray CT, SPECT, PET, and MRI. 6 / 71
  • 7. PNPG Algorithm Applications References References References Acronyms Goal Sense the significant components of .x/ using a small number of measurements. Define the noiseless measurement vector .x/, where . / W Rp 7! RN (N Ä p). Example: Linear model .x/ D ˆx where ˆ 2 RN p is a known sensing matrix. 7 / 71
  • 8. PNPG Algorithm Applications References References References Acronyms Penalized NLL objective function f .x/ D L.x/ C u k .x/k1 C IC .x/ ž r.x/ convex differentiable negative log-likelihood (NLL) convex penalty term u > 0 is a scalar tuning constant C Â cl dom L.x/ 8 / 71
  • 9. PNPG Algorithm Applications References References References Acronyms Comment Our objective function f .x/ is convex has a convex set as minimum (unique if L.x/ is strongly convex), not differentiable with respect to the signal x cannot apply usual gradient- or Newton-type algorithms, need proximal-gradient (PG) schemes. 9 / 71
  • 10. PNPG Algorithm Applications References References References Acronyms Goals Develop a fast algorithm with O k 2 convergence-rate and iterate convergence guarantees for minimizing f .x/ that is general (for a diverse set of NLLs), requires minimal tuning, and is matrix-free , a must for solving large-scale problems. involves only matrix-vector multiplications implementable using, e.g., function handle in Matlab 10 / 71
  • 11. PNPG Algorithm Applications References References References Acronyms Majorizing Function Define the quadratic approximation of the NLL L.x/: Qˇ x j xx D L.xx/ C .x xx/T rL.xx/ C 1 2ˇ kx xxk2 2 with ˇ chosen so that Qˇ x j xx majorizes L.x/ in the neighborhood of x D xx. ¯x(j)¯x(i) y 0 x L(x) Qβ(i) (x|¯x(i)) Qβ(j) (x|¯x(j)) 11 / 71
  • 12. PNPG Algorithm Applications References References References Acronyms ¯x(j)¯x(i) y 0 x L(x) Qβ(i) (x|¯x(i)) Qβ(j) (x|¯x(j)) ¯x(j)¯x(i) y 0 L(x) Qβ(i) (x|¯x(i)) Qβ(j) (x|¯x(j)) Q10β(j) (x|¯x(j)) Figure 1: Majorizing function: Impact of ˇ. No need for strict majorization, sufficient to majorize in the neighborhood of xx where we wish to move next! 12 / 71
  • 13. PNPG Algorithm Applications References References References Acronyms PNPG Method: Iteration i B.i/ D ˇ.i 1/ =ˇ.i/ Â.i/ D ˚ 1; i Ä 1 1 C q b C B.i/ Â.i 1/ 2 ; i > 1 xx.i/ D PC Â x.i 1/ C Â.i 1/ 1 Â.i/ x.i 1/ x.i 2/ Ã accel. step x.i/ D proxˇ.i/ur xx.i/ ˇ.i/ rL xx.i/ Á PG step where ˇ.i/ > 0 is an adaptive step size: satisfies L x.i/ Ä Qˇ.i/ x.i/ j xx.i/ majorization condition, is as large as possible. more 13 / 71
  • 14. PNPG Algorithm Applications References References References Acronyms PNPG Method: Iteration i allows general dom L B.i/ D ˇ.i 1/ =ˇ.i/ Â.i/ D ˚ i; i Ä 1 1 C q b C B.i/ Â.i 1/ 2 ; i > 1 xx.i/ D PC Â x.i 1/ C Â.i 1/ 1 Â.i/ x.i 1/ x.i 2/ Ã accel. step x.i/ D proxˇ.i/ur xx.i/ ˇ.i/ rL xx.i/ Á PG step where ˇ.i/ > 0 is an adaptive step size: satisfies L x.i/ Ä Qˇ.i/ x.i/ j xx.i/ majorization condition, is as large as possible. 14 / 71
  • 15. PNPG Algorithm Applications References References References Acronyms PNPG Method: Iteration i B.i/ D ˇ.i 1/ =ˇ.i/ Â.i/ D ˚ i; i Ä 1 1 C q b C B.i/ Â.i 1/ 2 ; i > 1 xx.i/ D PC Â x.i 1/ C Â.i 1/ 1 Â.i/ x.i 1/ x.i 2/ Ã accel. step x.i/ D proxˇ.i/ur xx.i/ ˇ.i/ rL xx.i/ Á PG step where ˇ.i/ > 0 is an adaptive step size: satisfies L x.i/ Ä Qˇ.i/ x.i/ j xx.i/ majorization condition, needs to hold for x.i/ , not for all x! L x — Qˇ.i/ x j xx.i/ in general, for an arbitrary x! more 15 / 71
  • 16. PNPG Algorithm Applications References References References Acronyms Algorithm 1: PNPG method Input: x. 1/ , u, n, m, , Á, and threshold Output: arg min x f .x/ Initialization: Â.0/ 0, x.0/ 0, i 0, Ä 0, ˇ.1/ by the BB method repeat i i C 1 and Ä Ä C 1 while true do // backtracking search evaluate B.i/ , Â.i/ , and xx.i/ if xx.i/ … dom L then // domain restart Â.i 1/ 1 and continue solve the PG step if majorization condition holds then break else if ˇ.i/ > ˇ.i 1/ then // increase n n n C m ˇ.i/ ˇ.i/ and Ä 0 if i > 1 and f x.i/ > f x.i 1/ then // function restart Â.i 1/ 1, i i 1, and continue if convergence condition holds then declare convergence if Ä n then // adapt step size Ä 0 and ˇ.iC1/ ˇ.i/ = else ˇ.iC1/ ˇ.i/ until convergence declared or maximum number of iterations exceeded 16 / 71
  • 17. PNPG Algorithm Applications References References References Acronyms 10 7 10 6 10 5 10 4 0 50 100 150 200 250 300 StepSizeˇ.i/ # of iterations PNPG .n D 4/ PNPG .n D 1/ n Figure 2: Illustration of step-size selection for Poisson generalized linear model (GLM) with identity link. 17 / 71
  • 18. PNPG Algorithm Applications References References References Acronyms Momentum Illustration xx.i/ D PC  yx .i/ ¥x.i 1/ C Â.i 1/ 1 Â.i/ x.i 1/ x.i 2/   momentum term prevents zigzagging à x(i−2) x(i−1) x(i) x(i) −∇L(x(i) ) C 18 / 71
  • 19. PNPG Algorithm Applications References References References Acronyms Comments on the extrapolation term Â.i/ I Â.i/ D 1 C q b C B.i/ Â.i 1/ 2 ; i 2 where 2; b 2 0; 1=4 are momentum tuning constants. To establish O k 2 convergence of PNPG, need Â.i/ Ä 1 2 C q 1 4 C B.i/ Â.i 1/ 2 ; i 2: controls the rate of increase of Â.i/ . 19 / 71
  • 20. PNPG Algorithm Applications References References References Acronyms Comments on the extrapolation term Â.i/ II Â.i/ x ? implies stronger momentum. Effect of step size In the “steady state” where ˇ.i 1/ D ˇ.i/ , Â.i/ x ? aproximately linearly with i, with slope 1= . Changes in the step size affect Â.i/ : ˇ.i/ < ˇ.i 1/ step size decrease, faster increase of Â.i/ , ˇ.i/ > ˇ.i 1/ step size increase, decrease or slower increase of Â.i/ than in the steady state. 20 / 71
  • 21. PNPG Algorithm Applications References References References Acronyms Proximal Mapping To compute prox r a D arg min x 1 2 kx ak2 2 C r.x/ use for `1-norm penalty with .x/ D ‰T x, alternating direction method of multipliers (ADMM) o iterative for total-variation (TV)-norm penalty with gradient map .x/, an inner iteration with the TV-based denoising method in (Beck and Teboulle 2009b). o iterative 21 / 71
  • 22. PNPG Algorithm Applications References References References Acronyms Inexact PG Steps B.i/ D ˇ.i 1/ =ˇ.i/ Â.i/ D ˚ i; i Ä 1 1 C q b C B.i/ Â.i 1/ 2 ; i > 1 xx.i/ D PC Â x.i 1/ C Â.i 1/ 1 Â.i/ x.i 1/ x.i 2/ Ã accel. step x.i/ Ñ".i/ proxˇ.i/ur xx.i/ ˇ.i/ rL xx.i/ Á PG step Because of their iterative nature, PG steps are inexact: ".i/ quantifies the precision of the PG step in Iteration i. more 22 / 71
  • 23. PNPG Algorithm Applications References References References Acronyms Remark (Monotonicity) The projected Nesterov’s proximal-gradient (PNPG) iteration with restart is non-increasing: f x.i/ Ä f x.i 1/ if the inexact PG steps are sufficiently accurate and satisfy ".i/ Ä p ı.i/ where ı.i/ , x.i/ x.i 1/ 2 2 is the local variation of signal iterates. 23 / 71
  • 24. PNPG Algorithm Applications References References References Acronyms Convergence Criterion p ı.i/ < x.i/ 2 where > 0 is the convergence threshold. more Restart The goal of function and domain restarts is to ensure that the PNPG iteration is monotonic and xx.i/ and x.i/ remain within dom f . more 24 / 71
  • 25. PNPG Algorithm Applications References References References Acronyms Summary of PNPG Approach Combine convex-set projection with Nesterov acceleration. Apply adaptive step size, restart. more 25 / 71
  • 26. PNPG Algorithm Applications References References References Acronyms Why? Thanks to step-size adaptation, no need for Lipschitz continuity of the gradient of the NLL. dom L does not have to be Rp . Extends the application of the Nesterov’s accelerationŽ to more general measurement models than those used previously. Ž Y. Nesterov, “A method of solving a convex programming problem with convergence rate O.1=k2 /,” Sov. Math. Dokl., vol. 27, 1983, pp. 372–376. 26 / 71
  • 27. PNPG Algorithm Applications References References References Acronyms Theorem (Convergence of the Objective Function) Assume NLL L.x/ is convex and differentiable and r.x/ is convex, C Â dom L: no need for domain restart. Consider the PNPG iteration without restart. 27 / 71
  • 28. PNPG Algorithm Applications References References References Acronyms Theorem (Convergence of the Objective Function) f x.k/ f x? Ä 2 x.0/ x? 2 2 C E.k/ 2 p ˇ.1/ C Pk iD1 p ˇ.i/ Á2 where E.k/ , kX iD1 Â.i/ ".i/ 2 error term, accounts for inexact PG steps x? , arg min x f .x/ 28 / 71
  • 29. PNPG Algorithm Applications References References References Acronyms Comments f x.k/ f x? Ä 2 x.0/ x? 2 2 C E.k/ 2 p ˇ.1/ C Pk iD1 p ˇ.i/ Á2 Step sizes ˇ.i/ x ?, convergence-rate upper bound ? y. better initialization, convergence-rate upper bound #. smaller prox-step approx. error, convergence-rate bound #. 29 / 71
  • 30. PNPG Algorithm Applications References References References Acronyms Corollary Under the condition of the Theorem, f x.k/ f x? Ä 2 x.0/ x? 2 2 C E.k/ 2k2ˇminŸ O k 2 if E.C1/ < C1 provided that ˇmin , C1 min kD1 ˇ.k/ > 0: The assumption that the step-size sequence is lower-bounded by a strictly positive quantity is weaker than Lipschitz continuity of rL.x/ because it is guaranteed to have ˇmin > =L if rL.x/ has a Lipschitz constant L. 30 / 71
  • 31. PNPG Algorithm Applications References References References Acronyms Theorem (Convergence of Iterates) Assume that 1 NLL L.x/ is convex and differentiable and r.x/ is convex, 2 C Â dom L, hence no need for domain restart, 3 cumulative error term E.k/ converges: E.C1/ < C1, 4 momentum tuning constants satisfy > 2 and b 2 0; 1= 2 , 5 the step-size sequence ˇ.i/ C1 iD1 is bounded within the range ˇmin; ˇmax ; .ˇmin > 0/. R The sequence of PNPG iterates x.i/ without restart converges weakly to a minimizer of f .x/. a minimizer of f .x/. 31 / 71
  • 32. PNPG Algorithm Applications References References References Acronyms Theorem (Convergence of Iterates) strict inequality Assume that 1 NLL L.x/ is convex and differentiable and r.x/ is convex, 2 C Â dom L, hence no need for domain restart, 3 cumulative error term E.k/ converges: E.C1/ < C1, 4 momentum tuning const. satisfy > 2 and b 2 0; 1= 2 , 5 the step-size sequence ˇ.i/ C1 iD1 is bounded within the range ˇmin; ˇmax ; .ˇmin > 0/. R The sequence of PNPG iterates x.i/ without restart converges weakly to a minimizer of f .x/. a minimizer of f .x/. narrower than 0; 1=4 32 / 71
  • 33. PNPG Algorithm Applications References References References Acronyms Idea of Proof. Recall the inequality Â.i/ D 1 C q b C B.i/ Â.i 1/ 2 Ä 1 2 C q 1 4 C B.i/ Â.i 1/ 2 used to establish convergence of the objective function. Assumption 4: > 2; b 2 0; 1= 2 creates a sufficient “gap” in this inequality that allows us to show faster convergence of the objective function than the previous theorem and establish the convergence of iterates. R Inspired by (Chambolle and Dossal 2015). 33 / 71
  • 34. PNPG Algorithm Applications References References References Acronyms Introduction Signal reconstruction from Poisson-distributed measurements with affine model for the mean-signal intensity is important for tomographic (Ollinger and Fessler 1997), astronomic, optical, microscopic (Bertero et al. 2009), hyperspectral (Willett et al. 2014) imaging. 34 / 71
  • 35. PNPG Algorithm Applications References References References Acronyms PET: Coincidence detection due to positron decay and annihilation (Prince and Links 2015). 35 / 71
  • 36. PNPG Algorithm Applications References References References Acronyms Measurement Model N independent measurements y D .yn/N nD1 follow the Poisson distribution with means Œˆx C bn where ˆ 2 R N p C ; b are the known sensing matrix and intercept term .  the intercept b models background radiation and scattering, obtained, e.g., by calibration before the measurements y have been collected 36 / 71
  • 37. PNPG Algorithm Applications References References References Acronyms Existing Work The sparse Poisson-intensity reconstruction algorithm (SPIRAL)§ approximates the logarithm function in the underlying NLL by adding a small positive term to it and then descends a regularized NLL objective function with proximal steps that employ Barzilai-Borwein (BB) step size in each iteration, followed by backtracking. § Z. T. Harmany et al., “This is SPIRAL-TAP: Sparse Poisson intensity reconstruction algorithms—theory and practice,” IEEE Trans. Image Process., vol. 21, no. 3, pp. 1084–1096, Mar. 2012. 37 / 71
  • 38. PNPG Algorithm Applications References References References Acronyms PET Image Reconstruction 128 128 concentration map x. Collect the photons from 90 equally spaced directions over 180ı , with 128 radial samples at each direction, Background radiation, scattering effect, and accidental coincidence combined to- gether lead to a known intercept term b. The elements of the intercept term are set to a constant equal to 10 % of the sample mean of ˆx: b D 1T ˆx 10N 1. The model, choices of parameters in the PET system setup, and concentration map have been adopted from Image Reconstruction Toolbox (IRT) (Fessler 2016). 38 / 71
  • 39. PNPG Algorithm Applications References References References Acronyms Numerical Example Main metric for assessing the performance of the compared algorithms is relative square error (RSE) RSE D kyx xtruek2 2 kxtruek2 2 where xtrue and yx are the true and reconstructed signal, respectively. All iterative methods use the convergence threshold D 10 6 and have the maximum number of iterations limited to 104 . Regularization constant u has the form u D 10a : We vary a in the range Œ 6; 3 with a grid size of 0.5 and search for the reconstructions with the best RSE performance. 39 / 71
  • 40. PNPG Algorithm Applications References References References Acronyms Compared Methods Filtered backprojection (FBP) (Ollinger and Fessler 1997) and PG methods that aim at minimizing f .x/ with nonnegative x: C D R p C: All iterative methods initialized by FBP reconstructions. Matlab implementation available at http://isucsp.github.io/imgRecSrc/npg.html. 40 / 71
  • 41. PNPG Algorithm Applications References References References Acronyms PG Methods PNPG with . ; b/ D .2; 1=4/. AT (Auslender and Teboulle 2006) implemented in the templates for first-order conic solvers (TFOCS) package (Becker et al. 2011) with a periodic restart every 200 iterations (tuned for its best performance) and our proximal mapping. SPIRAL, when possible. 41 / 71
  • 42. PNPG Algorithm Applications References References References Acronyms (a) radio-isotope concentration (b) attenuation map Figure 3: (a) True emission image and (b) density map. 42 / 71
  • 43. PNPG Algorithm Applications References References References Acronyms RSE=3.09 % (a) FBP RSE=0.66 % (b) `1 Figure 4: Reconstructions of the emission concentration map for expected total annihilation photon count (SNR) equal to 108 . 43 / 71
  • 44. PNPG Algorithm Applications References References References Acronyms RSE=0.66 % (a) `1 RSE=0.22 % (b) TV Figure 5: Comparison of the two sparsity regularizations. 44 / 71
  • 45. PNPG Algorithm Applications References References References Acronyms 10 1 100 101 102 103 104 105 106 0 20 40 60 80 100 120 140 .i/ Time/s AT PNPG .n D 1/ PNPG .n D m D 0/ PNPG .n D m D 4/ (a) `1, 1T E.y/ D 108 10 4 10 3 10 2 10 1 100 101 102 103 104 105 106 0 10 20 30 40 50 60 .i/ Time/s SPIRAL AT PNPG .n D 1/ PNPG .n D m D 0/ PNPG .n D m D 4/ (b) TV, 1T E.y/ D 107 Figure 6: Centered objectives as functions of CPU time. 45 / 71
  • 46. PNPG Algorithm Applications References 10 1 100 101 102 103 104 105 106 0 20 40 60 80 100 120 140 .i/ Time/s AT PNPG .n D 1/ PNPG .n D m D 0/ PNPG .n D m D 4/ (a) `1, 1T E.y/ D 108 10 4 10 3 10 2 10 1 100 101 102 103 104 105 106 0 10 20 30 40 50 60 .i/ Time/s SPIRAL AT PNPG .n D 1/ PNPG .n D m D 0/ PNPG .n D m D 4/ (b) TV, 1T E.y/ D 107 Figure 7: Centered objectives as functions of CPU time. 50 / 101
  • 47. PNPG Algorithm Applications References References References Acronyms Linear Model with Gaussian Noise L.x/ D 1 2 ky ˆxk2 2 where y 2 RN is the measurement vector and the elements of the sensing matrix ˆ are independent, identically distributed (i.i.d.), drawn from the standard normal distribution. We select the `1-norm sparsifying signal penalty with linear map: .x/ D ‰T x: 46 / 71
  • 48. PNPG Algorithm Applications References References References Acronyms −0.5 0 0.5 1 1.5 2 2.5 0 200 400 600 800 1000 Figure 7: True signal. 47 / 71
  • 49. PNPG Algorithm Applications References References References Acronyms Comments I More methods available for comparison: sparse reconstruction by separable approximation (SpaRSA) (Wright et al. 2009), generalized forward-backward (GFB) (Raguet et al. 2013), primal-dual splitting (PDS) (Condat 2013). Select the regularization parameter u as u D 10a U; U , ‰T rL.0/ 1 where a is an integer selected from the interval Œ 9; 1 and U is an upper bound on u of interest. Choose the nonnegativity convex set: C D R p C: 48 / 71
  • 50. PNPG Algorithm Applications References References References Acronyms Comments II If we remove the convex-set constraint by setting C D Rp , PNPG iteration reduces to the Nesterov’s proximal gradient iteration with adaptive step size that imposes signal sparsity only in the analysis form (termed NPGS). 49 / 71
  • 51. PNPG Algorithm Applications References References References Acronyms −0.5 0 0.5 1 1.5 2 2.5 0 200 400 600 800 1000 RSE=0.011 % (a) PNPG −0.5 0 0.5 1 1.5 2 2.5 0 200 400 600 800 1000 RSE=0.23 % (b) NPGS Figure 8: PNPG and NPGS reconstructions for N=p D 0:34. 50 / 71
  • 52. PNPG Algorithm Applications References References References Acronyms 10−8 10−6 10−4 10−2 100 102 104 0 10 20 30 40 50 60 70 80 90 ∆(i) time/s (a) a D 5; N=p D 0:34 10−8 10−6 10−4 10−2 100 102 104 0 10 20 30 40 50 60 70 80 90 ∆(i) time/s SPIRAL SpaRSA (cont.) PNPG PNPG (cont.) GFB PDS AT (b) a D 6; N=p D 0:49 Figure 9: Centered objectives as functions of CPU time. 51 / 71
  • 53. PNPG Algorithm Applications References References References Acronyms 10−8 10−6 10−4 10−2 100 102 104 0 10 20 30 40 50 60 70 80 time/s (a) a D 4; N=p D 0:34 10−8 10−6 10−4 10−2 100 102 104 0 10 20 30 40 50 60 70 80 time/s SPIRAL SpaRSA (cont.) PNPG PNPG (cont.) GFB PDS AT (b) a D 5; N=p D 0:49 Figure 10: Centered objectives as functions of CPU time. 52 / 71
  • 54. PNPG Algorithm Applications References References References Acronyms 10−8 10−6 10−4 10−2 100 102 104 0 5 10 15 20 25 30 time/s (a) a D 3; N=p D 0:34 10−8 10−6 10−4 10−2 100 102 104 0 5 10 15 20 time/s SPIRAL SpaRSA (cont.) PNPG PNPG (cont.) GFB PDS AT (b) a D 4; N=p D 0:49 Figure 11: Centered objectives as functions of CPU time. 53 / 71
  • 55. PNPG Algorithm Applications References References References Acronyms Publications R. G. and A. D. (May 2016), Projected Nesterov’s proximal-gradient algorithm for sparse signal reconstruction with a convex constraint, version 5. arXiv: 1502.02613 [stat.CO]. R. G. and A. D., “Projected Nesterov’s proximal-gradient signal recovery from compressive Poisson measurements,” Proc. Asilomar Conf. Signals, Syst. Comput., Pacific Grove, CA, Nov. 2015, pp. 1490–1495. 54 / 71
  • 56. PNPG Algorithm Applications References References References Acronyms Preliminary Work R. G. and A. D., “A fast proximal gradient algorithm for reconstructing nonnegative signals with sparse transform coefficients,” Proc. Asilomar Conf. Signals, Syst. Comput., Pacific Grove, CA, Nov. 2014, pp. 1662–1667. R. G. and A. D. (Mar. 2015), Reconstruction of nonnegative sparse signals using accelerated proximal-gradient algorithms, version 3. arXiv: 1502.02613v3 [stat.CO]. 55 / 71
  • 57. PNPG Algorithm Applications References References References Acronyms Conclusion Developed a fast framework for reconstructing signals that are sparse in a transform domain and belong to a closed convex set by employing a projected proximal-gradient scheme with Nesterov’s acceleration, restart and adaptive step size. Applied the proposed framework to construct the first Nesterov-accelerated Poisson compressed-sensing reconstruction algorithm. Derived convergence-rate upper-bound that accounts for inexactness of the proximal operator. Proved convergence of iterates. Our PNPG approach is computationally efficient compared with the state-of-the-art. 56 / 71
  • 58. PNPG Algorithm Applications References References References Acronyms References I A. Auslender and M. Teboulle, “Interior gradient and proximal methods for convex and conic optimization,” SIAM J. Optim., vol. 16, no. 3, pp. 697–725, 2006. A. Beck and M. Teboulle, “A fast iterative shrinkage-thresholding algorithm for linear inverse problems,” SIAM J. Imag. Sci., vol. 2, no. 1, pp. 183–202, 2009. A. Beck and M. Teboulle, “Fast gradient-based algorithms for constrained total variation image denoising and deblurring problems,” IEEE Trans. Image Process., vol. 18, no. 11, pp. 2419–2434, 2009. S. R. Becker, E. J. Candès, and M. C. Grant, “Templates for convex cone problems with applications to sparse signal recovery,” Math. Program. Comp., vol. 3, no. 3, pp. 165–218, 2011. [Online]. Available: http://cvxr.com/tfocs. M. Bertero, P. Boccacci, G. Desiderà, and G. Vicidomini, “Image deblurring with Poisson data: From cells to galaxies,” Inverse Prob., vol. 25, no. 12, pp. 123006-1–123006-26, 2009. S. Bonettini, I. Loris, F. Porta, and M. Prato, “Variable metric inexact line-search-based methods for nonsmooth optimization,” SIAM J. Optim., vol. 26, no. 2, pp. 891–921, 2016. 57 / 71
  • 59. PNPG Algorithm Applications References References References Acronyms References II S. Bonettini, F. Porta, and V. Ruggiero, “A variable metric forward-backward method with extrapolation,” SIAM J. Sci. Comput., vol. 38, no. 4, A2558–A2584, 2016. A. Chambolle and C. Dossal, “On the convergence of the iterates of the ‘fast iterative shrinkage/thresholding algorithm’,” J. Optim. Theory Appl., vol. 166, no. 3, pp. 968–982, 2015. L. Condat, “A primal–dual splitting method for convex optimization involving Lipschitzian, proximable and linear composite terms,” J. Optim. Theory Appl., vol. 158, no. 2, pp. 460–479, 2013. J. A. Fessler, Image reconstruction toolbox, [Online]. Available: http://www.eecs.umich.edu/~fessler/code (visited on 08/23/2016). Z. T. Harmany, R. F. Marcia, and R. M. Willett, “This is SPIRAL-TAP: Sparse Poisson intensity reconstruction algorithms—theory and practice,” IEEE Trans. Image Process., vol. 21, no. 3, pp. 1084–1096, Mar. 2012. Y. Nesterov, “A method of solving a convex programming problem with convergence rate O.1=k2 /,” Sov. Math. Dokl., vol. 27, 1983, pp. 372–376. 58 / 71
  • 60. PNPG Algorithm Applications References References References Acronyms References III B. O‘Donoghue and E. Candès, “Adaptive restart for accelerated gradient schemes,” Found. Comput. Math., vol. 15, no. 3, pp. 715–732, 2015. J. M. Ollinger and J. A. Fessler, “Positron-emission tomography,” IEEE Signal Process. Mag., vol. 14, no. 1, pp. 43–55, 1997. B. T. Polyak, “Some methods of speeding up the convergence of iteration methods,” USSR Comput. Math. Math. Phys., vol. 4, no. 5, pp. 1–17, 1964. B. T. Polyak, Introduction to Optimization. New York: Optimization Software, 1987. J. L. Prince and J. M. Links, Medical Imaging Signals and Systems, 2nd ed. Upper Saddle River, NJ: Pearson, 2015. H. Raguet, J. Fadili, and G. Peyré, “A generalized forward-backward splitting,” SIAM J. Imag. Sci., vol. 6, no. 3, pp. 1199–1226, 2013. R. T. Rockafellar, Convex Analysis. Princeton, NJ: Princeton Univ. Press, 1970. 59 / 71
  • 61. PNPG Algorithm Applications References References References Acronyms References IV S. Salzo. (May 2016), The variable metric forward-backward splitting algorithm under mild differentiability assumptions, arXiv: 1605.00952 [math.OC]. S. Villa, S. Salzo, L. Baldassarre, and A. Verri, “Accelerated and inexact forward-backward algorithms,” SIAM J. Optim., vol. 23, no. 3, pp. 1607–1633, 2013. R. M. Willett, M. F. Duarte, M. A. Davenport, and R. G. Baraniuk, “Sparsity and structure in hyperspectral imaging: Sensing, reconstruction, and target detection,” IEEE Signal Process. Mag., vol. 31, no. 1, pp. 116–126, Jan. 2014. S. J. Wright, R. D. Nowak, and M. A. T. Figueiredo, “Sparse reconstruction by separable approximation,” IEEE Trans. Signal Process., vol. 57, no. 7, pp. 2479–2493, 2009. 60 / 71
  • 62. PNPG Algorithm Applications References References References Acronyms Adaptive Step Size 1 if no step-size backtracking events or increase attempts for n consecutive iterations, start with a larger step size xˇ.i/ D ˇ.i 1/ (increase attempt) where 2 .0; 1/ is a step-size adaptation parameter; otherwise start with xˇ.i/ D ˇ.i 1/ I 2 (backtracking search) select ˇ.i/ D ti xˇ.i/ (2) where ti 0 is the smallest integer such that (2) satisfies the majorization condition (2); backtracking event corresponds to ti > 0. 3 if max ˇ.i/ ; ˇ.i 1/ < xˇ.i/ , increase n by a nonnegative integer m: n n C m: back 61 / 71
  • 63. PNPG Algorithm Applications References References References Acronyms Restart Whenever f x.i/ > f x.i 1/ or xx.i/ 2 C n dom L, we set Â.i 1/ D 1 (restart) and refer to this action as function restart (O‘Donoghue and Candès 2015) or domain restart respectively. back 62 / 71
  • 64. PNPG Algorithm Applications References References References Acronyms Inner Convergence Criteria TV: x.i;j/ x.i;j 1/ 2 Ä Á p ı.i 1/ (3a) `1: max s.i;j/ ‰T x.i;j/ 2 ; s.i;j/ s.i;j 1/ 2 Á Ä Á ‰T x.i 1/ x.i 2/ 2 (3b) where j is the inner-iteration index, x.i;j/ is the iterate of x in the jth inner iteration step within the ith step of the (outer) PNPG iteration, and Á 2 .0; 1/ is the convergence tuning constant chosen to trade off the accuracy and speed of the inner iterations and provide sufficiently accurate solutions to the proximal mapping. back 63 / 71
  • 65. PNPG Algorithm Applications References References References Acronyms Definition (Inexact Proximal Operator (Villa et al. 2013)) We say that x is an approximation of proxur a with "-precision, denoted by x Ñ" proxur a if a x u 2 @"2 2u r.x/: Note: This definition implies kx proxur ak2 2 Ä "2 : back 64 / 71
  • 66. PNPG Algorithm Applications References References References Acronyms Relationship with FISTA I PNPG can be thought of as a generalized fast iterative shrinkage-thresholding algorithm (FISTA) (Beck and Teboulle 2009a) that accommodates convex constraints, more general NLLs,‘ and (increasing) adaptive step size thanks to this step-size adaptation, PNPG does not require Lipschitz continuity of the gradient of the NLL. back ‘ FISTA has been developed for the linear Gaussian model. 65 / 71
  • 67. PNPG Algorithm Applications References References References Acronyms Relationship with FISTA II Need B.i/ to derive theoretical guarantee for convergence speed of the PNPG iteration. In contrast with PNPG, FISTA has a non-increasing step size ˇ.i/ , which allows for setting B.i/ D 1 for all i: Â.i/ D 1 2 Ä 1 C q 1 C 4 Â.i 1/ 2 : A simpler version of FISTA is Â.i/ D 1 2 C Â.i 1/ D i C 1 2 for i 1, which corresponds to . ; b/ D .2; 0/. back Y. Nesterov, “A method of solving a convex programming problem with convergence rate O.1=k2 /,” Sov. Math. Dokl., vol. 27, 1983, pp. 372–376. 66 / 71
  • 68. PNPG Algorithm Applications References References References Acronyms Relationship with AT (Auslender and Teboulle 2006): Â.i/ D 1 2 Ä 1 C q 1 C 4 Â.i 1/ 2 xx.i/ D Â 1 1 Â.i/ Ã x.i 1/ C 1 Â.i/ zx.i 1/ zx.i/ D proxÂ.i/ˇ.i/ur zx.i 1/ Â.i/ ˇ.i/ rL xx.i/ Á x.i/ D Â 1 1 Â.i/ Ã x.i 1/ C 1 Â.i/ zx.i/ Other variants with infinite memory are available at (Becker et al. 2011). back 67 / 71
  • 69. PNPG Algorithm Applications References References References Acronyms Heavy-ball Method (Polyak 1964; Polyak 1987): x.i/ D proxˇ.i/ur x.i 1/ ˇ.i/ rL x.i 1/ Á C ‚.i/ x.i 1/ x.i 2/ : back 68 / 71
  • 70. PNPG Algorithm Applications References References References Acronyms Glossary I ADMM alternating direction method of multipliers. 21 BB Barzilai-Borwein. 37 CPU central processing unit. 45 CT computed tomography. 6 FBP filtered backprojection. 40 FISTA fast iterative shrinkage-thresholding algorithm. 66, 67 GFB generalized forward-backward. 49 GLM generalized linear model. 17 i.i.d. independent, identically distributed. 47 IRT Image Reconstruction Toolbox. 38 MRI magnetic resonance imaging. 6 70 / 71
  • 71. PNPG Algorithm Applications References References References Acronyms Glossary II NLL negative log-likelihood. 10, 11, 26, 27, 31, 32, 37, 66 NPGS Nesterov’s proximal-gradient sparse. 50, 51 PDS primal-dual splitting. 49 PET positron emission tomography. 6, 35, 38 PG proximal-gradient. 9, 16, 22, 23, 28, 40 PNPG projected Nesterov’s proximal-gradient. 23, 24, 27, 31, 32, 41, 51, 57, 64, 66, 67 RSE relative square error. 39 SpaRSA sparse reconstruction by separable approximation. 49 SPECT single photon emission computed tomography. 6 SPIRAL sparse Poisson-intensity reconstruction algorithm. 37, 41 TFOCS templates for first-order conic solvers. 41 TV total-variation. 21 71 / 71