SlideShare a Scribd company logo
A new (more intuitive?) interpretation of inertial
algorithms
Silvia Villa
Politecnico di Milano
https://www.mate.polimi.it/analysis/?settore=people
SAMSI Workshop: Operator Splitting Methods in Data Analysis
Raleigh, March 21th, 2018
S. Villa (Polimi) 1 / 20
Introduction
Problem setting
Let
H be a Hilbert space
f : H → R convex, differentiable, with an L-Lipschitz continuous
gradient
Assume that argmin f = ∅.
We consider the problem of computing
min
x∈H
f (x)
S. Villa (Polimi) 2 / 20
Introduction
Gradient method
Given x0 ∈ H and γ ∈ ]0, 2/L[
For k ≥ 0 define
xk+1 = xk − γ f (xk)
Classic convergence results:
f (xk) − min f = O(1/k)
Convergence of the iterates: xk ¯x ∈ argmin f
S. Villa (Polimi) 3 / 20
Introduction
Accelerated/Inertial gradient method
Given x0 = y0 ∈ H and γ ∈ ]0, 1/L], αk ∈ ]0, 1[:
For k ≥ 0, define
xk+1 = yk − γ f (yk)
yk+1 = xk+1 + αk(xk+1 − xk)
Convergence results:
f (xk) − min f = O(1/k2)
S. Villa (Polimi) 4 / 20
Introduction
Accelerated/Inertial gradient method
Given x0 = y0 ∈ H and γ ∈ ]0, 1/L], αk ∈ ]0, 1[:
For k ≥ 0, define
xk+1 = yk − γ f (yk)
yk+1 = xk+1 + αk(xk+1 − xk)
Convergence results:
f (xk) − min f = O(1/k2)
xk ¯x ∈ argmin f
S. Villa (Polimi) 4 / 20
Introduction
Convergence results and remarks
Convergence and its rate depend on the choice of αk
S. Villa (Polimi) 5 / 20
Introduction
Convergence results and remarks
Convergence and its rate depend on the choice of αk
Nesterov rule for choosing αk [Nesterov, A method for solving a convex
programming problem with rate of convergence O(1/k2
), 1983]
S. Villa (Polimi) 5 / 20
Introduction
Convergence results and remarks
Convergence and its rate depend on the choice of αk
Nesterov rule for choosing αk [Nesterov, A method for solving a convex
programming problem with rate of convergence O(1/k2
), 1983]
Extension to proximal point algorithm [G¨uler, New proximal point
algorithms for convex minimization, 1992] and to forward-backward
[Beck-Teboulle, FISTA, 2009]
S. Villa (Polimi) 5 / 20
Introduction
Convergence results and remarks
Convergence and its rate depend on the choice of αk
Nesterov rule for choosing αk [Nesterov, A method for solving a convex
programming problem with rate of convergence O(1/k2
), 1983]
Extension to proximal point algorithm [G¨uler, New proximal point
algorithms for convex minimization, 1992] and to forward-backward
[Beck-Teboulle, FISTA, 2009]
Other choices for αk [Chambolle-Dossal, 2014],[Attouch-Peypouquet-Redont
2014], [Apidopoulus-Aujol-Dossal 2017]
αk = 1 −
α
k + 2
For convergence: α > 3.
S. Villa (Polimi) 5 / 20
Introduction
In the rest of the talk
Brief review of some approaches to show convergence
“Algebraic proof” [Nesterov; G¨uler; Beck-Teboulle; Chambolle-Dossal]
Estimate sequences [Nesterov; G¨uler; Salzo-Villa]
Primal and mirror descent combination [Zhu-Orecchia, Linear coupling: An
ultimate unification of gradient and mirror descent, 2017
ODE approach [Su-Boyd-Candes; Apidopoulos-Aujol-Dossal;
Attouch-Cabot-Peypouquet-Redont-Chbani...]
S. Villa (Polimi) 6 / 20
Introduction
In the rest of the talk
Brief review of some approaches to show convergence
“Algebraic proof” [Nesterov; G¨uler; Beck-Teboulle; Chambolle-Dossal]
Estimate sequences [Nesterov; G¨uler; Salzo-Villa]
Primal and mirror descent combination [Zhu-Orecchia, Linear coupling: An
ultimate unification of gradient and mirror descent, 2017
ODE approach [Su-Boyd-Candes; Apidopoulos-Aujol-Dossal;
Attouch-Cabot-Peypouquet-Redont-Chbani...]
Many results...but something is still missing :-)
S. Villa (Polimi) 6 / 20
Introduction
In the rest of the talk
Brief review of some approaches to show convergence
“Algebraic proof” [Nesterov; G¨uler; Beck-Teboulle; Chambolle-Dossal]
Estimate sequences [Nesterov; G¨uler; Salzo-Villa]
Primal and mirror descent combination [Zhu-Orecchia, Linear coupling: An
ultimate unification of gradient and mirror descent, 2017
ODE approach [Su-Boyd-Candes; Apidopoulos-Aujol-Dossal;
Attouch-Cabot-Peypouquet-Redont-Chbani...]
Many results...but something is still missing :-)
S. Villa (Polimi) 6 / 20
Classic approaches
“Algebraic proof” - Nesterov’s choice of αk
αk =
(tk − 1)
tk+1
For an arbitrary y ∈ H, it holds:
2
L
(t2
k+1 − tk+1)(f (xk) − min f ) + tk+1y − (tk+1 − 1)xk − ¯x 2
≥
2
L
t2
k+1(f (xk+1) − min f ) + tk+1xk+1 − (tk+1 − 1)xk − ¯x 2
If
t2
k ≥ t2
k+1 − tk+1
and tk+1y = tk+1xk + (tk − 1)(xk − xk−1)
S. Villa (Polimi) 7 / 20
Classic approaches
then
2
L
t2
k (f (xk) − min f ) + tkxk − (tk − 1)xk−1 − ¯x 2
≥
2
L
t2
k+1(f (xk+1) − min f ) + tk+1xk+1 − (tk+1 − 1)xk − ¯x 2
At the end, there exists c > 0 such that
f (xk) − min f ≤
c
t2
k
S. Villa (Polimi) 8 / 20
Classic approaches
then
2
L
t2
k (f (xk) − min f ) + tkxk − (tk − 1)xk−1 − ¯x 2
≥
2
L
t2
k+1(f (xk+1) − min f ) + tk+1xk+1 − (tk+1 − 1)xk − ¯x 2
At the end, there exists c > 0 such that
f (xk) − min f ≤
c
t2
k
=⇒ f (xk) − min f ≤
c
(k + 1)2
S. Villa (Polimi) 8 / 20
Classic approaches
then
2
L
t2
k (f (xk) − min f ) + tkxk − (tk − 1)xk−1 − ¯x 2
≥
2
L
t2
k+1(f (xk+1) − min f ) + tk+1xk+1 − (tk+1 − 1)xk − ¯x 2
At the end, there exists c > 0 such that
f (xk) − min f ≤
c
t2
k
=⇒ f (xk) − min f ≤
c
(k + 1)2
**SIMILAR** proof for convergence of (xk) ([Chambolle-Dossal, On the weak
convergence of the iterates of “FISTA”, 2014])
S. Villa (Polimi) 8 / 20
Classic approaches
Nesterov’s estimate sequences
Definition
A pair of sequences ϕk : X → R and βk ≥ 0 is called an estimate
sequence of the function f if for any x ∈ X and all k ≥ 0 we have
ϕk(x) − f (x) ≤ βk(ϕ0(x) − f (x)) and βk → 0.
S. Villa (Polimi) 9 / 20
Classic approaches
Nesterov’s estimate sequences
Definition
A pair of sequences ϕk : X → R and βk ≥ 0 is called an estimate
sequence of the function f if for any x ∈ X and all k ≥ 0 we have
ϕk(x) − f (x) ≤ βk(ϕ0(x) − f (x)) and βk → 0.
Theorem (Nesterov ’83)
Let (ϕk, βk) be an estimate sequence of f and let min f = f (¯x). If for
some xk ∈ X we have
f (xk) ≤ min ϕk
then
f (xk) − min f ≤ βk(ϕ0(¯x) − f (¯x))
βk gives a convergence rate for f (xk) − min f
S. Villa (Polimi) 9 / 20
Classic approaches
A method to build estimate sequences
Given a quadratic function ϕ0 = f (u0) + A
2 · −u0
2, tk > 1, xk+1 , we
define recursively
ϕk+1(x) := (1 − t−1
k )ϕk(x) + t−1
k (f (yk+1) + x − yk+1, f (yk+1)
≤f (x), linear term
).
We have:
ϕk = ¯ϕk + Ak
2 x − uk
2, is a quadratic function for every k
ϕk+1(x) − f (x) ≤ (1 − t−1
k )(ϕk(x) − f (x))
≤
k
i=0
(1 − t−1
i )
=βk+1
(ϕ0(x) − f (x)).
S. Villa (Polimi) 10 / 20
Classic approaches
Choice of tk and yk
The parameters ¯ϕk, Ak, uk can be updated recursively and depend on the
choice of tk and yk.
Lemma
Suppose that xk is such that ¯ϕk ≥ f (xk). Set yk = (1 − t−1
k )xk + t−1
k uk,
t2
k = t2
k+1 − tk+1, and xk+1 = yk − γ f (yk). Then
¯ϕk+1 ≥ f (xk+1)
S. Villa (Polimi) 11 / 20
More recent approaches
Linear coupling of gradient and mirror steps
Initialize x0 = y0 = u0.



yk+1 = (1 − t−1
k )xk + t−1
k uk Linear coupling
xk+1 = yk+1 − 1
L f (yk+1) Gradient step
uk+1 = uk − αk f (xk+1) Mirror step
S. Villa (Polimi) 12 / 20
More recent approaches
Linear coupling of gradient and mirror steps
Initialize x0 = y0 = u0.



yk+1 = (1 − t−1
k )xk + t−1
k uk Linear coupling
xk+1 = yk+1 − 1
L f (yk+1) Gradient step
uk+1 = uk − αk f (xk+1) Mirror step
Link with estimate sequences?
S. Villa (Polimi) 12 / 20
More recent approaches
Linear coupling of gradient and mirror steps
Initialize x0 = y0 = u0.



yk+1 = (1 − t−1
k )xk + t−1
k uk Linear coupling
xk+1 = yk+1 − 1
L f (yk+1) Gradient step
uk+1 = uk − αk f (xk+1) Mirror step
Link with estimate sequences?
Bregman estimate sequences version?
S. Villa (Polimi) 12 / 20
More recent approaches
ODE approach
The convergence properties of the algorithm are “derived” from the
asymptotic behavior, as t → +∞, of the evolution system



¨x(t) + α
t ˙x(t) + f (x(t)) = 0
x(t0) = x0
˙x(t0) = v0.
A finite difference discretization, with tk = t0 + kh, gives
1
h2
(xk+1 − 2xk + xk−1) +
α
kh2
(xk − xk−1) + f (yk) = 0,
with yk to be determined later.
S. Villa (Polimi) 13 / 20
More recent approaches
ODE approach
We **NATURALLY** obtain
xk+1 = xk + 1 −
α
k
(xk − xk−1) − h2
f (yk),
Setting γ = h2:
yk = xk + 1 − α
k (xk − xk−1)
xk+1 = yk − γ f (yk),
S. Villa (Polimi) 14 / 20
More recent approaches
The heavy ball
Choosing yk = xk we obtain the “heavy ball method”
xk+1 = xk + 1 −
α
k
(xk − xk−1) − γ f (xk),
The continuous dynamic is the same, but the convergence behavior of the
algorithm is different
S. Villa (Polimi) 15 / 20
More recent approaches
The problem
What is the difference between the heavy ball and the inertial
algorithms?
S. Villa (Polimi) 16 / 20
More recent approaches
The problem
What is the difference between the heavy ball and the inertial
algorithms?
Why choosing yk in such a way accelerate the gradient method?
S. Villa (Polimi) 16 / 20
New directions
Another point of view: link with numerical analysis
[Scieur-Roulet-Bach-D’Aspremont, Integration methods and Accelerated Optimization
Algorithms, 2017]
Use numerical integration schemes to approximate the solution of the
problem:
˙x(t) + f (x(t)) = 0
x(t0) = x0
Linear two-steps schemes
Fix h > 0. Given x0, x1 ∈ H, for k ≥ 0:
xk+1 = −ρ1xk − ρ0xk−1 + γ(σ0 f (xk−1) + σ1 f (xk)).
Restrictions on ρ0, ρ1, σ0, σ1 in R to guarantee convergence, i.e.
lim
γ→0
xk − x(t0 + kγ) = 0, ∀k ∈ [1, K]
under appropriate initial conditions x0, x1.
S. Villa (Polimi) 17 / 20
New directions
Quadratic strongly convex case
Let H = Rn, f (x) = Ax, with A symmetric positive definite. Then the
inertial coefficient can be chosen constant, equal to α. In Polyak’s method:
xk+1 = xk + (1 − α)(xk − xk−1) − γ f (xk),
The heavy ball method is a multi-step method
Recall
xk+1 = −ρ1xk − ρ0xk−1 + γ(σ0 f (xk−1) + σ1 f (xk)).
Then, it is enough to choose:
ρ0 = 1 − α
ρ1 = 2 − α
σ0 = 0
σ1 = −1
S. Villa (Polimi) 18 / 20
New directions
Quadratic strongly convex case
Let H = Rn, f (x) = Ax, with A symmetric positive definite. Then the
inertial coefficient in the accelerated algorithm can be chosen constant,
equal to α.
yk = xk + (1 − α) (xk − xk−1)
xk+1 = yk − γ f (yk),
Accelerated method
Nesterov’s method is a multi-step method
ρ0 = 1 − α
ρ1 = α − 2
σ0 = 1 − α
σ1 = α − 1
S. Villa (Polimi) 19 / 20
New directions
Quadratic strongly convex case
Let H = Rn, f (x) = Ax, with A symmetric positive definite. Then the
inertial coefficient in the accelerated algorithm can be chosen constant,
equal to α.
yk = xk + (1 − α) (xk − xk−1)
xk+1 = yk − γ f (yk),
Accelerated method
Nesterov’s method is a multi-step method
ρ0 = 1 − α
ρ1 = α − 2
σ0 = 1 − α
σ1 = α − 1
It suffices to write
yk+1 = yk − γ f (yk) + (1 − α) (yk − γ f (yk) − yk−1 + γ f (yk−1))
S. Villa (Polimi) 19 / 20
New directions
Conclusions/Open problems
The Polyak method turns out to be “optimal” for quadratic strongly
convex functions
S. Villa (Polimi) 20 / 20
New directions
Conclusions/Open problems
The Polyak method turns out to be “optimal” for quadratic strongly
convex functions
What about the general case? The coefficients are not constant, but
depend on k if f is only convex.
S. Villa (Polimi) 20 / 20
New directions
Conclusions/Open problems
The Polyak method turns out to be “optimal” for quadratic strongly
convex functions
What about the general case? The coefficients are not constant, but
depend on k if f is only convex.
Better choices of αk? Maybe driven by the geometry of the function?
S. Villa (Polimi) 20 / 20
New directions
Conclusions/Open problems
The Polyak method turns out to be “optimal” for quadratic strongly
convex functions
What about the general case? The coefficients are not constant, but
depend on k if f is only convex.
Better choices of αk? Maybe driven by the geometry of the function?
Can this framework encompass restarting techniques?
[O’Donoghue-Cand`es, Adaptive restart for accelerated gradient schemes, Found.
Comput. Math., 2013.]
S. Villa (Polimi) 20 / 20
New directions
Conclusions/Open problems
The Polyak method turns out to be “optimal” for quadratic strongly
convex functions
What about the general case? The coefficients are not constant, but
depend on k if f is only convex.
Better choices of αk? Maybe driven by the geometry of the function?
Can this framework encompass restarting techniques?
[O’Donoghue-Cand`es, Adaptive restart for accelerated gradient schemes, Found.
Comput. Math., 2013.]
New algorithms? New integration methods?
S. Villa (Polimi) 20 / 20

More Related Content

What's hot

Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
The Statistical and Applied Mathematical Sciences Institute
 
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
The Statistical and Applied Mathematical Sciences Institute
 
Signal Processing Course : Inverse Problems Regularization
Signal Processing Course : Inverse Problems RegularizationSignal Processing Course : Inverse Problems Regularization
Signal Processing Course : Inverse Problems Regularization
Gabriel Peyré
 
Patch Matching with Polynomial Exponential Families and Projective Divergences
Patch Matching with Polynomial Exponential Families and Projective DivergencesPatch Matching with Polynomial Exponential Families and Projective Divergences
Patch Matching with Polynomial Exponential Families and Projective Divergences
Frank Nielsen
 
Divergence clustering
Divergence clusteringDivergence clustering
Divergence clustering
Frank Nielsen
 
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
The Statistical and Applied Mathematical Sciences Institute
 
CLIM Fall 2017 Course: Statistics for Climate Research, Estimating Curves and...
CLIM Fall 2017 Course: Statistics for Climate Research, Estimating Curves and...CLIM Fall 2017 Course: Statistics for Climate Research, Estimating Curves and...
CLIM Fall 2017 Course: Statistics for Climate Research, Estimating Curves and...
The Statistical and Applied Mathematical Sciences Institute
 
QMC Opening Workshop, High Accuracy Algorithms for Interpolating and Integrat...
QMC Opening Workshop, High Accuracy Algorithms for Interpolating and Integrat...QMC Opening Workshop, High Accuracy Algorithms for Interpolating and Integrat...
QMC Opening Workshop, High Accuracy Algorithms for Interpolating and Integrat...
The Statistical and Applied Mathematical Sciences Institute
 
H2O World - Consensus Optimization and Machine Learning - Stephen Boyd
H2O World - Consensus Optimization and Machine Learning - Stephen BoydH2O World - Consensus Optimization and Machine Learning - Stephen Boyd
H2O World - Consensus Optimization and Machine Learning - Stephen Boyd
Sri Ambati
 
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
The Statistical and Applied Mathematical Sciences Institute
 
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
The Statistical and Applied Mathematical Sciences Institute
 
Computational Information Geometry: A quick review (ICMS)
Computational Information Geometry: A quick review (ICMS)Computational Information Geometry: A quick review (ICMS)
Computational Information Geometry: A quick review (ICMS)
Frank Nielsen
 
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
The Statistical and Applied Mathematical Sciences Institute
 
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
The Statistical and Applied Mathematical Sciences Institute
 
BEADS : filtrage asymétrique de ligne de base (tendance) et débruitage pour d...
BEADS : filtrage asymétrique de ligne de base (tendance) et débruitage pour d...BEADS : filtrage asymétrique de ligne de base (tendance) et débruitage pour d...
BEADS : filtrage asymétrique de ligne de base (tendance) et débruitage pour d...
Laurent Duval
 
Divergence center-based clustering and their applications
Divergence center-based clustering and their applicationsDivergence center-based clustering and their applications
Divergence center-based clustering and their applications
Frank Nielsen
 
A T(1)-type theorem for entangled multilinear Calderon-Zygmund operators
A T(1)-type theorem for entangled multilinear Calderon-Zygmund operatorsA T(1)-type theorem for entangled multilinear Calderon-Zygmund operators
A T(1)-type theorem for entangled multilinear Calderon-Zygmund operators
VjekoslavKovac1
 
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
The Statistical and Applied Mathematical Sciences Institute
 
Classification with mixtures of curved Mahalanobis metrics
Classification with mixtures of curved Mahalanobis metricsClassification with mixtures of curved Mahalanobis metrics
Classification with mixtures of curved Mahalanobis metrics
Frank Nielsen
 
Optimal L-shaped matrix reordering, aka graph's core-periphery
Optimal L-shaped matrix reordering, aka graph's core-peripheryOptimal L-shaped matrix reordering, aka graph's core-periphery
Optimal L-shaped matrix reordering, aka graph's core-periphery
Francesco Tudisco
 

What's hot (20)

Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
 
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
 
Signal Processing Course : Inverse Problems Regularization
Signal Processing Course : Inverse Problems RegularizationSignal Processing Course : Inverse Problems Regularization
Signal Processing Course : Inverse Problems Regularization
 
Patch Matching with Polynomial Exponential Families and Projective Divergences
Patch Matching with Polynomial Exponential Families and Projective DivergencesPatch Matching with Polynomial Exponential Families and Projective Divergences
Patch Matching with Polynomial Exponential Families and Projective Divergences
 
Divergence clustering
Divergence clusteringDivergence clustering
Divergence clustering
 
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
 
CLIM Fall 2017 Course: Statistics for Climate Research, Estimating Curves and...
CLIM Fall 2017 Course: Statistics for Climate Research, Estimating Curves and...CLIM Fall 2017 Course: Statistics for Climate Research, Estimating Curves and...
CLIM Fall 2017 Course: Statistics for Climate Research, Estimating Curves and...
 
QMC Opening Workshop, High Accuracy Algorithms for Interpolating and Integrat...
QMC Opening Workshop, High Accuracy Algorithms for Interpolating and Integrat...QMC Opening Workshop, High Accuracy Algorithms for Interpolating and Integrat...
QMC Opening Workshop, High Accuracy Algorithms for Interpolating and Integrat...
 
H2O World - Consensus Optimization and Machine Learning - Stephen Boyd
H2O World - Consensus Optimization and Machine Learning - Stephen BoydH2O World - Consensus Optimization and Machine Learning - Stephen Boyd
H2O World - Consensus Optimization and Machine Learning - Stephen Boyd
 
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
 
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
 
Computational Information Geometry: A quick review (ICMS)
Computational Information Geometry: A quick review (ICMS)Computational Information Geometry: A quick review (ICMS)
Computational Information Geometry: A quick review (ICMS)
 
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
 
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
 
BEADS : filtrage asymétrique de ligne de base (tendance) et débruitage pour d...
BEADS : filtrage asymétrique de ligne de base (tendance) et débruitage pour d...BEADS : filtrage asymétrique de ligne de base (tendance) et débruitage pour d...
BEADS : filtrage asymétrique de ligne de base (tendance) et débruitage pour d...
 
Divergence center-based clustering and their applications
Divergence center-based clustering and their applicationsDivergence center-based clustering and their applications
Divergence center-based clustering and their applications
 
A T(1)-type theorem for entangled multilinear Calderon-Zygmund operators
A T(1)-type theorem for entangled multilinear Calderon-Zygmund operatorsA T(1)-type theorem for entangled multilinear Calderon-Zygmund operators
A T(1)-type theorem for entangled multilinear Calderon-Zygmund operators
 
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
 
Classification with mixtures of curved Mahalanobis metrics
Classification with mixtures of curved Mahalanobis metricsClassification with mixtures of curved Mahalanobis metrics
Classification with mixtures of curved Mahalanobis metrics
 
Optimal L-shaped matrix reordering, aka graph's core-periphery
Optimal L-shaped matrix reordering, aka graph's core-peripheryOptimal L-shaped matrix reordering, aka graph's core-periphery
Optimal L-shaped matrix reordering, aka graph's core-periphery
 

Similar to QMC: Operator Splitting Workshop, A New (More Intuitive?) Interpretation of Inertial Algorithms - Silvia Villa, Mar 21, 2018

Diffusion Schrödinger bridges for score-based generative modeling
Diffusion Schrödinger bridges for score-based generative modelingDiffusion Schrödinger bridges for score-based generative modeling
Diffusion Schrödinger bridges for score-based generative modeling
JeremyHeng10
 
Diffusion Schrödinger bridges for score-based generative modeling
Diffusion Schrödinger bridges for score-based generative modelingDiffusion Schrödinger bridges for score-based generative modeling
Diffusion Schrödinger bridges for score-based generative modeling
JeremyHeng10
 
Bayesian inference on mixtures
Bayesian inference on mixturesBayesian inference on mixtures
Bayesian inference on mixtures
Christian Robert
 
cheb_conf_aksenov.pdf
cheb_conf_aksenov.pdfcheb_conf_aksenov.pdf
cheb_conf_aksenov.pdf
Alexey Vasyukov
 
Fast parallelizable scenario-based stochastic optimization
Fast parallelizable scenario-based stochastic optimizationFast parallelizable scenario-based stochastic optimization
Fast parallelizable scenario-based stochastic optimization
Pantelis Sopasakis
 
Solution to schrodinger equation with dirac comb potential
Solution to schrodinger equation with dirac comb potential Solution to schrodinger equation with dirac comb potential
Solution to schrodinger equation with dirac comb potential
slides
 
QMC: Transition Workshop - Applying Quasi-Monte Carlo Methods to a Stochastic...
QMC: Transition Workshop - Applying Quasi-Monte Carlo Methods to a Stochastic...QMC: Transition Workshop - Applying Quasi-Monte Carlo Methods to a Stochastic...
QMC: Transition Workshop - Applying Quasi-Monte Carlo Methods to a Stochastic...
The Statistical and Applied Mathematical Sciences Institute
 
Calculus of variations
Calculus of variationsCalculus of variations
Calculus of variations
Solo Hermelin
 
NIPS2010: optimization algorithms in machine learning
NIPS2010: optimization algorithms in machine learningNIPS2010: optimization algorithms in machine learning
NIPS2010: optimization algorithms in machine learningzukun
 
Convexity in the Theory of the Gamma Function.pdf
Convexity in the Theory of the Gamma Function.pdfConvexity in the Theory of the Gamma Function.pdf
Convexity in the Theory of the Gamma Function.pdf
PeterEsnayderBarranz
 
Hierarchical matrices for approximating large covariance matries and computin...
Hierarchical matrices for approximating large covariance matries and computin...Hierarchical matrices for approximating large covariance matries and computin...
Hierarchical matrices for approximating large covariance matries and computin...
Alexander Litvinenko
 
QMC Error SAMSI Tutorial Aug 2017
QMC Error SAMSI Tutorial Aug 2017QMC Error SAMSI Tutorial Aug 2017
QMC Error SAMSI Tutorial Aug 2017
Fred J. Hickernell
 
QMC: Operator Splitting Workshop, Forward-Backward Splitting Algorithm withou...
QMC: Operator Splitting Workshop, Forward-Backward Splitting Algorithm withou...QMC: Operator Splitting Workshop, Forward-Backward Splitting Algorithm withou...
QMC: Operator Splitting Workshop, Forward-Backward Splitting Algorithm withou...
The Statistical and Applied Mathematical Sciences Institute
 
Computer Controlled Systems (solutions manual). Astrom. 3rd edition 1997
Computer Controlled Systems (solutions manual). Astrom. 3rd edition 1997Computer Controlled Systems (solutions manual). Astrom. 3rd edition 1997
Computer Controlled Systems (solutions manual). Astrom. 3rd edition 1997
JOAQUIN REA
 
The lattice Boltzmann equation: background, boundary conditions, and Burnett-...
The lattice Boltzmann equation: background, boundary conditions, and Burnett-...The lattice Boltzmann equation: background, boundary conditions, and Burnett-...
The lattice Boltzmann equation: background, boundary conditions, and Burnett-...
Tim Reis
 
QMC: Transition Workshop - Probabilistic Integrators for Deterministic Differ...
QMC: Transition Workshop - Probabilistic Integrators for Deterministic Differ...QMC: Transition Workshop - Probabilistic Integrators for Deterministic Differ...
QMC: Transition Workshop - Probabilistic Integrators for Deterministic Differ...
The Statistical and Applied Mathematical Sciences Institute
 
Probabilistic Control of Uncertain Linear Systems Using Stochastic Reachability
Probabilistic Control of Uncertain Linear Systems Using Stochastic ReachabilityProbabilistic Control of Uncertain Linear Systems Using Stochastic Reachability
Probabilistic Control of Uncertain Linear Systems Using Stochastic Reachability
Leo Asselborn
 
Robust Control of Uncertain Switched Linear Systems based on Stochastic Reach...
Robust Control of Uncertain Switched Linear Systems based on Stochastic Reach...Robust Control of Uncertain Switched Linear Systems based on Stochastic Reach...
Robust Control of Uncertain Switched Linear Systems based on Stochastic Reach...
Leo Asselborn
 
Hiroyuki Sato
Hiroyuki SatoHiroyuki Sato
Hiroyuki Sato
Suurist
 
Harmonic Analysis and Deep Learning
Harmonic Analysis and Deep LearningHarmonic Analysis and Deep Learning
Harmonic Analysis and Deep Learning
Sungbin Lim
 

Similar to QMC: Operator Splitting Workshop, A New (More Intuitive?) Interpretation of Inertial Algorithms - Silvia Villa, Mar 21, 2018 (20)

Diffusion Schrödinger bridges for score-based generative modeling
Diffusion Schrödinger bridges for score-based generative modelingDiffusion Schrödinger bridges for score-based generative modeling
Diffusion Schrödinger bridges for score-based generative modeling
 
Diffusion Schrödinger bridges for score-based generative modeling
Diffusion Schrödinger bridges for score-based generative modelingDiffusion Schrödinger bridges for score-based generative modeling
Diffusion Schrödinger bridges for score-based generative modeling
 
Bayesian inference on mixtures
Bayesian inference on mixturesBayesian inference on mixtures
Bayesian inference on mixtures
 
cheb_conf_aksenov.pdf
cheb_conf_aksenov.pdfcheb_conf_aksenov.pdf
cheb_conf_aksenov.pdf
 
Fast parallelizable scenario-based stochastic optimization
Fast parallelizable scenario-based stochastic optimizationFast parallelizable scenario-based stochastic optimization
Fast parallelizable scenario-based stochastic optimization
 
Solution to schrodinger equation with dirac comb potential
Solution to schrodinger equation with dirac comb potential Solution to schrodinger equation with dirac comb potential
Solution to schrodinger equation with dirac comb potential
 
QMC: Transition Workshop - Applying Quasi-Monte Carlo Methods to a Stochastic...
QMC: Transition Workshop - Applying Quasi-Monte Carlo Methods to a Stochastic...QMC: Transition Workshop - Applying Quasi-Monte Carlo Methods to a Stochastic...
QMC: Transition Workshop - Applying Quasi-Monte Carlo Methods to a Stochastic...
 
Calculus of variations
Calculus of variationsCalculus of variations
Calculus of variations
 
NIPS2010: optimization algorithms in machine learning
NIPS2010: optimization algorithms in machine learningNIPS2010: optimization algorithms in machine learning
NIPS2010: optimization algorithms in machine learning
 
Convexity in the Theory of the Gamma Function.pdf
Convexity in the Theory of the Gamma Function.pdfConvexity in the Theory of the Gamma Function.pdf
Convexity in the Theory of the Gamma Function.pdf
 
Hierarchical matrices for approximating large covariance matries and computin...
Hierarchical matrices for approximating large covariance matries and computin...Hierarchical matrices for approximating large covariance matries and computin...
Hierarchical matrices for approximating large covariance matries and computin...
 
QMC Error SAMSI Tutorial Aug 2017
QMC Error SAMSI Tutorial Aug 2017QMC Error SAMSI Tutorial Aug 2017
QMC Error SAMSI Tutorial Aug 2017
 
QMC: Operator Splitting Workshop, Forward-Backward Splitting Algorithm withou...
QMC: Operator Splitting Workshop, Forward-Backward Splitting Algorithm withou...QMC: Operator Splitting Workshop, Forward-Backward Splitting Algorithm withou...
QMC: Operator Splitting Workshop, Forward-Backward Splitting Algorithm withou...
 
Computer Controlled Systems (solutions manual). Astrom. 3rd edition 1997
Computer Controlled Systems (solutions manual). Astrom. 3rd edition 1997Computer Controlled Systems (solutions manual). Astrom. 3rd edition 1997
Computer Controlled Systems (solutions manual). Astrom. 3rd edition 1997
 
The lattice Boltzmann equation: background, boundary conditions, and Burnett-...
The lattice Boltzmann equation: background, boundary conditions, and Burnett-...The lattice Boltzmann equation: background, boundary conditions, and Burnett-...
The lattice Boltzmann equation: background, boundary conditions, and Burnett-...
 
QMC: Transition Workshop - Probabilistic Integrators for Deterministic Differ...
QMC: Transition Workshop - Probabilistic Integrators for Deterministic Differ...QMC: Transition Workshop - Probabilistic Integrators for Deterministic Differ...
QMC: Transition Workshop - Probabilistic Integrators for Deterministic Differ...
 
Probabilistic Control of Uncertain Linear Systems Using Stochastic Reachability
Probabilistic Control of Uncertain Linear Systems Using Stochastic ReachabilityProbabilistic Control of Uncertain Linear Systems Using Stochastic Reachability
Probabilistic Control of Uncertain Linear Systems Using Stochastic Reachability
 
Robust Control of Uncertain Switched Linear Systems based on Stochastic Reach...
Robust Control of Uncertain Switched Linear Systems based on Stochastic Reach...Robust Control of Uncertain Switched Linear Systems based on Stochastic Reach...
Robust Control of Uncertain Switched Linear Systems based on Stochastic Reach...
 
Hiroyuki Sato
Hiroyuki SatoHiroyuki Sato
Hiroyuki Sato
 
Harmonic Analysis and Deep Learning
Harmonic Analysis and Deep LearningHarmonic Analysis and Deep Learning
Harmonic Analysis and Deep Learning
 

More from The Statistical and Applied Mathematical Sciences Institute

Causal Inference Opening Workshop - Latent Variable Models, Causal Inference,...
Causal Inference Opening Workshop - Latent Variable Models, Causal Inference,...Causal Inference Opening Workshop - Latent Variable Models, Causal Inference,...
Causal Inference Opening Workshop - Latent Variable Models, Causal Inference,...
The Statistical and Applied Mathematical Sciences Institute
 
2019 Fall Series: Special Guest Lecture - 0-1 Phase Transitions in High Dimen...
2019 Fall Series: Special Guest Lecture - 0-1 Phase Transitions in High Dimen...2019 Fall Series: Special Guest Lecture - 0-1 Phase Transitions in High Dimen...
2019 Fall Series: Special Guest Lecture - 0-1 Phase Transitions in High Dimen...
The Statistical and Applied Mathematical Sciences Institute
 
Causal Inference Opening Workshop - Causal Discovery in Neuroimaging Data - F...
Causal Inference Opening Workshop - Causal Discovery in Neuroimaging Data - F...Causal Inference Opening Workshop - Causal Discovery in Neuroimaging Data - F...
Causal Inference Opening Workshop - Causal Discovery in Neuroimaging Data - F...
The Statistical and Applied Mathematical Sciences Institute
 
Causal Inference Opening Workshop - Smooth Extensions to BART for Heterogeneo...
Causal Inference Opening Workshop - Smooth Extensions to BART for Heterogeneo...Causal Inference Opening Workshop - Smooth Extensions to BART for Heterogeneo...
Causal Inference Opening Workshop - Smooth Extensions to BART for Heterogeneo...
The Statistical and Applied Mathematical Sciences Institute
 
Causal Inference Opening Workshop - A Bracketing Relationship between Differe...
Causal Inference Opening Workshop - A Bracketing Relationship between Differe...Causal Inference Opening Workshop - A Bracketing Relationship between Differe...
Causal Inference Opening Workshop - A Bracketing Relationship between Differe...
The Statistical and Applied Mathematical Sciences Institute
 
Causal Inference Opening Workshop - Testing Weak Nulls in Matched Observation...
Causal Inference Opening Workshop - Testing Weak Nulls in Matched Observation...Causal Inference Opening Workshop - Testing Weak Nulls in Matched Observation...
Causal Inference Opening Workshop - Testing Weak Nulls in Matched Observation...
The Statistical and Applied Mathematical Sciences Institute
 
Causal Inference Opening Workshop - Difference-in-differences: more than meet...
Causal Inference Opening Workshop - Difference-in-differences: more than meet...Causal Inference Opening Workshop - Difference-in-differences: more than meet...
Causal Inference Opening Workshop - Difference-in-differences: more than meet...
The Statistical and Applied Mathematical Sciences Institute
 
Causal Inference Opening Workshop - New Statistical Learning Methods for Esti...
Causal Inference Opening Workshop - New Statistical Learning Methods for Esti...Causal Inference Opening Workshop - New Statistical Learning Methods for Esti...
Causal Inference Opening Workshop - New Statistical Learning Methods for Esti...
The Statistical and Applied Mathematical Sciences Institute
 
Causal Inference Opening Workshop - Bipartite Causal Inference with Interfere...
Causal Inference Opening Workshop - Bipartite Causal Inference with Interfere...Causal Inference Opening Workshop - Bipartite Causal Inference with Interfere...
Causal Inference Opening Workshop - Bipartite Causal Inference with Interfere...
The Statistical and Applied Mathematical Sciences Institute
 
Causal Inference Opening Workshop - Bridging the Gap Between Causal Literatur...
Causal Inference Opening Workshop - Bridging the Gap Between Causal Literatur...Causal Inference Opening Workshop - Bridging the Gap Between Causal Literatur...
Causal Inference Opening Workshop - Bridging the Gap Between Causal Literatur...
The Statistical and Applied Mathematical Sciences Institute
 
Causal Inference Opening Workshop - Some Applications of Reinforcement Learni...
Causal Inference Opening Workshop - Some Applications of Reinforcement Learni...Causal Inference Opening Workshop - Some Applications of Reinforcement Learni...
Causal Inference Opening Workshop - Some Applications of Reinforcement Learni...
The Statistical and Applied Mathematical Sciences Institute
 
Causal Inference Opening Workshop - Bracketing Bounds for Differences-in-Diff...
Causal Inference Opening Workshop - Bracketing Bounds for Differences-in-Diff...Causal Inference Opening Workshop - Bracketing Bounds for Differences-in-Diff...
Causal Inference Opening Workshop - Bracketing Bounds for Differences-in-Diff...
The Statistical and Applied Mathematical Sciences Institute
 
Causal Inference Opening Workshop - Assisting the Impact of State Polcies: Br...
Causal Inference Opening Workshop - Assisting the Impact of State Polcies: Br...Causal Inference Opening Workshop - Assisting the Impact of State Polcies: Br...
Causal Inference Opening Workshop - Assisting the Impact of State Polcies: Br...
The Statistical and Applied Mathematical Sciences Institute
 
Causal Inference Opening Workshop - Experimenting in Equilibrium - Stefan Wag...
Causal Inference Opening Workshop - Experimenting in Equilibrium - Stefan Wag...Causal Inference Opening Workshop - Experimenting in Equilibrium - Stefan Wag...
Causal Inference Opening Workshop - Experimenting in Equilibrium - Stefan Wag...
The Statistical and Applied Mathematical Sciences Institute
 
Causal Inference Opening Workshop - Targeted Learning for Causal Inference Ba...
Causal Inference Opening Workshop - Targeted Learning for Causal Inference Ba...Causal Inference Opening Workshop - Targeted Learning for Causal Inference Ba...
Causal Inference Opening Workshop - Targeted Learning for Causal Inference Ba...
The Statistical and Applied Mathematical Sciences Institute
 
Causal Inference Opening Workshop - Bayesian Nonparametric Models for Treatme...
Causal Inference Opening Workshop - Bayesian Nonparametric Models for Treatme...Causal Inference Opening Workshop - Bayesian Nonparametric Models for Treatme...
Causal Inference Opening Workshop - Bayesian Nonparametric Models for Treatme...
The Statistical and Applied Mathematical Sciences Institute
 
2019 Fall Series: Special Guest Lecture - Adversarial Risk Analysis of the Ge...
2019 Fall Series: Special Guest Lecture - Adversarial Risk Analysis of the Ge...2019 Fall Series: Special Guest Lecture - Adversarial Risk Analysis of the Ge...
2019 Fall Series: Special Guest Lecture - Adversarial Risk Analysis of the Ge...
The Statistical and Applied Mathematical Sciences Institute
 
2019 Fall Series: Professional Development, Writing Academic Papers…What Work...
2019 Fall Series: Professional Development, Writing Academic Papers…What Work...2019 Fall Series: Professional Development, Writing Academic Papers…What Work...
2019 Fall Series: Professional Development, Writing Academic Papers…What Work...
The Statistical and Applied Mathematical Sciences Institute
 
2019 GDRR: Blockchain Data Analytics - Machine Learning in/for Blockchain: Fu...
2019 GDRR: Blockchain Data Analytics - Machine Learning in/for Blockchain: Fu...2019 GDRR: Blockchain Data Analytics - Machine Learning in/for Blockchain: Fu...
2019 GDRR: Blockchain Data Analytics - Machine Learning in/for Blockchain: Fu...
The Statistical and Applied Mathematical Sciences Institute
 
2019 GDRR: Blockchain Data Analytics - QuTrack: Model Life Cycle Management f...
2019 GDRR: Blockchain Data Analytics - QuTrack: Model Life Cycle Management f...2019 GDRR: Blockchain Data Analytics - QuTrack: Model Life Cycle Management f...
2019 GDRR: Blockchain Data Analytics - QuTrack: Model Life Cycle Management f...
The Statistical and Applied Mathematical Sciences Institute
 

More from The Statistical and Applied Mathematical Sciences Institute (20)

Causal Inference Opening Workshop - Latent Variable Models, Causal Inference,...
Causal Inference Opening Workshop - Latent Variable Models, Causal Inference,...Causal Inference Opening Workshop - Latent Variable Models, Causal Inference,...
Causal Inference Opening Workshop - Latent Variable Models, Causal Inference,...
 
2019 Fall Series: Special Guest Lecture - 0-1 Phase Transitions in High Dimen...
2019 Fall Series: Special Guest Lecture - 0-1 Phase Transitions in High Dimen...2019 Fall Series: Special Guest Lecture - 0-1 Phase Transitions in High Dimen...
2019 Fall Series: Special Guest Lecture - 0-1 Phase Transitions in High Dimen...
 
Causal Inference Opening Workshop - Causal Discovery in Neuroimaging Data - F...
Causal Inference Opening Workshop - Causal Discovery in Neuroimaging Data - F...Causal Inference Opening Workshop - Causal Discovery in Neuroimaging Data - F...
Causal Inference Opening Workshop - Causal Discovery in Neuroimaging Data - F...
 
Causal Inference Opening Workshop - Smooth Extensions to BART for Heterogeneo...
Causal Inference Opening Workshop - Smooth Extensions to BART for Heterogeneo...Causal Inference Opening Workshop - Smooth Extensions to BART for Heterogeneo...
Causal Inference Opening Workshop - Smooth Extensions to BART for Heterogeneo...
 
Causal Inference Opening Workshop - A Bracketing Relationship between Differe...
Causal Inference Opening Workshop - A Bracketing Relationship between Differe...Causal Inference Opening Workshop - A Bracketing Relationship between Differe...
Causal Inference Opening Workshop - A Bracketing Relationship between Differe...
 
Causal Inference Opening Workshop - Testing Weak Nulls in Matched Observation...
Causal Inference Opening Workshop - Testing Weak Nulls in Matched Observation...Causal Inference Opening Workshop - Testing Weak Nulls in Matched Observation...
Causal Inference Opening Workshop - Testing Weak Nulls in Matched Observation...
 
Causal Inference Opening Workshop - Difference-in-differences: more than meet...
Causal Inference Opening Workshop - Difference-in-differences: more than meet...Causal Inference Opening Workshop - Difference-in-differences: more than meet...
Causal Inference Opening Workshop - Difference-in-differences: more than meet...
 
Causal Inference Opening Workshop - New Statistical Learning Methods for Esti...
Causal Inference Opening Workshop - New Statistical Learning Methods for Esti...Causal Inference Opening Workshop - New Statistical Learning Methods for Esti...
Causal Inference Opening Workshop - New Statistical Learning Methods for Esti...
 
Causal Inference Opening Workshop - Bipartite Causal Inference with Interfere...
Causal Inference Opening Workshop - Bipartite Causal Inference with Interfere...Causal Inference Opening Workshop - Bipartite Causal Inference with Interfere...
Causal Inference Opening Workshop - Bipartite Causal Inference with Interfere...
 
Causal Inference Opening Workshop - Bridging the Gap Between Causal Literatur...
Causal Inference Opening Workshop - Bridging the Gap Between Causal Literatur...Causal Inference Opening Workshop - Bridging the Gap Between Causal Literatur...
Causal Inference Opening Workshop - Bridging the Gap Between Causal Literatur...
 
Causal Inference Opening Workshop - Some Applications of Reinforcement Learni...
Causal Inference Opening Workshop - Some Applications of Reinforcement Learni...Causal Inference Opening Workshop - Some Applications of Reinforcement Learni...
Causal Inference Opening Workshop - Some Applications of Reinforcement Learni...
 
Causal Inference Opening Workshop - Bracketing Bounds for Differences-in-Diff...
Causal Inference Opening Workshop - Bracketing Bounds for Differences-in-Diff...Causal Inference Opening Workshop - Bracketing Bounds for Differences-in-Diff...
Causal Inference Opening Workshop - Bracketing Bounds for Differences-in-Diff...
 
Causal Inference Opening Workshop - Assisting the Impact of State Polcies: Br...
Causal Inference Opening Workshop - Assisting the Impact of State Polcies: Br...Causal Inference Opening Workshop - Assisting the Impact of State Polcies: Br...
Causal Inference Opening Workshop - Assisting the Impact of State Polcies: Br...
 
Causal Inference Opening Workshop - Experimenting in Equilibrium - Stefan Wag...
Causal Inference Opening Workshop - Experimenting in Equilibrium - Stefan Wag...Causal Inference Opening Workshop - Experimenting in Equilibrium - Stefan Wag...
Causal Inference Opening Workshop - Experimenting in Equilibrium - Stefan Wag...
 
Causal Inference Opening Workshop - Targeted Learning for Causal Inference Ba...
Causal Inference Opening Workshop - Targeted Learning for Causal Inference Ba...Causal Inference Opening Workshop - Targeted Learning for Causal Inference Ba...
Causal Inference Opening Workshop - Targeted Learning for Causal Inference Ba...
 
Causal Inference Opening Workshop - Bayesian Nonparametric Models for Treatme...
Causal Inference Opening Workshop - Bayesian Nonparametric Models for Treatme...Causal Inference Opening Workshop - Bayesian Nonparametric Models for Treatme...
Causal Inference Opening Workshop - Bayesian Nonparametric Models for Treatme...
 
2019 Fall Series: Special Guest Lecture - Adversarial Risk Analysis of the Ge...
2019 Fall Series: Special Guest Lecture - Adversarial Risk Analysis of the Ge...2019 Fall Series: Special Guest Lecture - Adversarial Risk Analysis of the Ge...
2019 Fall Series: Special Guest Lecture - Adversarial Risk Analysis of the Ge...
 
2019 Fall Series: Professional Development, Writing Academic Papers…What Work...
2019 Fall Series: Professional Development, Writing Academic Papers…What Work...2019 Fall Series: Professional Development, Writing Academic Papers…What Work...
2019 Fall Series: Professional Development, Writing Academic Papers…What Work...
 
2019 GDRR: Blockchain Data Analytics - Machine Learning in/for Blockchain: Fu...
2019 GDRR: Blockchain Data Analytics - Machine Learning in/for Blockchain: Fu...2019 GDRR: Blockchain Data Analytics - Machine Learning in/for Blockchain: Fu...
2019 GDRR: Blockchain Data Analytics - Machine Learning in/for Blockchain: Fu...
 
2019 GDRR: Blockchain Data Analytics - QuTrack: Model Life Cycle Management f...
2019 GDRR: Blockchain Data Analytics - QuTrack: Model Life Cycle Management f...2019 GDRR: Blockchain Data Analytics - QuTrack: Model Life Cycle Management f...
2019 GDRR: Blockchain Data Analytics - QuTrack: Model Life Cycle Management f...
 

Recently uploaded

aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa
aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa
aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa
siemaillard
 
The approach at University of Liverpool.pptx
The approach at University of Liverpool.pptxThe approach at University of Liverpool.pptx
The approach at University of Liverpool.pptx
Jisc
 
Additional Benefits for Employee Website.pdf
Additional Benefits for Employee Website.pdfAdditional Benefits for Employee Website.pdf
Additional Benefits for Employee Website.pdf
joachimlavalley1
 
Digital Tools and AI for Teaching Learning and Research
Digital Tools and AI for Teaching Learning and ResearchDigital Tools and AI for Teaching Learning and Research
Digital Tools and AI for Teaching Learning and Research
Vikramjit Singh
 
Home assignment II on Spectroscopy 2024 Answers.pdf
Home assignment II on Spectroscopy 2024 Answers.pdfHome assignment II on Spectroscopy 2024 Answers.pdf
Home assignment II on Spectroscopy 2024 Answers.pdf
Tamralipta Mahavidyalaya
 
2024.06.01 Introducing a competency framework for languag learning materials ...
2024.06.01 Introducing a competency framework for languag learning materials ...2024.06.01 Introducing a competency framework for languag learning materials ...
2024.06.01 Introducing a competency framework for languag learning materials ...
Sandy Millin
 
Phrasal Verbs.XXXXXXXXXXXXXXXXXXXXXXXXXX
Phrasal Verbs.XXXXXXXXXXXXXXXXXXXXXXXXXXPhrasal Verbs.XXXXXXXXXXXXXXXXXXXXXXXXXX
Phrasal Verbs.XXXXXXXXXXXXXXXXXXXXXXXXXX
MIRIAMSALINAS13
 
How to Create Map Views in the Odoo 17 ERP
How to Create Map Views in the Odoo 17 ERPHow to Create Map Views in the Odoo 17 ERP
How to Create Map Views in the Odoo 17 ERP
Celine George
 
Ethnobotany and Ethnopharmacology ......
Ethnobotany and Ethnopharmacology ......Ethnobotany and Ethnopharmacology ......
Ethnobotany and Ethnopharmacology ......
Ashokrao Mane college of Pharmacy Peth-Vadgaon
 
Mule 4.6 & Java 17 Upgrade | MuleSoft Mysore Meetup #46
Mule 4.6 & Java 17 Upgrade | MuleSoft Mysore Meetup #46Mule 4.6 & Java 17 Upgrade | MuleSoft Mysore Meetup #46
Mule 4.6 & Java 17 Upgrade | MuleSoft Mysore Meetup #46
MysoreMuleSoftMeetup
 
Chapter 3 - Islamic Banking Products and Services.pptx
Chapter 3 - Islamic Banking Products and Services.pptxChapter 3 - Islamic Banking Products and Services.pptx
Chapter 3 - Islamic Banking Products and Services.pptx
Mohd Adib Abd Muin, Senior Lecturer at Universiti Utara Malaysia
 
Thesis Statement for students diagnonsed withADHD.ppt
Thesis Statement for students diagnonsed withADHD.pptThesis Statement for students diagnonsed withADHD.ppt
Thesis Statement for students diagnonsed withADHD.ppt
EverAndrsGuerraGuerr
 
Sha'Carri Richardson Presentation 202345
Sha'Carri Richardson Presentation 202345Sha'Carri Richardson Presentation 202345
Sha'Carri Richardson Presentation 202345
beazzy04
 
Overview on Edible Vaccine: Pros & Cons with Mechanism
Overview on Edible Vaccine: Pros & Cons with MechanismOverview on Edible Vaccine: Pros & Cons with Mechanism
Overview on Edible Vaccine: Pros & Cons with Mechanism
DeeptiGupta154
 
CLASS 11 CBSE B.St Project AIDS TO TRADE - INSURANCE
CLASS 11 CBSE B.St Project AIDS TO TRADE - INSURANCECLASS 11 CBSE B.St Project AIDS TO TRADE - INSURANCE
CLASS 11 CBSE B.St Project AIDS TO TRADE - INSURANCE
BhavyaRajput3
 
Unit 8 - Information and Communication Technology (Paper I).pdf
Unit 8 - Information and Communication Technology (Paper I).pdfUnit 8 - Information and Communication Technology (Paper I).pdf
Unit 8 - Information and Communication Technology (Paper I).pdf
Thiyagu K
 
Welcome to TechSoup New Member Orientation and Q&A (May 2024).pdf
Welcome to TechSoup   New Member Orientation and Q&A (May 2024).pdfWelcome to TechSoup   New Member Orientation and Q&A (May 2024).pdf
Welcome to TechSoup New Member Orientation and Q&A (May 2024).pdf
TechSoup
 
How to Make a Field invisible in Odoo 17
How to Make a Field invisible in Odoo 17How to Make a Field invisible in Odoo 17
How to Make a Field invisible in Odoo 17
Celine George
 
The French Revolution Class 9 Study Material pdf free download
The French Revolution Class 9 Study Material pdf free downloadThe French Revolution Class 9 Study Material pdf free download
The French Revolution Class 9 Study Material pdf free download
Vivekanand Anglo Vedic Academy
 
Sectors of the Indian Economy - Class 10 Study Notes pdf
Sectors of the Indian Economy - Class 10 Study Notes pdfSectors of the Indian Economy - Class 10 Study Notes pdf
Sectors of the Indian Economy - Class 10 Study Notes pdf
Vivekanand Anglo Vedic Academy
 

Recently uploaded (20)

aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa
aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa
aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa
 
The approach at University of Liverpool.pptx
The approach at University of Liverpool.pptxThe approach at University of Liverpool.pptx
The approach at University of Liverpool.pptx
 
Additional Benefits for Employee Website.pdf
Additional Benefits for Employee Website.pdfAdditional Benefits for Employee Website.pdf
Additional Benefits for Employee Website.pdf
 
Digital Tools and AI for Teaching Learning and Research
Digital Tools and AI for Teaching Learning and ResearchDigital Tools and AI for Teaching Learning and Research
Digital Tools and AI for Teaching Learning and Research
 
Home assignment II on Spectroscopy 2024 Answers.pdf
Home assignment II on Spectroscopy 2024 Answers.pdfHome assignment II on Spectroscopy 2024 Answers.pdf
Home assignment II on Spectroscopy 2024 Answers.pdf
 
2024.06.01 Introducing a competency framework for languag learning materials ...
2024.06.01 Introducing a competency framework for languag learning materials ...2024.06.01 Introducing a competency framework for languag learning materials ...
2024.06.01 Introducing a competency framework for languag learning materials ...
 
Phrasal Verbs.XXXXXXXXXXXXXXXXXXXXXXXXXX
Phrasal Verbs.XXXXXXXXXXXXXXXXXXXXXXXXXXPhrasal Verbs.XXXXXXXXXXXXXXXXXXXXXXXXXX
Phrasal Verbs.XXXXXXXXXXXXXXXXXXXXXXXXXX
 
How to Create Map Views in the Odoo 17 ERP
How to Create Map Views in the Odoo 17 ERPHow to Create Map Views in the Odoo 17 ERP
How to Create Map Views in the Odoo 17 ERP
 
Ethnobotany and Ethnopharmacology ......
Ethnobotany and Ethnopharmacology ......Ethnobotany and Ethnopharmacology ......
Ethnobotany and Ethnopharmacology ......
 
Mule 4.6 & Java 17 Upgrade | MuleSoft Mysore Meetup #46
Mule 4.6 & Java 17 Upgrade | MuleSoft Mysore Meetup #46Mule 4.6 & Java 17 Upgrade | MuleSoft Mysore Meetup #46
Mule 4.6 & Java 17 Upgrade | MuleSoft Mysore Meetup #46
 
Chapter 3 - Islamic Banking Products and Services.pptx
Chapter 3 - Islamic Banking Products and Services.pptxChapter 3 - Islamic Banking Products and Services.pptx
Chapter 3 - Islamic Banking Products and Services.pptx
 
Thesis Statement for students diagnonsed withADHD.ppt
Thesis Statement for students diagnonsed withADHD.pptThesis Statement for students diagnonsed withADHD.ppt
Thesis Statement for students diagnonsed withADHD.ppt
 
Sha'Carri Richardson Presentation 202345
Sha'Carri Richardson Presentation 202345Sha'Carri Richardson Presentation 202345
Sha'Carri Richardson Presentation 202345
 
Overview on Edible Vaccine: Pros & Cons with Mechanism
Overview on Edible Vaccine: Pros & Cons with MechanismOverview on Edible Vaccine: Pros & Cons with Mechanism
Overview on Edible Vaccine: Pros & Cons with Mechanism
 
CLASS 11 CBSE B.St Project AIDS TO TRADE - INSURANCE
CLASS 11 CBSE B.St Project AIDS TO TRADE - INSURANCECLASS 11 CBSE B.St Project AIDS TO TRADE - INSURANCE
CLASS 11 CBSE B.St Project AIDS TO TRADE - INSURANCE
 
Unit 8 - Information and Communication Technology (Paper I).pdf
Unit 8 - Information and Communication Technology (Paper I).pdfUnit 8 - Information and Communication Technology (Paper I).pdf
Unit 8 - Information and Communication Technology (Paper I).pdf
 
Welcome to TechSoup New Member Orientation and Q&A (May 2024).pdf
Welcome to TechSoup   New Member Orientation and Q&A (May 2024).pdfWelcome to TechSoup   New Member Orientation and Q&A (May 2024).pdf
Welcome to TechSoup New Member Orientation and Q&A (May 2024).pdf
 
How to Make a Field invisible in Odoo 17
How to Make a Field invisible in Odoo 17How to Make a Field invisible in Odoo 17
How to Make a Field invisible in Odoo 17
 
The French Revolution Class 9 Study Material pdf free download
The French Revolution Class 9 Study Material pdf free downloadThe French Revolution Class 9 Study Material pdf free download
The French Revolution Class 9 Study Material pdf free download
 
Sectors of the Indian Economy - Class 10 Study Notes pdf
Sectors of the Indian Economy - Class 10 Study Notes pdfSectors of the Indian Economy - Class 10 Study Notes pdf
Sectors of the Indian Economy - Class 10 Study Notes pdf
 

QMC: Operator Splitting Workshop, A New (More Intuitive?) Interpretation of Inertial Algorithms - Silvia Villa, Mar 21, 2018

  • 1. A new (more intuitive?) interpretation of inertial algorithms Silvia Villa Politecnico di Milano https://www.mate.polimi.it/analysis/?settore=people SAMSI Workshop: Operator Splitting Methods in Data Analysis Raleigh, March 21th, 2018 S. Villa (Polimi) 1 / 20
  • 2. Introduction Problem setting Let H be a Hilbert space f : H → R convex, differentiable, with an L-Lipschitz continuous gradient Assume that argmin f = ∅. We consider the problem of computing min x∈H f (x) S. Villa (Polimi) 2 / 20
  • 3. Introduction Gradient method Given x0 ∈ H and γ ∈ ]0, 2/L[ For k ≥ 0 define xk+1 = xk − γ f (xk) Classic convergence results: f (xk) − min f = O(1/k) Convergence of the iterates: xk ¯x ∈ argmin f S. Villa (Polimi) 3 / 20
  • 4. Introduction Accelerated/Inertial gradient method Given x0 = y0 ∈ H and γ ∈ ]0, 1/L], αk ∈ ]0, 1[: For k ≥ 0, define xk+1 = yk − γ f (yk) yk+1 = xk+1 + αk(xk+1 − xk) Convergence results: f (xk) − min f = O(1/k2) S. Villa (Polimi) 4 / 20
  • 5. Introduction Accelerated/Inertial gradient method Given x0 = y0 ∈ H and γ ∈ ]0, 1/L], αk ∈ ]0, 1[: For k ≥ 0, define xk+1 = yk − γ f (yk) yk+1 = xk+1 + αk(xk+1 − xk) Convergence results: f (xk) − min f = O(1/k2) xk ¯x ∈ argmin f S. Villa (Polimi) 4 / 20
  • 6. Introduction Convergence results and remarks Convergence and its rate depend on the choice of αk S. Villa (Polimi) 5 / 20
  • 7. Introduction Convergence results and remarks Convergence and its rate depend on the choice of αk Nesterov rule for choosing αk [Nesterov, A method for solving a convex programming problem with rate of convergence O(1/k2 ), 1983] S. Villa (Polimi) 5 / 20
  • 8. Introduction Convergence results and remarks Convergence and its rate depend on the choice of αk Nesterov rule for choosing αk [Nesterov, A method for solving a convex programming problem with rate of convergence O(1/k2 ), 1983] Extension to proximal point algorithm [G¨uler, New proximal point algorithms for convex minimization, 1992] and to forward-backward [Beck-Teboulle, FISTA, 2009] S. Villa (Polimi) 5 / 20
  • 9. Introduction Convergence results and remarks Convergence and its rate depend on the choice of αk Nesterov rule for choosing αk [Nesterov, A method for solving a convex programming problem with rate of convergence O(1/k2 ), 1983] Extension to proximal point algorithm [G¨uler, New proximal point algorithms for convex minimization, 1992] and to forward-backward [Beck-Teboulle, FISTA, 2009] Other choices for αk [Chambolle-Dossal, 2014],[Attouch-Peypouquet-Redont 2014], [Apidopoulus-Aujol-Dossal 2017] αk = 1 − α k + 2 For convergence: α > 3. S. Villa (Polimi) 5 / 20
  • 10. Introduction In the rest of the talk Brief review of some approaches to show convergence “Algebraic proof” [Nesterov; G¨uler; Beck-Teboulle; Chambolle-Dossal] Estimate sequences [Nesterov; G¨uler; Salzo-Villa] Primal and mirror descent combination [Zhu-Orecchia, Linear coupling: An ultimate unification of gradient and mirror descent, 2017 ODE approach [Su-Boyd-Candes; Apidopoulos-Aujol-Dossal; Attouch-Cabot-Peypouquet-Redont-Chbani...] S. Villa (Polimi) 6 / 20
  • 11. Introduction In the rest of the talk Brief review of some approaches to show convergence “Algebraic proof” [Nesterov; G¨uler; Beck-Teboulle; Chambolle-Dossal] Estimate sequences [Nesterov; G¨uler; Salzo-Villa] Primal and mirror descent combination [Zhu-Orecchia, Linear coupling: An ultimate unification of gradient and mirror descent, 2017 ODE approach [Su-Boyd-Candes; Apidopoulos-Aujol-Dossal; Attouch-Cabot-Peypouquet-Redont-Chbani...] Many results...but something is still missing :-) S. Villa (Polimi) 6 / 20
  • 12. Introduction In the rest of the talk Brief review of some approaches to show convergence “Algebraic proof” [Nesterov; G¨uler; Beck-Teboulle; Chambolle-Dossal] Estimate sequences [Nesterov; G¨uler; Salzo-Villa] Primal and mirror descent combination [Zhu-Orecchia, Linear coupling: An ultimate unification of gradient and mirror descent, 2017 ODE approach [Su-Boyd-Candes; Apidopoulos-Aujol-Dossal; Attouch-Cabot-Peypouquet-Redont-Chbani...] Many results...but something is still missing :-) S. Villa (Polimi) 6 / 20
  • 13. Classic approaches “Algebraic proof” - Nesterov’s choice of αk αk = (tk − 1) tk+1 For an arbitrary y ∈ H, it holds: 2 L (t2 k+1 − tk+1)(f (xk) − min f ) + tk+1y − (tk+1 − 1)xk − ¯x 2 ≥ 2 L t2 k+1(f (xk+1) − min f ) + tk+1xk+1 − (tk+1 − 1)xk − ¯x 2 If t2 k ≥ t2 k+1 − tk+1 and tk+1y = tk+1xk + (tk − 1)(xk − xk−1) S. Villa (Polimi) 7 / 20
  • 14. Classic approaches then 2 L t2 k (f (xk) − min f ) + tkxk − (tk − 1)xk−1 − ¯x 2 ≥ 2 L t2 k+1(f (xk+1) − min f ) + tk+1xk+1 − (tk+1 − 1)xk − ¯x 2 At the end, there exists c > 0 such that f (xk) − min f ≤ c t2 k S. Villa (Polimi) 8 / 20
  • 15. Classic approaches then 2 L t2 k (f (xk) − min f ) + tkxk − (tk − 1)xk−1 − ¯x 2 ≥ 2 L t2 k+1(f (xk+1) − min f ) + tk+1xk+1 − (tk+1 − 1)xk − ¯x 2 At the end, there exists c > 0 such that f (xk) − min f ≤ c t2 k =⇒ f (xk) − min f ≤ c (k + 1)2 S. Villa (Polimi) 8 / 20
  • 16. Classic approaches then 2 L t2 k (f (xk) − min f ) + tkxk − (tk − 1)xk−1 − ¯x 2 ≥ 2 L t2 k+1(f (xk+1) − min f ) + tk+1xk+1 − (tk+1 − 1)xk − ¯x 2 At the end, there exists c > 0 such that f (xk) − min f ≤ c t2 k =⇒ f (xk) − min f ≤ c (k + 1)2 **SIMILAR** proof for convergence of (xk) ([Chambolle-Dossal, On the weak convergence of the iterates of “FISTA”, 2014]) S. Villa (Polimi) 8 / 20
  • 17. Classic approaches Nesterov’s estimate sequences Definition A pair of sequences ϕk : X → R and βk ≥ 0 is called an estimate sequence of the function f if for any x ∈ X and all k ≥ 0 we have ϕk(x) − f (x) ≤ βk(ϕ0(x) − f (x)) and βk → 0. S. Villa (Polimi) 9 / 20
  • 18. Classic approaches Nesterov’s estimate sequences Definition A pair of sequences ϕk : X → R and βk ≥ 0 is called an estimate sequence of the function f if for any x ∈ X and all k ≥ 0 we have ϕk(x) − f (x) ≤ βk(ϕ0(x) − f (x)) and βk → 0. Theorem (Nesterov ’83) Let (ϕk, βk) be an estimate sequence of f and let min f = f (¯x). If for some xk ∈ X we have f (xk) ≤ min ϕk then f (xk) − min f ≤ βk(ϕ0(¯x) − f (¯x)) βk gives a convergence rate for f (xk) − min f S. Villa (Polimi) 9 / 20
  • 19. Classic approaches A method to build estimate sequences Given a quadratic function ϕ0 = f (u0) + A 2 · −u0 2, tk > 1, xk+1 , we define recursively ϕk+1(x) := (1 − t−1 k )ϕk(x) + t−1 k (f (yk+1) + x − yk+1, f (yk+1) ≤f (x), linear term ). We have: ϕk = ¯ϕk + Ak 2 x − uk 2, is a quadratic function for every k ϕk+1(x) − f (x) ≤ (1 − t−1 k )(ϕk(x) − f (x)) ≤ k i=0 (1 − t−1 i ) =βk+1 (ϕ0(x) − f (x)). S. Villa (Polimi) 10 / 20
  • 20. Classic approaches Choice of tk and yk The parameters ¯ϕk, Ak, uk can be updated recursively and depend on the choice of tk and yk. Lemma Suppose that xk is such that ¯ϕk ≥ f (xk). Set yk = (1 − t−1 k )xk + t−1 k uk, t2 k = t2 k+1 − tk+1, and xk+1 = yk − γ f (yk). Then ¯ϕk+1 ≥ f (xk+1) S. Villa (Polimi) 11 / 20
  • 21. More recent approaches Linear coupling of gradient and mirror steps Initialize x0 = y0 = u0.    yk+1 = (1 − t−1 k )xk + t−1 k uk Linear coupling xk+1 = yk+1 − 1 L f (yk+1) Gradient step uk+1 = uk − αk f (xk+1) Mirror step S. Villa (Polimi) 12 / 20
  • 22. More recent approaches Linear coupling of gradient and mirror steps Initialize x0 = y0 = u0.    yk+1 = (1 − t−1 k )xk + t−1 k uk Linear coupling xk+1 = yk+1 − 1 L f (yk+1) Gradient step uk+1 = uk − αk f (xk+1) Mirror step Link with estimate sequences? S. Villa (Polimi) 12 / 20
  • 23. More recent approaches Linear coupling of gradient and mirror steps Initialize x0 = y0 = u0.    yk+1 = (1 − t−1 k )xk + t−1 k uk Linear coupling xk+1 = yk+1 − 1 L f (yk+1) Gradient step uk+1 = uk − αk f (xk+1) Mirror step Link with estimate sequences? Bregman estimate sequences version? S. Villa (Polimi) 12 / 20
  • 24. More recent approaches ODE approach The convergence properties of the algorithm are “derived” from the asymptotic behavior, as t → +∞, of the evolution system    ¨x(t) + α t ˙x(t) + f (x(t)) = 0 x(t0) = x0 ˙x(t0) = v0. A finite difference discretization, with tk = t0 + kh, gives 1 h2 (xk+1 − 2xk + xk−1) + α kh2 (xk − xk−1) + f (yk) = 0, with yk to be determined later. S. Villa (Polimi) 13 / 20
  • 25. More recent approaches ODE approach We **NATURALLY** obtain xk+1 = xk + 1 − α k (xk − xk−1) − h2 f (yk), Setting γ = h2: yk = xk + 1 − α k (xk − xk−1) xk+1 = yk − γ f (yk), S. Villa (Polimi) 14 / 20
  • 26. More recent approaches The heavy ball Choosing yk = xk we obtain the “heavy ball method” xk+1 = xk + 1 − α k (xk − xk−1) − γ f (xk), The continuous dynamic is the same, but the convergence behavior of the algorithm is different S. Villa (Polimi) 15 / 20
  • 27. More recent approaches The problem What is the difference between the heavy ball and the inertial algorithms? S. Villa (Polimi) 16 / 20
  • 28. More recent approaches The problem What is the difference between the heavy ball and the inertial algorithms? Why choosing yk in such a way accelerate the gradient method? S. Villa (Polimi) 16 / 20
  • 29. New directions Another point of view: link with numerical analysis [Scieur-Roulet-Bach-D’Aspremont, Integration methods and Accelerated Optimization Algorithms, 2017] Use numerical integration schemes to approximate the solution of the problem: ˙x(t) + f (x(t)) = 0 x(t0) = x0 Linear two-steps schemes Fix h > 0. Given x0, x1 ∈ H, for k ≥ 0: xk+1 = −ρ1xk − ρ0xk−1 + γ(σ0 f (xk−1) + σ1 f (xk)). Restrictions on ρ0, ρ1, σ0, σ1 in R to guarantee convergence, i.e. lim γ→0 xk − x(t0 + kγ) = 0, ∀k ∈ [1, K] under appropriate initial conditions x0, x1. S. Villa (Polimi) 17 / 20
  • 30. New directions Quadratic strongly convex case Let H = Rn, f (x) = Ax, with A symmetric positive definite. Then the inertial coefficient can be chosen constant, equal to α. In Polyak’s method: xk+1 = xk + (1 − α)(xk − xk−1) − γ f (xk), The heavy ball method is a multi-step method Recall xk+1 = −ρ1xk − ρ0xk−1 + γ(σ0 f (xk−1) + σ1 f (xk)). Then, it is enough to choose: ρ0 = 1 − α ρ1 = 2 − α σ0 = 0 σ1 = −1 S. Villa (Polimi) 18 / 20
  • 31. New directions Quadratic strongly convex case Let H = Rn, f (x) = Ax, with A symmetric positive definite. Then the inertial coefficient in the accelerated algorithm can be chosen constant, equal to α. yk = xk + (1 − α) (xk − xk−1) xk+1 = yk − γ f (yk), Accelerated method Nesterov’s method is a multi-step method ρ0 = 1 − α ρ1 = α − 2 σ0 = 1 − α σ1 = α − 1 S. Villa (Polimi) 19 / 20
  • 32. New directions Quadratic strongly convex case Let H = Rn, f (x) = Ax, with A symmetric positive definite. Then the inertial coefficient in the accelerated algorithm can be chosen constant, equal to α. yk = xk + (1 − α) (xk − xk−1) xk+1 = yk − γ f (yk), Accelerated method Nesterov’s method is a multi-step method ρ0 = 1 − α ρ1 = α − 2 σ0 = 1 − α σ1 = α − 1 It suffices to write yk+1 = yk − γ f (yk) + (1 − α) (yk − γ f (yk) − yk−1 + γ f (yk−1)) S. Villa (Polimi) 19 / 20
  • 33. New directions Conclusions/Open problems The Polyak method turns out to be “optimal” for quadratic strongly convex functions S. Villa (Polimi) 20 / 20
  • 34. New directions Conclusions/Open problems The Polyak method turns out to be “optimal” for quadratic strongly convex functions What about the general case? The coefficients are not constant, but depend on k if f is only convex. S. Villa (Polimi) 20 / 20
  • 35. New directions Conclusions/Open problems The Polyak method turns out to be “optimal” for quadratic strongly convex functions What about the general case? The coefficients are not constant, but depend on k if f is only convex. Better choices of αk? Maybe driven by the geometry of the function? S. Villa (Polimi) 20 / 20
  • 36. New directions Conclusions/Open problems The Polyak method turns out to be “optimal” for quadratic strongly convex functions What about the general case? The coefficients are not constant, but depend on k if f is only convex. Better choices of αk? Maybe driven by the geometry of the function? Can this framework encompass restarting techniques? [O’Donoghue-Cand`es, Adaptive restart for accelerated gradient schemes, Found. Comput. Math., 2013.] S. Villa (Polimi) 20 / 20
  • 37. New directions Conclusions/Open problems The Polyak method turns out to be “optimal” for quadratic strongly convex functions What about the general case? The coefficients are not constant, but depend on k if f is only convex. Better choices of αk? Maybe driven by the geometry of the function? Can this framework encompass restarting techniques? [O’Donoghue-Cand`es, Adaptive restart for accelerated gradient schemes, Found. Comput. Math., 2013.] New algorithms? New integration methods? S. Villa (Polimi) 20 / 20