SlideShare a Scribd company logo
Convex Optimization — Boyd & Vandenberghe
5. Duality
• Lagrange dual problem
• weak and strong duality
• geometric interpretation
• optimality conditions
• perturbation and sensitivity analysis
• examples
• generalized inequalities
5–1
Lagrangian
standard form problem (not necessarily convex)
minimize f0(x)
subject to fi(x) ≤ 0, i = 1, . . . , m
hi(x) = 0, i = 1, . . . , p
variable x ∈ Rn
, domain D, optimal value p⋆
Lagrangian: L : Rn
× Rm
× Rp
→ R, with dom L = D × Rm
× Rp
,
L(x, λ, ν) = f0(x) +
m
X
i=1
λifi(x) +
p
X
i=1
νihi(x)
• weighted sum of objective and constraint functions
• λi is Lagrange multiplier associated with fi(x) ≤ 0
• νi is Lagrange multiplier associated with hi(x) = 0
Duality 5–2
Lagrange dual function
Lagrange dual function: g : Rm
× Rp
→ R,
g(λ, ν) = inf
x∈D
L(x, λ, ν)
= inf
x∈D
f0(x) +
m
X
i=1
λifi(x) +
p
X
i=1
νihi(x)
!
g is concave, can be −∞ for some λ, ν
lower bound property: if λ  0, then g(λ, ν) ≤ p⋆
proof: if x̃ is feasible and λ  0, then
f0(x̃) ≥ L(x̃, λ, ν) ≥ inf
x∈D
L(x, λ, ν) = g(λ, ν)
minimizing over all feasible x̃ gives p⋆
≥ g(λ, ν)
Duality 5–3
Least-norm solution of linear equations
minimize xT
x
subject to Ax = b
dual function
• Lagrangian is L(x, ν) = xT
x + νT
(Ax − b)
• to minimize L over x, set gradient equal to zero:
∇xL(x, ν) = 2x + AT
ν = 0 =⇒ x = −(1/2)AT
ν
• plug in in L to obtain g:
g(ν) = L((−1/2)AT
ν, ν) = −
1
4
νT
AAT
ν − bT
ν
a concave function of ν
lower bound property: p⋆
≥ −(1/4)νT
AAT
ν − bT
ν for all ν
Duality 5–4
Standard form LP
minimize cT
x
subject to Ax = b, x  0
dual function
• Lagrangian is
L(x, λ, ν) = cT
x + νT
(Ax − b) − λT
x
= −bT
ν + (c + AT
ν − λ)T
x
• L is affine in x, hence
g(λ, ν) = inf
x
L(x, λ, ν) =

−bT
ν AT
ν − λ + c = 0
−∞ otherwise
g is linear on affine domain {(λ, ν) | AT
ν − λ + c = 0}, hence concave
lower bound property: p⋆
≥ −bT
ν if AT
ν + c  0
Duality 5–5
Equality constrained norm minimization
minimize kxk
subject to Ax = b
dual function
g(ν) = inf
x
(kxk − νT
Ax + bT
ν) =

bT
ν kAT
νk∗ ≤ 1
−∞ otherwise
where kvk∗ = supkuk≤1 uT
v is dual norm of k · k
proof: follows from infx(kxk − yT
x) = 0 if kyk∗ ≤ 1, −∞ otherwise
• if kyk∗ ≤ 1, then kxk − yT
x ≥ 0 for all x, with equality if x = 0
• if kyk∗  1, choose x = tu where kuk ≤ 1, uT
y = kyk∗  1:
kxk − yT
x = t(kuk − kyk∗) → −∞ as t → ∞
lower bound property: p⋆
≥ bT
ν if kAT
νk∗ ≤ 1
Duality 5–6
Two-way partitioning
minimize xT
Wx
subject to x2
i = 1, i = 1, . . . , n
• a nonconvex problem; feasible set contains 2n
discrete points
• interpretation: partition {1, . . . , n} in two sets; Wij is cost of assigning
i, j to the same set; −Wij is cost of assigning to different sets
dual function
g(ν) = inf
x
(xT
Wx +
X
i
νi(x2
i − 1)) = inf
x
xT
(W + diag(ν))x − 1T
ν
=

−1T
ν W + diag(ν)  0
−∞ otherwise
lower bound property: p⋆
≥ −1T
ν if W + diag(ν)  0
example: ν = −λmin(W)1 gives bound p⋆
≥ nλmin(W)
Duality 5–7
Lagrange dual and conjugate function
minimize f0(x)
subject to Ax  b, Cx = d
dual function
g(λ, ν) = inf
x∈dom f0
f0(x) + (AT
λ + CT
ν)T
x − bT
λ − dT
ν

= −f∗
0 (−AT
λ − CT
ν) − bT
λ − dT
ν
• recall definition of conjugate f∗
(y) = supx∈dom f(yT
x − f(x))
• simplifies derivation of dual if conjugate of f0 is known
example: entropy maximization
f0(x) =
n
X
i=1
xi log xi, f∗
0 (y) =
n
X
i=1
eyi−1
Duality 5–8
The dual problem
Lagrange dual problem
maximize g(λ, ν)
subject to λ  0
• finds best lower bound on p⋆
, obtained from Lagrange dual function
• a convex optimization problem; optimal value denoted d⋆
• λ, ν are dual feasible if λ  0, (λ, ν) ∈ dom g
• often simplified by making implicit constraint (λ, ν) ∈ dom g explicit
example: standard form LP and its dual (page 5–5)
minimize cT
x
subject to Ax = b
x  0
maximize −bT
ν
subject to AT
ν + c  0
Duality 5–9
Weak and strong duality
weak duality: d⋆
≤ p⋆
• always holds (for convex and nonconvex problems)
• can be used to find nontrivial lower bounds for difficult problems
for example, solving the SDP
maximize −1T
ν
subject to W + diag(ν)  0
gives a lower bound for the two-way partitioning problem on page 5–7
strong duality: d⋆
= p⋆
• does not hold in general
• (usually) holds for convex problems
• conditions that guarantee strong duality in convex problems are called
constraint qualifications
Duality 5–10
Slater’s constraint qualification
strong duality holds for a convex problem
minimize f0(x)
subject to fi(x) ≤ 0, i = 1, . . . , m
Ax = b
if it is strictly feasible, i.e.,
∃x ∈ int D : fi(x)  0, i = 1, . . . , m, Ax = b
• also guarantees that the dual optimum is attained (if p⋆
 −∞)
• can be sharpened: e.g., can replace int D with relint D (interior
relative to affine hull); linear inequalities do not need to hold with strict
inequality, . . .
• there exist many other types of constraint qualifications
Duality 5–11
Inequality form LP
primal problem
minimize cT
x
subject to Ax  b
dual function
g(λ) = inf
x
(c + AT
λ)T
x − bT
λ

=

−bT
λ AT
λ + c = 0
−∞ otherwise
dual problem
maximize −bT
λ
subject to AT
λ + c = 0, λ  0
• from Slater’s condition: p⋆
= d⋆
if Ax̃ ≺ b for some x̃
• in fact, p⋆
= d⋆
except when primal and dual are infeasible
Duality 5–12
Quadratic program
primal problem (assume P ∈ Sn
++)
minimize xT
Px
subject to Ax  b
dual function
g(λ) = inf
x
xT
Px + λT
(Ax − b)

= −
1
4
λT
AP−1
AT
λ − bT
λ
dual problem
maximize −(1/4)λT
AP−1
AT
λ − bT
λ
subject to λ  0
• from Slater’s condition: p⋆
= d⋆
if Ax̃ ≺ b for some x̃
• in fact, p⋆
= d⋆
always
Duality 5–13
A nonconvex problem with strong duality
minimize xT
Ax + 2bT
x
subject to xT
x ≤ 1
A 6 0, hence nonconvex
dual function: g(λ) = infx(xT
(A + λI)x + 2bT
x − λ)
• unbounded below if A + λI 6 0 or if A + λI  0 and b 6∈ R(A + λI)
• minimized by x = −(A + λI)†
b otherwise: g(λ) = −bT
(A + λI)†
b − λ
dual problem and equivalent SDP:
maximize −bT
(A + λI)†
b − λ
subject to A + λI  0
b ∈ R(A + λI)
maximize −t − λ
subject to

A + λI b
bT
t

 0
strong duality although primal problem is not convex (not easy to show)
Duality 5–14
Geometric interpretation
for simplicity, consider problem with one constraint f1(x) ≤ 0
interpretation of dual function:
g(λ) = inf
(u,t)∈G
(t + λu), where G = {(f1(x), f0(x)) | x ∈ D}
G
p⋆
g(λ)
λu + t = g(λ)
t
u
G
p⋆
d⋆
t
u
• λu + t = g(λ) is (non-vertical) supporting hyperplane to G
• hyperplane intersects t-axis at t = g(λ)
Duality 5–15
epigraph variation: same interpretation if G is replaced with
A = {(u, t) | f1(x) ≤ u, f0(x) ≤ t for some x ∈ D}
A
p⋆
g(λ)
λu + t = g(λ)
t
u
strong duality
• holds if there is a non-vertical supporting hyperplane to A at (0, p⋆
)
• for convex problem, A is convex, hence has supp. hyperplane at (0, p⋆
)
• Slater’s condition: if there exist (ũ, t̃) ∈ A with ũ  0, then supporting
hyperplanes at (0, p⋆
) must be non-vertical
Duality 5–16
Complementary slackness
assume strong duality holds, x⋆
is primal optimal, (λ⋆
, ν⋆
) is dual optimal
f0(x⋆
) = g(λ⋆
, ν⋆
) = inf
x
f0(x) +
m
X
i=1
λ⋆
i fi(x) +
p
X
i=1
ν⋆
i hi(x)
!
≤ f0(x⋆
) +
m
X
i=1
λ⋆
i fi(x⋆
) +
p
X
i=1
ν⋆
i hi(x⋆
)
≤ f0(x⋆
)
hence, the two inequalities hold with equality
• x⋆
minimizes L(x, λ⋆
, ν⋆
)
• λ⋆
i fi(x⋆
) = 0 for i = 1, . . . , m (known as complementary slackness):
λ⋆
i  0 =⇒ fi(x⋆
) = 0, fi(x⋆
)  0 =⇒ λ⋆
i = 0
Duality 5–17
Karush-Kuhn-Tucker (KKT) conditions
the following four conditions are called KKT conditions (for a problem with
differentiable fi, hi):
1. primal constraints: fi(x) ≤ 0, i = 1, . . . , m, hi(x) = 0, i = 1, . . . , p
2. dual constraints: λ  0
3. complementary slackness: λifi(x) = 0, i = 1, . . . , m
4. gradient of Lagrangian with respect to x vanishes:
∇f0(x) +
m
X
i=1
λi∇fi(x) +
p
X
i=1
νi∇hi(x) = 0
from page 5–17: if strong duality holds and x, λ, ν are optimal, then they
must satisfy the KKT conditions
Duality 5–18
KKT conditions for convex problem
if x̃, λ̃, ν̃ satisfy KKT for a convex problem, then they are optimal:
• from complementary slackness: f0(x̃) = L(x̃, λ̃, ν̃)
• from 4th condition (and convexity): g(λ̃, ν̃) = L(x̃, λ̃, ν̃)
hence, f0(x̃) = g(λ̃, ν̃)
if Slater’s condition is satisfied:
x is optimal if and only if there exist λ, ν that satisfy KKT conditions
• recall that Slater implies strong duality, and dual optimum is attained
• generalizes optimality condition ∇f0(x) = 0 for unconstrained problem
Duality 5–19
example: water-filling (assume αi  0)
minimize −
Pn
i=1 log(xi + αi)
subject to x  0, 1T
x = 1
x is optimal iff x  0, 1T
x = 1, and there exist λ ∈ Rn
, ν ∈ R such that
λ  0, λixi = 0,
1
xi + αi
+ λi = ν
• if ν  1/αi: λi = 0 and xi = 1/ν − αi
• if ν ≥ 1/αi: λi = ν − 1/αi and xi = 0
• determine ν from 1T
x =
Pn
i=1 max{0, 1/ν − αi} = 1
interpretation
• n patches; level of patch i is at height αi
• flood area with unit amount of water
• resulting level is 1/ν⋆
i
1/ν⋆
xi
αi
Duality 5–20
Perturbation and sensitivity analysis
(unperturbed) optimization problem and its dual
minimize f0(x)
subject to fi(x) ≤ 0, i = 1, . . . , m
hi(x) = 0, i = 1, . . . , p
maximize g(λ, ν)
subject to λ  0
perturbed problem and its dual
min. f0(x)
s.t. fi(x) ≤ ui, i = 1, . . . , m
hi(x) = vi, i = 1, . . . , p
max. g(λ, ν) − uT
λ − vT
ν
s.t. λ  0
• x is primal variable; u, v are parameters
• p⋆
(u, v) is optimal value as a function of u, v
• we are interested in information about p⋆
(u, v) that we can obtain from
the solution of the unperturbed problem and its dual
Duality 5–21
global sensitivity result
assume strong duality holds for unperturbed problem, and that λ⋆
, ν⋆
are
dual optimal for unperturbed problem
apply weak duality to perturbed problem:
p⋆
(u, v) ≥ g(λ⋆
, ν⋆
) − uT
λ⋆
− vT
ν⋆
= p⋆
(0, 0) − uT
λ⋆
− vT
ν⋆
sensitivity interpretation
• if λ⋆
i large: p⋆
increases greatly if we tighten constraint i (ui  0)
• if λ⋆
i small: p⋆
does not decrease much if we loosen constraint i (ui  0)
• if ν⋆
i large and positive: p⋆
increases greatly if we take vi  0;
if ν⋆
i large and negative: p⋆
increases greatly if we take vi  0
• if ν⋆
i small and positive: p⋆
does not decrease much if we take vi  0;
if ν⋆
i small and negative: p⋆
does not decrease much if we take vi  0
Duality 5–22
local sensitivity: if (in addition) p⋆
(u, v) is differentiable at (0, 0), then
λ⋆
i = −
∂p⋆
(0, 0)
∂ui
, ν⋆
i = −
∂p⋆
(0, 0)
∂vi
proof (for λ⋆
i ): from global sensitivity result,
∂p⋆
(0, 0)
∂ui
= lim
tց0
p⋆
(tei, 0) − p⋆
(0, 0)
t
≥ −λ⋆
i
∂p⋆
(0, 0)
∂ui
= lim
tր0
p⋆
(tei, 0) − p⋆
(0, 0)
t
≤ −λ⋆
i
hence, equality
p⋆
(u) for a problem with one (inequality)
constraint: u
p⋆
(u)
p⋆
(0) − λ⋆
u
u = 0
Duality 5–23
Duality and problem reformulations
• equivalent formulations of a problem can lead to very different duals
• reformulating the primal problem can be useful when the dual is difficult
to derive, or uninteresting
common reformulations
• introduce new variables and equality constraints
• make explicit constraints implicit or vice-versa
• transform objective or constraint functions
e.g., replace f0(x) by φ(f0(x)) with φ convex, increasing
Duality 5–24
Introducing new variables and equality constraints
minimize f0(Ax + b)
• dual function is constant: g = infx L(x) = infx f0(Ax + b) = p⋆
• we have strong duality, but dual is quite useless
reformulated problem and its dual
minimize f0(y)
subject to Ax + b − y = 0
maximize bT
ν − f∗
0 (ν)
subject to AT
ν = 0
dual function follows from
g(ν) = inf
x,y
(f0(y) − νT
y + νT
Ax + bT
ν)
=

−f∗
0 (ν) + bT
ν AT
ν = 0
−∞ otherwise
Duality 5–25
norm approximation problem: minimize kAx − bk
minimize kyk
subject to y = Ax − b
can look up conjugate of k · k, or derive dual directly
g(ν) = inf
x,y
(kyk + νT
y − νT
Ax + bT
ν)
=

bT
ν + infy(kyk + νT
y) AT
ν = 0
−∞ otherwise
=

bT
ν AT
ν = 0, kνk∗ ≤ 1
−∞ otherwise
(see page 5–4)
dual of norm approximation problem
maximize bT
ν
subject to AT
ν = 0, kνk∗ ≤ 1
Duality 5–26
Implicit constraints
LP with box constraints: primal and dual problem
minimize cT
x
subject to Ax = b
−1  x  1
maximize −bT
ν − 1T
λ1 − 1T
λ2
subject to c + AT
ν + λ1 − λ2 = 0
λ1  0, λ2  0
reformulation with box constraints made implicit
minimize f0(x) =

cT
x −1  x  1
∞ otherwise
subject to Ax = b
dual function
g(ν) = inf
−1x1
(cT
x + νT
(Ax − b))
= −bT
ν − kAT
ν + ck1
dual problem: maximize −bT
ν − kAT
ν + ck1
Duality 5–27
Problems with generalized inequalities
minimize f0(x)
subject to fi(x) Ki
0, i = 1, . . . , m
hi(x) = 0, i = 1, . . . , p
Ki
is generalized inequality on Rki
definitions are parallel to scalar case:
• Lagrange multiplier for fi(x) Ki
0 is vector λi ∈ Rki
• Lagrangian L : Rn
× Rk1
× · · · × Rkm
× Rp
→ R, is defined as
L(x, λ1, · · · , λm, ν) = f0(x) +
m
X
i=1
λT
i fi(x) +
p
X
i=1
νihi(x)
• dual function g : Rk1
× · · · × Rkm
× Rp
→ R, is defined as
g(λ1, . . . , λm, ν) = inf
x∈D
L(x, λ1, · · · , λm, ν)
Duality 5–28
lower bound property: if λi K∗
i
0, then g(λ1, . . . , λm, ν) ≤ p⋆
proof: if x̃ is feasible and λ K∗
i
0, then
f0(x̃) ≥ f0(x̃) +
m
X
i=1
λT
i fi(x̃) +
p
X
i=1
νihi(x̃)
≥ inf
x∈D
L(x, λ1, . . . , λm, ν)
= g(λ1, . . . , λm, ν)
minimizing over all feasible x̃ gives p⋆
≥ g(λ1, . . . , λm, ν)
dual problem
maximize g(λ1, . . . , λm, ν)
subject to λi K∗
i
0, i = 1, . . . , m
• weak duality: p⋆
≥ d⋆
always
• strong duality: p⋆
= d⋆
for convex problem with constraint qualification
(for example, Slater’s: primal problem is strictly feasible)
Duality 5–29
Semidefinite program
primal SDP (Fi, G ∈ Sk
)
minimize cT
x
subject to x1F1 + · · · + xnFn  G
• Lagrange multiplier is matrix Z ∈ Sk
• Lagrangian L(x, Z) = cT
x + tr (Z(x1F1 + · · · + xnFn − G))
• dual function
g(Z) = inf
x
L(x, Z) =

− tr(GZ) tr(FiZ) + ci = 0, i = 1, . . . , n
−∞ otherwise
dual SDP
maximize − tr(GZ)
subject to Z  0, tr(FiZ) + ci = 0, i = 1, . . . , n
p⋆
= d⋆
if primal SDP is strictly feasible (∃x with x1F1 + · · · + xnFn ≺ G)
Duality 5–30

More Related Content

What's hot

Usage of regular expressions in nlp
Usage of regular expressions in nlpUsage of regular expressions in nlp
Usage of regular expressions in nlp
eSAT Journals
 
Dynamic programming
Dynamic programmingDynamic programming
Dynamic programming
Shakil Ahmed
 
Theory of first order logic
Theory of first order logicTheory of first order logic
Theory of first order logic
Devaddd
 
The LaTeX Workshop: Typesetting Mathematics with LaTeX
The LaTeX Workshop: Typesetting Mathematics with LaTeXThe LaTeX Workshop: Typesetting Mathematics with LaTeX
The LaTeX Workshop: Typesetting Mathematics with LaTeX
Suddhasheel GHOSH, PhD
 
Sequences and Series (Mathematics)
Sequences and Series (Mathematics) Sequences and Series (Mathematics)
Sequences and Series (Mathematics)
Dhrumil Maniar
 
Limits and Continuity of Functions
Limits and Continuity of Functions Limits and Continuity of Functions
Limits and Continuity of Functions
OlooPundit
 
Homomorphisms
HomomorphismsHomomorphisms
Homomorphisms
Aishwarya Kannan
 
Master theorem
Master theoremMaster theorem
Master theorem
fika sweety
 
What is analytic functions
What is analytic functionsWhat is analytic functions
What is analytic functions
Tarun Gehlot
 
Natural Language Processing
Natural Language ProcessingNatural Language Processing
Natural Language Processing
Saurabh Kaushik
 
ディジタル信号処理の課題解説 その3
ディジタル信号処理の課題解説 その3ディジタル信号処理の課題解説 その3
ディジタル信号処理の課題解説 その3noname409
 
Matlab code for crout method
Matlab code for crout methodMatlab code for crout method
Matlab code for crout method
Taimoor Muzaffar Gondal
 
Vector differentiation, the ∇ operator,
Vector differentiation, the ∇ operator,Vector differentiation, the ∇ operator,
Vector differentiation, the ∇ operator,
Tarun Gehlot
 
Registers and its type DLD
Registers and its type DLDRegisters and its type DLD
Registers and its type DLD
Azeemaj101
 
improper integrals
improper integralsimproper integrals
improper integrals
2461998
 
Exponential and logarithmic functions
Exponential and logarithmic functionsExponential and logarithmic functions
Exponential and logarithmic functions
awesomepossum7676
 
Análise de Algoritmos - As classes P e NP
Análise de Algoritmos - As classes P e NPAnálise de Algoritmos - As classes P e NP
Análise de Algoritmos - As classes P e NP
Delacyr Ferreira
 
#4 formal methods – predicate logic
#4 formal methods – predicate logic#4 formal methods – predicate logic
#4 formal methods – predicate logic
Sharif Omar Salem
 
Introduction to prolog
Introduction to prologIntroduction to prolog
Introduction to prolog
Rakhi Sinha
 
Linear Algebra PowerPoint
Linear Algebra PowerPointLinear Algebra PowerPoint
Linear Algebra PowerPoint
Ashley Carter
 

What's hot (20)

Usage of regular expressions in nlp
Usage of regular expressions in nlpUsage of regular expressions in nlp
Usage of regular expressions in nlp
 
Dynamic programming
Dynamic programmingDynamic programming
Dynamic programming
 
Theory of first order logic
Theory of first order logicTheory of first order logic
Theory of first order logic
 
The LaTeX Workshop: Typesetting Mathematics with LaTeX
The LaTeX Workshop: Typesetting Mathematics with LaTeXThe LaTeX Workshop: Typesetting Mathematics with LaTeX
The LaTeX Workshop: Typesetting Mathematics with LaTeX
 
Sequences and Series (Mathematics)
Sequences and Series (Mathematics) Sequences and Series (Mathematics)
Sequences and Series (Mathematics)
 
Limits and Continuity of Functions
Limits and Continuity of Functions Limits and Continuity of Functions
Limits and Continuity of Functions
 
Homomorphisms
HomomorphismsHomomorphisms
Homomorphisms
 
Master theorem
Master theoremMaster theorem
Master theorem
 
What is analytic functions
What is analytic functionsWhat is analytic functions
What is analytic functions
 
Natural Language Processing
Natural Language ProcessingNatural Language Processing
Natural Language Processing
 
ディジタル信号処理の課題解説 その3
ディジタル信号処理の課題解説 その3ディジタル信号処理の課題解説 その3
ディジタル信号処理の課題解説 その3
 
Matlab code for crout method
Matlab code for crout methodMatlab code for crout method
Matlab code for crout method
 
Vector differentiation, the ∇ operator,
Vector differentiation, the ∇ operator,Vector differentiation, the ∇ operator,
Vector differentiation, the ∇ operator,
 
Registers and its type DLD
Registers and its type DLDRegisters and its type DLD
Registers and its type DLD
 
improper integrals
improper integralsimproper integrals
improper integrals
 
Exponential and logarithmic functions
Exponential and logarithmic functionsExponential and logarithmic functions
Exponential and logarithmic functions
 
Análise de Algoritmos - As classes P e NP
Análise de Algoritmos - As classes P e NPAnálise de Algoritmos - As classes P e NP
Análise de Algoritmos - As classes P e NP
 
#4 formal methods – predicate logic
#4 formal methods – predicate logic#4 formal methods – predicate logic
#4 formal methods – predicate logic
 
Introduction to prolog
Introduction to prologIntroduction to prolog
Introduction to prolog
 
Linear Algebra PowerPoint
Linear Algebra PowerPointLinear Algebra PowerPoint
Linear Algebra PowerPoint
 

Similar to Duality.pdf

03 convexfunctions
03 convexfunctions03 convexfunctions
03 convexfunctions
Sufyan Sahoo
 
5.4 Saddle-point interpretation, 5.5 Optimality conditions, 5.6 Perturbation ...
5.4 Saddle-point interpretation, 5.5 Optimality conditions, 5.6 Perturbation ...5.4 Saddle-point interpretation, 5.5 Optimality conditions, 5.6 Perturbation ...
5.4 Saddle-point interpretation, 5.5 Optimality conditions, 5.6 Perturbation ...
RyotaroTsukada
 
QMC: Operator Splitting Workshop, Progressive Decoupling of Linkages in Optim...
QMC: Operator Splitting Workshop, Progressive Decoupling of Linkages in Optim...QMC: Operator Splitting Workshop, Progressive Decoupling of Linkages in Optim...
QMC: Operator Splitting Workshop, Progressive Decoupling of Linkages in Optim...
The Statistical and Applied Mathematical Sciences Institute
 
Convex Analysis and Duality (based on "Functional Analysis and Optimization" ...
Convex Analysis and Duality (based on "Functional Analysis and Optimization" ...Convex Analysis and Duality (based on "Functional Analysis and Optimization" ...
Convex Analysis and Duality (based on "Functional Analysis and Optimization" ...
Katsuya Ito
 
Lesson 4: Calculating Limits (slides)
Lesson 4: Calculating Limits (slides)Lesson 4: Calculating Limits (slides)
Lesson 4: Calculating Limits (slides)
Mel Anthony Pepito
 
Lesson 4: Calcuating Limits (slides)
Lesson 4: Calcuating Limits (slides)Lesson 4: Calcuating Limits (slides)
Lesson 4: Calcuating Limits (slides)
Matthew Leingang
 
IVR - Chapter 1 - Introduction
IVR - Chapter 1 - IntroductionIVR - Chapter 1 - Introduction
IVR - Chapter 1 - Introduction
Charles Deledalle
 
Lecture 2: linear SVM in the Dual
Lecture 2: linear SVM in the DualLecture 2: linear SVM in the Dual
Lecture 2: linear SVM in the Dual
Stéphane Canu
 
Lecture 2: linear SVM in the dual
Lecture 2: linear SVM in the dualLecture 2: linear SVM in the dual
Lecture 2: linear SVM in the dual
Stéphane Canu
 
lecture on support vector machine in the field of ML
lecture on support vector machine in the field of MLlecture on support vector machine in the field of ML
lecture on support vector machine in the field of ML
AkashVerma639361
 
Introduction to inverse problems
Introduction to inverse problemsIntroduction to inverse problems
Introduction to inverse problems
Delta Pi Systems
 
Lesson 14: Derivatives of Logarithmic and Exponential Functions (slides)
Lesson 14: Derivatives of Logarithmic and Exponential Functions (slides)Lesson 14: Derivatives of Logarithmic and Exponential Functions (slides)
Lesson 14: Derivatives of Logarithmic and Exponential Functions (slides)
Matthew Leingang
 
Lesson 14: Derivatives of Logarithmic and Exponential Functions (slides)
Lesson 14: Derivatives of Logarithmic and Exponential Functions (slides)Lesson 14: Derivatives of Logarithmic and Exponential Functions (slides)
Lesson 14: Derivatives of Logarithmic and Exponential Functions (slides)
Mel Anthony Pepito
 
A043001006
A043001006A043001006
A043001006
inventy
 
A043001006
A043001006A043001006
A043001006
inventy
 
A043001006
A043001006A043001006
A043001006
inventy
 
boyd 3.1
boyd 3.1boyd 3.1
boyd 3.1
Koki Isokawa
 
QMC: Operator Splitting Workshop, Thresholdings, Robustness, and Generalized ...
QMC: Operator Splitting Workshop, Thresholdings, Robustness, and Generalized ...QMC: Operator Splitting Workshop, Thresholdings, Robustness, and Generalized ...
QMC: Operator Splitting Workshop, Thresholdings, Robustness, and Generalized ...
The Statistical and Applied Mathematical Sciences Institute
 
subdiff_prox.pdf
subdiff_prox.pdfsubdiff_prox.pdf
subdiff_prox.pdf
diehardgamer1
 
Lesson 28: The Fundamental Theorem of Calculus
Lesson 28: The Fundamental Theorem of CalculusLesson 28: The Fundamental Theorem of Calculus
Lesson 28: The Fundamental Theorem of Calculus
Matthew Leingang
 

Similar to Duality.pdf (20)

03 convexfunctions
03 convexfunctions03 convexfunctions
03 convexfunctions
 
5.4 Saddle-point interpretation, 5.5 Optimality conditions, 5.6 Perturbation ...
5.4 Saddle-point interpretation, 5.5 Optimality conditions, 5.6 Perturbation ...5.4 Saddle-point interpretation, 5.5 Optimality conditions, 5.6 Perturbation ...
5.4 Saddle-point interpretation, 5.5 Optimality conditions, 5.6 Perturbation ...
 
QMC: Operator Splitting Workshop, Progressive Decoupling of Linkages in Optim...
QMC: Operator Splitting Workshop, Progressive Decoupling of Linkages in Optim...QMC: Operator Splitting Workshop, Progressive Decoupling of Linkages in Optim...
QMC: Operator Splitting Workshop, Progressive Decoupling of Linkages in Optim...
 
Convex Analysis and Duality (based on "Functional Analysis and Optimization" ...
Convex Analysis and Duality (based on "Functional Analysis and Optimization" ...Convex Analysis and Duality (based on "Functional Analysis and Optimization" ...
Convex Analysis and Duality (based on "Functional Analysis and Optimization" ...
 
Lesson 4: Calculating Limits (slides)
Lesson 4: Calculating Limits (slides)Lesson 4: Calculating Limits (slides)
Lesson 4: Calculating Limits (slides)
 
Lesson 4: Calcuating Limits (slides)
Lesson 4: Calcuating Limits (slides)Lesson 4: Calcuating Limits (slides)
Lesson 4: Calcuating Limits (slides)
 
IVR - Chapter 1 - Introduction
IVR - Chapter 1 - IntroductionIVR - Chapter 1 - Introduction
IVR - Chapter 1 - Introduction
 
Lecture 2: linear SVM in the Dual
Lecture 2: linear SVM in the DualLecture 2: linear SVM in the Dual
Lecture 2: linear SVM in the Dual
 
Lecture 2: linear SVM in the dual
Lecture 2: linear SVM in the dualLecture 2: linear SVM in the dual
Lecture 2: linear SVM in the dual
 
lecture on support vector machine in the field of ML
lecture on support vector machine in the field of MLlecture on support vector machine in the field of ML
lecture on support vector machine in the field of ML
 
Introduction to inverse problems
Introduction to inverse problemsIntroduction to inverse problems
Introduction to inverse problems
 
Lesson 14: Derivatives of Logarithmic and Exponential Functions (slides)
Lesson 14: Derivatives of Logarithmic and Exponential Functions (slides)Lesson 14: Derivatives of Logarithmic and Exponential Functions (slides)
Lesson 14: Derivatives of Logarithmic and Exponential Functions (slides)
 
Lesson 14: Derivatives of Logarithmic and Exponential Functions (slides)
Lesson 14: Derivatives of Logarithmic and Exponential Functions (slides)Lesson 14: Derivatives of Logarithmic and Exponential Functions (slides)
Lesson 14: Derivatives of Logarithmic and Exponential Functions (slides)
 
A043001006
A043001006A043001006
A043001006
 
A043001006
A043001006A043001006
A043001006
 
A043001006
A043001006A043001006
A043001006
 
boyd 3.1
boyd 3.1boyd 3.1
boyd 3.1
 
QMC: Operator Splitting Workshop, Thresholdings, Robustness, and Generalized ...
QMC: Operator Splitting Workshop, Thresholdings, Robustness, and Generalized ...QMC: Operator Splitting Workshop, Thresholdings, Robustness, and Generalized ...
QMC: Operator Splitting Workshop, Thresholdings, Robustness, and Generalized ...
 
subdiff_prox.pdf
subdiff_prox.pdfsubdiff_prox.pdf
subdiff_prox.pdf
 
Lesson 28: The Fundamental Theorem of Calculus
Lesson 28: The Fundamental Theorem of CalculusLesson 28: The Fundamental Theorem of Calculus
Lesson 28: The Fundamental Theorem of Calculus
 

Recently uploaded

在线办理(英国UCA毕业证书)创意艺术大学毕业证在读证明一模一样
在线办理(英国UCA毕业证书)创意艺术大学毕业证在读证明一模一样在线办理(英国UCA毕业证书)创意艺术大学毕业证在读证明一模一样
在线办理(英国UCA毕业证书)创意艺术大学毕业证在读证明一模一样
v7oacc3l
 
STATATHON: Unleashing the Power of Statistics in a 48-Hour Knowledge Extravag...
STATATHON: Unleashing the Power of Statistics in a 48-Hour Knowledge Extravag...STATATHON: Unleashing the Power of Statistics in a 48-Hour Knowledge Extravag...
STATATHON: Unleashing the Power of Statistics in a 48-Hour Knowledge Extravag...
sameer shah
 
End-to-end pipeline agility - Berlin Buzzwords 2024
End-to-end pipeline agility - Berlin Buzzwords 2024End-to-end pipeline agility - Berlin Buzzwords 2024
End-to-end pipeline agility - Berlin Buzzwords 2024
Lars Albertsson
 
Experts live - Improving user adoption with AI
Experts live - Improving user adoption with AIExperts live - Improving user adoption with AI
Experts live - Improving user adoption with AI
jitskeb
 
一比一原版(UMN文凭证书)明尼苏达大学毕业证如何办理
一比一原版(UMN文凭证书)明尼苏达大学毕业证如何办理一比一原版(UMN文凭证书)明尼苏达大学毕业证如何办理
一比一原版(UMN文凭证书)明尼苏达大学毕业证如何办理
nyfuhyz
 
The Ipsos - AI - Monitor 2024 Report.pdf
The  Ipsos - AI - Monitor 2024 Report.pdfThe  Ipsos - AI - Monitor 2024 Report.pdf
The Ipsos - AI - Monitor 2024 Report.pdf
Social Samosa
 
一比一原版(爱大毕业证书)爱丁堡大学毕业证如何办理
一比一原版(爱大毕业证书)爱丁堡大学毕业证如何办理一比一原版(爱大毕业证书)爱丁堡大学毕业证如何办理
一比一原版(爱大毕业证书)爱丁堡大学毕业证如何办理
g4dpvqap0
 
ViewShift: Hassle-free Dynamic Policy Enforcement for Every Data Lake
ViewShift: Hassle-free Dynamic Policy Enforcement for Every Data LakeViewShift: Hassle-free Dynamic Policy Enforcement for Every Data Lake
ViewShift: Hassle-free Dynamic Policy Enforcement for Every Data Lake
Walaa Eldin Moustafa
 
Challenges of Nation Building-1.pptx with more important
Challenges of Nation Building-1.pptx with more importantChallenges of Nation Building-1.pptx with more important
Challenges of Nation Building-1.pptx with more important
Sm321
 
Everything you wanted to know about LIHTC
Everything you wanted to know about LIHTCEverything you wanted to know about LIHTC
Everything you wanted to know about LIHTC
Roger Valdez
 
一比一原版(牛布毕业证书)牛津布鲁克斯大学毕业证如何办理
一比一原版(牛布毕业证书)牛津布鲁克斯大学毕业证如何办理一比一原版(牛布毕业证书)牛津布鲁克斯大学毕业证如何办理
一比一原版(牛布毕业证书)牛津布鲁克斯大学毕业证如何办理
74nqk8xf
 
The Building Blocks of QuestDB, a Time Series Database
The Building Blocks of QuestDB, a Time Series DatabaseThe Building Blocks of QuestDB, a Time Series Database
The Building Blocks of QuestDB, a Time Series Database
javier ramirez
 
一比一原版(Coventry毕业证书)考文垂大学毕业证如何办理
一比一原版(Coventry毕业证书)考文垂大学毕业证如何办理一比一原版(Coventry毕业证书)考文垂大学毕业证如何办理
一比一原版(Coventry毕业证书)考文垂大学毕业证如何办理
74nqk8xf
 
Predictably Improve Your B2B Tech Company's Performance by Leveraging Data
Predictably Improve Your B2B Tech Company's Performance by Leveraging DataPredictably Improve Your B2B Tech Company's Performance by Leveraging Data
Predictably Improve Your B2B Tech Company's Performance by Leveraging Data
Kiwi Creative
 
State of Artificial intelligence Report 2023
State of Artificial intelligence Report 2023State of Artificial intelligence Report 2023
State of Artificial intelligence Report 2023
kuntobimo2016
 
一比一原版(Glasgow毕业证书)格拉斯哥大学毕业证如何办理
一比一原版(Glasgow毕业证书)格拉斯哥大学毕业证如何办理一比一原版(Glasgow毕业证书)格拉斯哥大学毕业证如何办理
一比一原版(Glasgow毕业证书)格拉斯哥大学毕业证如何办理
g4dpvqap0
 
Beyond the Basics of A/B Tests: Highly Innovative Experimentation Tactics You...
Beyond the Basics of A/B Tests: Highly Innovative Experimentation Tactics You...Beyond the Basics of A/B Tests: Highly Innovative Experimentation Tactics You...
Beyond the Basics of A/B Tests: Highly Innovative Experimentation Tactics You...
Aggregage
 
Learn SQL from basic queries to Advance queries
Learn SQL from basic queries to Advance queriesLearn SQL from basic queries to Advance queries
Learn SQL from basic queries to Advance queries
manishkhaire30
 
Udemy_2024_Global_Learning_Skills_Trends_Report (1).pdf
Udemy_2024_Global_Learning_Skills_Trends_Report (1).pdfUdemy_2024_Global_Learning_Skills_Trends_Report (1).pdf
Udemy_2024_Global_Learning_Skills_Trends_Report (1).pdf
Fernanda Palhano
 
Population Growth in Bataan: The effects of population growth around rural pl...
Population Growth in Bataan: The effects of population growth around rural pl...Population Growth in Bataan: The effects of population growth around rural pl...
Population Growth in Bataan: The effects of population growth around rural pl...
Bill641377
 

Recently uploaded (20)

在线办理(英国UCA毕业证书)创意艺术大学毕业证在读证明一模一样
在线办理(英国UCA毕业证书)创意艺术大学毕业证在读证明一模一样在线办理(英国UCA毕业证书)创意艺术大学毕业证在读证明一模一样
在线办理(英国UCA毕业证书)创意艺术大学毕业证在读证明一模一样
 
STATATHON: Unleashing the Power of Statistics in a 48-Hour Knowledge Extravag...
STATATHON: Unleashing the Power of Statistics in a 48-Hour Knowledge Extravag...STATATHON: Unleashing the Power of Statistics in a 48-Hour Knowledge Extravag...
STATATHON: Unleashing the Power of Statistics in a 48-Hour Knowledge Extravag...
 
End-to-end pipeline agility - Berlin Buzzwords 2024
End-to-end pipeline agility - Berlin Buzzwords 2024End-to-end pipeline agility - Berlin Buzzwords 2024
End-to-end pipeline agility - Berlin Buzzwords 2024
 
Experts live - Improving user adoption with AI
Experts live - Improving user adoption with AIExperts live - Improving user adoption with AI
Experts live - Improving user adoption with AI
 
一比一原版(UMN文凭证书)明尼苏达大学毕业证如何办理
一比一原版(UMN文凭证书)明尼苏达大学毕业证如何办理一比一原版(UMN文凭证书)明尼苏达大学毕业证如何办理
一比一原版(UMN文凭证书)明尼苏达大学毕业证如何办理
 
The Ipsos - AI - Monitor 2024 Report.pdf
The  Ipsos - AI - Monitor 2024 Report.pdfThe  Ipsos - AI - Monitor 2024 Report.pdf
The Ipsos - AI - Monitor 2024 Report.pdf
 
一比一原版(爱大毕业证书)爱丁堡大学毕业证如何办理
一比一原版(爱大毕业证书)爱丁堡大学毕业证如何办理一比一原版(爱大毕业证书)爱丁堡大学毕业证如何办理
一比一原版(爱大毕业证书)爱丁堡大学毕业证如何办理
 
ViewShift: Hassle-free Dynamic Policy Enforcement for Every Data Lake
ViewShift: Hassle-free Dynamic Policy Enforcement for Every Data LakeViewShift: Hassle-free Dynamic Policy Enforcement for Every Data Lake
ViewShift: Hassle-free Dynamic Policy Enforcement for Every Data Lake
 
Challenges of Nation Building-1.pptx with more important
Challenges of Nation Building-1.pptx with more importantChallenges of Nation Building-1.pptx with more important
Challenges of Nation Building-1.pptx with more important
 
Everything you wanted to know about LIHTC
Everything you wanted to know about LIHTCEverything you wanted to know about LIHTC
Everything you wanted to know about LIHTC
 
一比一原版(牛布毕业证书)牛津布鲁克斯大学毕业证如何办理
一比一原版(牛布毕业证书)牛津布鲁克斯大学毕业证如何办理一比一原版(牛布毕业证书)牛津布鲁克斯大学毕业证如何办理
一比一原版(牛布毕业证书)牛津布鲁克斯大学毕业证如何办理
 
The Building Blocks of QuestDB, a Time Series Database
The Building Blocks of QuestDB, a Time Series DatabaseThe Building Blocks of QuestDB, a Time Series Database
The Building Blocks of QuestDB, a Time Series Database
 
一比一原版(Coventry毕业证书)考文垂大学毕业证如何办理
一比一原版(Coventry毕业证书)考文垂大学毕业证如何办理一比一原版(Coventry毕业证书)考文垂大学毕业证如何办理
一比一原版(Coventry毕业证书)考文垂大学毕业证如何办理
 
Predictably Improve Your B2B Tech Company's Performance by Leveraging Data
Predictably Improve Your B2B Tech Company's Performance by Leveraging DataPredictably Improve Your B2B Tech Company's Performance by Leveraging Data
Predictably Improve Your B2B Tech Company's Performance by Leveraging Data
 
State of Artificial intelligence Report 2023
State of Artificial intelligence Report 2023State of Artificial intelligence Report 2023
State of Artificial intelligence Report 2023
 
一比一原版(Glasgow毕业证书)格拉斯哥大学毕业证如何办理
一比一原版(Glasgow毕业证书)格拉斯哥大学毕业证如何办理一比一原版(Glasgow毕业证书)格拉斯哥大学毕业证如何办理
一比一原版(Glasgow毕业证书)格拉斯哥大学毕业证如何办理
 
Beyond the Basics of A/B Tests: Highly Innovative Experimentation Tactics You...
Beyond the Basics of A/B Tests: Highly Innovative Experimentation Tactics You...Beyond the Basics of A/B Tests: Highly Innovative Experimentation Tactics You...
Beyond the Basics of A/B Tests: Highly Innovative Experimentation Tactics You...
 
Learn SQL from basic queries to Advance queries
Learn SQL from basic queries to Advance queriesLearn SQL from basic queries to Advance queries
Learn SQL from basic queries to Advance queries
 
Udemy_2024_Global_Learning_Skills_Trends_Report (1).pdf
Udemy_2024_Global_Learning_Skills_Trends_Report (1).pdfUdemy_2024_Global_Learning_Skills_Trends_Report (1).pdf
Udemy_2024_Global_Learning_Skills_Trends_Report (1).pdf
 
Population Growth in Bataan: The effects of population growth around rural pl...
Population Growth in Bataan: The effects of population growth around rural pl...Population Growth in Bataan: The effects of population growth around rural pl...
Population Growth in Bataan: The effects of population growth around rural pl...
 

Duality.pdf

  • 1. Convex Optimization — Boyd & Vandenberghe 5. Duality • Lagrange dual problem • weak and strong duality • geometric interpretation • optimality conditions • perturbation and sensitivity analysis • examples • generalized inequalities 5–1
  • 2. Lagrangian standard form problem (not necessarily convex) minimize f0(x) subject to fi(x) ≤ 0, i = 1, . . . , m hi(x) = 0, i = 1, . . . , p variable x ∈ Rn , domain D, optimal value p⋆ Lagrangian: L : Rn × Rm × Rp → R, with dom L = D × Rm × Rp , L(x, λ, ν) = f0(x) + m X i=1 λifi(x) + p X i=1 νihi(x) • weighted sum of objective and constraint functions • λi is Lagrange multiplier associated with fi(x) ≤ 0 • νi is Lagrange multiplier associated with hi(x) = 0 Duality 5–2
  • 3. Lagrange dual function Lagrange dual function: g : Rm × Rp → R, g(λ, ν) = inf x∈D L(x, λ, ν) = inf x∈D f0(x) + m X i=1 λifi(x) + p X i=1 νihi(x) ! g is concave, can be −∞ for some λ, ν lower bound property: if λ 0, then g(λ, ν) ≤ p⋆ proof: if x̃ is feasible and λ 0, then f0(x̃) ≥ L(x̃, λ, ν) ≥ inf x∈D L(x, λ, ν) = g(λ, ν) minimizing over all feasible x̃ gives p⋆ ≥ g(λ, ν) Duality 5–3
  • 4. Least-norm solution of linear equations minimize xT x subject to Ax = b dual function • Lagrangian is L(x, ν) = xT x + νT (Ax − b) • to minimize L over x, set gradient equal to zero: ∇xL(x, ν) = 2x + AT ν = 0 =⇒ x = −(1/2)AT ν • plug in in L to obtain g: g(ν) = L((−1/2)AT ν, ν) = − 1 4 νT AAT ν − bT ν a concave function of ν lower bound property: p⋆ ≥ −(1/4)νT AAT ν − bT ν for all ν Duality 5–4
  • 5. Standard form LP minimize cT x subject to Ax = b, x 0 dual function • Lagrangian is L(x, λ, ν) = cT x + νT (Ax − b) − λT x = −bT ν + (c + AT ν − λ)T x • L is affine in x, hence g(λ, ν) = inf x L(x, λ, ν) = −bT ν AT ν − λ + c = 0 −∞ otherwise g is linear on affine domain {(λ, ν) | AT ν − λ + c = 0}, hence concave lower bound property: p⋆ ≥ −bT ν if AT ν + c 0 Duality 5–5
  • 6. Equality constrained norm minimization minimize kxk subject to Ax = b dual function g(ν) = inf x (kxk − νT Ax + bT ν) = bT ν kAT νk∗ ≤ 1 −∞ otherwise where kvk∗ = supkuk≤1 uT v is dual norm of k · k proof: follows from infx(kxk − yT x) = 0 if kyk∗ ≤ 1, −∞ otherwise • if kyk∗ ≤ 1, then kxk − yT x ≥ 0 for all x, with equality if x = 0 • if kyk∗ 1, choose x = tu where kuk ≤ 1, uT y = kyk∗ 1: kxk − yT x = t(kuk − kyk∗) → −∞ as t → ∞ lower bound property: p⋆ ≥ bT ν if kAT νk∗ ≤ 1 Duality 5–6
  • 7. Two-way partitioning minimize xT Wx subject to x2 i = 1, i = 1, . . . , n • a nonconvex problem; feasible set contains 2n discrete points • interpretation: partition {1, . . . , n} in two sets; Wij is cost of assigning i, j to the same set; −Wij is cost of assigning to different sets dual function g(ν) = inf x (xT Wx + X i νi(x2 i − 1)) = inf x xT (W + diag(ν))x − 1T ν = −1T ν W + diag(ν) 0 −∞ otherwise lower bound property: p⋆ ≥ −1T ν if W + diag(ν) 0 example: ν = −λmin(W)1 gives bound p⋆ ≥ nλmin(W) Duality 5–7
  • 8. Lagrange dual and conjugate function minimize f0(x) subject to Ax b, Cx = d dual function g(λ, ν) = inf x∈dom f0 f0(x) + (AT λ + CT ν)T x − bT λ − dT ν = −f∗ 0 (−AT λ − CT ν) − bT λ − dT ν • recall definition of conjugate f∗ (y) = supx∈dom f(yT x − f(x)) • simplifies derivation of dual if conjugate of f0 is known example: entropy maximization f0(x) = n X i=1 xi log xi, f∗ 0 (y) = n X i=1 eyi−1 Duality 5–8
  • 9. The dual problem Lagrange dual problem maximize g(λ, ν) subject to λ 0 • finds best lower bound on p⋆ , obtained from Lagrange dual function • a convex optimization problem; optimal value denoted d⋆ • λ, ν are dual feasible if λ 0, (λ, ν) ∈ dom g • often simplified by making implicit constraint (λ, ν) ∈ dom g explicit example: standard form LP and its dual (page 5–5) minimize cT x subject to Ax = b x 0 maximize −bT ν subject to AT ν + c 0 Duality 5–9
  • 10. Weak and strong duality weak duality: d⋆ ≤ p⋆ • always holds (for convex and nonconvex problems) • can be used to find nontrivial lower bounds for difficult problems for example, solving the SDP maximize −1T ν subject to W + diag(ν) 0 gives a lower bound for the two-way partitioning problem on page 5–7 strong duality: d⋆ = p⋆ • does not hold in general • (usually) holds for convex problems • conditions that guarantee strong duality in convex problems are called constraint qualifications Duality 5–10
  • 11. Slater’s constraint qualification strong duality holds for a convex problem minimize f0(x) subject to fi(x) ≤ 0, i = 1, . . . , m Ax = b if it is strictly feasible, i.e., ∃x ∈ int D : fi(x) 0, i = 1, . . . , m, Ax = b • also guarantees that the dual optimum is attained (if p⋆ −∞) • can be sharpened: e.g., can replace int D with relint D (interior relative to affine hull); linear inequalities do not need to hold with strict inequality, . . . • there exist many other types of constraint qualifications Duality 5–11
  • 12. Inequality form LP primal problem minimize cT x subject to Ax b dual function g(λ) = inf x (c + AT λ)T x − bT λ = −bT λ AT λ + c = 0 −∞ otherwise dual problem maximize −bT λ subject to AT λ + c = 0, λ 0 • from Slater’s condition: p⋆ = d⋆ if Ax̃ ≺ b for some x̃ • in fact, p⋆ = d⋆ except when primal and dual are infeasible Duality 5–12
  • 13. Quadratic program primal problem (assume P ∈ Sn ++) minimize xT Px subject to Ax b dual function g(λ) = inf x xT Px + λT (Ax − b) = − 1 4 λT AP−1 AT λ − bT λ dual problem maximize −(1/4)λT AP−1 AT λ − bT λ subject to λ 0 • from Slater’s condition: p⋆ = d⋆ if Ax̃ ≺ b for some x̃ • in fact, p⋆ = d⋆ always Duality 5–13
  • 14. A nonconvex problem with strong duality minimize xT Ax + 2bT x subject to xT x ≤ 1 A 6 0, hence nonconvex dual function: g(λ) = infx(xT (A + λI)x + 2bT x − λ) • unbounded below if A + λI 6 0 or if A + λI 0 and b 6∈ R(A + λI) • minimized by x = −(A + λI)† b otherwise: g(λ) = −bT (A + λI)† b − λ dual problem and equivalent SDP: maximize −bT (A + λI)† b − λ subject to A + λI 0 b ∈ R(A + λI) maximize −t − λ subject to A + λI b bT t 0 strong duality although primal problem is not convex (not easy to show) Duality 5–14
  • 15. Geometric interpretation for simplicity, consider problem with one constraint f1(x) ≤ 0 interpretation of dual function: g(λ) = inf (u,t)∈G (t + λu), where G = {(f1(x), f0(x)) | x ∈ D} G p⋆ g(λ) λu + t = g(λ) t u G p⋆ d⋆ t u • λu + t = g(λ) is (non-vertical) supporting hyperplane to G • hyperplane intersects t-axis at t = g(λ) Duality 5–15
  • 16. epigraph variation: same interpretation if G is replaced with A = {(u, t) | f1(x) ≤ u, f0(x) ≤ t for some x ∈ D} A p⋆ g(λ) λu + t = g(λ) t u strong duality • holds if there is a non-vertical supporting hyperplane to A at (0, p⋆ ) • for convex problem, A is convex, hence has supp. hyperplane at (0, p⋆ ) • Slater’s condition: if there exist (ũ, t̃) ∈ A with ũ 0, then supporting hyperplanes at (0, p⋆ ) must be non-vertical Duality 5–16
  • 17. Complementary slackness assume strong duality holds, x⋆ is primal optimal, (λ⋆ , ν⋆ ) is dual optimal f0(x⋆ ) = g(λ⋆ , ν⋆ ) = inf x f0(x) + m X i=1 λ⋆ i fi(x) + p X i=1 ν⋆ i hi(x) ! ≤ f0(x⋆ ) + m X i=1 λ⋆ i fi(x⋆ ) + p X i=1 ν⋆ i hi(x⋆ ) ≤ f0(x⋆ ) hence, the two inequalities hold with equality • x⋆ minimizes L(x, λ⋆ , ν⋆ ) • λ⋆ i fi(x⋆ ) = 0 for i = 1, . . . , m (known as complementary slackness): λ⋆ i 0 =⇒ fi(x⋆ ) = 0, fi(x⋆ ) 0 =⇒ λ⋆ i = 0 Duality 5–17
  • 18. Karush-Kuhn-Tucker (KKT) conditions the following four conditions are called KKT conditions (for a problem with differentiable fi, hi): 1. primal constraints: fi(x) ≤ 0, i = 1, . . . , m, hi(x) = 0, i = 1, . . . , p 2. dual constraints: λ 0 3. complementary slackness: λifi(x) = 0, i = 1, . . . , m 4. gradient of Lagrangian with respect to x vanishes: ∇f0(x) + m X i=1 λi∇fi(x) + p X i=1 νi∇hi(x) = 0 from page 5–17: if strong duality holds and x, λ, ν are optimal, then they must satisfy the KKT conditions Duality 5–18
  • 19. KKT conditions for convex problem if x̃, λ̃, ν̃ satisfy KKT for a convex problem, then they are optimal: • from complementary slackness: f0(x̃) = L(x̃, λ̃, ν̃) • from 4th condition (and convexity): g(λ̃, ν̃) = L(x̃, λ̃, ν̃) hence, f0(x̃) = g(λ̃, ν̃) if Slater’s condition is satisfied: x is optimal if and only if there exist λ, ν that satisfy KKT conditions • recall that Slater implies strong duality, and dual optimum is attained • generalizes optimality condition ∇f0(x) = 0 for unconstrained problem Duality 5–19
  • 20. example: water-filling (assume αi 0) minimize − Pn i=1 log(xi + αi) subject to x 0, 1T x = 1 x is optimal iff x 0, 1T x = 1, and there exist λ ∈ Rn , ν ∈ R such that λ 0, λixi = 0, 1 xi + αi + λi = ν • if ν 1/αi: λi = 0 and xi = 1/ν − αi • if ν ≥ 1/αi: λi = ν − 1/αi and xi = 0 • determine ν from 1T x = Pn i=1 max{0, 1/ν − αi} = 1 interpretation • n patches; level of patch i is at height αi • flood area with unit amount of water • resulting level is 1/ν⋆ i 1/ν⋆ xi αi Duality 5–20
  • 21. Perturbation and sensitivity analysis (unperturbed) optimization problem and its dual minimize f0(x) subject to fi(x) ≤ 0, i = 1, . . . , m hi(x) = 0, i = 1, . . . , p maximize g(λ, ν) subject to λ 0 perturbed problem and its dual min. f0(x) s.t. fi(x) ≤ ui, i = 1, . . . , m hi(x) = vi, i = 1, . . . , p max. g(λ, ν) − uT λ − vT ν s.t. λ 0 • x is primal variable; u, v are parameters • p⋆ (u, v) is optimal value as a function of u, v • we are interested in information about p⋆ (u, v) that we can obtain from the solution of the unperturbed problem and its dual Duality 5–21
  • 22. global sensitivity result assume strong duality holds for unperturbed problem, and that λ⋆ , ν⋆ are dual optimal for unperturbed problem apply weak duality to perturbed problem: p⋆ (u, v) ≥ g(λ⋆ , ν⋆ ) − uT λ⋆ − vT ν⋆ = p⋆ (0, 0) − uT λ⋆ − vT ν⋆ sensitivity interpretation • if λ⋆ i large: p⋆ increases greatly if we tighten constraint i (ui 0) • if λ⋆ i small: p⋆ does not decrease much if we loosen constraint i (ui 0) • if ν⋆ i large and positive: p⋆ increases greatly if we take vi 0; if ν⋆ i large and negative: p⋆ increases greatly if we take vi 0 • if ν⋆ i small and positive: p⋆ does not decrease much if we take vi 0; if ν⋆ i small and negative: p⋆ does not decrease much if we take vi 0 Duality 5–22
  • 23. local sensitivity: if (in addition) p⋆ (u, v) is differentiable at (0, 0), then λ⋆ i = − ∂p⋆ (0, 0) ∂ui , ν⋆ i = − ∂p⋆ (0, 0) ∂vi proof (for λ⋆ i ): from global sensitivity result, ∂p⋆ (0, 0) ∂ui = lim tց0 p⋆ (tei, 0) − p⋆ (0, 0) t ≥ −λ⋆ i ∂p⋆ (0, 0) ∂ui = lim tր0 p⋆ (tei, 0) − p⋆ (0, 0) t ≤ −λ⋆ i hence, equality p⋆ (u) for a problem with one (inequality) constraint: u p⋆ (u) p⋆ (0) − λ⋆ u u = 0 Duality 5–23
  • 24. Duality and problem reformulations • equivalent formulations of a problem can lead to very different duals • reformulating the primal problem can be useful when the dual is difficult to derive, or uninteresting common reformulations • introduce new variables and equality constraints • make explicit constraints implicit or vice-versa • transform objective or constraint functions e.g., replace f0(x) by φ(f0(x)) with φ convex, increasing Duality 5–24
  • 25. Introducing new variables and equality constraints minimize f0(Ax + b) • dual function is constant: g = infx L(x) = infx f0(Ax + b) = p⋆ • we have strong duality, but dual is quite useless reformulated problem and its dual minimize f0(y) subject to Ax + b − y = 0 maximize bT ν − f∗ 0 (ν) subject to AT ν = 0 dual function follows from g(ν) = inf x,y (f0(y) − νT y + νT Ax + bT ν) = −f∗ 0 (ν) + bT ν AT ν = 0 −∞ otherwise Duality 5–25
  • 26. norm approximation problem: minimize kAx − bk minimize kyk subject to y = Ax − b can look up conjugate of k · k, or derive dual directly g(ν) = inf x,y (kyk + νT y − νT Ax + bT ν) = bT ν + infy(kyk + νT y) AT ν = 0 −∞ otherwise = bT ν AT ν = 0, kνk∗ ≤ 1 −∞ otherwise (see page 5–4) dual of norm approximation problem maximize bT ν subject to AT ν = 0, kνk∗ ≤ 1 Duality 5–26
  • 27. Implicit constraints LP with box constraints: primal and dual problem minimize cT x subject to Ax = b −1 x 1 maximize −bT ν − 1T λ1 − 1T λ2 subject to c + AT ν + λ1 − λ2 = 0 λ1 0, λ2 0 reformulation with box constraints made implicit minimize f0(x) = cT x −1 x 1 ∞ otherwise subject to Ax = b dual function g(ν) = inf −1x1 (cT x + νT (Ax − b)) = −bT ν − kAT ν + ck1 dual problem: maximize −bT ν − kAT ν + ck1 Duality 5–27
  • 28. Problems with generalized inequalities minimize f0(x) subject to fi(x) Ki 0, i = 1, . . . , m hi(x) = 0, i = 1, . . . , p Ki is generalized inequality on Rki definitions are parallel to scalar case: • Lagrange multiplier for fi(x) Ki 0 is vector λi ∈ Rki • Lagrangian L : Rn × Rk1 × · · · × Rkm × Rp → R, is defined as L(x, λ1, · · · , λm, ν) = f0(x) + m X i=1 λT i fi(x) + p X i=1 νihi(x) • dual function g : Rk1 × · · · × Rkm × Rp → R, is defined as g(λ1, . . . , λm, ν) = inf x∈D L(x, λ1, · · · , λm, ν) Duality 5–28
  • 29. lower bound property: if λi K∗ i 0, then g(λ1, . . . , λm, ν) ≤ p⋆ proof: if x̃ is feasible and λ K∗ i 0, then f0(x̃) ≥ f0(x̃) + m X i=1 λT i fi(x̃) + p X i=1 νihi(x̃) ≥ inf x∈D L(x, λ1, . . . , λm, ν) = g(λ1, . . . , λm, ν) minimizing over all feasible x̃ gives p⋆ ≥ g(λ1, . . . , λm, ν) dual problem maximize g(λ1, . . . , λm, ν) subject to λi K∗ i 0, i = 1, . . . , m • weak duality: p⋆ ≥ d⋆ always • strong duality: p⋆ = d⋆ for convex problem with constraint qualification (for example, Slater’s: primal problem is strictly feasible) Duality 5–29
  • 30. Semidefinite program primal SDP (Fi, G ∈ Sk ) minimize cT x subject to x1F1 + · · · + xnFn G • Lagrange multiplier is matrix Z ∈ Sk • Lagrangian L(x, Z) = cT x + tr (Z(x1F1 + · · · + xnFn − G)) • dual function g(Z) = inf x L(x, Z) = − tr(GZ) tr(FiZ) + ci = 0, i = 1, . . . , n −∞ otherwise dual SDP maximize − tr(GZ) subject to Z 0, tr(FiZ) + ci = 0, i = 1, . . . , n p⋆ = d⋆ if primal SDP is strictly feasible (∃x with x1F1 + · · · + xnFn ≺ G) Duality 5–30