Professional Documents
Culture Documents
Convex Problems, Separation Theorems: September 17, 2008
Convex Problems, Separation Theorems: September 17, 2008
Outline
• Duality
Motivation
Visualization of Primal-Dual Framework
Examples
Convex Optimization 1
Lecture 7
Some Terminology
x0 ∈ H ⇐⇒ a T x0 = b
Convex Optimization 2
Lecture 7
sup aT z ≤ aT x0
z∈C
Convex Optimization 3
Lecture 7
Convex Optimization 4
Lecture 7
sup aT x ≤ inf aT z
x∈C z∈D
Convex Optimization 5
Lecture 7
sup aT x ≤ inf aT z
x∈C z∈D
Convex Optimization 6
Lecture 7
Convex Optimization 7
Lecture 7
Duality Theory
Convex Optimization 8
Lecture 7
• Investigate:
• Is there a general relation between the primal and its associated dual
problem?
• Under which conditions the primal and the dual problems have the
same optimal values?
• Under which conditions the primal and dual optimal solutions exist?
• What are the relations between primal and dual optimal solutions?
Convex Optimization 9
Lecture 7
Convex Optimization 10
Lecture 7
Geometric Primal
minimize w
subject to (0, w) ∈ V
Convex Optimization 11
Lecture 7
Nonvertical Hyperplanes
A hyperplane in Rm × R: {(u, w) | µT u + µ0w = ξ}, µ ∈ Rm, µ0, ξ ∈ R
• We say that a hyperplane is nonvertical when µ0 6= 0
• Let Hµ,ξ denote a nonvertical hyperplane in Rm × R, i.e.,
Hµ,ξ = {(u, w) | µT u + w = ξ} with µ ∈ Rm, ξ ∈ R
• Let q(µ) be the minimum value of µT u + w for (u, w) ∈ V , i.e.,
q(µ) = inf (u,w)∈V {µT u + w}
• A nonvertical hyperplane Hµ,ξ̂ supports a set V when ξ̂ = q(µ)
Convex Optimization 12
Lecture 7
maximize q(µ)
subject to µ ∈ Rm
• Note: q(µ) = inf (u,w)∈V {µT u + w}, q(µ) = −∞ for µ 6 0
Convex Optimization 13
Lecture 7
Observations
Convex Optimization 14
Lecture 7
Convex Optimization 15
Lecture 7
Strong Duality
Convex Optimization 16
Lecture 7
minimize f (x)
subject to gj (x) ≤ 0, j = 1, . . . , m
x∈X
Convex Optimization 17
Lecture 7
General Case
Primal Problem (not necessarily convex)
minimize f (x)
subject to gj (x) ≤ 0, j = 1, . . . , m
hj (x) = 0, j = 1, . . . , p
x∈X
variable x ∈ Rn, feasible, optimal value f ∗ > −∞
Lagrangian Function: L : Rn × Rm × Rp → R given by
m
X p
X
L(x, µ, λ) = f (x) + µj gj (x) + λj hj (x)
j=1 j=1
Convex Optimization 18
Lecture 7
Dual Problem
Lagrangian Function:
T T
q(µ, λ) = inf x∈X L(x, µ, λ) = inf x∈X f (x) + µ g(x) + λ h(x)
The infimum above has an implicit constraint on the primal problem domain
• Dual Problem:
maximize q(µ, λ)
subject to µ 0, λ ∈ Rp
Convex Optimization 19
Lecture 7
T 1 T
∇xL(x, λ) = 2x + A λ = 0 =⇒ xλ = − A λ
2
1 T
q(λ) = L(xλ, λ) = − λ AAT λ − bT λ
4
a concave function of λ
Lower Bound Property: −(1/4)λT AAT λ − bT λ ≤ f ∗ for all λ
Convex Optimization 20
Lecture 7
Standard Form LP
minimize cT x
subject to Ax = b, x0
Dual Function:
• Lagrangian is:
L(x, µ, λ) = cT x + λT (Ax − b) − µT x
= −bT λ + (c + AT λ − µ)T x
• L is linear in x, hence
(
−bT λ when AT λ − µ + c = 0
q(µ, λ) = inf L(x, µ, λ) =
x −∞ otherwise
Convex Optimization 21
Lecture 7
Two-Way Partitioning
minimize xT W x
subject to x2i = 1, i = 1, . . . , n
• A nonconvex problem: feasible set contains 2n discrete points
• Interpretation: partition {1, . . . , n} in two sets; Wij is cost of assigning
i, j to the same set; −Wij is cost of assigning to different sets
Dual Function:
( )
q(λ) = inf xT W x + λi(x2i − 1) = inf xT [W + diag(λ)]x − 1T λ
X
x x
i (
−1T λ when W + diag(λ) 0
=
−∞ otherwise
Convex Optimization 22
Lecture 7
maximize −1T λ
subject to W + diag(λ) 0
Convex Optimization 23
Lecture 7
Convex Optimization 24
Lecture 7
• Discrete Optimization:
minimize −x
subject to x ≤ 1, x ∈ {0, 2}
• Convex Optimization:
√
− x1 x2
minimize e
subject to x21 ≤ 0, x1 ≥ 0, x2 ≥ 0
Convex Optimization 25