Professional Documents
Culture Documents
CONVEX OPTIMIZATION 1 / 34
1 ONE-DIMENSIONAL
2 MULTI-DIMENSIONAL CASE
CONVEX OPTIMIZATION 2 / 34
Consider a smooth function f : Ω → R, where Ω is a nonempty open
subset of Rn . We focus on the problem of finding x0 ∈ Ω so that
f (x0 ) = maxf (x) or f (x0 ) = minf (x), denoted by
x∈Ω x∈Ω
CONVEX OPTIMIZATION 3 / 34
ONE-DIMENSIONAL
Outlines
1 ONE-DIMENSIONAL
2 MULTI-DIMENSIONAL CASE
CONVEX OPTIMIZATION 4 / 34
ONE-DIMENSIONAL
CONVEX OPTIMIZATION 5 / 34
ONE-DIMENSIONAL
and
f (y ) ≥ f (z) + f 0 (z) (y − z) . (3)
Multiply both sides of (2) by θ, (3) by 1 − θ, and add, side by side,
we get
CONVEX OPTIMIZATION 9 / 34
ONE-DIMENSIONAL
CONVEX OPTIMIZATION 10 / 34
MULTI-DIMENSIONAL CASE
Outlines
1 ONE-DIMENSIONAL
2 MULTI-DIMENSIONAL CASE
CONVEX OPTIMIZATION 11 / 34
MULTI-DIMENSIONAL CASE
MULTI-DIMENSIONAL CASE
Let f : Ω → R, where Ω is a nonempty subset of Rn .
First, we set some notations and terminologies.
Definition 6. A nonempty subset Ω ⊂ Rn is said to be an open
convex set in Rn if for all x ∈ Ω and h ∈ Rn , {t ∈ R| x + th ∈ Ω}
is an open interval (a, b) in R, for some −∞ ≤ a < b ≤ ∞.
CONVEX OPTIMIZATION 12 / 34
MULTI-DIMENSIONAL CASE
λ1 0 · · · 0 y1
T
T T
0 λ2 · · · 0 y2
Q h D Q h = y1 y2 · · · yn .. .. .. ..
. . . .
0 0 · · · λn yn
n
X
= λi yi2
i=1
T
which shows that h Ah ≥ 0 when λi ≥ 0, for all i = 1, · · · , n, and it
is ≤ 0 when λi ≤ 0, for all i = 1, · · · , n.
Example 5. Consider again the matrix A of example 4. The
characteristic equation
1−λ 2
det (A − λI2 ) = = (1 − λ) (5 − λ)−4 = λ2 −6λ+5 = (
2 5−λ
shows that A has two positive eigenvalues, 1 and 5. Therefore, A is
positive semidefinite. CONVEX OPTIMIZATION 15 / 34
MULTI-DIMENSIONAL CASE
CONVEX OPTIMIZATION 17 / 34
MULTI-DIMENSIONAL CASE
CONVEX OPTIMIZATION 18 / 34
MULTI-DIMENSIONAL CASE
2x1 − x2 + x3
∇f (x1 , x2 , x3 ) = 2x2 − x1 and
2x3 + x1
2 −1 1
2
∇ f (x1 , x2 , x3 ) = −1 2 0 .
1 0 2
Proposition 11. Let f : Ω → R be of class C 2 , where Ω is a
nonempty open convex set in Rn . For each x ∈ Ω and
h = (h1 , h2 , · · · , hn ) ∈ Rn , let I = { t ∈ R| x + th ∈ Ω} and let
gx,h : I → R be defined. We have
0
gx,h (t) = ∇f (x + th) h và gx,h 00
(t) = hT ∇2 f (x + th) h.
Proof. The differentability of f gives the existence of a function
ε (h), defined on khk < r , such that ε (h) → 0 as h → 0, and
CONVEX OPTIMIZATION 20 / 34
MULTI-DIMENSIONAL CASE
∂f
Now, for each i = 1, 2, · · · , n, let gi (t) = ∂xi
(x + th). Apply the
∂f
above result for ∂x i
in place of f, we get
n n
∂2
X ∂ ∂f X
gi0 (t) = (x + th) hj = (x + th) hj
j=1
∂x j ∂x i
j=1
∂x j ∂xi
and then
Pn 2
00
hi gi0 (t) = ni=1 hi nj=1 ∂x∂j ∂xi (x + th) hj
P P
gx,h (t) = i=1
∂2
Pn
j=1 ∂xj ∂x1 (x + th) hj
Pn 2
j=1 ∂x∂j ∂x2 (x + th) hj
= h1 h2 · · · hn ..
.
∂2
Pn
j=1 ∂xj ∂xn (x + th) h j
as desired.
CONVEX OPTIMIZATION 21 / 34
MULTI-DIMENSIONAL CASE
∂2 ∂2 ∂2
∂x12
(x + th) ∂x2 ∂x1
(x + th) ··· ∂xn ∂x1
(
∂ 2 ∂ 2 ∂2
(x + th) (x + th) ··· (
∂x1 ∂x2 ∂x22 ∂xn ∂x2
= h1 h2 · · · hn .. ..
. .
∂2 ∂2 ∂2
∂x1 ∂xn
(x + th) ∂x2 ∂xn
(x + th) · · · ∂xn2
(x
T
= h ∇2 f (x + th) h.
as desired.
We have the following multidimensional version of Theorem 2.
Theorem 12. Let f : Ω → R be a function of class C 2 on an open
convex set Ω in Rn . The following statements are equivalent.
(i) f is a convex function.
(ii) f (y) ≥ f (x) + ∇f (x)T (y − x), for all x, y ∈ Ω.
(iii) ∇2 f (x) ≥ 0, for all x ∈ Ω.
CONVEX OPTIMIZATION 22 / 34
MULTI-DIMENSIONAL CASE
CONVEX OPTIMIZATION 23 / 34
MULTI-DIMENSIONAL CASE
CONVEX OPTIMIZATION 25 / 34
MULTI-DIMENSIONAL CASE
CONVEX OPTIMIZATION 26 / 34
MULTI-DIMENSIONAL CASE
CONVEX OPTIMIZATION 27 / 34
MULTI-DIMENSIONAL CASE
1 − 21 12
A = − 12 1 0 .
1
2
0 1
∂xi
(x) = qi , for all i = 1, 2, · · · , n, and ∂x∂ j ∂x
q
i
(x) = 0, for all
i, j = 1, 2, · · · , n.
(ii) Since ∂x∂ k (xi xj ) = 0, when k 6= i, j, ∂x∂ k (xk xj ) = xj , when
i = k 6= j, and ∂x∂ k (xk xk ) = 2xk , it follows from the identity
fA (x) = ni=1 nj=1 aij xi xj that, for each k = 1, 2, · · · , n,
P P
CONVEX OPTIMIZATION 29 / 34
MULTI-DIMENSIONAL CASE
P
∂fA
(x) = ni=1 nj=1 aij ∂x∂ k (xi xj ) = ni=1 aik ∂x∂ k (xi xk ) + j6=k aij ∂x∂ k
P P P
∂xk
P P
= ni=1 aik ∂x∂ k (xi xk ) + ni=1 ∂
P P
j6=k a ij ∂xk
(x i x j ) = 2akk xk + i6=k aik
= 2akk xk + i6=k aik xi + j6=k akj xj = 2 ni=1 aki xi
P P P
(4)
which is the k-th row of 2Ax. Hence, ∇fA (x) = 2Ax. Now, by (4),
∂fA
∂xk
(x) is a linear function with q = (2ak1 , 2ak2 , · · · , 2akn ), the k-th
2
fA
row of 2A. By (i), ∂x∂l ∂x k
(x) = alk , for each l, k = 1, 2, · · · , n, which
2
gives ∇ fA (x) = 2Aas desired.
CONVEX OPTIMIZATION 30 / 34
MULTI-DIMENSIONAL CASE
CONVEX OPTIMIZATION 31 / 34
MULTI-DIMENSIONAL CASE
−1 T
Hence, the solution x = AT A A b of the equation ∇f (x) = 0 is
the global minimum of f.
We state this result in the following theorem for future applications.
Theorem 17. Let a matrix A ∈ RN×n , with n ≤ N, so that
rank A = n = min (N, n), and let b ∈ RN . We have
−1 T
AT A A b = arg min kAx − bk2 .
x∈Rn
1 4
It is easy to see that
dim (col (A)) = 2 = rank (A) = min (4, 2),
4 10
AT A = is an invertible matrix, and
10 30
CONVEX OPTIMIZATION 32 / 34
MULTI-DIMENSIONAL CASE
−1
T
−1 T 4 10 17 −1
A A A b= = . Hence
10 30 53 2.1
Now, with
12
1 1
1 2 22
∈ R4×3 .
A=
1 3 32
1 4 42
We havedim (col (A)) =
3 = rank (A) = min (4, 3),
4 10 30
AT A = 10 30 100 is an invertible matrix, and
30 100 354
CONVEX OPTIMIZATION 33 / 34
MULTI-DIMENSIONAL CASE
−1
−1 T 4 10 30 17 2.75
AT A A b = 10 30 100 53 = −1.65 .
30 100 354 183 0.75
Hence
(2.75, −1.65, 0.75) = arg min kAx − bk2 .
x∈R3
CONVEX OPTIMIZATION 34 / 34