Professional Documents
Culture Documents
Lecture 12
Lecture 12
Vassilis Kekatos
Lecture 10 V. Kekatos 1
Unconstrained minimization
p∗ := min f (x)
x
∇f (x∗ ) = 0
1
0 ≤ f (x) − p∗ ≤ k∇f (x)k22
2m
Lecture 10 V. Kekatos 2
Examples
Lecture 10 V. Kekatos 3
Descent method
Recall for convex f , we have f (xt+1 ) ≥ f (xt ) + (∇f (x))> (xt+1 − xt ). Then:
f (xt+1 ) < f (xt ) ⇒ descent direction satisfies (∇f (xt ))> ∆xt < 0
Lecture 10 V. Kekatos 4
Gradient descent
Lecture 10 V. Kekatos 5
Example
where M > 0
• initialize at x0 = (M, 1)
• slow, zig-zagging if M 1 or M 1
Lecture 10 V. Kekatos 6
Example 2
For m = 100 and n = 50, use gradient method (exact line search)
m
X
min c> x − log(a>
i x − bi )
x
i=1
Figure: Function value convergence for gradient method [Z.-Q. Luo’s slides]
Lecture 10 V. Kekatos 7
Steepest descent direction
f (x + z) ≈ f (x) + ∇f (x)> z
√
Quadratic norm kzkP := z> Pz for some P 0
−1/2
zsd = − ∇f (x)> P−1 ∇f (x) P−1 ∇f (x)
Lecture 10 V. Kekatos 8
Geometric interpretation
move as far as possible in direction −∇f (x), while staying inside the unit ball
Lecture 10 V. Kekatos 9
Choosing the norm
• steepest descent with backtracking line search for two quadratic norms
• ellipses show {x : kx − xt kP = 1}
Lecture 10 V. Kekatos 10
Pure Newton step and interpretations
Newton update: x+ = x + v
∇f (x) + ∇2 f (x)(x+ − x) = 0
Lecture 10 V. Kekatos 11
Global behavior of Newton iterations
Lecture 10 V. Kekatos 12
Newton method
• affine invariance:
Newton iterates for minx f (x) and minz f (Tz) for invertible T
are equivalent and xt = Tzt
Lecture 10 V. Kekatos 13
Convergence results
Lecture 10 V. Kekatos 14
Example
10,000 100,000
X X
f (x) = − log(1 − x2n ) − log(bi − a>
i x)
n=1 i=1
Lecture 10 V. Kekatos 15
Minimization with linear equality constraints
min{f (x) : Ax = b}
x
min f (Fz + x0 )
z
s.to A(x + v) = b
Lecture 10 V. Kekatos 16