Professional Documents
Culture Documents
1
Content of this course (tentative)
Linear programming
Lecture 2
Linear
Dual Theory – Part I
Lecture 3
Linearization
technique Programming
Lecture 7 Unconstrained optimization Lecture 13
Lecture 5 & 6
3
Basic concept
Global optimum. Let 𝑓𝑓(𝑥𝑥) be the objective function, 𝒳𝒳 be the feasible region, and
𝑥𝑥0 ∈ 𝒳𝒳. Then 𝑥𝑥0 is the global optimum if and only if 𝑓𝑓 𝑥𝑥 ≥ 𝑓𝑓 𝑥𝑥0 , ∀𝑥𝑥 ∈ 𝒳𝒳.
Local optimum. Let 𝑓𝑓(𝑥𝑥) be the objective function, 𝒳𝒳 be the feasible region , and
𝑥𝑥0 ∈ 𝒳𝒳. If there is a neighborhood of 𝑥𝑥0 with radius 𝜀𝜀 > 0:
Such that ∀𝑥𝑥 ∈ 𝜒𝜒 ∩ 𝑁𝑁𝜀𝜀 (𝑥𝑥0 ), we have 𝑓𝑓 𝑥𝑥 ≥ 𝑓𝑓 𝑥𝑥0 . Then 𝑥𝑥0 is a local optimum.
4
Pictures from Google image
Recall the single variable optimization
Remark-1:
it is a necessary condition, not a
sufficient condition.
Remark-2:
It is possible that the derivative of 𝑓𝑓(𝑥𝑥) at its 𝑓𝑓 𝑥𝑥 = |𝑥𝑥|
maximum or minimum point does not exist.
6
Recall the single variable optimization
7
Example
Solution:
For 𝑥𝑥 = 0, we have 𝑓𝑓 ′ 0 = 0, 𝑓𝑓 ′′ 0 = 0, 𝑓𝑓 ′′′ 0 = 30
For 𝑥𝑥 = 1, we have 𝑓𝑓 ′ 1 = 0, 𝑓𝑓 ′′ 1 = 20 − 60 + 30 = −10
For 𝑥𝑥 = 3, we have 𝑓𝑓 ′ 3 = 0, 𝑓𝑓 ′′ 3 = 90
Therefore
n=3 • 𝑥𝑥 = 0 is neither a maximum nor a minimum point
• 𝑥𝑥 = 1 is a maximum point with 𝑓𝑓 1 = 2
• 𝑥𝑥 = 3 is a minimum point with 𝑓𝑓 3 = −26
9
Multivariable optimization – Review of mathematics
Gradient
The gradient of a scalar-valued differentiable function 𝑓𝑓 of several variables is
the vector field (or vector-valued function) ∇𝑓𝑓, whose value at a point 𝑥𝑥 is the
vector, whose components are the partial derivatives of 𝑓𝑓 at 𝑥𝑥. i.e.
𝑇𝑇
𝜕𝜕𝜕𝜕 𝜕𝜕𝜕𝜕
∇𝑓𝑓 𝑥𝑥 = ,…,
𝜕𝜕𝑥𝑥1 𝜕𝜕𝑥𝑥𝑁𝑁
𝜕𝜕𝜕𝜕 𝜕𝜕𝜕𝜕
= 6𝑥𝑥1 + 𝑥𝑥2 , = 5 + 𝑥𝑥1
𝜕𝜕𝑥𝑥1 𝜕𝜕𝑥𝑥2
Therefore
𝑇𝑇
∇𝑓𝑓 𝑥𝑥 = 6𝑥𝑥1 + 𝑥𝑥2 , 5 + 𝑥𝑥1
10
Multivariable optimization – Review of mathematics
Hessian
the Hessian matrix or Hessian is a square matrix of second-order
partial derivatives of a scalar-valued function, or scalar field
where
𝑇𝑇
𝜕𝜕𝜕𝜕 𝜕𝜕𝜕𝜕
∇𝑓𝑓 𝑥𝑥 = ,…,
𝜕𝜕𝑥𝑥1 𝜕𝜕𝑥𝑥𝑁𝑁
12
Multivariable optimization – Review of mathematics
Positive/negative-definite matrix
13
Multivariable optimization – Review of mathematics
𝑀𝑀 − 𝜆𝜆𝐼𝐼 = 0
Then
• A matrix 𝑀𝑀 is positive definite if all its eigenvalues are positive.
• A matrix 𝑀𝑀 is negative definite if its eigenvalues are negative.
• A matrix 𝑀𝑀 is positive semi-definite if all its eigenvalues are positive or zero.
• A matrix 𝑀𝑀 is negative semi-definite if its eigenvalues are negative or zero.
14
Multivariable optimization – Review of mathematics
Criterion 2: Let
• The matrix 𝑀𝑀 will be positive definite if and only if all the values 𝑀𝑀1 , 𝑀𝑀2 , … , 𝑀𝑀𝑛𝑛
are positive
• The matrix 𝑀𝑀 will be negative definite if and only if the sign of 𝑀𝑀𝑗𝑗 is −1 𝑗𝑗 for 𝑗𝑗 =
1,2, … , 𝑛𝑛
• If some of the 𝑀𝑀𝑗𝑗 are positive and the remaining 𝑀𝑀𝑗𝑗 are zero, the matrix 𝑀𝑀 will be
positive semidefinite
15
Example
1 0
Consider matrix 𝑀𝑀 = , we try to prove that it is positive definite in three ways.
0 1
1. By definition
For any non-zero vector 𝑧𝑧 = [𝑥𝑥, 𝑦𝑦]𝑇𝑇 , we have
𝑇𝑇 1 0 𝑥𝑥 𝑥𝑥 2 2
𝑧𝑧 𝑀𝑀𝑀𝑀 = 𝑥𝑥, 𝑦𝑦 = 𝑥𝑥 𝑦𝑦 = 𝑥𝑥 + 𝑦𝑦 >0
0 1 𝑦𝑦 𝑦𝑦
1 0
3. 𝑀𝑀1 = 1, 𝑀𝑀2 = = 1.
0 1
16
Multivariable optimization
17
Pictures from Google image
Multivariable optimization
𝑯𝑯(𝒙𝒙∗ )
18
Multivariable optimization
Sufficient condition: If 𝑓𝑓(𝑥𝑥) is twice-differentiable at 𝑥𝑥 ∗ , ∇𝑓𝑓 𝑥𝑥 ∗ = 0 and its
Hessian 𝐻𝐻(𝑥𝑥 ∗ ) is positive definite, then 𝑥𝑥 = 𝑥𝑥 ∗ is a strict local minimum point.
∇𝑓𝑓 𝑥𝑥 ∗ = 0
20
Example
21
Multivariable optimization
Necessary and sufficient condition: If 𝑓𝑓(𝑥𝑥) is twice-differentiable at 𝑥𝑥 ∗ and is
a convex function, then 𝑥𝑥 ∗ is a global minimum if and only if ∇𝑓𝑓 𝑥𝑥 ∗ = 0.
Proof:
⇒ If 𝑥𝑥 ∗ is a global minimum, then it is a local minimum. According to the first-
order necessary condition, we have ∇𝑓𝑓 𝑥𝑥 ∗ = 0.
22
Summary
First-order necessary condition: If 𝑓𝑓(𝑥𝑥) has an extreme point at 𝑥𝑥 = 𝑥𝑥 ∗ ,
and its gradient exists at point 𝑥𝑥 ∗ , then ∇𝑓𝑓 𝑥𝑥 ∗ = 0.
𝒇𝒇(𝒙𝒙) is optimal → ?
Second-order necessary condition: If 𝑓𝑓(𝑥𝑥) has a minimum point at 𝑥𝑥 = 𝑥𝑥 ∗ ,
and it is twice-differentiable at 𝑥𝑥 ∗ , then ∇𝑓𝑓 𝑥𝑥 ∗ = 0 and its Hessian 𝐻𝐻(𝑥𝑥 ∗ )
is semi-positive definite.
?→ 𝒇𝒇(𝒙𝒙) is optimal
Sufficient condition: If 𝑓𝑓(𝑥𝑥) is twice-differentiable at 𝑥𝑥 ∗ , ∇𝑓𝑓 𝑥𝑥 ∗ = 0 and
its Hessian 𝐻𝐻(𝑥𝑥 ∗ ) is positive definite, then 𝑥𝑥 = 𝑥𝑥 ∗ is a strict minimum point.
23
Thanks!
24