Professional Documents
Culture Documents
Lecture 12
Quadratic Forms, Positive Definiteness and Singular Value
Decomposition
Lecture 12 1/26
Outline
1 Quadratic Forms
Principal Axis Theorem
2 Positive Definiteness
Negative Definiteness
Singular Case and Saddle Point
Higher Dimensions
Quadratic Form for a Random Function
Lecture 12 2/26
Quadrartic Forms
Lecture 12 3/26
Quadrartic Forms
Lecture 12 3/26
Quadrartic Forms
Lecture 12 3/26
Quadratic Forms
If A is not symmetric, we can take off-diagonal terms together in pairs and write the
result as a sum of two equal terms, which is equivalent to replacing A with (A+AT )/2,
as illustrated in the example below.
Lecture 12 4/26
Quadratic Forms
If A is not symmetric, we can take off-diagonal terms together in pairs and write the
result as a sum of two equal terms, which is equivalent to replacing A with (A+AT )/2,
as illustrated in the example below.
Example: Consider
3 4 x1
xT Ax = [x1 x2 ]
6 2 x2
Lecture 12 4/26
Quadratic Forms
If A is not symmetric, we can take off-diagonal terms together in pairs and write the
result as a sum of two equal terms, which is equivalent to replacing A with (A+AT )/2,
as illustrated in the example below.
Example: Consider
3 4 x1
xT Ax = [x1 x2 ] = 3x21 + 4x1 x2 + 6x1 x2 + 2x22
6 2 x2
Lecture 12 4/26
Quadratic Forms
If A is not symmetric, we can take off-diagonal terms together in pairs and write the
result as a sum of two equal terms, which is equivalent to replacing A with (A+AT )/2,
as illustrated in the example below.
Example: Consider
3 4 x1
xT Ax = [x1 x2 ] = 3x21 + 4x1 x2 + 6x1 x2 + 2x22
6 2 x2
= 3x21 + 10x1 x2 + 2x22
Lecture 12 4/26
Quadratic Forms
If A is not symmetric, we can take off-diagonal terms together in pairs and write the
result as a sum of two equal terms, which is equivalent to replacing A with (A+AT )/2,
as illustrated in the example below.
Example: Consider
3 4 x1
xT Ax = [x1 x2 ] = 3x21 + 4x1 x2 + 6x1 x2 + 2x22
6 2 x2
= 3x21 + 10x1 x2 + 2x22 = 3x21 + 5x1 x2 + 5x1 x2 + 2x22 .
Lecture 12 4/26
Quadratic Forms
If A is not symmetric, we can take off-diagonal terms together in pairs and write the
result as a sum of two equal terms, which is equivalent to replacing A with (A+AT )/2,
as illustrated in the example below.
Example: Consider
3 4 x1
xT Ax = [x1 x2 ] = 3x21 + 4x1 x2 + 6x1 x2 + 2x22
6 2 x2
= 3x21 + 10x1 x2 + 2x22 = 3x21 + 5x1 x2 + 5x1 x2 + 2x22 .
Then,
T 3 5 x1
x Cx = [x1 x2 ]
5 2 x2
Lecture 12 4/26
Quadratic Forms
If A is not symmetric, we can take off-diagonal terms together in pairs and write the
result as a sum of two equal terms, which is equivalent to replacing A with (A+AT )/2,
as illustrated in the example below.
Example: Consider
3 4 x1
xT Ax = [x1 x2 ] = 3x21 + 4x1 x2 + 6x1 x2 + 2x22
6 2 x2
= 3x21 + 10x1 x2 + 2x22 = 3x21 + 5x1 x2 + 5x1 x2 + 2x22 .
Then,
A + AT
T 3 5 x1
x Cx = [x1 x2 ] = xT Ax, C= .
5 2 x2 2
Lecture 12 4/26
Quadratic Form – Principal Axis Theorem
Statement:
The substitution
x = Xy, (3)
Lecture 12 5/26
Quadratic Form – Principal Axis Theorem
Statement:
The substitution
x = Xy, (3)
Lecture 12 5/26
Quadratic Form – Principal Axis Theorem
Statement:
The substitution
x = Xy, (3)
Lecture 12 5/26
Quadratic Form – Principal Axis Theorem
Statement:
The substitution
x = Xy, (3)
Lecture 12 5/26
Quadratic Form – Principal Axis Theorem
Proof:
It is known that
D = X −1 AX = X T AX, X = [x1 x2 . . . xn ]
T T −1
⇒ Q = x Ax = x XDX x = xT XDX T x.
Let
X T x , y ⇒ X −1 x = y ⇒ x = Xy
T T
⇒ xT X = X T x = X −1 x = yT .
Hence,
Lecture 12 6/26
Quadratic Form – Principal Axis Theorem
Proof:
It is known that
D = X −1 AX = X T AX, X = [x1 x2 . . . xn ]
T T −1
⇒ Q = x Ax = x XDX x = xT XDX T x.
Let
X T x , y ⇒ X −1 x = y ⇒ x = Xy
T T
⇒ xT X = X T x = X −1 x = yT .
Hence,
Lecture 12 6/26
Quadratic Form – Principal Axis Theorem
Proof:
It is known that
D = X −1 AX = X T AX, X = [x1 x2 . . . xn ]
T T −1
⇒ Q = x Ax = x XDX x = xT XDX T x.
Let
X T x , y ⇒ X −1 x = y ⇒ x = Xy
T T
⇒ xT X = X T x = X −1 x = yT .
Hence,
Lecture 12 6/26
Quadratic Form – Principal Axis Theorem
Proof:
It is known that
D = X −1 AX = X T AX, X = [x1 x2 . . . xn ]
T T −1
⇒ Q = x Ax = x XDX x = xT XDX T x.
Let
X T x , y ⇒ X −1 x = y ⇒ x = Xy
T T
⇒ xT X = X T x = X −1 x = yT .
Hence,
Lecture 12 6/26
Quadratic Form – Principal Axis Theorem
Proof:
It is known that
D = X −1 AX = X T AX, X = [x1 x2 . . . xn ]
T T −1
⇒ Q = x Ax = x XDX x = xT XDX T x.
Let
X T x , y ⇒ X −1 x = y ⇒ x = Xy
T T
⇒ xT X = X T x = X −1 x = yT .
Hence,
Lecture 12 6/26
Quadratic Form – Principal Axis Theorem
Proof:
It is known that
D = X −1 AX = X T AX, X = [x1 x2 . . . xn ]
T T −1
⇒ Q = x Ax = x XDX x = xT XDX T x.
Let
X T x , y ⇒ X −1 x = y ⇒ x = Xy
T T
⇒ xT X = X T x = X −1 x = yT .
Hence,
Lecture 12 6/26
Quadratic Form – Principal Axis Theorem
Proof:
It is known that
D = X −1 AX = X T AX, X = [x1 x2 . . . xn ]
T T −1
⇒ Q = x Ax = x XDX x = xT XDX T x.
Let
X T x , y ⇒ X −1 x = y ⇒ x = Xy
T T
⇒ xT X = X T x = X −1 x = yT .
Hence,
Lecture 12 6/26
Quadratic Form – Principal Axis Theorem
Proof:
It is known that
D = X −1 AX = X T AX, X = [x1 x2 . . . xn ]
T T −1
⇒ Q = x Ax = x XDX x = xT XDX T x.
Let
X T x , y ⇒ X −1 x = y ⇒ x = Xy
T T
⇒ xT X = X T x = X −1 x = yT .
Hence,
Lecture 12 6/26
Quadratic Form – Principal Axis Theorem
Proof:
It is known that
D = X −1 AX = X T AX, X = [x1 x2 . . . xn ]
T T −1
⇒ Q = x Ax = x XDX x = xT XDX T x.
Let
X T x , y ⇒ X −1 x = y ⇒ x = Xy
T T
⇒ xT X = X T x = X −1 x = yT .
Hence,
Lecture 12 6/26
Quadratic Form – Principal Axis Theorem
Proof:
It is known that
D = X −1 AX = X T AX, X = [x1 x2 . . . xn ]
T T −1
⇒ Q = x Ax = x XDX x = xT XDX T x.
Let
X T x , y ⇒ X −1 x = y ⇒ x = Xy
T T
⇒ xT X = X T x = X −1 x = yT .
Hence,
Lecture 12 6/26
Quadratic Form – Principal Axis Theorem
Proof:
It is known that
D = X −1 AX = X T AX, X = [x1 x2 . . . xn ]
T T −1
⇒ Q = x Ax = x XDX x = xT XDX T x.
Let
X T x , y ⇒ X −1 x = y ⇒ x = Xy
T T
⇒ xT X = X T x = X −1 x = yT .
Hence,
Lecture 12 6/26
Quadratic Form – Principal Axis Theorem
Proof:
It is known that
D = X −1 AX = X T AX, X = [x1 x2 . . . xn ]
T T −1
⇒ Q = x Ax = x XDX x = xT XDX T x.
Let
X T x , y ⇒ X −1 x = y ⇒ x = Xy
T T
⇒ xT X = X T x = X −1 x = yT .
Hence,
Lecture 12 6/26
Quadratic Form – Principal Axis Theorem
Proof:
It is known that
D = X −1 AX = X T AX, X = [x1 x2 . . . xn ]
T T −1
⇒ Q = x Ax = x XDX x = xT XDX T x.
Let
X T x , y ⇒ X −1 x = y ⇒ x = Xy
T T
⇒ xT X = X T x = X −1 x = yT .
Hence,
Lecture 12 6/26
Positive Definiteness
b2 b2
f (x1 , x2 ) = ax21 + x22 + 2bx1 x2 − x22 + cx22
a a
2
b2
2 b 2 b
= a x1 + 2 x2 + 2 x1 x2 + c − x22
a a a
2
b2
b
= a x1 + x2 + c − x22 . (8)
a a
1 Origin is the stationary point of the function, where the function has its minimum.
Lecture 12 7/26
Positive Definiteness
b2 b2
f (x1 , x2 ) = ax21 + x22 + 2bx1 x2 − x22 + cx22
a a
2
b2
2 b 2 b
= a x1 + 2 x2 + 2 x1 x2 + c − x22
a a a
2
b2
b
= a x1 + x2 + c − x22 . (8)
a a
1 Origin is the stationary point of the function, where the function has its minimum.
Lecture 12 7/26
Positive Definiteness
b2 b2
f (x1 , x2 ) = ax21 + x22 + 2bx1 x2 − x22 + cx22
a a
2
b2
2 b 2 b
= a x1 + 2 x2 + 2 x1 x2 + c − x22
a a a
2
b2
b
= a x1 + x2 + c − x22 . (8)
a a
1 Origin is the stationary point of the function, where the function has its minimum.
Lecture 12 7/26
Positive Definiteness
b2 b2
f (x1 , x2 ) = ax21 + x22 + 2bx1 x2 − x22 + cx22
a a
2
b2
2 b 2 b
= a x1 + 2 x2 + 2 x1 x2 + c − x22
a a a
2
b2
b
= a x1 + x2 + c − x22 . (8)
a a
1 Origin is the stationary point of the function, where the function has its minimum.
Lecture 12 7/26
Positive Definiteness
b2 b2
f (x1 , x2 ) = ax21 + x22 + 2bx1 x2 − x22 + cx22
a a
2
b2
2 b 2 b
= a x1 + 2 x2 + 2 x1 x2 + c − x22
a a a
2
b2
b
= a x1 + x2 + c − x22 . (8)
a a
1 Origin is the stationary point of the function, where the function has its minimum.
Lecture 12 7/26
Positive Definiteness
b2 b2
f (x1 , x2 ) = ax21 + x22 + 2bx1 x2 − x22 + cx22
a a
2
b2
2 b 2 b
= a x1 + 2 x2 + 2 x1 x2 + c − x22
a a a
2
b2
b
= a x1 + x2 + c − x22 . (8)
a a
1 Origin is the stationary point of the function, where the function has its minimum.
Lecture 12 7/26
Positive Definiteness
Because of the complete square in the two terms, f (x1 , x2 ) > 0 iff
b2
a > 0, c− > 0 ⇒ ac − b2 > 0 ⇒ ac > b2 ⇒ c > 0. (9)
a
For coefficients a, b, c defined as in (9), f (x1 , x2 ) > 0 at all points except at the
origin where it has a minimum2 , i.e.,
∂f ∂f
= 2ax1 + 2bx2 = 0, = 2bx1 + 2cx2 = 0,
∂x1 ∂x2
(0,0) (0,0) (10)
∂2f ∂2f 3 ∂ f
2
∂2f
= 2a > 0, = = 2b, = 2c > 0.
∂x21 ∂x1 ∂x2 ∂x2 ∂x1 ∂x22
Because of the complete square in the two terms, f (x1 , x2 ) > 0 iff
b2
a > 0, c− > 0 ⇒ ac − b2 > 0 ⇒ ac > b2 ⇒ c > 0. (9)
a
For coefficients a, b, c defined as in (9), f (x1 , x2 ) > 0 at all points except at the
origin where it has a minimum2 , i.e.,
∂f ∂f
= 2ax1 + 2bx2 = 0, = 2bx1 + 2cx2 = 0,
∂x1 ∂x2
(0,0) (0,0) (10)
∂2f ∂2f 3 ∂ f
2
∂2f
= 2a > 0, = = 2b, = 2c > 0.
∂x21 ∂x1 ∂x2 ∂x2 ∂x1 ∂x22
Because of the complete square in the two terms, f (x1 , x2 ) > 0 iff
b2
a > 0, c− > 0 ⇒ ac − b2 > 0 ⇒ ac > b2 ⇒ c > 0. (9)
a
For coefficients a, b, c defined as in (9), f (x1 , x2 ) > 0 at all points except at the
origin where it has a minimum2 , i.e.,
∂f ∂f
= 2ax1 + 2bx2 = 0, = 2bx1 + 2cx2 = 0,
∂x1 ∂x2
(0,0) (0,0) (10)
∂2f ∂2f 3 ∂ f
2
∂2f
= 2a > 0, = = 2b, = 2c > 0.
∂x21 ∂x1 ∂x2 ∂x2 ∂x1 ∂x22
Because of the complete square in the two terms, f (x1 , x2 ) > 0 iff
b2
a > 0, c− > 0 ⇒ ac − b2 > 0 ⇒ ac > b2 ⇒ c > 0. (9)
a
For coefficients a, b, c defined as in (9), f (x1 , x2 ) > 0 at all points except at the
origin where it has a minimum2 , i.e.,
∂f ∂f
= 2ax1 + 2bx2 = 0, = 2bx1 + 2cx2 = 0,
∂x1 ∂x2
(0,0) (0,0) (10)
∂2f ∂2f 3 ∂ f
2
∂2f
= 2a > 0, = = 2b, = 2c > 0.
∂x21 ∂x1 ∂x2 ∂x2 ∂x1 ∂x22
Because of the complete square in the two terms, f (x1 , x2 ) > 0 iff
b2
a > 0, c− > 0 ⇒ ac − b2 > 0 ⇒ ac > b2 ⇒ c > 0. (9)
a
For coefficients a, b, c defined as in (9), f (x1 , x2 ) > 0 at all points except at the
origin where it has a minimum2 , i.e.,
∂f ∂f
= 2ax1 + 2bx2 = 0, = 2bx1 + 2cx2 = 0,
∂x1 ∂x2
(0,0) (0,0) (10)
∂2f ∂2f 3 ∂ f
2
∂2f
= 2a > 0, = = 2b, = 2c > 0.
∂x21 ∂x1 ∂x2 ∂x2 ∂x1 ∂x22
then
2
b2
b
f (x1 , x2 ) = ax21 + 2bx1 x2 + cx22 = a x1 + x2 + c− x22 < 0 (13)
a a
at all points except at the origin, where it has a maximum, and is called negative
definite.
Lecture 12 9/26
Negative Definiteness
Test for Maximum:
f (x1 , x2 ) has a maximum whenever −f (x1 , x2 ) has a minimum. Hence,
then
2
b2
b
f (x1 , x2 ) = ax21 + 2bx1 x2 + cx22 = a x1 + x2 + c− x22 < 0 (13)
a a
at all points except at the origin, where it has a maximum, and is called negative
definite.
Lecture 12 9/26
Negative Definiteness
Test for Maximum:
f (x1 , x2 ) has a maximum whenever −f (x1 , x2 ) has a minimum. Hence,
then
2
b2
b
f (x1 , x2 ) = ax21 + 2bx1 x2 + cx22 = a x1 + x2 + c− x22 < 0 (13)
a a
at all points except at the origin, where it has a maximum, and is called negative
definite.
Lecture 12 9/26
Negative Definiteness
Test for Maximum:
f (x1 , x2 ) has a maximum whenever −f (x1 , x2 ) has a minimum. Hence,
then
2
b2
b
f (x1 , x2 ) = ax21 + 2bx1 x2 + cx22 = a x1 + x2 + c− x22 < 0 (13)
a a
at all points except at the origin, where it has a maximum, and is called negative
definite.
Lecture 12 9/26
Negative Definiteness
Test for Maximum:
f (x1 , x2 ) has a maximum whenever −f (x1 , x2 ) has a minimum. Hence,
then
2
b2
b
f (x1 , x2 ) = ax21 + 2bx1 x2 + cx22 = a x1 + x2 + c− x22 < 0 (13)
a a
at all points except at the origin, where it has a maximum, and is called negative
definite.
Lecture 12 9/26
Negative Definiteness
Test for Maximum:
f (x1 , x2 ) has a maximum whenever −f (x1 , x2 ) has a minimum. Hence,
then
2
b2
b
f (x1 , x2 ) = ax21 + 2bx1 x2 + cx22 = a x1 + x2 + c− x22 < 0 (13)
a a
at all points except at the origin, where it has a maximum, and is called negative
definite.
Lecture 12 9/26
Singular Case and Saddle Point
which is:
2
1. positive semidefinite if a > 0 ⇒ c > 0, ∵ c = ba ,
2
2. negative semidefinite if a < 0 ⇒ c < 0, ∵ c = ba .
The prefix semi indicates that f (x1 , x2 ) can be zero at points other than the
origin, e.g., at x1 = b, x2 = −a.
If ac < b2 or ac − b2 < 0, which can also happen when a and c have opposite
signs, then the stationary point (in general the critical point4 ) is called the saddle
point of f (x1 , x2 ), which is neither positive (semi)definite nor negative (semi)
definite.
which is:
2
1. positive semidefinite if a > 0 ⇒ c > 0, ∵ c = ba ,
2
2. negative semidefinite if a < 0 ⇒ c < 0, ∵ c = ba .
The prefix semi indicates that f (x1 , x2 ) can be zero at points other than the
origin, e.g., at x1 = b, x2 = −a.
If ac < b2 or ac − b2 < 0, which can also happen when a and c have opposite
signs, then the stationary point (in general the critical point4 ) is called the saddle
point of f (x1 , x2 ), which is neither positive (semi)definite nor negative (semi)
definite.
which is:
2
1. positive semidefinite if a > 0 ⇒ c > 0, ∵ c = ba ,
2
2. negative semidefinite if a < 0 ⇒ c < 0, ∵ c = ba .
The prefix semi indicates that f (x1 , x2 ) can be zero at points other than the
origin, e.g., at x1 = b, x2 = −a.
If ac < b2 or ac − b2 < 0, which can also happen when a and c have opposite
signs, then the stationary point (in general the critical point4 ) is called the saddle
point of f (x1 , x2 ), which is neither positive (semi)definite nor negative (semi)
definite.
which is:
2
1. positive semidefinite if a > 0 ⇒ c > 0, ∵ c = ba ,
2
2. negative semidefinite if a < 0 ⇒ c < 0, ∵ c = ba .
The prefix semi indicates that f (x1 , x2 ) can be zero at points other than the
origin, e.g., at x1 = b, x2 = −a.
If ac < b2 or ac − b2 < 0, which can also happen when a and c have opposite
signs, then the stationary point (in general the critical point4 ) is called the saddle
point of f (x1 , x2 ), which is neither positive (semi)definite nor negative (semi)
definite.
which is:
2
1. positive semidefinite if a > 0 ⇒ c > 0, ∵ c = ba ,
2
2. negative semidefinite if a < 0 ⇒ c < 0, ∵ c = ba .
The prefix semi indicates that f (x1 , x2 ) can be zero at points other than the
origin, e.g., at x1 = b, x2 = −a.
If ac < b2 or ac − b2 < 0, which can also happen when a and c have opposite
signs, then the stationary point (in general the critical point4 ) is called the saddle
point of f (x1 , x2 ), which is neither positive (semi)definite nor negative (semi)
definite.
which is:
2
1. positive semidefinite if a > 0 ⇒ c > 0, ∵ c = ba ,
2
2. negative semidefinite if a < 0 ⇒ c < 0, ∵ c = ba .
The prefix semi indicates that f (x1 , x2 ) can be zero at points other than the
origin, e.g., at x1 = b, x2 = −a.
If ac < b2 or ac − b2 < 0, which can also happen when a and c have opposite
signs, then the stationary point (in general the critical point4 ) is called the saddle
point of f (x1 , x2 ), which is neither positive (semi)definite nor negative (semi)
definite.
in terms of which, we can rewrite the conditions for definiteness, by realizing that
as follows:
1. Positive definiteness: |A1 | > 0, |A| > 0 ⇒ λ1 , λ2 > 0 because c > 0 and |A| = λ1 λ2 >
0, trace(A) = λ1 + λ2 = a + c > 0.
2. Negative definiteness: |A1 | < 0, |A| > 0 ⇒ λ1 , λ2 < 0 because c < 0 and |A| = λ1 λ2 <
0, trace(A) = λ1 + λ2 = a + c < 0.
3. Positive semidefiniteness: |A1 | > 0, |A| = 0 ⇒ λ1 = 0, λ2 > 0 or λ1 > 0, λ2 = 0
because c > 0 and |A| = λ1 λ2 = 0, trace(A) = λ1 + λ2 = a + c > 0.
4. Negative semidefiniteness: |A1 | < 0, |A| = 0 ⇒ λ1 = 0, λ2 < 0 or λ1 < 0, λ2 = 0
because c < 0 and |A| = λ1 λ2 = 0, trace(A) = λ1 + λ2 = a + c < 0.
5. Saddle point: |A| < 0, which indicates that the two eigenvalues have opposite signs.
The corresponding matrix A is called positive definite, negative definite, positive semidefinite,
negative semidefinite, and indefinite respectively.
Lecture 12 11/26
Positive Definiteness in Higher Dimensions
The function
in terms of which, we can rewrite the conditions for definiteness, by realizing that
as follows:
1. Positive definiteness: |A1 | > 0, |A| > 0 ⇒ λ1 , λ2 > 0 because c > 0 and |A| = λ1 λ2 >
0, trace(A) = λ1 + λ2 = a + c > 0.
2. Negative definiteness: |A1 | < 0, |A| > 0 ⇒ λ1 , λ2 < 0 because c < 0 and |A| = λ1 λ2 <
0, trace(A) = λ1 + λ2 = a + c < 0.
3. Positive semidefiniteness: |A1 | > 0, |A| = 0 ⇒ λ1 = 0, λ2 > 0 or λ1 > 0, λ2 = 0
because c > 0 and |A| = λ1 λ2 = 0, trace(A) = λ1 + λ2 = a + c > 0.
4. Negative semidefiniteness: |A1 | < 0, |A| = 0 ⇒ λ1 = 0, λ2 < 0 or λ1 < 0, λ2 = 0
because c < 0 and |A| = λ1 λ2 = 0, trace(A) = λ1 + λ2 = a + c < 0.
5. Saddle point: |A| < 0, which indicates that the two eigenvalues have opposite signs.
The corresponding matrix A is called positive definite, negative definite, positive semidefinite,
negative semidefinite, and indefinite respectively.
Lecture 12 11/26
Positive Definiteness in Higher Dimensions
The function
in terms of which, we can rewrite the conditions for definiteness, by realizing that
as follows:
1. Positive definiteness: |A1 | > 0, |A| > 0 ⇒ λ1 , λ2 > 0 because c > 0 and |A| = λ1 λ2 >
0, trace(A) = λ1 + λ2 = a + c > 0.
2. Negative definiteness: |A1 | < 0, |A| > 0 ⇒ λ1 , λ2 < 0 because c < 0 and |A| = λ1 λ2 <
0, trace(A) = λ1 + λ2 = a + c < 0.
3. Positive semidefiniteness: |A1 | > 0, |A| = 0 ⇒ λ1 = 0, λ2 > 0 or λ1 > 0, λ2 = 0
because c > 0 and |A| = λ1 λ2 = 0, trace(A) = λ1 + λ2 = a + c > 0.
4. Negative semidefiniteness: |A1 | < 0, |A| = 0 ⇒ λ1 = 0, λ2 < 0 or λ1 < 0, λ2 = 0
because c < 0 and |A| = λ1 λ2 = 0, trace(A) = λ1 + λ2 = a + c < 0.
5. Saddle point: |A| < 0, which indicates that the two eigenvalues have opposite signs.
The corresponding matrix A is called positive definite, negative definite, positive semidefinite,
negative semidefinite, and indefinite respectively.
Lecture 12 11/26
Positive Definiteness in Higher Dimensions
The function
in terms of which, we can rewrite the conditions for definiteness, by realizing that
as follows:
1. Positive definiteness: |A1 | > 0, |A| > 0 ⇒ λ1 , λ2 > 0 because c > 0 and |A| = λ1 λ2 >
0, trace(A) = λ1 + λ2 = a + c > 0.
2. Negative definiteness: |A1 | < 0, |A| > 0 ⇒ λ1 , λ2 < 0 because c < 0 and |A| = λ1 λ2 <
0, trace(A) = λ1 + λ2 = a + c < 0.
3. Positive semidefiniteness: |A1 | > 0, |A| = 0 ⇒ λ1 = 0, λ2 > 0 or λ1 > 0, λ2 = 0
because c > 0 and |A| = λ1 λ2 = 0, trace(A) = λ1 + λ2 = a + c > 0.
4. Negative semidefiniteness: |A1 | < 0, |A| = 0 ⇒ λ1 = 0, λ2 < 0 or λ1 < 0, λ2 = 0
because c < 0 and |A| = λ1 λ2 = 0, trace(A) = λ1 + λ2 = a + c < 0.
5. Saddle point: |A| < 0, which indicates that the two eigenvalues have opposite signs.
The corresponding matrix A is called positive definite, negative definite, positive semidefinite,
negative semidefinite, and indefinite respectively.
Lecture 12 11/26
Positive Definiteness in Higher Dimensions
The function
in terms of which, we can rewrite the conditions for definiteness, by realizing that
as follows:
1. Positive definiteness: |A1 | > 0, |A| > 0 ⇒ λ1 , λ2 > 0 because c > 0 and |A| = λ1 λ2 >
0, trace(A) = λ1 + λ2 = a + c > 0.
2. Negative definiteness: |A1 | < 0, |A| > 0 ⇒ λ1 , λ2 < 0 because c < 0 and |A| = λ1 λ2 <
0, trace(A) = λ1 + λ2 = a + c < 0.
3. Positive semidefiniteness: |A1 | > 0, |A| = 0 ⇒ λ1 = 0, λ2 > 0 or λ1 > 0, λ2 = 0
because c > 0 and |A| = λ1 λ2 = 0, trace(A) = λ1 + λ2 = a + c > 0.
4. Negative semidefiniteness: |A1 | < 0, |A| = 0 ⇒ λ1 = 0, λ2 < 0 or λ1 < 0, λ2 = 0
because c < 0 and |A| = λ1 λ2 = 0, trace(A) = λ1 + λ2 = a + c < 0.
5. Saddle point: |A| < 0, which indicates that the two eigenvalues have opposite signs.
The corresponding matrix A is called positive definite, negative definite, positive semidefinite,
negative semidefinite, and indefinite respectively.
Lecture 12 11/26
Positive Definiteness in Higher Dimensions
The function
in terms of which, we can rewrite the conditions for definiteness, by realizing that
as follows:
1. Positive definiteness: |A1 | > 0, |A| > 0 ⇒ λ1 , λ2 > 0 because c > 0 and |A| = λ1 λ2 >
0, trace(A) = λ1 + λ2 = a + c > 0.
2. Negative definiteness: |A1 | < 0, |A| > 0 ⇒ λ1 , λ2 < 0 because c < 0 and |A| = λ1 λ2 <
0, trace(A) = λ1 + λ2 = a + c < 0.
3. Positive semidefiniteness: |A1 | > 0, |A| = 0 ⇒ λ1 = 0, λ2 > 0 or λ1 > 0, λ2 = 0
because c > 0 and |A| = λ1 λ2 = 0, trace(A) = λ1 + λ2 = a + c > 0.
4. Negative semidefiniteness: |A1 | < 0, |A| = 0 ⇒ λ1 = 0, λ2 < 0 or λ1 < 0, λ2 = 0
because c < 0 and |A| = λ1 λ2 = 0, trace(A) = λ1 + λ2 = a + c < 0.
5. Saddle point: |A| < 0, which indicates that the two eigenvalues have opposite signs.
The corresponding matrix A is called positive definite, negative definite, positive semidefinite,
negative semidefinite, and indefinite respectively.
Lecture 12 11/26
Positive Definiteness in Higher Dimensions
The function
in terms of which, we can rewrite the conditions for definiteness, by realizing that
as follows:
1. Positive definiteness: |A1 | > 0, |A| > 0 ⇒ λ1 , λ2 > 0 because c > 0 and |A| = λ1 λ2 >
0, trace(A) = λ1 + λ2 = a + c > 0.
2. Negative definiteness: |A1 | < 0, |A| > 0 ⇒ λ1 , λ2 < 0 because c < 0 and |A| = λ1 λ2 <
0, trace(A) = λ1 + λ2 = a + c < 0.
3. Positive semidefiniteness: |A1 | > 0, |A| = 0 ⇒ λ1 = 0, λ2 > 0 or λ1 > 0, λ2 = 0
because c > 0 and |A| = λ1 λ2 = 0, trace(A) = λ1 + λ2 = a + c > 0.
4. Negative semidefiniteness: |A1 | < 0, |A| = 0 ⇒ λ1 = 0, λ2 < 0 or λ1 < 0, λ2 = 0
because c < 0 and |A| = λ1 λ2 = 0, trace(A) = λ1 + λ2 = a + c < 0.
5. Saddle point: |A| < 0, which indicates that the two eigenvalues have opposite signs.
The corresponding matrix A is called positive definite, negative definite, positive semidefinite,
negative semidefinite, and indefinite respectively.
Lecture 12 11/26
Positive Definiteness in Higher Dimensions
The function
in terms of which, we can rewrite the conditions for definiteness, by realizing that
as follows:
1. Positive definiteness: |A1 | > 0, |A| > 0 ⇒ λ1 , λ2 > 0 because c > 0 and |A| = λ1 λ2 >
0, trace(A) = λ1 + λ2 = a + c > 0.
2. Negative definiteness: |A1 | < 0, |A| > 0 ⇒ λ1 , λ2 < 0 because c < 0 and |A| = λ1 λ2 <
0, trace(A) = λ1 + λ2 = a + c < 0.
3. Positive semidefiniteness: |A1 | > 0, |A| = 0 ⇒ λ1 = 0, λ2 > 0 or λ1 > 0, λ2 = 0
because c > 0 and |A| = λ1 λ2 = 0, trace(A) = λ1 + λ2 = a + c > 0.
4. Negative semidefiniteness: |A1 | < 0, |A| = 0 ⇒ λ1 = 0, λ2 < 0 or λ1 < 0, λ2 = 0
because c < 0 and |A| = λ1 λ2 = 0, trace(A) = λ1 + λ2 = a + c < 0.
5. Saddle point: |A| < 0, which indicates that the two eigenvalues have opposite signs.
The corresponding matrix A is called positive definite, negative definite, positive semidefinite,
negative semidefinite, and indefinite respectively.
Lecture 12 11/26
Positive Definiteness in Higher Dimensions
The function
in terms of which, we can rewrite the conditions for definiteness, by realizing that
as follows:
1. Positive definiteness: |A1 | > 0, |A| > 0 ⇒ λ1 , λ2 > 0 because c > 0 and |A| = λ1 λ2 >
0, trace(A) = λ1 + λ2 = a + c > 0.
2. Negative definiteness: |A1 | < 0, |A| > 0 ⇒ λ1 , λ2 < 0 because c < 0 and |A| = λ1 λ2 <
0, trace(A) = λ1 + λ2 = a + c < 0.
3. Positive semidefiniteness: |A1 | > 0, |A| = 0 ⇒ λ1 = 0, λ2 > 0 or λ1 > 0, λ2 = 0
because c > 0 and |A| = λ1 λ2 = 0, trace(A) = λ1 + λ2 = a + c > 0.
4. Negative semidefiniteness: |A1 | < 0, |A| = 0 ⇒ λ1 = 0, λ2 < 0 or λ1 < 0, λ2 = 0
because c < 0 and |A| = λ1 λ2 = 0, trace(A) = λ1 + λ2 = a + c < 0.
5. Saddle point: |A| < 0, which indicates that the two eigenvalues have opposite signs.
The corresponding matrix A is called positive definite, negative definite, positive semidefinite,
negative semidefinite, and indefinite respectively.
Lecture 12 11/26
Positive Definiteness in Higher Dimensions
The function
in terms of which, we can rewrite the conditions for definiteness, by realizing that
as follows:
1. Positive definiteness: |A1 | > 0, |A| > 0 ⇒ λ1 , λ2 > 0 because c > 0 and |A| = λ1 λ2 >
0, trace(A) = λ1 + λ2 = a + c > 0.
2. Negative definiteness: |A1 | < 0, |A| > 0 ⇒ λ1 , λ2 < 0 because c < 0 and |A| = λ1 λ2 <
0, trace(A) = λ1 + λ2 = a + c < 0.
3. Positive semidefiniteness: |A1 | > 0, |A| = 0 ⇒ λ1 = 0, λ2 > 0 or λ1 > 0, λ2 = 0
because c > 0 and |A| = λ1 λ2 = 0, trace(A) = λ1 + λ2 = a + c > 0.
4. Negative semidefiniteness: |A1 | < 0, |A| = 0 ⇒ λ1 = 0, λ2 < 0 or λ1 < 0, λ2 = 0
because c < 0 and |A| = λ1 λ2 = 0, trace(A) = λ1 + λ2 = a + c < 0.
5. Saddle point: |A| < 0, which indicates that the two eigenvalues have opposite signs.
The corresponding matrix A is called positive definite, negative definite, positive semidefinite,
negative semidefinite, and indefinite respectively.
Lecture 12 11/26
Positive Definiteness in Higher Dimensions
in which
ac − b2
d1 = a, d2 = (18)
a
are called the pivots of the matrix A.
Hence,
2
ac − b2
b
f (x1 , x2 ) = a x1 + x2 + x22 = xT Ax
a a
Lecture 12 12/26
Positive Definiteness in Higher Dimensions
in which
ac − b2
d1 = a, d2 = (18)
a
are called the pivots of the matrix A.
Hence,
2
ac − b2
b
f (x1 , x2 ) = a x1 + x2 + x22 = xT Ax
a a
Lecture 12 12/26
Positive Definiteness in Higher Dimensions
in which
ac − b2
d1 = a, d2 = (18)
a
are called the pivots of the matrix A.
Hence,
2
ac − b2
b
f (x1 , x2 ) = a x1 + x2 + x22 = xT Ax
a a
Lecture 12 12/26
Positive Definiteness in Higher Dimensions
in which
ac − b2
d1 = a, d2 = (18)
a
are called the pivots of the matrix A.
Hence,
2
ac − b2
b
f (x1 , x2 ) = a x1 + x2 + x22 = xT Ax
a a
Lecture 12 12/26
Positive Definiteness in Higher Dimensions
in which
ac − b2
d1 = a, d2 = (18)
a
are called the pivots of the matrix A.
Hence,
2
ac − b2
b
f (x1 , x2 ) = a x1 + x2 + x22 = xT Ax
a a
Lecture 12 12/26
Positive Definiteness in Higher Dimensions
in which
ac − b2
d1 = a, d2 = (18)
a
are called the pivots of the matrix A.
Hence,
2
ac − b2
b
f (x1 , x2 ) = a x1 + x2 + x22 = xT Ax
a a
Lecture 12 12/26
Positive Definiteness in Higher Dimensions
in which
ac − b2
d1 = a, d2 = (18)
a
are called the pivots of the matrix A.
Hence,
2
ac − b2
b
f (x1 , x2 ) = a x1 + x2 + x22 = xT Ax
a a
Lecture 12 12/26
Positive Definiteness in Higher Dimensions
Extending these ideas to a function of n variables, i.e.,
a11 a12 ... x1n x1
a12
a22 ... x2n x2
f (x1 , x2 , . . . , xn ) = xT Ax = x1 x2 . . . xn .
.. .. .. ..
.. . . . .
a1n a2n ... ann xn
n X
X n
= ajk xj xk , (19)
j=1 k=1
quadratic function f and the symmetric matrix A are positive definite if each of the
following necessary and sufficient conditions holds:
1. xT Ax > 0 for all non-zero vectors x (definition).
2. Eigenvalues of A are positive, i.e., λj > 0, j = 1, 2, . . . , n.
2 2
Since Axj = λj xj ⇒ xT T
j Axj = λj xj xj = λj kxj k > 0. Since kxj k > 0 for
non-zero vector xj , λj > 0, j = 1, 2, . . . , n.
Conversely, if λj > 0, j = 1, 2, . . . , n, then since symmetric matrices have full set
of orthonormal eigenvectors, any x ∈ Rn can be written as
Xn X n n
X
x= cj xj ⇒ Ax = cj Axj = cj λj xj
j=1 j=1 j=1
Xn
xT Ax = c2j λj > 0 (∵ xT
j xk = δj,k ).
j=1
Lecture 12 13/26
Positive Definiteness in Higher Dimensions
Extending these ideas to a function of n variables, i.e.,
a11 a12 ... x1n x1
a12
a22 ... x2n x2
f (x1 , x2 , . . . , xn ) = xT Ax = x1 x2 . . . xn .
.. .. .. ..
.. . . . .
a1n a2n ... ann xn
n X
X n
= ajk xj xk , (19)
j=1 k=1
quadratic function f and the symmetric matrix A are positive definite if each of the
following necessary and sufficient conditions holds:
1. xT Ax > 0 for all non-zero vectors x (definition).
2. Eigenvalues of A are positive, i.e., λj > 0, j = 1, 2, . . . , n.
2 2
Since Axj = λj xj ⇒ xT T
j Axj = λj xj xj = λj kxj k > 0. Since kxj k > 0 for
non-zero vector xj , λj > 0, j = 1, 2, . . . , n.
Conversely, if λj > 0, j = 1, 2, . . . , n, then since symmetric matrices have full set
of orthonormal eigenvectors, any x ∈ Rn can be written as
Xn X n n
X
x= cj xj ⇒ Ax = cj Axj = cj λj xj
j=1 j=1 j=1
Xn
xT Ax = c2j λj > 0 (∵ xT
j xk = δj,k ).
j=1
Lecture 12 13/26
Positive Definiteness in Higher Dimensions
Extending these ideas to a function of n variables, i.e.,
a11 a12 ... x1n x1
a12
a22 ... x2n x2
f (x1 , x2 , . . . , xn ) = xT Ax = x1 x2 . . . xn .
.. .. .. ..
.. . . . .
a1n a2n ... ann xn
n X
X n
= ajk xj xk , (19)
j=1 k=1
quadratic function f and the symmetric matrix A are positive definite if each of the
following necessary and sufficient conditions holds:
1. xT Ax > 0 for all non-zero vectors x (definition).
2. Eigenvalues of A are positive, i.e., λj > 0, j = 1, 2, . . . , n.
2 2
Since Axj = λj xj ⇒ xT T
j Axj = λj xj xj = λj kxj k > 0. Since kxj k > 0 for
non-zero vector xj , λj > 0, j = 1, 2, . . . , n.
Conversely, if λj > 0, j = 1, 2, . . . , n, then since symmetric matrices have full set
of orthonormal eigenvectors, any x ∈ Rn can be written as
Xn X n n
X
x= cj xj ⇒ Ax = cj Axj = cj λj xj
j=1 j=1 j=1
Xn
xT Ax = c2j λj > 0 (∵ xT
j xk = δj,k ).
j=1
Lecture 12 13/26
Positive Definiteness in Higher Dimensions
Extending these ideas to a function of n variables, i.e.,
a11 a12 ... x1n x1
a12
a22 ... x2n x2
f (x1 , x2 , . . . , xn ) = xT Ax = x1 x2 . . . xn .
.. .. .. ..
.. . . . .
a1n a2n ... ann xn
n X
X n
= ajk xj xk , (19)
j=1 k=1
quadratic function f and the symmetric matrix A are positive definite if each of the
following necessary and sufficient conditions holds:
1. xT Ax > 0 for all non-zero vectors x (definition).
2. Eigenvalues of A are positive, i.e., λj > 0, j = 1, 2, . . . , n.
2 2
Since Axj = λj xj ⇒ xT T
j Axj = λj xj xj = λj kxj k > 0. Since kxj k > 0 for
non-zero vector xj , λj > 0, j = 1, 2, . . . , n.
Conversely, if λj > 0, j = 1, 2, . . . , n, then since symmetric matrices have full set
of orthonormal eigenvectors, any x ∈ Rn can be written as
Xn X n n
X
x= cj xj ⇒ Ax = cj Axj = cj λj xj
j=1 j=1 j=1
Xn
xT Ax = c2j λj > 0 (∵ xT
j xk = δj,k ).
j=1
Lecture 12 13/26
Positive Definiteness in Higher Dimensions
Extending these ideas to a function of n variables, i.e.,
a11 a12 ... x1n x1
a12
a22 ... x2n x2
f (x1 , x2 , . . . , xn ) = xT Ax = x1 x2 . . . xn .
.. .. .. ..
.. . . . .
a1n a2n ... ann xn
n X
X n
= ajk xj xk , (19)
j=1 k=1
quadratic function f and the symmetric matrix A are positive definite if each of the
following necessary and sufficient conditions holds:
1. xT Ax > 0 for all non-zero vectors x (definition).
2. Eigenvalues of A are positive, i.e., λj > 0, j = 1, 2, . . . , n.
2 2
Since Axj = λj xj ⇒ xT T
j Axj = λj xj xj = λj kxj k > 0. Since kxj k > 0 for
non-zero vector xj , λj > 0, j = 1, 2, . . . , n.
Conversely, if λj > 0, j = 1, 2, . . . , n, then since symmetric matrices have full set
of orthonormal eigenvectors, any x ∈ Rn can be written as
Xn X n n
X
x= cj xj ⇒ Ax = cj Axj = cj λj xj
j=1 j=1 j=1
Xn
xT Ax = c2j λj > 0 (∵ xT
j xk = δj,k ).
j=1
Lecture 12 13/26
Positive Definiteness in Higher Dimensions
Extending these ideas to a function of n variables, i.e.,
a11 a12 ... x1n x1
a12
a22 ... x2n x2
f (x1 , x2 , . . . , xn ) = xT Ax = x1 x2 . . . xn .
.. .. .. ..
.. . . . .
a1n a2n ... ann xn
n X
X n
= ajk xj xk , (19)
j=1 k=1
quadratic function f and the symmetric matrix A are positive definite if each of the
following necessary and sufficient conditions holds:
1. xT Ax > 0 for all non-zero vectors x (definition).
2. Eigenvalues of A are positive, i.e., λj > 0, j = 1, 2, . . . , n.
2 2
Since Axj = λj xj ⇒ xT T
j Axj = λj xj xj = λj kxj k > 0. Since kxj k > 0 for
non-zero vector xj , λj > 0, j = 1, 2, . . . , n.
Conversely, if λj > 0, j = 1, 2, . . . , n, then since symmetric matrices have full set
of orthonormal eigenvectors, any x ∈ Rn can be written as
Xn X n n
X
x= cj xj ⇒ Ax = cj Axj = cj λj xj
j=1 j=1 j=1
Xn
xT Ax = c2j λj > 0 (∵ xT
j xk = δj,k ).
j=1
Lecture 12 13/26
Positive Definiteness in Higher Dimensions
Extending these ideas to a function of n variables, i.e.,
a11 a12 ... x1n x1
a12
a22 ... x2n x2
f (x1 , x2 , . . . , xn ) = xT Ax = x1 x2 . . . xn .
.. .. .. ..
.. . . . .
a1n a2n ... ann xn
n X
X n
= ajk xj xk , (19)
j=1 k=1
quadratic function f and the symmetric matrix A are positive definite if each of the
following necessary and sufficient conditions holds:
1. xT Ax > 0 for all non-zero vectors x (definition).
2. Eigenvalues of A are positive, i.e., λj > 0, j = 1, 2, . . . , n.
2 2
Since Axj = λj xj ⇒ xT T
j Axj = λj xj xj = λj kxj k > 0. Since kxj k > 0 for
non-zero vector xj , λj > 0, j = 1, 2, . . . , n.
Conversely, if λj > 0, j = 1, 2, . . . , n, then since symmetric matrices have full set
of orthonormal eigenvectors, any x ∈ Rn can be written as
Xn X n n
X
x= cj xj ⇒ Ax = cj Axj = cj λj xj
j=1 j=1 j=1
Xn
xT Ax = c2j λj > 0 (∵ xT
j xk = δj,k ).
j=1
Lecture 12 13/26
Positive Definiteness in Higher Dimensions
3. All upperleft submatrices Ak have positive determinants.
Qn
For xT Ax > 0, λj > 0, j = 1, 2, . . . , n ⇒ |A| = j=1 λj > 0.
Now, considering those vectors whoe last k components are zero, their quadratic
form is written as
Ak 0 x T
T k ≡ xT A x > 0
T
x Ax = xk 0 k k k
0 0 0
Conversely, we note that pivots are numbers that multiply complete squares to make
up xT Ax. Hence, dk > 0, k = 1, 2, . . . , n to render the quadratic form positive
everywhere except at origin.
T 2
xT Ax = xT RT Rx = (Rx) (Rx) = kRxk > 0,
because x 6= 0 and
n
X
Rx = xj rj = 0
j=1
Lecture 12 15/26
Positive Definiteness in Higher Dimensions
Conversely, we note that pivots are numbers that multiply complete squares to make
up xT Ax. Hence, dk > 0, k = 1, 2, . . . , n to render the quadratic form positive
everywhere except at origin.
T 2
xT Ax = xT RT Rx = (Rx) (Rx) = kRxk > 0,
because x 6= 0 and
n
X
Rx = xj rj = 0
j=1
Lecture 12 15/26
Positive Definiteness in Higher Dimensions
Conversely, we note that pivots are numbers that multiply complete squares to make
up xT Ax. Hence, dk > 0, k = 1, 2, . . . , n to render the quadratic form positive
everywhere except at origin.
T 2
xT Ax = xT RT Rx = (Rx) (Rx) = kRxk > 0,
because x 6= 0 and
n
X
Rx = xj rj = 0
j=1
Lecture 12 15/26
Positive Definiteness in Higher Dimensions
Conversely, we note that pivots are numbers that multiply complete squares to make
up xT Ax. Hence, dk > 0, k = 1, 2, . . . , n to render the quadratic form positive
everywhere except at origin.
T 2
xT Ax = xT RT Rx = (Rx) (Rx) = kRxk > 0,
because x 6= 0 and
n
X
Rx = xj rj = 0
j=1
Lecture 12 15/26
Positive Definiteness in Higher Dimensions
Conversely, we note that pivots are numbers that multiply complete squares to make
up xT Ax. Hence, dk > 0, k = 1, 2, . . . , n to render the quadratic form positive
everywhere except at origin.
T 2
xT Ax = xT RT Rx = (Rx) (Rx) = kRxk > 0,
because x 6= 0 and
n
X
Rx = xj rj = 0
j=1
Lecture 12 15/26
Positive Semidefiniteness in Higher Dimensions
The symmetric matrix in the quadrartic form given in (19) is positive semidefinite, if
each of the following necessary and sufficient conditions hold:
1. xT Ax ≥ 0 for all x 6= 0. (definition)
2. Eigenvalues of A are non-negative, i.e., λj ≥ 0, j = 1, 2, . . . , n.
3. No principal submatrices have negative determinants5 .
Yet both of these matrices have zero determinants for all upperleft submatrices.
5 A principal submatrix is a square matrix obtained by removing certain rows and columns of a matrix.
Lecture 12 16/26
Positive Semidefiniteness in Higher Dimensions
The symmetric matrix in the quadrartic form given in (19) is positive semidefinite, if
each of the following necessary and sufficient conditions hold:
1. xT Ax ≥ 0 for all x 6= 0. (definition)
2. Eigenvalues of A are non-negative, i.e., λj ≥ 0, j = 1, 2, . . . , n.
3. No principal submatrices have negative determinants5 .
Yet both of these matrices have zero determinants for all upperleft submatrices.
5 A principal submatrix is a square matrix obtained by removing certain rows and columns of a matrix.
Lecture 12 16/26
Positive Semidefiniteness in Higher Dimensions
The symmetric matrix in the quadrartic form given in (19) is positive semidefinite, if
each of the following necessary and sufficient conditions hold:
1. xT Ax ≥ 0 for all x 6= 0. (definition)
2. Eigenvalues of A are non-negative, i.e., λj ≥ 0, j = 1, 2, . . . , n.
3. No principal submatrices have negative determinants5 .
Yet both of these matrices have zero determinants for all upperleft submatrices.
5 A principal submatrix is a square matrix obtained by removing certain rows and columns of a matrix.
Lecture 12 16/26
Positive Semidefiniteness in Higher Dimensions
The symmetric matrix in the quadrartic form given in (19) is positive semidefinite, if
each of the following necessary and sufficient conditions hold:
1. xT Ax ≥ 0 for all x 6= 0. (definition)
2. Eigenvalues of A are non-negative, i.e., λj ≥ 0, j = 1, 2, . . . , n.
3. No principal submatrices have negative determinants5 .
Yet both of these matrices have zero determinants for all upperleft submatrices.
5 A principal submatrix is a square matrix obtained by removing certain rows and columns of a matrix.
Lecture 12 16/26
Positive Semidefiniteness in Higher Dimensions
The symmetric matrix in the quadrartic form given in (19) is positive semidefinite, if
each of the following necessary and sufficient conditions hold:
1. xT Ax ≥ 0 for all x 6= 0. (definition)
2. Eigenvalues of A are non-negative, i.e., λj ≥ 0, j = 1, 2, . . . , n.
3. No principal submatrices have negative determinants5 .
Yet both of these matrices have zero determinants for all upperleft submatrices.
5 A principal submatrix is a square matrix obtained by removing certain rows and columns of a matrix.
Lecture 12 16/26
Positive Semidefiniteness in Higher Dimensions
The symmetric matrix in the quadrartic form given in (19) is positive semidefinite, if
each of the following necessary and sufficient conditions hold:
1. xT Ax ≥ 0 for all x 6= 0. (definition)
2. Eigenvalues of A are non-negative, i.e., λj ≥ 0, j = 1, 2, . . . , n.
3. No principal submatrices have negative determinants5 .
Yet both of these matrices have zero determinants for all upperleft submatrices.
5 A principal submatrix is a square matrix obtained by removing certain rows and columns of a matrix.
Lecture 12 16/26
Positive Semidefiniteness in Higher Dimensions
The symmetric matrix in the quadrartic form given in (19) is positive semidefinite, if
each of the following necessary and sufficient conditions hold:
1. xT Ax ≥ 0 for all x 6= 0. (definition)
2. Eigenvalues of A are non-negative, i.e., λj ≥ 0, j = 1, 2, . . . , n.
3. No principal submatrices have negative determinants5 .
Yet both of these matrices have zero determinants for all upperleft submatrices.
5 A principal submatrix is a square matrix obtained by removing certain rows and columns of a matrix.
Lecture 12 16/26
Positive Semidefiniteness in Higher Dimensions
The symmetric matrix in the quadrartic form given in (19) is positive semidefinite, if
each of the following necessary and sufficient conditions hold:
1. xT Ax ≥ 0 for all x 6= 0. (definition)
2. Eigenvalues of A are non-negative, i.e., λj ≥ 0, j = 1, 2, . . . , n.
3. No principal submatrices have negative determinants5 .
Yet both of these matrices have zero determinants for all upperleft submatrices.
5 A principal submatrix is a square matrix obtained by removing certain rows and columns of a matrix.
Lecture 12 16/26
Positive Semidefiniteness in Higher Dimensions
The symmetric matrix in the quadrartic form given in (19) is positive semidefinite, if
each of the following necessary and sufficient conditions hold:
1. xT Ax ≥ 0 for all x 6= 0. (definition)
2. Eigenvalues of A are non-negative, i.e., λj ≥ 0, j = 1, 2, . . . , n.
3. No principal submatrices have negative determinants5 .
Yet both of these matrices have zero determinants for all upperleft submatrices.
5 A principal submatrix is a square matrix obtained by removing certain rows and columns of a matrix.
Lecture 12 16/26
Quadratic Form for a Random Function
Finally, we note that any function F (x1 , x2 , . . . , xn ) has a quadratic form whose
matrix A = [ajk ] is such that
∂2F
ajk = = akj . (20)
∂xj ∂xk
The function F has a minimum6 when the resulting quadratic form is positive
definite.
6 at the stationary point where all of its first derivatives are zero.
Lecture 12 17/26
Quadratic Form for a Random Function
Finally, we note that any function F (x1 , x2 , . . . , xn ) has a quadratic form whose
matrix A = [ajk ] is such that
∂2F
ajk = = akj . (20)
∂xj ∂xk
The function F has a minimum6 when the resulting quadratic form is positive
definite.
6 at the stationary point where all of its first derivatives are zero.
Lecture 12 17/26
Singular Value Decomposition
Any m × n matrix A can be factored into
where Um×m and Vn×n (and thus V T ) are orthogonal and Σm×n is a diagonal matrix.
Colummns of U are eigenvectors of AAT and columns of V are eigenvectors of AT A.
The r singular values on the diagonal of Σ are square roots of the non-zero eigenvalues
of both AAT and AT A.
Proof:
Let Am×n = [aij ] be a matrix with Rank(A) = r. Then, AT A is symmetric and
positive semidefinite, i.e.,
T
AT A = AT A (AT A is symmetric) and
T T T T T 2 2
x (A A)x = x A Ax = (Ax) (Ax) = kAxk = kyk ≥ 0,
where
y = x1 a1 + x2 a2 + · · · + xn an ,
which may equal 0 for non-zero xj if the columns of A, i.e., aj , are linearly
dependent.
Lecture 12 18/26
Singular Value Decomposition
Any m × n matrix A can be factored into
where Um×m and Vn×n (and thus V T ) are orthogonal and Σm×n is a diagonal matrix.
Colummns of U are eigenvectors of AAT and columns of V are eigenvectors of AT A.
The r singular values on the diagonal of Σ are square roots of the non-zero eigenvalues
of both AAT and AT A.
Proof:
Let Am×n = [aij ] be a matrix with Rank(A) = r. Then, AT A is symmetric and
positive semidefinite, i.e.,
T
AT A = AT A (AT A is symmetric) and
T T T T T 2 2
x (A A)x = x A Ax = (Ax) (Ax) = kAxk = kyk ≥ 0,
where
y = x1 a1 + x2 a2 + · · · + xn an ,
which may equal 0 for non-zero xj if the columns of A, i.e., aj , are linearly
dependent.
Lecture 12 18/26
Singular Value Decomposition
Any m × n matrix A can be factored into
where Um×m and Vn×n (and thus V T ) are orthogonal and Σm×n is a diagonal matrix.
Colummns of U are eigenvectors of AAT and columns of V are eigenvectors of AT A.
The r singular values on the diagonal of Σ are square roots of the non-zero eigenvalues
of both AAT and AT A.
Proof:
Let Am×n = [aij ] be a matrix with Rank(A) = r. Then, AT A is symmetric and
positive semidefinite, i.e.,
T
AT A = AT A (AT A is symmetric) and
T T T T T 2 2
x (A A)x = x A Ax = (Ax) (Ax) = kAxk = kyk ≥ 0,
where
y = x1 a1 + x2 a2 + · · · + xn an ,
which may equal 0 for non-zero xj if the columns of A, i.e., aj , are linearly
dependent.
Lecture 12 18/26
Singular Value Decomposition
Any m × n matrix A can be factored into
where Um×m and Vn×n (and thus V T ) are orthogonal and Σm×n is a diagonal matrix.
Colummns of U are eigenvectors of AAT and columns of V are eigenvectors of AT A.
The r singular values on the diagonal of Σ are square roots of the non-zero eigenvalues
of both AAT and AT A.
Proof:
Let Am×n = [aij ] be a matrix with Rank(A) = r. Then, AT A is symmetric and
positive semidefinite, i.e.,
T
AT A = AT A (AT A is symmetric) and
T T T T T 2 2
x (A A)x = x A Ax = (Ax) (Ax) = kAxk = kyk ≥ 0,
where
y = x1 a1 + x2 a2 + · · · + xn an ,
which may equal 0 for non-zero xj if the columns of A, i.e., aj , are linearly
dependent.
Lecture 12 18/26
Singular Value Decomposition
Any m × n matrix A can be factored into
where Um×m and Vn×n (and thus V T ) are orthogonal and Σm×n is a diagonal matrix.
Colummns of U are eigenvectors of AAT and columns of V are eigenvectors of AT A.
The r singular values on the diagonal of Σ are square roots of the non-zero eigenvalues
of both AAT and AT A.
Proof:
Let Am×n = [aij ] be a matrix with Rank(A) = r. Then, AT A is symmetric and
positive semidefinite, i.e.,
T
AT A = AT A (AT A is symmetric) and
T T T T T 2 2
x (A A)x = x A Ax = (Ax) (Ax) = kAxk = kyk ≥ 0,
where
y = x1 a1 + x2 a2 + · · · + xn an ,
which may equal 0 for non-zero xj if the columns of A, i.e., aj , are linearly
dependent.
Lecture 12 18/26
Singular Value Decomposition
Any m × n matrix A can be factored into
where Um×m and Vn×n (and thus V T ) are orthogonal and Σm×n is a diagonal matrix.
Colummns of U are eigenvectors of AAT and columns of V are eigenvectors of AT A.
The r singular values on the diagonal of Σ are square roots of the non-zero eigenvalues
of both AAT and AT A.
Proof:
Let Am×n = [aij ] be a matrix with Rank(A) = r. Then, AT A is symmetric and
positive semidefinite, i.e.,
T
AT A = AT A (AT A is symmetric) and
T T T T T 2 2
x (A A)x = x A Ax = (Ax) (Ax) = kAxk = kyk ≥ 0,
where
y = x1 a1 + x2 a2 + · · · + xn an ,
which may equal 0 for non-zero xj if the columns of A, i.e., aj , are linearly
dependent.
Lecture 12 18/26
Singular Value Decomposition
Any m × n matrix A can be factored into
where Um×m and Vn×n (and thus V T ) are orthogonal and Σm×n is a diagonal matrix.
Colummns of U are eigenvectors of AAT and columns of V are eigenvectors of AT A.
The r singular values on the diagonal of Σ are square roots of the non-zero eigenvalues
of both AAT and AT A.
Proof:
Let Am×n = [aij ] be a matrix with Rank(A) = r. Then, AT A is symmetric and
positive semidefinite, i.e.,
T
AT A = AT A (AT A is symmetric) and
T T T T T 2 2
x (A A)x = x A Ax = (Ax) (Ax) = kAxk = kyk ≥ 0,
where
y = x1 a1 + x2 a2 + · · · + xn an ,
which may equal 0 for non-zero xj if the columns of A, i.e., aj , are linearly
dependent.
Lecture 12 18/26
Singular Value Decomposition
Any m × n matrix A can be factored into
where Um×m and Vn×n (and thus V T ) are orthogonal and Σm×n is a diagonal matrix.
Colummns of U are eigenvectors of AAT and columns of V are eigenvectors of AT A.
The r singular values on the diagonal of Σ are square roots of the non-zero eigenvalues
of both AAT and AT A.
Proof:
Let Am×n = [aij ] be a matrix with Rank(A) = r. Then, AT A is symmetric and
positive semidefinite, i.e.,
T
AT A = AT A (AT A is symmetric) and
T T T T T 2 2
x (A A)x = x A Ax = (Ax) (Ax) = kAxk = kyk ≥ 0,
where
y = x1 a1 + x2 a2 + · · · + xn an ,
which may equal 0 for non-zero xj if the columns of A, i.e., aj , are linearly
dependent.
Lecture 12 18/26
Singular Value Decomposition
Any m × n matrix A can be factored into
where Um×m and Vn×n (and thus V T ) are orthogonal and Σm×n is a diagonal matrix.
Colummns of U are eigenvectors of AAT and columns of V are eigenvectors of AT A.
The r singular values on the diagonal of Σ are square roots of the non-zero eigenvalues
of both AAT and AT A.
Proof:
Let Am×n = [aij ] be a matrix with Rank(A) = r. Then, AT A is symmetric and
positive semidefinite, i.e.,
T
AT A = AT A (AT A is symmetric) and
T T T T T 2 2
x (A A)x = x A Ax = (Ax) (Ax) = kAxk = kyk ≥ 0,
where
y = x1 a1 + x2 a2 + · · · + xn an ,
which may equal 0 for non-zero xj if the columns of A, i.e., aj , are linearly
dependent.
Lecture 12 18/26
Singular Value Decomposition
Any m × n matrix A can be factored into
where Um×m and Vn×n (and thus V T ) are orthogonal and Σm×n is a diagonal matrix.
Colummns of U are eigenvectors of AAT and columns of V are eigenvectors of AT A.
The r singular values on the diagonal of Σ are square roots of the non-zero eigenvalues
of both AAT and AT A.
Proof:
Let Am×n = [aij ] be a matrix with Rank(A) = r. Then, AT A is symmetric and
positive semidefinite, i.e.,
T
AT A = AT A (AT A is symmetric) and
T T T T T 2 2
x (A A)x = x A Ax = (Ax) (Ax) = kAxk = kyk ≥ 0,
where
y = x1 a1 + x2 a2 + · · · + xn an ,
which may equal 0 for non-zero xj if the columns of A, i.e., aj , are linearly
dependent.
Lecture 12 18/26
Singular Value Decomposition
Any m × n matrix A can be factored into
where Um×m and Vn×n (and thus V T ) are orthogonal and Σm×n is a diagonal matrix.
Colummns of U are eigenvectors of AAT and columns of V are eigenvectors of AT A.
The r singular values on the diagonal of Σ are square roots of the non-zero eigenvalues
of both AAT and AT A.
Proof:
Let Am×n = [aij ] be a matrix with Rank(A) = r. Then, AT A is symmetric and
positive semidefinite, i.e.,
T
AT A = AT A (AT A is symmetric) and
T T T T T 2 2
x (A A)x = x A Ax = (Ax) (Ax) = kAxk = kyk ≥ 0,
where
y = x1 a1 + x2 a2 + · · · + xn an ,
which may equal 0 for non-zero xj if the columns of A, i.e., aj , are linearly
dependent.
Lecture 12 18/26
Singular Value Decomposition
Any m × n matrix A can be factored into
where Um×m and Vn×n (and thus V T ) are orthogonal and Σm×n is a diagonal matrix.
Colummns of U are eigenvectors of AAT and columns of V are eigenvectors of AT A.
The r singular values on the diagonal of Σ are square roots of the non-zero eigenvalues
of both AAT and AT A.
Proof:
Let Am×n = [aij ] be a matrix with Rank(A) = r. Then, AT A is symmetric and
positive semidefinite, i.e.,
T
AT A = AT A (AT A is symmetric) and
T T T T T 2 2
x (A A)x = x A Ax = (Ax) (Ax) = kAxk = kyk ≥ 0,
where
y = x1 a1 + x2 a2 + · · · + xn an ,
which may equal 0 for non-zero xj if the columns of A, i.e., aj , are linearly
dependent.
Lecture 12 18/26
Singular Value Decomposition
AT A vj = λj vj = σj2 vj , j = 1, 2, . . . , n,
(22)
λj = σj2 ∵ λj ≥ 0, j = 1, 2, . . . , n. (23)
AT A n×n = V DV −1 = V DV T
λ1 v1
λ2 v2
= v1 v2 . . . vn . . , (24)
.. ..
λn vn
7 because AT A is symmetric!
Lecture 12 19/26
Singular Value Decomposition
AT A vj = λj vj = σj2 vj , j = 1, 2, . . . , n,
(22)
λj = σj2 ∵ λj ≥ 0, j = 1, 2, . . . , n. (23)
AT A n×n = V DV −1 = V DV T
λ1 v1
λ2 v2
= v1 v2 . . . vn . . , (24)
.. ..
λn vn
7 because AT A is symmetric!
Lecture 12 19/26
Singular Value Decomposition
AT A vj = λj vj = σj2 vj , j = 1, 2, . . . , n,
(22)
λj = σj2 ∵ λj ≥ 0, j = 1, 2, . . . , n. (23)
AT A n×n = V DV −1 = V DV T
λ1 v1
λ2 v2
= v1 v2 . . . vn . . , (24)
.. ..
λn vn
7 because AT A is symmetric!
Lecture 12 19/26
Singular Value Decomposition
AT A vj = λj vj = σj2 vj , j = 1, 2, . . . , n,
(22)
λj = σj2 ∵ λj ≥ 0, j = 1, 2, . . . , n. (23)
AT A n×n = V DV −1 = V DV T
λ1 v1
λ2 v2
= v1 v2 . . . vn . . , (24)
.. ..
λn vn
7 because AT A is symmetric!
Lecture 12 19/26
Singular Value Decomposition
It must be noted that A and AT A have the same rank, equal to r, because if Ax = 0,
then AT Ax = 0, implying that A and AT A have the same null space. Hence,
(25)
T
Because AAT = AAT , u0i are orthogonal eigenvectors of AAT .Furthermore,
since AAT is a square matrix of size m × m, i = 1, 2, . . . , m in (27).
Lecture 12 20/26
Singular Value Decomposition
It must be noted that A and AT A have the same rank, equal to r, because if Ax = 0,
then AT Ax = 0, implying that A and AT A have the same null space. Hence,
(25)
T
Because AAT = AAT , u0i are orthogonal eigenvectors of AAT .Furthermore,
since AAT is a square matrix of size m × m, i = 1, 2, . . . , m in (27).
Lecture 12 20/26
Singular Value Decomposition
It must be noted that A and AT A have the same rank, equal to r, because if Ax = 0,
then AT Ax = 0, implying that A and AT A have the same null space. Hence,
(25)
T
Because AAT = AAT , u0i are orthogonal eigenvectors of AAT .Furthermore,
since AAT is a square matrix of size m × m, i = 1, 2, . . . , m in (27).
Lecture 12 20/26
Singular Value Decomposition
It must be noted that A and AT A have the same rank, equal to r, because if Ax = 0,
then AT Ax = 0, implying that A and AT A have the same null space. Hence,
(25)
T
Because AAT = AAT , u0i are orthogonal eigenvectors of AAT .Furthermore,
since AAT is a square matrix of size m × m, i = 1, 2, . . . , m in (27).
Lecture 12 20/26
Singular Value Decomposition
It must be noted that A and AT A have the same rank, equal to r, because if Ax = 0,
then AT Ax = 0, implying that A and AT A have the same null space. Hence,
(25)
T
Because AAT = AAT , u0i are orthogonal eigenvectors of AAT .Furthermore,
since AAT is a square matrix of size m × m, i = 1, 2, . . . , m in (27).
Lecture 12 20/26
Singular Value Decomposition
It must be noted that A and AT A have the same rank, equal to r, because if Ax = 0,
then AT Ax = 0, implying that A and AT A have the same null space. Hence,
(25)
T
Because AAT = AAT , u0i are orthogonal eigenvectors of AAT .Furthermore,
since AAT is a square matrix of size m × m, i = 1, 2, . . . , m in (27).
Lecture 12 20/26
Singular Value Decomposition
It must be noted that A and AT A have the same rank, equal to r, because if Ax = 0,
then AT Ax = 0, implying that A and AT A have the same null space. Hence,
(25)
T
Because AAT = AAT , u0i are orthogonal eigenvectors of AAT .Furthermore,
since AAT is a square matrix of size m × m, i = 1, 2, . . . , m in (27).
Lecture 12 20/26
Singular Value Decomposition
It must be noted that A and AT A have the same rank, equal to r, because if Ax = 0,
then AT Ax = 0, implying that A and AT A have the same null space. Hence,
(25)
T
Because AAT = AAT , u0i are orthogonal eigenvectors of AAT .Furthermore,
since AAT is a square matrix of size m × m, i = 1, 2, . . . , m in (27).
Lecture 12 20/26
Singular Value Decomposition
The inner product between u0 s can be found as
T T
hu0i , u0i i = u0i u0i = (Avi ) Avi = viT AT Avi = viT λi vi = λi , (28)
where
which may equal 0 for non-zero xi if the rows of A, i.e., âi , are linearly dependent.
Lecture 12 21/26
Singular Value Decomposition
The inner product between u0 s can be found as
T T
hu0i , u0i i = u0i u0i = (Avi ) Avi = viT AT Avi = viT λi vi = λi , (28)
where
which may equal 0 for non-zero xi if the rows of A, i.e., âi , are linearly dependent.
Lecture 12 21/26
Singular Value Decomposition
The inner product between u0 s can be found as
T T
hu0i , u0i i = u0i u0i = (Avi ) Avi = viT AT Avi = viT λi vi = λi , (28)
where
which may equal 0 for non-zero xi if the rows of A, i.e., âi , are linearly dependent.
Lecture 12 21/26
Singular Value Decomposition
The inner product between u0 s can be found as
T T
hu0i , u0i i = u0i u0i = (Avi ) Avi = viT AT Avi = viT λi vi = λi , (28)
where
which may equal 0 for non-zero xi if the rows of A, i.e., âi , are linearly dependent.
Lecture 12 21/26
Singular Value Decomposition
The inner product between u0 s can be found as
T T
hu0i , u0i i = u0i u0i = (Avi ) Avi = viT AT Avi = viT λi vi = λi , (28)
where
which may equal 0 for non-zero xi if the rows of A, i.e., âi , are linearly dependent.
Lecture 12 21/26
Singular Value Decomposition
The inner product between u0 s can be found as
T T
hu0i , u0i i = u0i u0i = (Avi ) Avi = viT AT Avi = viT λi vi = λi , (28)
where
which may equal 0 for non-zero xi if the rows of A, i.e., âi , are linearly dependent.
Lecture 12 21/26
Singular Value Decomposition
The inner product between u0 s can be found as
T T
hu0i , u0i i = u0i u0i = (Avi ) Avi = viT AT Avi = viT λi vi = λi , (28)
where
which may equal 0 for non-zero xi if the rows of A, i.e., âi , are linearly dependent.
Lecture 12 21/26
Singular Value Decomposition
The inner product between u0 s can be found as
T T
hu0i , u0i i = u0i u0i = (Avi ) Avi = viT AT Avi = viT λi vi = λi , (28)
where
which may equal 0 for non-zero xi if the rows of A, i.e., âi , are linearly dependent.
Lecture 12 21/26
Singular Value Decomposition
AAT n×n = U DU −1 = U DU T
λ1 u1
λ 2
u2
= u1 u2 . . . um .. , (31)
..
. .
λm um
It must be noted that A and AAT have the same rank, equal to r, because if AT y = 0,
then AAT y = 0, implying that A and AAT have the same null space. Hence,
Lecture 12 22/26
Singular Value Decomposition
AAT n×n = U DU −1 = U DU T
λ1 u1
λ 2
u2
= u1 u2 . . . um .. , (31)
..
. .
λm um
It must be noted that A and AAT have the same rank, equal to r, because if AT y = 0,
then AAT y = 0, implying that A and AAT have the same null space. Hence,
Lecture 12 22/26
Singular Value Decomposition
AAT n×n = U DU −1 = U DU T
λ1 u1
λ 2
u2
= u1 u2 . . . um .. , (31)
..
. .
λm um
It must be noted that A and AAT have the same rank, equal to r, because if AT y = 0,
then AAT y = 0, implying that A and AAT have the same null space. Hence,
Lecture 12 22/26
Singular Value Decomposition
AAT n×n = U DU −1 = U DU T
λ1 u1
λ 2
u2
= u1 u2 . . . um .. , (31)
..
. .
λm um
It must be noted that A and AAT have the same rank, equal to r, because if AT y = 0,
then AAT y = 0, implying that A and AAT have the same null space. Hence,
Lecture 12 22/26
Singular Value Decomposition
If Rank(A) = r = Rank(AT A) = Rank(AAT ), then
(
T σj2 vj , j = 1, 2, . . . , r
A Avj = (33)
0, j = r + 1, . . . , n,
and
(
σi ui , i = 1, 2, . . . , r
∵ Avi = σi ui ⇒ AAT ui = σi2 ui .
Avi =
0, i = r + 1, . . . , m,
(34)
Lecture 12 23/26
Singular Value Decomposition
If Rank(A) = r = Rank(AT A) = Rank(AAT ), then
(
T σj2 vj , j = 1, 2, . . . , r
A Avj = (33)
0, j = r + 1, . . . , n,
and
(
σi ui , i = 1, 2, . . . , r
∵ Avi = σi ui ⇒ AAT ui = σi2 ui .
Avi =
0, i = r + 1, . . . , m,
(34)
Lecture 12 23/26
Singular Value Decomposition
If Rank(A) = r = Rank(AT A) = Rank(AAT ), then
(
T σj2 vj , j = 1, 2, . . . , r
A Avj = (33)
0, j = r + 1, . . . , n,
and
(
σi ui , i = 1, 2, . . . , r
∵ Avi = σi ui ⇒ AAT ui = σi2 ui .
Avi =
0, i = r + 1, . . . , m,
(34)
Lecture 12 23/26
Singular Value Decomposition
which can be rearranged to obtain the signular value decomposition as
A = U ΣV −1 = U ΣV T . (36)
Lecture 12 24/26
Singular Value Decomposition
which can be rearranged to obtain the signular value decomposition as
A = U ΣV −1 = U ΣV T . (36)
Intuitive Picture
SVD of an m × n matrix A represents linear transformation of a vector x ∈ Rn into
Ax ∈ Rm , i.e., for any vector x ∈ Rn and an orthonormal basis set vj ∈ Rn , j =
1, 2, . . . , n, we can write
n
X
x= cj vj ,
j=1
n
X n
X n
X r
X
Ax = A hx, vj i vj = xT vj Avj = xT vj σj uj = vjT xσj uj
j=1 j=1 j=1 j=1
Xr r
X
= uj σj vjT x ⇒ A = uj σj vjT = U ΣV T .
j=1 j=1
Lecture 12 24/26
Singular Value Decomposition
which can be rearranged to obtain the signular value decomposition as
A = U ΣV −1 = U ΣV T . (36)
Intuitive Picture
SVD of an m × n matrix A represents linear transformation of a vector x ∈ Rn into
Ax ∈ Rm , i.e., for any vector x ∈ Rn and an orthonormal basis set vj ∈ Rn , j =
1, 2, . . . , n, we can write
n
X
x= cj vj ,
j=1
n
X n
X n
X r
X
Ax = A hx, vj i vj = xT vj Avj = xT vj σj uj = vjT xσj uj
j=1 j=1 j=1 j=1
Xr r
X
= uj σj vjT x ⇒ A = uj σj vjT = U ΣV T .
j=1 j=1
Lecture 12 24/26
Singular Value Decomposition
which can be rearranged to obtain the signular value decomposition as
A = U ΣV −1 = U ΣV T . (36)
Intuitive Picture
SVD of an m × n matrix A represents linear transformation of a vector x ∈ Rn into
Ax ∈ Rm , i.e., for any vector x ∈ Rn and an orthonormal basis set vj ∈ Rn , j =
1, 2, . . . , n, we can write
n
X
x= cj vj ,
j=1
n
X n
X n
X r
X
Ax = A hx, vj i vj = xT vj Avj = xT vj σj uj = vjT xσj uj
j=1 j=1 j=1 j=1
Xr r
X
= uj σj vjT x ⇒ A = uj σj vjT = U ΣV T .
j=1 j=1
Lecture 12 24/26
Singular Value Decomposition
which can be rearranged to obtain the signular value decomposition as
A = U ΣV −1 = U ΣV T . (36)
Intuitive Picture
SVD of an m × n matrix A represents linear transformation of a vector x ∈ Rn into
Ax ∈ Rm , i.e., for any vector x ∈ Rn and an orthonormal basis set vj ∈ Rn , j =
1, 2, . . . , n, we can write
n
X
x= cj vj ,
j=1
n
X n
X n
X r
X
Ax = A hx, vj i vj = xT vj Avj = xT vj σj uj = vjT xσj uj
j=1 j=1 j=1 j=1
Xr r
X
= uj σj vjT x ⇒ A = uj σj vjT = U ΣV T .
j=1 j=1
Lecture 12 24/26
Remarks on SVD
and
Lecture 12 25/26
Remarks on SVD
and
Lecture 12 25/26
Remarks on SVD
and
Lecture 12 25/26
Remarks on SVD
and
Lecture 12 25/26
Remarks on SVD
and
Lecture 12 25/26
Remarks on SVD
and
Lecture 12 25/26
Remarks on SVD
and
Lecture 12 25/26
Remarks on SVD
Lecture 12 26/26
Remarks on SVD
Lecture 12 26/26
Remarks on SVD
Lecture 12 26/26
Remarks on SVD
Lecture 12 26/26
Remarks on SVD
Lecture 12 26/26