Professional Documents
Culture Documents
U hU = U U h = I
U tU = U U t = I
θ(U x, U y) = θ(x, y)
Example 8.1
It can easily be verified that
1 1 −1
R= √
2 1 1
is an orthogonal matrix.
Let
x1 y1 1 x1 − x2
x= , y = Rx = = √
x2 y2 2 x1 + x2
Then
(x1 − x2 )2 + (x1 + x2 )2
k y k2 = = x21 + x22 = k x k2
2
269
270 Unitary and Hermitian Matrices
Expressing R as
cos θ − sin θ π
R= , θ=
sin θ cos θ 4
P1 = [ v1 V1 ]
8.1 Unitary Matrices 271
Since
we have
λ∗1 λ1 0 1 0
=
0 U2h U2 0 I
which implies that λ∗1 λ1 = 1, that is, | λ1 |2 = 1, and that U2 is unitary. Let U2 have the
eigenvalues λ2 , . . . , λn . Then they are also eigenvalues of U = U1 . Since U2 is of order
n − 1, by induction hypothesis | λ2 |2 = · · · = | λn |2 = 1 and there exists a unitary matrix
P2 such that
P2h U2 P2 = D2 = diag [ λ2 , . . . , λn ]
Let
1
P = P1
P2
Then
h 1 1
P UP = P1h U1 P1
P2h P2
1 λ1 0 1
=
P2h 0 U2 P2
λ1 0
= = D1
0 D2
Theorem 8.1 simply states that eigenvalues of a unitary (orthogonal) matrix are
located on the unit circle in the complex plane, that such a matrix can always be
diagonalized (even if it has multiple eigenvalues), and that a modal matrix can be
chosen to be unitary (orthogonal).
Example 8.2
The matrix
1 1 i
U= √
2 i 1
272 Unitary and Hermitian Matrices
is unitary as
1 1 −i 1 i 1 0
U hU = =
2 −i 1 i 1 0 1
v1 = col [ 1, 1 ]
v1 = col [ −1, 1 ]
Note that we need not look specifically for an eigenvector v2 that is orthogonal to v1 ;
eigenvectors of a unitary matrix associated with distinct eigenvalues are orthogonal (see
Exercise 8.11).
Normalizing the eigenvectors, we obtain a unitary modal matrix
1 1 −1
P = √
2 1 1
P h AP = D = diag [ λ1 , . . . , λn ]
8.2 Hermitian Matrices 273
and
vh A = vh Ah = (Av)h = (λv)h = λ∗ vh
Premultiplying both sides of the first equality by vh , postmultiplying both sides of the
second equality by v, and noting that vh v = k v k2 = 1, we get
vh Av = λ = λ∗
Hence, like unitary matrices, Hermitian (symmetric) matrices can always be di-
agonalized by means of a unitary (orthogonal) modal matrix.
Example 8.3
The real symmetrix matrix
5 2 2
" #
A= 2 2 1
2 1 2
has the characteristic polynomial d(s) = (s − 1)2 (s − 7). We observe that the eigenvalues
are real.
Two linearly independent eigenvectors associated with the multiple eigenvalue λ1 = 1
can be found by solving
4 2 2
" #
(A − λ1 I)v = 2 1 1 v=0
2 1 1
as
−1 −1
" # " #
v11 = 2 , v22 = 0
0 2
Applying the Gram-Schmidt process to {v11 , v12 }, and normalizing the orthogonal eigen-
vector generated by the process, we obtain two orthonormal eigenvectors associated with
λ1 = 1 as
−1 −2
" # " #
1 1
u11 = √ 2 , u12 = √ −1
5 0 30 5
2
" #
v2 = 1
1
274 Unitary and Hermitian Matrices
√2 − √130 √1
P = [ u11 u12 u2 ] =
5 6
0 √5 √1
30 6
is called a quadratic form in x. Note that q(x) is a scalar for every x ∈ Rn×1 .
Clearly q(0) = 0. If q(x) > 0 for all x 6= 0, then q(x) is said to be positive
definite. If q(x) ≥ 0 for all x, and q(y) = 0 for at least one y 6= 0, then q(x) is
said to be positive semi-definite. q(x) is said to be negative definite (negative
semi-definite) if −q(x) is positive definite (positive semi-definite), and indefinite if it
is neither positive definite nor negative definite.
A real symmetric matrix S is said to be positive definite (positive semi-definite,
negative definite, negative semi-definite, indefinite) if the associated quadratic form
q(x) = xt Sx is positive definite (positive semi-definite, negative definite, negative
semi-definite, indefinite).
Example 8.4
The quadratic form
q1 (x1 , x2 ) = x21 + 2x1 x2 + 4x22
involving the real variables x1 and x2 can be written as
1 1 x1
q1 (x) = [ x1 x2 ] = x t S1 x
1 4 x2
Note that the diagonal elements of S are the coefficients of the square terms x21 and x22 ,
and the symmetrically located off-diagonal elements are one half of the coefficient of the
cross-product term x1 x2 . Since
q1 (x1 , x2 ) = (x1 + x2 )2 + 3x22
8.3 Quadratic Forms 275
is indefinite, because q3 (1, 0) = 1 > 0 and q3 (1, −1) = −1 < 0. Thus its matrix
1 3
S3 =
3 4
is indefinite.
x = P x̃ (8.2)
Example 8.5
The matrix S1 in Example 8.4 has the eigenvalues
√ √
λ1 = (5 + 13)/2 ≈ 4.3028 , λ2 = (5 − 13)/2 ≈ 0.6972
Since both eigenvalues are positive, S1 and hence q1 are positive definite.
S2 has the eigenvalues
λ1 = 4 , λ2 = 0
we have
n
X n
X
k x k2 = k αi vi k2 = | αi |2
i=1 i=1
Considering
n X
X n n X
X n n
X
xt Sx = αi αj vit Svj = αi αj λi vit vj = λi | αi |2
i=1 j=1 i=1 j=1 i=1
we find that
n
X n
X
λmin | αi |2 ≤ xt Sx ≤ λmax | αi |2
i=1 i=1
or equivalently
where λmin and λmax are the minimum and and maximum eigenvalues of S. Clearly,
equality on either side holds if x equals the corresponding eigenvector. (8.4) estab-
lishes bounds on a quadratic form q(x) = xt Sx. It also provides an alternative
explanation to the relation between sign-definiteness of a quadratic form and the
eigenvalues of its symmetric matrix.
8.3 Quadratic Forms 277
Expressions similar to (8.3) and (8.4) can easily be derived for a quadratic form
in complex variables. With z = P z̃, where P is a unitary modal matrix of H, the
quadratic form q(z) = zh Hz is transformed into
n
X
q(z) = z̃h P h HP z̃ = z̃h Dz̃ = q̃(z̃) = λi | z̃i |2 (8.6)
i=1
Again, the sign properties of q(z) can be deduced from the eigenvalues of H. For
example, the Hermitian matrix of the quadratic form in Example 8.6 has the real
eigenvalues
√ √
λ1 = (3 + 5)/2 ≈ 2.6180 , λ1 = (3 − 5)/2 ≈ 0.3280
278 Unitary and Hermitian Matrices
Since both eigenvalues are positive we conclude that the quadratic form (and hence
its matrix) are positive definite.
Similarly, for q(z) = zh Hz (8.4) becomes
λmin k z k2 ≤ zh Hz ≤ λmax k z k2 (8.7)
where λmin and λmax are the minimum and and maximum eigenvalues of H.
Finally, we observe from Example 8.6 that a quadratic form in complex variables
is equivalent to a quadratic form in real variables (which are the real and imaginary
parts of the complex variables). To prove this statement in general, let
z = x + iy , H = S + iK
in (8.5), where x and y are real, S is a real symmetric matrix, and K is a real
skew-symmetric matrix. Then, noting that
xt Kx = yt Ky = 0 , xt Sy = yt Sx
q(z) can be expressed as
q(z) = (xt − iyt )(S + iK)(x + iy)
= xt Sx − xt Ky + yt Kx + yt Sy
(8.8)
t t S −K x
= [x y ] = Q(x, y)
K S y
involving real quantities only.
Since the quadratic forms q(x) and Q(y, z) in (8.8) are equivalent, the eigenvalues
of their matrices H = S + iK and
S −K
H̃ = (8.9)
K S
must be related. This relation is studied in Exercise 8.12.
√ √
a) λ1 ≥ λ2 > 0 (S positive definite): Letting a1 = 1/ λ1 , a2 = 1/ λ2 , the
equation of the conic takes the form
x̃21 x̃22
+ =1
a21 a22
which represents an ellipse in the x̃1 x̃2 plane with axes of length a1 and a2 .
√
b) λ1 > λ2 = 0 (S positive semi-definite): Again, letting a1 = 1/ λ1 , the equa-
tion becomes
x̃21 = a21
which represents two straight lines x̃1 = ∓a1 .
√ √
c) λ1 > 0 > λ2 (S indefinite): With a1 = 1/ λ1 , a2 = 1/ −λ2 , we have
x̃21 x̃2
2 − 22 = 1
a1 a2
which represents a hyperbola.
d) 0 ≥ λ1 ≥ λ2 (S negative definite or negative semi-definite): In these cases no
point in the x̃1 x̃2 plane (and therefore, no point in the x1 x2 plane) satisfies the
equation.
Note that the parabola, which is another conic, does not occur in any of the cases
considered above, because it is not a central conic. A more general equation, which
also includes the parabola, is considered in Exercise 8.28.
Example 8.7
Consider the equation
(10 − c)x21 + 2(6 + c)x1 x2 + (10 − c)x22 = xt Sx = 1
where c is a real parameter, and
10 − c 6+c x1
S= , x=
6+c 10 − c x2
The matrix S has the eigenvalues
λ1 = 16 , λ2 = 4 − 2c
and an orthogonal modal matrix
1 1 −1
P = √
2 1 1
which is independent of the parameter c. Thus the change of basis as in (8.2) transforms
the equation of the conic into
16x̃21 + (4 − 2c)x̃22 = 1
For c = 0, 2, 4, the equation becomes and represents
c=0: 16x̃21 + 4x̃22 = 1 : an elipse
c=2: 16x̃21 = 1 : two parallel lines
c=4: 16x̃21 − 4x̃22 = 1 : a hyporbala
280 Unitary and Hermitian Matrices
0.8
c=2 c=4
0.4 c=0
−0.4
−0.8
−0.8 −0.4 0 0.4 0.8
The loci of the points that satisfy the given equation for each case are shown in Figure
8.1. Note that the transformation in (8.2) corresponds to a counter-clock-wise rotation
of the coordinate axes by 45◦ .
(a) (b)
(c)
x̃21 x̃2
2 − 23 = 1
a1 a3
The quadric surfaces corresponding to the above cases are illustrated in Figure 8.2,
where (a) contains an ellipsoid and a hyperboloid of two sheets, (b) a hyperboloid of
one sheet, and (c) an elliptic and a hyperbolic cylinder.
where
Σ1 = diag [ σ1 , . . . , σk ]
and let
V = [ V1 V2 ]
and
Let
U1 = AV1 Σ−1
1
8.4 The Singular Value Decomposition 283
Since
−1 2 −1
U1h U1 = Σ−1 h h −1
1 V1 A AV1 Σ1 = Σ1 Σ1 Σ1 = I
U = [ U1 U2 ]
(a) follows from (8.11) on noting that U and V are nonsingular, and (b) is a direct
consequence of (a). (c) is the definition used in the proof of Theorem 8.3. (d) follows
from
AAh U = U ΣV h V Σh U h U = U ΣΣh
(e) and (f) are obvious. Finally, (g) is a result of the fact that if A is Hermitian with
h 2 2
eigenvalues
p λi then A A = A has the eigenvalues λi , so that singular values of A
2
are σi = λi = | λi |.
The following corollary of Theorem 8.3 characterizes various subspaces associated
with a matrix. The proof is left to the reader (see Exercise 8.20).
284 Unitary and Hermitian Matrices
Then
a) im (U1 ) = im (A)
b) im (U2 ) = ker (Ah )
c) im (V1 ) = im (Ah )
d) im (V2 ) = ker (A)
Thus U1 U1h is the orthogonal projection matrix on im (A), U2 U2h is the orthogonal
projection matrix on ker(Ah ), etc.
As a consequence of Corollary 8.3.1 we also have
⊥ ⊥
Cm×1 = im (U1 ) ⊕ im (U2 ) = im (A) ⊕ ker (Ah )
Example 8.8
Let us find the singular value decomposition of
2 0
" #
A= 1 1
0 2
Eigenvalues of
5 1
At A =
1 5
Letting
√ √
1/√3 1/ 2
U1 = AV1 Σ−1
1 = 1/√3 √0
1/ 3 −1/ 2
8.4 The Singular Value Decomposition 285
and completing the columns of U1 to an orthonormal basis for R3×1 , the singular value
decomposition of A is obtained as
A = U ΣV t
√ √ √ √
1/√3 1/ 2 1/√6 6 0 √ √
1/√2 −1/√2
= 1/ 3
√ 0
√ −2/√6 0 2
1/ 2 1/ 2
1/ 3 −1/ 2 1/ 6 0 0
we observe that im (A) = im (U1 ), verifying the result of Corollary 8.3.1(a). Other results
of the corollary can be verified similarly.
In other words, the least-squares solutions of (8.12) are precisely the solutions of
(8.15). In particular,
x = φLS = V1 Σ−1 h †
1 U1 y = A y (8.16)
is a least-squares solution with minimum 2-norm (Exercise 8.22). Note that φLS can
be written in open form as
uh1 y uh y
φLS = v1 + · · · + r vr (8.17)
σ1 σr
Consider the matrix
A† = V1 Σ−1 h
1 U1 (8.18)
A† A and AA† are both Hermitian.1 Moreover, A† reduces to a left inverse when
r(A) = n (in which case V1 V1h = In ), to a right inverse when r(A) = m (in which
case U1 U1h = Im ), and to A−1 when A is square and nonsingular.
Example 8.9
Consider a linear system with
5 0 5 4
" # " #
A= 1 1 2 , y= −15
0 5 5 2
5 0 1
" # " #
t −1 t
R= 1 1 , yA = R(R R) R = 0
0 5 −1
0.2 −1
" # " #
x= −0.2 +c −1
0.0 1
1 In fact,  = A† is the unique generalized inverse such that  A and A are both Hermitian.
G G G
We will not prove this fact.
8.4 The Singular Value Decomposition 287
0.2
" #
x = φLS = −0.2
0.0
Σ = diag [ 9, 5, 0 ]
√ √ √ √
5 3 3 √2 1 √3 √2
1 1
U= √ 2 √0 −5√2 , V = √ 1 − 3 √2
3 6 5 −3 3 2 6 2 0 − 2
106 10 −56
" #
1
= −56 10 106
810
50 20 50
and the expression in (8.16) gives the same minimum norm least-squares solution as
obtained above. (Alternatively, (8.17) produces the same solution.)
The results can be checked using the MATLAB commands [U,S,V]=svd(A), which
produces the singular value decomposition of A, and API=pinv(A), which computes the
pseudoinverse of A. The reader is also urged to verify that A† A and AA† are both
symmetric.
k A kp = max { k Ax kp }
k x kp =1
with equality holding for the eigenvector corresponding to λmax (Ah A), we find that
q
k A k2 = λmax (Ah A) = σ1 (8.19)
This is a significant result, which states that the matrix norm subordinate to the
Euclidean vector norm is its largest singular value.
An equally significant result, which is left to the reader to prove (Exercise 8.25),
is that
q
k A kF = σ12 + · · · + σr2 (8.20)
288 Unitary and Hermitian Matrices
Then B has rank r(B) = 2 (which can be verified by MATLAB), and is the closest
rank-2 matrix to A in Frobenius norm. The command norm(A-B,’fro’) computes the
Frobenius norm of the difference of A and B as
k A − B kF = 0.0063
as expected.
Example 8.11
The matrix
0.2676 0.5111 0.7627
" #
A= 0.6467 0.6931 0.5241
0.7371 0.6137 0.1690
µ = 1.6305e + 004
implies that A is badly conditioned, and that A−1 has large element values of the order
of 104 . The reader can verify this observation by computing A−1 using MATLAB.
8.5 Exercises
1. (a) Verify that the following matrices are unitary.
(b) Find a unitary modal matrix for each of these matrices and diagonalize them.
(c) Use MATLAB command eig to verify your results in part (b).
" √ # "
0 0 1
#
1/2 − 3/2 0 i
A= √ , B= , C= 0 1 0
3/2 1/2 −i 0 1 0 0
3 The expression in (8.22) is similar to the singular value decomposition of A−1 except that its
2. Prove that the product of two unitary matrices is unitary. Verify this result for the
matrices A an B in Exercise 8.1.
3. In the xyz space a counterclockwise rotation about the z axis by an angle of θz is
represented by the matrix
cos θz − sin θz 0
" #
Rz = sin θz cos θz 0
0 0 1
(a) Determine the structure of the rotation matrices Ry and Rx about the y and x
axes.
(b) Show that Rx , Ry and Rz are orthogonal.
(c) Characterize an invariant subspace for each of Rx , Ry and Rz .
4. Let
0 1 0 ··· 0
0 0 1 ··· 0
A= .. .. .. .. ..
. . . . .
0 0 0 ··· 1
1 0 0 ··· 0 n×n
5. Use MATLAB command eig to verify your answer to Exercise 8.4 for n = 2, 3, 4.
6. Let b be a unit vector. Show that A = I − bbt is a symmetric projection matrix.
7. Prove the Schur’s theorem: Let A be an n×n complex matrix. There exists a unitary
matrix U such that U h AU is an upper triangular matrix with diagonal elements being
the eigenvalues of A. Hint: Refer to Exercise 5.13.
8. A ∈ C n×n is said to be an involution if A2 = I. Show that any two of the following
imply the third.
(a) A is Hermitian
(b) A is unitary
(c) A is an involution.
9. Verify the result of Exercise 8.8 for the matrices B and C in Exercise 8.1.
10. Prove that a Hermitian matrix A has a unitary modal matrix, and thus complete the
proof of Theorem 8.2. Hint: Let v1 be a unit eigenvector of A = A1 associated with
some eigenvalue λ1 . Choose V1 such that P1 = [ v1 V1 ] is unitary, and consider
P1h AP1 .
11. (a) Show that eigenvectors of a unitary matrix associated with distinct eigenvalues
are orthogonal.
(b) Show that eigenvectors of a Hermitian matrix associated with distinct eigenvalues
are orthogonal.
8.5 Exercises 291
(a) by hand,
(b) by using MATLAB.
⊥ ⊥
20. Prove Corollary 8.1. Hint: Cm×1 = im (U1 ) ⊕ im (U2 ) and Cn×1 = im (V1 ) ⊕ im (V2 ).
21. Let
−1 i
A=
−i −1
it follows that
r
X ∗
aij = σk uik vjk
k=1
8.5 Exercises 293
where
u1k v1k
.. ..
uk = . and vk = .
umk vnk
are the kth left and right singular vectors of A. Manipulate the expression
m n m n r r
X X X X X X
k A kF = aij a∗ij = ( ∗
σk uik vjk )( ∗
σl uil vjl )
i=1 j=1 i=1 j=1 k=1 l=1
27. Let
0.5055 0.6412 0.8035
" #
A= 0.1693 0.0162 0.6978
0.5247 0.8369 0.4619
Using MATLAB
xt Sx + 2rt x = 1
x = P x̃ − xo , x̃ = P t (x + xo ) (8.23)
x2
6
α2 M@
I
@ B
@
x̃2 B@
MB α̃@
2 B @ x̃1
B @ B
x̃ @
@
B x @ B α̃1
B @B
P x̃ B @B
o
x 1
B 0̃
B -
0 α1 x1
Figure 8.3: The coordinate transformation in (8.23)