Professional Documents
Culture Documents
Appendix3 Matrices
Appendix3 Matrices
B.1 Definition
In this tutorial, we only consider Linear Operators between finite-dimensional Complex Vector
Spaces CN and CM with α ∈ C. Consequently, a Linear Operator A is represented by a Matrix
A : CN → CM .
235
Derivation.
′ M
Assume {un }N 1
n=1 and {um }m=1 is an Orthonormal Basis of an abstract N -dimensional vector space
S and the abstract M -dimensional vector space S , respectively, and A : S → S′ is a Linear Operator
′
* N
+ N
X X
ym = hy, u′m i = hAx, u′m i = A xn un , u′m = hAun , u′m ixn (B.3)
n=1 n=1
△ △
[A]m,n = hAun , u′m i and [AT ]n,m = hAun , u′m i. (B.4)
236
A : S → S′ , u1 7→ Au1 = A1,1 u′1 + A2,1 u′2
△ △
A1,1 = hAu1 , u′1 i, A2,1 = hAu1 , u′2 i
u1
A1,1 u′1
S S′
Au1
u2 A2,1
u′2
237
B.2 Elementwise Perspective on Matrices
and
a1,1 a1,2 ··· a1,N x1
a2,1 a2,2 ··· a2,N
x2
Ax = .. .. .. ..
. . . .
aM,1 aM,2 · · · aM,N xN
a1,1 x1 + a1,2 x2 + · · · + a1,N xN
a2,1 x1 + a2,2 x2 + · · · + a2,N xN
= ..
.
aM,1 x1 + aM,2 x2 + · · · + aM,N xN
PN
n=1 a 1,n x n
PN a x
n=1 2,n n
= .. .
.
PN
n=1 aM,n xn
238
Definition. Matrix, Transpose Matrix, Adjoint Matrix
a1,1 a1,2 · · · a1,N
a2,1 a2,2 · · · a2,N
A = .. .. .. ,
. . .
aM,1 aM,2 · · · aM,N
a1,1 a2,1 ··· aM,1
T
a1,2 a2,2 ··· aM,2
A = .. .. .. ,
. . .
a1,N a2,N · · · aM,N
a∗1,1 a∗2,1 ··· a∗M,1
a∗1,2 a∗2,2 ··· a∗M,2
AH =
.. .. .. .
. . .
a∗1,N a∗2,N · · · a∗M,N
239
The Adjoint Matrix AH ∈ CN ×M of a matrix A ∈ CM ×N has the constituting property
H
hAx, yi = x, AH y ⇔ y H Ax = AH y x, ∀x ∈ CN , y ∈ CM . (B.6)
T H
with AT = A and AH = A.
240
Definition. Identity Matrix, Zero Matrix, Inverse Matrix, Selection Vector
1 0 ··· 0
0 1
··· 0
I = .. . . . .. = diag [1 1 · · · 1]
. .
0 0 ··· 1
0 0 ··· 0
0 0 ··· 0
0= .. .. ..
. . .
0 0 ··· 0
· · 0} 1 0| ·{z
ei = [ 0| ·{z · · 0} ]T ∈ CN with
i−1 zeros N −i zeros
0
a1,1 a1,2 ··· a1,N . a1,i
.. a2,i
a2,1 a2,2 ··· a2,N
Aei = .. .. .. 1 = .. = ai .
. . . . .
..
aM,1 aM,2 · · · aM,N aM,i
0
241
B.3 Fundamental Subspace Perspective on Matrices
Range Space of A : range [A] = Ax| x ∈ CN (B.7)
Null Space of A : null [A] = x ∈ CN Ax = 0 (B.8)
Orthogonal Complement of range [A] : range [A]⊥ = null A H
(B.9)
Orthogonal Complement of null [A] : null [A]⊥ = range AH . (B.10)
CN
range [A]
null [A] 0
0
CM
242
B.4 Eigenvectors and Eigenvalues
The constant λ is the Eigenvalue to the Eigenvector v. The v and its corresponding λ form a
so-called Eigenpair of the matrix A.
For finite dimensional matrices the Set of Eigenvalues is discrete, the maximum number of different
eigenvalues is M .
since λi and λj are real-valued (1st part) and different for i 6= j by assumption.
243
B.5 Orthogonal Projectors
x ∈ CM
P : CM → U
U⊥
U
x̂ = P x
244
According to the Closest Point Theorem2 we obtain
Note.
Projection Operators are always Indempotent.
Orthogonal Projectors must be Indempotent and Self-Adjoint.
Oblique Projectors (non-orthogonal projectors) are only indempotent, i.e., P 2 = P , but Self-
Adjointness does not hold.
Oblique Projectors are not considered in this tutorial.
2 The Closest Point Theorem applies for the category of Convex Sets which Linear Subspaces are belonging to.
245
Orthogonal Projection on Subspaces.
Given a pair of matrices A : CN → CM and B : CM → CN and A is a Left-Inverse of B, i.e.,
AB = I, then
Example. Assume the column vectors of U form an ONB of U, then P = U U H is the Orthogonal
Projector onto U. For a proof consider the plugins
A = UH and B = U . (B.19)
246
Orthogonal Projection via Pseudo-Inverse.
If (AH A)−1 exists, then A+ H −1 H
left = (A A) A is a Left-Inverse of A, i.e.,
−1 H
(1) AA+ H
left = A(A A) A is the Orthogonal Projector onto range [A] (B.20)
H −1
H
(2) A+ H
right A = A (AA ) A is the Orthogonal Projector onto range A (B.21)
247
Weighted Sum of Disjoint Orthogonal Projectors
P
with rank R = Pi=1 Ri and range [P i ] ⊥ range [P j ] for all i 6= j, i = 1, . . . , P , i.e., all projectors are
mutually orthogonal to each other.
AAH = AH A. (B.23)
248
B.6 Special Matrices
AH = A. (B.24)
249
Eigenpairs of Normal Matrices
Consequently, any uj ∈ range [P j ] and its corresponding weight wj form an Eigenpair of A, since for
every element of range [P j ] we obtain P j uj = uj by definition and P i uj = 0 for i 6= j due to the
orthogonality of projectors, and thus
P
!
X
Auj = w i P i uj (B.26)
i=1
P
X
= wi (P i uj ) = wj uj . (B.27)
i=1
250
Spectral Theorem of Normal Matrices
Since (1) Normal Matrices A are Linear Combinations of Orthogonal Projectors P i and
any uj ∈ range [P j ] and its corresponding weight wj form an Eigenpair of A, and
(2) since any P i can be represented as a Linear Combination of rank-1 projectors ui,j uH
i,j with respect
Ri
to elements of an ONB {ui,j }j=1 , with
with U i = [ui,1 , . . . , ui,Ri ] and Ri is the dimension of range [P i ], we obtain the main result of the
Spectral Theorem of Normal Matrices,
P
X
A= wi P i (B.30)
i=1
P
X Ri
P X
X
= wi U i U H
i = wi ui,j uH
i,j (B.31)
i=1 i=1 j=1
H
w1 I R1 ×R1 · · · 0 U1
.. . .. .
.. .
.. .
= U1 · · · UP . (B.32)
0 · · · wP I RP ×RP UHP
251
B.7 Singular Value Decomposition
Any Compact Linear Operator A can be decomposed by means of the Singular Value Decom-
position (SVD) and any Finite Dimensional Matrix A : CN → CM is a compact linear operator.
(1) Since the product of matrices AH A—the so-called Gramian Matrix—is Normal, Self-Adjoint
and Positive Definite,4 the Spectral Theorem can be applied such that
P
X R
X
H
A A= λi P i = λi v i v H
i , (B.33)
i=1 i=1
H
4 AH A = AH A and v H AH Av = v H λv = λv H v = λ kvk2 ≥ 0.
252
(2) Given the Orthonormal Vectors {v i }R R
i=1 , the vectors {ui }i=1 , with
△ −1
ui = λi 2 Av i , (B.34)
−1 −1 H
= λj 2 λi 2 v H
j A Av i
−1 −1
= λj 2 λi 2 λi v H
j vi
1 ; i=j
= .
0 ; otherwise
Definition.
H
The elements of an ONB {v i }R R
i=1 with Eigenvalues {λi }i=1 of a Gramian Matrix A A are called
Right Singular Vectors of A.
√
The elements of the ONB {ui = 1/ λi Av i }R R
i=1 with Eigenvalues {λi }i=1 of a Gramian Matrix
H
A A are called corresponding Left Singular Vectors of A.
253
The Singular Value Decomposition (SVD) of an arbitrary matrix A : CN → CM is constituted
by
R
X
A= σi ui v H
i , with matrix rank R (B.35)
i=1
uH
j ui = δi,j and v H
j v i = δi,j and σi > 0 i, j = 1, . . . , Q. (B.36)
254
Derivation.
Applying the matrix A to a vector x can be equivalently expressed by
Ax = A(x− + x⊥ ) = Ax− ,
with x− ∈ range AH ≡ null [A]⊥ and x⊥ ∈ null [A].
H
Since {v i }R
i=1 forms an ONB for range A ≡ null [A]⊥ , we obtain
Ax = Ax−
R
X
=A vivH
i x−
i=1
R
X
= Av i v H
i x−
i=1
XR p
= λi ui v H
i x−
i=1
R p
X
= λi ui v H
i x,
i=1
√
with σi = λi .
255
Sorted Matrix Representation
A = U ΣV H , with (B.39)
U = [u1 u2 · · · uR ] ∈ CM ×R , (B.40)
σ1
Σ= ...
= diag [σ1 σ2 · · · σR ] ∈ C
R×R
(B.41)
σR
vH 1
VH = ... ∈ CR×N .
(B.42)
vH R
256
Extended Sorted Matrix Representation. By accomplishing the ONB of the Left Singular Vec-
tors (column vectors of U ) and Right Singular Vectors (column vectors of V ) by their Orthog-
onal Complements, and taking into account that Q = min(M, N ),
U⊥ ⊥
R<Q = [uR+1 uR+2 · · · uQ ] and U Q<M = [uQ+1 uQ+2 · · · uM ] (B.43)
V⊥ ⊥
R<Q = [v R+1 v R+2 · · · v Q ] and V Q<N = [v Q+1 v Q+2 · · · v N ] (B.44)
⊥
range [U ] ⊕ range U ⊥ R<Q ⊕ range U Q<M = C
M
and (B.45)
⊥ ⊥ N
range [V ] ⊕ range V R<Q ⊕ range V Q<N = C , (B.46)
we obtain the Extended Sorted Matrix Representation of the SVD of matrices with maxi-
mum rank (R = Q)
⊥
Σ
U , U Q<M VH ; M ≥ N (Tall)
0
H
A=
U Σ VH ; M = N (Square) (B.47)
V
U Σ 0 ⊥,H ; M ≤ N (Wide).
V Q<N
257
SVD of Matrices without maximum Rank.
For Tall Matrices AM ≥N with linear dependent columns or Wide Matrices AM ≤N with linear
dependent rows, i.e., with
Σ R×R 0R×N −R
VH
AM >N = U, U⊥ ⊥
R<Q , U Q<M
0N −R×R 0N −R×N −R (B.49)
V ⊥,H
R<Q
0M −N ×R 0M −N ×N −R
Σ R×R 0R×N −R VH
AM =N = U, U⊥
R<Q ⊥,H (B.50)
0N −R×R 0N −R×N −R V R<Q
VH
Σ R×R 0R×M −R 0R×N −M V ⊥,H
AM <N = U, U⊥
R<Q R<Q
. (B.51)
0M −R×R 0M −R×M −R 0M −R×N −M ⊥,H
V Q<N
258
Fundamental Subspaces (continued)
Image Space of A : range [A] = Ax| x ∈ CN (B.52)
Null Space of A : null [A] = x ∈ CN Ax = 0 (B.53)
⊥ H
Orthogonal Complement of range [A] : range [A] = null A (B.54)
Orthogonal Complement of null [A] : null [A]⊥ = range AH . (B.55)
CN
range [A]
null [A] 0
0
CM
259
ONB.
The column vectors of U ext = U , U ⊥ with U ⊥ = U ⊥ R<Q , U⊥
Q<M and V ext = V , V ⊥ with
V⊥ = V⊥ R<Q , V ⊥
Q<N from the Extended SVD of a rank-R matrix A ∈ CM ×N constitute the
Orthonormal Bases for the Four Fundamental Subspaces range [A], range [A]⊥ , null [A], and
null [A]⊥ , viz.
260
References
A. Papoulis and S. U. Pillai. Probability, Random Variables and Stochastic Processes, 4th edition, Mc
Graw Hill, 2002.
H. Stark and J. W. Woods. Probability, Random Processes, and Estimation Theory for Engineers, 2nd
edition, Prentice Hall, 1994.
W. Utschick. Stochastische Signale, Aktuelles Manuskript zur gleichnamigen Vorlesung an der Technis-
chen Universität München, 2012.
G. Strang. Linear Algebra and Its Applications, Harcourt Brace Jovanovich, Publishers, 1988.
L. N. Trefethen and D. Bau III. Numerical Linear Algebra, Society for Industrial and Applied Mathematic,
1997.
W. Utschick. Mathematische Methoden der Signalverarbeitung, Aktuelles Manuskript zur gleichnamigen
Vorlesung an der Technischen Universität München, 2012.
265