Professional Documents
Culture Documents
Appendix B
Appendix B
LINEAR ALGEBRA
x1 + x2 ∈ X for all x1 , x2 ∈ X
and
αx ∈ X for all x ∈ X and for all α ∈ F
1. Commutativity
x1 + x2 = x2 + x1
2. Associativity
a. (x1 + x2 ) + x3 = x1 + (x2 + x3 )
b. (α1 α2 )x = α1 (α2 x)
3. Distributivity
a. α(x1 + x2 ) = αx1 + αx2
b. (α1 + α2 )x = α1 x + α2 x
4. Identity elements
a. There is a zero vector 0 ∈ X such that x + 0 = x
b. For additive and multiplicative identity elements in F, denoted 0
and 1, respectively,
0x = 0 and 1x = x
Example B.1
Fn = {x = (x1 , x2 , . . . , xn ), xi ∈ F, i = 1, . . . , n}
i.e., the set of all n-tuples with elements in F. When F is either the real field
R or the complex field C, Rn and Cn denote real and complex Euclidean
space, respectively.
Example B.2
span{x1 , x2 , . . . , xk } := {x = α1 x1 + α2 x2 + · · · + αk xk , αi ∈ F}
implies that α1 = α2 = · · · = αk = 0.
x = α1 x1 + α2 x2 + · · · + αk xk
is unique.
0=x−x
= (α1 x1 + α2 x2 + · · · + αk xk ) − (β1 x1 + β2 x2 + · · · + βk xk )
= (α1 − β1 )x1 + (α2 − β2 )x2 + · · · + (αk − βk )xk
B.2 SUBSPACES
S = span{x1 , x2 , . . . , xk }
A basis for S is not unique; however, all bases for S have the same
number of elements, which defines the dimension of the linear subspace
S. Since a linear vector space X can be viewed as a subspace of itself, the
concepts of basis and dimension apply to X as well.
420 LINEAR ALGEBRA
On R3 :
because
a11 a12
A=
a21 a22
1 0 0 1 0 0 0 0
= a11 + a12 + a21 + a22
0 0 0 0 1 0 0 1
{1, x, x 2 , . . . , x k }
0
..
.
0
ei = 1 ← i th position i = 1, . . . , n
0
.
..
0
x = x1 e1 + x2 e2 + x3 e3
1 0 0
= x1 0 + x2 1 + x3 0
0 0 1
x1
= x2
x3
422 LINEAR ALGEBRA
tij ∈ F for i, j = 1, · · · , n
x = α1 x1 + α2 x2 + · · · + αn xn
= β1 y1 + β2 y2 + · · · + βn yn
n
= αi xi
i=1
In matrix form,
α1 t11 t12 ··· t1n β1
α2 t21 t22 ··· t2n β2
. = . .. .. .
.. .. .
..
. . ..
αn tn1 tn2 ··· tnn βn
so that
x = (10)y1 + (−8)y2 + (13)y3
n
∗
x, y := x y = x i yi
i=1
ORTHOGONALITY AND ORTHOGONAL COMPLEMENTS 425
1
1
n 2
||x|| = x, x 2 = |xi |2
i=1
For a subspace S ⊂ Cn
S⊥ = {y ∈ Cn |y, x = 0 ∀x ∈ S}
S⊥ = {y ∈ Cn |y, xi = 0; i = 1, . . . , k}
S⊥ = span{y1 , y2 , . . . , yn−k }
for any linearly independent set of vectors {y1 , y2 , . . . , yn−k } that satisfy
yj , xi = 0 i = 1, . . . , k j = 1, . . . , n − k
426 LINEAR ALGEBRA
Example B.7 On R3 ,
Definition B.8 Let X and Y be linear vector spaces over the same field
F. A transformation A : X → Y is linear if
As we have seen previously, the m-tuple (a1j , a2j , . . . , amj ) defines the
coordinates of Axj in the basis {y1 , y2 , . . . , ym }.
Next, for every x ∈ Cn and y := Ax, we have the unique representation
in the appropriate basis
x = α1 x1 + α2 x2 + · · · + αn xn
y = β1 y1 + β2 y2 + · · · + βm ym
n
= αj Axj
j =1
n
m
= αj aij yi
j =1 i=1
m
n
= aij αj yi
i=1 j =1
Thus, with respect to the basis {x1 , x2 , . . . , xn } for Cn and the basis
{y1 , y2 , . . . , ym } for Cm , the linear transformation A : Cn → Cm has the
m × n matrix representation
a11 a12 ··· a1n
a21 a22 ··· a2n
A=
... .. .. .
.
..
. .
am1 am2 ··· amn
If different bases are used for Cn and/or Cm , the same linear transformation
will have a different matrix representation. Often a matrix is used to
represent a linear transformation with the implicit understanding that the
standard bases for Cn and Cm are being used.
That is,
1
1 1 0
A 0 = = (1) + (4)
4 0 1
0
0
2 1 0
A 1 = = (2) + (5)
5 0 1
0
0
3 1 0
A0 = = (3) + (6) .
6 0 1
1
1 0 0
Ax2 = A (0) 0 + (2) 1 + (2) 0
0 0 1
1 0 0
= (0)A 0 + (2)A 1 + (2)A 0
0 0 1
1 2 3
= (0) + (2) + (2)
4 5 6
10
=
22
1 0 0
Ax3 = A (0) 0 + (0) 1 + (3) 0
0 0 1
1 0 0
= (0)A 0 + (0)A 1 + (3)A 0
0 0 1
1 2 3
= (0) + (0) + (3)
4 5 6
9
=
18
430 LINEAR ALGEBRA
Thus, with a different basis for R3 , the same linear transformation now
has the matrix representation
6 10 9
A= .
15 22 18
which illustrates that the matrix representation for the linear transforma-
tion with respect to the new {x1 , x2 , x3 } basis for R3 and the standard basis
for R2 is given by the matrix representation for the linear transformation
with respect to the original standard basis for R3 and the standard basis
for R2 multiplied on the right by a 3 × 3 matrix characterizing the change
of basis on R3 from the {x1 , x2 , x3 } basis to the standard basis {e1 , e2 , e3 }
as described in Section B.4. If, in addition, we had considered a change
of basis on R2 , the original matrix representation for the linear transfor-
mation would have been multiplied on the left by a 2 × 2 matrix relating
the two bases for R2 to yield the new matrix representation with respect
to the different bases for both R2 and R3 .
R(A) = Im A = span{a1 , a2 , · · · , an }
rank(A) + nullity(A) = n
We seek to find the rank and nullity of A along with bases for Im A and
Ker A. A key computational tool is the application of elementary row
operations:
432 LINEAR ALGEBRA
We then have
nullity(A) = n − rank(A)
=4−2
=2
x1
0 1 0 1 x2 + x4 0
0 x
0 1 −2 2 = x3 − 2x4 = 0
x3
0 0 0 0 0 0
x4
The set
1 0
0 −1
0, 2
0 1
Since for every z ∈ Im A, there must exist some x ∈ Cn such that z = Ax,
it follows that y, z = 0, that is, y ∈ [Im A]⊥ . The desired containment
Ker A∗ ⊂ [Im A]⊥ follows because y ∈ Ker A∗ was arbitrary. Having
established the required subspace containments, we may conclude that
[Im A]⊥ = Ker A∗ .
Note that since the orthogonal complements satisfy S = T if and only if
S⊥ = T⊥ and (S⊥ )⊥ = S, the second subspace identity in Proposition B.4
is equivalent to [ImA∗ ]⊥ = Ker A. This, in turn, follows by applying
the first subspace identity in Proposition B.4 to the linear transformation
A∗ : Cm → Cn .
|λI − A| = 0
σ (A) = {λ1 , λ2 , . . . , λn }
1
As indicated, the characteristic polynomial of an n × n matrix is guaranteed to be a
monic degree-n polynomial that, by the fundamental theorem of algebra, must have n
roots. If A is real, the roots still may be complex but must occur in conjugate pairs.
436 LINEAR ALGEBRA
w ∗ (λi I − A) = 0 equivalently w ∗ A = λi w ∗
1
w3 = w 2 = 1
−2 + j 2
1 1
for i = 1, . . . , d.
As a consequence, for each distinct eigenvalue λi , i = 1, . . . , d, there are
ni linearly independent solutions to the homogeneous equation (λi I −
A)v = 0, which we denote by {v1i , v2i , . . . , vni i }. This set of vectors, in turn,
forms a so-called eigenbasis for the eigenspace Ker(λi I − A) associated
with the eigenvalue λi .
The following facts establish some useful relationships.
B = T −1 AT equivalently A = TBT −1
440 LINEAR ALGEBRA
Proposition B.18 If A and B are similar matrices, then they have the
same eigenvalues.
This result is proven by using determinant properties to show that similar
matrices have identical characteristic polynomials.
A diagonal matrix necessarily displays its eigenvalues on the main
diagonal. For a general matrix A, diagonalization refers to the process of
constructing a similarity transformation yielding a diagonal matrix that,
by virtue of the preceding proposition, displays the eigenvalues of A. The
ability to diagonalize the matrix A via similarity transformation is closely
connected to its underlying eigenstructure.
To investigate further, note that, by definition, the distinct eigenvalues
of A ∈ Cn×n have associated algebraic multiplicities that satisfy
m1 + m2 + · · · + md = n
AT = T
n 1 + n 2 + · · · + n d = m1 + m2 + · · · + md = n
EIGENVALUES, EIGENVECTORS, AND RELATED TOPICS 441
= T −1 AT
displaying the parameter λ on the main diagonal and 1s on the first super-
diagonal. An n × n Jordan matrix is a block diagonal matrix constructed
from Jordan block matrices:
Jk1 (λ1 ) 0 ··· 0
0 Jk2 (λ2 ) ··· 0
J =
... .. ..
with k1 + k2 + · · · + kr = n
. .
0 0 ··· Jkr (λr )
J = T −1 AT
in which there are ni Jordan blocks associated with the eigenvalue λi , and
the sum of the sizes of these blocks equals the algebraic multiplicity mi ,
for i = 1, . . . , d. The matrix J is unique up to a reordering of the Jordan
blocks and defines the Jordan canonical form of A.
Note that the algebraic and geometric multiplicities alone do not
completely determine the Jordan canonical form. For example, the Jordan
matrices
λ 1 00 0 λ 1 0 00
0 λ 1 0 0 0 λ 1 00
J1 =
0 0 λ 0 0
J2 =
0 0 λ 10
0 0 0 λ 1 0 0 0 λ0
0 0 00 λ 0 0 0 0λ
eigenvalue is scalar, then the geometric and algebraic multiplicities for this
eigenvalue must be equal. If the geometric multiplicity equals the algebraic
multiplicity for each distinct eigenvalue, then the Jordan canonical form
is diagonal. This is consistent with Theorem B.19.
Cayley-Hamilton Theorem
Theorem B.22 For any matrix A ∈ Cn×n with characteristic polynomial
there holds
An + an−1 An−1 + · · · + a1 A + a0 I = 0 (n × n)
Example B.12 For the matrix A studied in Example B.10, the charac-
teristic polynomial was found to be
λ3 − 5λ2 + 9λ − 5
Then
1 0 0 1 0 0
A3 − 5A2 + 9A − 5I = 12 13 −11 − 5 · 6 7 −4
22 22 −9 8 8 −1
1 0 0 1 0 0
+ 9 · 2 3 −1 − 5 · 0 1 0
2 2 1 0 0 1
0 0 0
= 0 0 0
0 0 0
444 LINEAR ALGEBRA
1
||x||p = (|x1 |p + |x2 |p + · · · + |xn |p ) p for all p ∈ [1, ∞)
||x||∞ = max |xi |
1≤i≤n
1
||x||2 = (|x1 |2 + |x2 |2 + · · · + |xn |2 ) 2
1
= (x ∗ x) 2
which is the familiar Euclidean norm. These p− norms satisfy the so-
called Hölder inequality
||Ax||
||A|| = sup
x =0 ||x||
in which sup stands for supremum, or least upper bound, here taken over
all nonzero vectors in Cn . It can be shown that this definition is equiva-
lent to
||A|| = max ||Ax||
||x||=1
m
||A||1 = max |aij |
1≤j ≤n
i=1
1
||A||2 = (λmax (AT A)) 2
n
||A||∞ = max |aij |
1≤i≤m
j =1
1
2
m n
||A||F = |aij |
2
i=1 j =1
This can be seen to correspond to the Euclidean vector norm of the (nm) ×
1-dimensional vector obtained by stacking the columns of A on top of
one another.