Professional Documents
Culture Documents
Disclaimer: These notes have not been subjected to the usual scrutiny reserved for formal publications.
They may be distributed outside this class only with the permission of the Instructor.
Remark 1.1 (Why do we care? ) Eigenvalues and eigenvectors are important for
• assessing stability; recall the example
ẋ = −λx
whose solution is x(t) = x0 e−λt and if λ > 0 solution decays to zero, otherwise it blows up. λ is an
’eigenvalue’ for this scalar system.
• as we will see, it is also important for assessing controllability and observability
Av = λv
Definition 1.1 (Spectrum.) The set of all eigenvalues of A is known as the spectrum of A, denoted by
σ(A). The spectral radius of A is defined as
(A − λI)x = 0
Thus, λ is an eigenvalue of A iff the linear system (A − λI)x = 0 has a nontrivial—i.e. x 6= 0—solution.
4-1
4-2 Lecture 4: Eigenvalues and Eigenvectors
Example 1.1 It is well known that even real matrices can have complex eigenvalues. For instance,
0 1
A=
−1 0
Fact 1.2 if A is symmetric (or Hermitian), then all its eigenvalues are real. That is, if A = A∗ = ĀT
(aij = āji ).
Suppose we define
χA (X) = α0 I + α1 X + α2 X 2 + · · · + αn X n
Theorem 1.1 (Cayley-Hamilton.) For any square matrix A with elements in a field F ,
χA (A) = 0
i.e.
0 = α0 I + α1 A + α2 A2 + · · · + αn An
Remark 2.1 (Why do we care? ) Range and null spaces are going to be important for characterizing
the reachable/controllable and observable subspaces.
Consider the linear map A : U → V where dim(U ) = n and dim(V ) = m. Let A be its matrix representation.
R(A) = {w ∈ Cm | w = Av, v ∈ Cm } ⊂ Cm
Lecture 4: Eigenvalues and Eigenvectors 4-3
N (A) = {v ∈ Cn | Av = 0} ⊂ Cn
Fact 2.1 The range space and the null space are both linear subspaces. As a consequence, they have
well-defined dimensions denoted by dim(R(A)) and dim(N (A)).
and thus
α1 v1 + α2 v2 ∈ N (A)
which shows that N (A) is a subspace.
w1 = Av1
w2 = Av2
T (x, y, z, w) = (0, 0)
T (x, y, z) = x + y + z
T (a, 0, 0) = a + 0 + 0 = a, ∀ a ∈ R
Therefore, we can get any real number as an output of T . What do we call maps like this? surjective
4-4 Lecture 4: Eigenvalues and Eigenvectors
T (a, b, c) = a + b + c = 0.
In other words, if we solve for c in terms of the other two variables, it’s the collection
{(a, b, −a − b) ∈ R3 | a, b ∈ R} ⊂ R3
Given the notion of ‘null space’, we can define an eigenbasis for a linear map A.
To find an eigenbasis, you find a basis for each eigenspace of A where the null space of A − λI is called the
eigenspace associated with λ. That is, the vector space Eλ of all eigenvectors corresponding to λ:
Eλ = span{x| Ax = λx, λ ∈ C}
Example 2.2 Let (v, λ) be an eigenvalue/eigenvector pair for A ∈ Cn×n . The nullity of λI − A is called
the geometric multiplicity of λ. In particular, the null space of λI − A (or the eigenspace of A for λ) is
the space of all eigenvectors of A for λ and the zero vector. Its dimension is the geometric multiplicity of
λ.
Note: It is called geometric because it refers to the dimension of a particular space. On the other hand,
the algebraic multiplicity of λ is the number of times λ is a root of the characteristic polynomial for A—
i.e. det(A − λI) = 0. We will return to these concepts as they are important for understanding stability as
well as characterizing controllability and observability.
and thus the geometric multiplicity of λ = 1 is only 2 since all vectors in the null space can be written in
terms of the two basis vectors given.
Definition 2.4 (Injectivity ) A function f : X → Y is injective if, for every pair x1 6= x2 of elements of
X, f (x1 ) 6= f (x2 ).
v ∈ ker(T ) =⇒ v = 0
i.e., ker(T ) = {0}. Conversely, suppose that ker(T ) = {0}. If T (v) = T (w), then T (vw) = 0, and hence
vw = 0. So v = w. Hence,
T (v) = T (w) =⇒ v = w
i.e., T is injective.
Theorem 2.1 (Rank-Nullity Theorem.) For a linear map A : F n → F m , the following holds
Proof Sketch. Note that N (A) ⊂ F n while R(A) ⊂ F m . Consider a basis for N (A) where dim N (A) = k.
Complete that basis (i.e. generate n − k linearly independent vectors independent of the original basis for
N (A)). Consider the representation of x w.r.t. this completed basis.
To prove the Rank-Nullity theorem, you need to know that you can extend a basis of a subspace to a basis
for the whole space.
Lemma 2.1 Let (U, F ) be a finite-dimensional linear space with dim U = n. Suppose {u1 , u2 , . . . , um } is
a set of m linearly independent vectors. There exists n − m additional vectors {ũm+1 , . . . , ũn } such that
forms a basis for U —i.e. any set of linearly independent vectors can be extended to a basis.
Proof. If m = n, no extension is needed. Hence we assume m < n. It follows that {u1 , . . . , um } is not a
basis for U so there must exists a v ∈ U such that
m
X
αi ui 6= v, ∀ {αi }m
i=1 ⊂ F
i=1
Then,
{u1 , . . . , um , v}
is linearly independent so that we can take ũm+1 = v. The lemma follows by induction.
Fact 2.3 Let A ∈ F m×n be a matrix. The rank-nullity theorem gives us that
0 ≤ rankA ≤ min{m, n}
That is, the following fundamental result from linear algebra holds:
where row rank is the row rank of A (i.e. maximum number of linearly independent row vectors of A) and
col rank is the column rank of A (i.e. maximum number of linearly independent column vectors of A).
The column rank of A is the dimension of the column space of A, while the row rank of A is the dimension
of the row space of A.
Theorem 2.2 (Sylvester’s Inequality.) Let A ∈ F m×n and B ∈ F n×p , then AB ∈ F m×p and
Proof. DIY Exercise. Hint: Let W be the co-domain of A and let A|R(B) : R(B) → W be the restriction of
A to R(B). Apply the rank-nullity theorem to A|R(B) .
Remark 2.2 (Why important? ) Sylvester’s inequality and the rank-nullity theorem are the main
tools to prove the following fact, which is very useful for our analysis since we will see that transformation,
e.g., to the controllable canonical form does not change key facts about the system which allow us to
analyze its properties.
Fact 2.4 Rank and nullity are invariant under equivalence (i.e. if two matrices are equivalent via similarity
transformation, then they have the same rank and nullity).
Fact 2.5 In fact, two similar matrices A and Ā = P −1 AP have the same spectrum (i.e. same eigenvalues).
This is important because we can transform our system and preserve key properties like controllability,
observability, stability.
Fact 2.6 Similar matrices have the same characteristic polynomial, eigenvalues, algebraic and geometric
multiplicities. The eigenvectors, however, are in general different.
Theorem 2.3 (A) matrix A ∈ Cm×m is nondefective—i.e., has no defective eigenvalues meaning the
geometric and algebraic multiplicites are equal—if and only if A is similar to a diagonal matrix, i.e.,
A = XΛX −1
where
X = x1 x2 ··· xm
is the matrix formed with the eigenvectors of A as its columns, and Λ = diag(λ1 , . . . , λm ).
we will talk more about this when we discuss Jordan form, but essentially there is a full set of eigenvectors.