Professional Documents
Culture Documents
Ch19-Lecture Notes
Ch19-Lecture Notes
where the coefficients aij on the left side are assumed to be given. The unknowns are x1, x2, ..., xn, and
an additional parameter . There are (n+1) variables but n equations. Thus we know from the start
that one variable will be independent, i.e. one may assign an arbitrary value to one of the variables.
We have already given an example of such equations at the end of Ch17. Here we will discuss Eq 1 in
more detail.
Eq 1 may be written as
Non-trivial solutions of these homogeneous linear equations exist if and only if the determinant of
the coefficients is zero.
When expanded, the determinant on the left gives a polynomial of degree n in . By finding the roots
of this polynomial we obtain the allowed values of which will permit non-trivial solutions of Eq 2.
There will be n roots: 1, 2, ..., n, including the possibility that some roots may repeat. This is
always the first step in solving Eq 2. In the second step, for each of these values, we solve Eq 2 for
the variables x1, x2, ..., xn. Note that for each we will obtain a different set of values for x1, x2, ..., xn.
The linear homogeneous equation system Eq 1 can be written in matrix notation as follows. One
forms a n×n matrix A from the coefficients in Eq 1, and defines a n×1 column vector x of the
unknowns
… 𝑎1𝑛 𝑥1
𝑎11 𝑎12
… 𝑎2𝑛 𝑥2
𝑎21 𝑎22
𝑨=( ⋮ ⋮ ), 𝒙 = 𝑥3 Eq 4
⋮
𝑎𝑛1 𝑎𝑛2 … 𝑎𝑛𝑛 𝑥
( 𝑛)
Left side of the equations in Eq 1 can be expressed as a n×1 column vector by the product, Ax,
according to the rules of matrix multiplication, and this vector is equal to the vector x.
This matrix equation is very special; because in general, when one multiplies two matrices (here A
and x) one expects to find a resultant matrix that is different from the matrices involved in the
product. Yet, according to Eq 5, the resultant matrix of the product Ax is proportional to the column
matrix x involved in the product, being the proportionality constant.
Please note that in Eq 5, the matrix A is the known part. The unknowns are and x. The allowed
values of the scalar , and the associated vectors x all depend on what A is. Hence the central
quantity of Eq 5 is the coefficient matrix A.
The determinant in Eq 3 is called the characteristic determinant of the matrix A, and Eq 3 itself is
called the characteristic equation of the matrix A. The set of values obtained from Eq 3 are called
the eigenvalues or characteristic values of the matrix A. For each value of , say i, solution of the
linear system Eq 2 gives a corresponding vector xi. There are n values of , and the same number of
vectors x. These vectors are called eigenvectors of the matrix A.
Before we discuss several examples, let us note a simple but important property of the eigenvalue-
eigenvector Eq 5. If a vector x is an eigenvector of A with eigenvalue (i.e. if x satisfies Eq 5), then
any multiple of it, kx, also satisfies Eq 5 with the same eigenvalue, where k is an arbitrary constant.
Thus (remember that multiplication of a matrix by a scalar is commutative; i.e. kA = Ak),
Eq 6
A(kx) = kAx = kx = (kx)
Hence, eigenvectors of the matrix A are defined only within an arbitrary factor k. This is not
surprising since Eq 5 is equivalent to Eq 2, and as remarked above there are (n+1) variables in Eq 2,
but n equations with the consequence that one of the (nonzero) variables x1, x2, ..., xn is independent
and hence may be assigned any arbitrary value. The constant k in Eq 6 corresponds to the value of
the independent variable chosen.
8 −8 −2
𝑨 = (4 −3 −2) Eq 7
3 −4 1
(8 − 𝜆) −8 −2
| 4 (−3 − 𝜆) −2 | = 0
3 −4 (1 − 𝜆)
Note that trace of matrix A is: 8 – 3 + 1 = 6, and sum of its eigenvalues is: 1 + 2 + 3 =1 +2 + 3 = 6.
This is a general property valid for any square matrix A of order n
tr A = 1 + 2 + 3 + ... + n Eq 8
det A = 1 2 3 ... n Eq 9
Please note from Eq 9 that if det A is zero, we immediately know that (at least) one of the
eigenvalues of A is zero; in this case, expansion of the characteristic determinant of A will not contain
a constant additive term, i.e, the nth degree polynomial in that results from Eq 3 will have as a
factor.
e.g.2 Find the eigenvalues of
6 −8 −2
𝑩 = (4 −5 −2 ) Eq 10
3 −4 −1
Before we construct the characteristic determinant, we notice that first row of B is twice row 3, and
therefore det B is zero. Hence we conclude that one of the eigenvalues will be zero. Let us continue
with Eq 3 for matrix B
(6 − 𝜆) −8 −2
| 4 (−5 − 𝜆) −2 | = −𝜆3 + 𝜆 = −𝜆(𝜆2 − 1) = 0
3 −4 (−1 − 𝜆)
As expected, there is no additive constant term in the polynomial, and is a factor. Eigenvalues of B
are: 1= -1, 2 = 0, 3 = 1. Sum of eigenvalues is zero which is the same as tr B.
The first step in such a question is to find the eigenvalues because non-trivial solutions of these
equations can be found only if is one of the eigenvalues. We already found the eigenvalues in e.g.1;
they are: 1 = 1, 2 = 2, and 3 =3. For each of these eigenvalues we solve Eq 11 to obtain the
corresponding eigenvectors.
For = 1 = 1, Eq 11 becomes
Equations 2 and 3 are the same, so we discard Eq (3). There are 3 variables but 2 equations so that
one of the variables can be assigned any arbitrary value. We choose x2, and assign x2=3a where a is
an arbitrary number. Then from Eq (2), x1=4a, and from Eq (1), x3=2a. We index each eigenvector
with the index of the eigenvalue used in solving Eq 11. So the index is 1 for this solution, and the
eigenvector x1 is
4𝑎 4
𝒙1 = (3𝑎) = 𝑎 (3)
2𝑎 2
The arbitrary constant a corresponds to k in Eq 6; its value can not be determined from the
eigenvalue-eigenvector equation Eq 5 (which means Eq 3 plus Eq 2). It is customary to fix the value of
a by an additional condition called the “normalization” condition: eigenvectors are forced to be unit
vectors.
Either sign may be selected for the square root, we have chosen the plus sign. Hence the normalized
eigenvector x1 corresponding to the eigenvalue 1 = 1 is
14
𝒙1 = (3)
√29 2
We repeat the above procedure for finding the eigenvectors of A corresponding to 2 and 3.
For 2 =2, Eq 11 is
Setting x2 = 2b where b is an arbitrary constant, one gets x1 = 3b from Eq (1), and then x3 = b from
Eq (2). The eigenvector corresponding to 2 =2 is thus
3𝑏 3 3
1
𝒙2 = (2𝑏) = 𝑏 (2) 𝒙2 = (2) normalized eigenvector
√14
𝑏 1 1
The eigenvector x3 corresponding to 3=3 is found similarly from Eq 11. I leave the details to you. You
should obtain
2𝑐 2 2
1
𝒙3 = ( 𝑐 ) = 𝑐 (1) 𝒙3 = (1) normalized eigenvector
√6
𝑐 1 1
Which of the following vectors are eigenvectors of B? For those that are eigenvectors, state the
corresponding eigenvalues.
1 8 −2 3
𝒙1 = (2), 𝒙2 = (6), 𝒙𝟑 = ( 1 ), 𝒙𝟒 = ( 2)
1 4 1 1
The eigenvalue-eigenvector equation is Bx = x. We test this equation for each of the given vectors.
For x1
6 −8 −2 1 −12 1 1
𝑩𝒙𝟏 = (4 −5 −2 ) (2) = ( −8 ) = −12 (8/12) ≠ 𝑐𝑜𝑛𝑠𝑡 (2)
3 −4 −1 1 −6 6/12 1
6 −8 −2 8 −8 8
𝑩𝒙𝟐 = (4 −5 −2 ) (6) = (−6) = (−1) (6)
3 −4 −1 4 −4 4
6 −8 −2 −2 −22 −2 −2
𝑩𝒙𝟑 = (4 −5 −2 ) ( 1 ) = (−15) = (11) (−15/11) ≠ 𝑐𝑜𝑛𝑠𝑡 ( 1 )
3 −4 −1 1 −11 −1 1
6 −8 −2 3 0 3
𝑩𝒙𝟒 = (4 −5 −2 ) ( 2 ) = (0) = (0) (2)
3 −4 −1 1 0 1
Ax = x
we can easily see that x is also an eigenvector of, for example, A2.
Thus x is an eigenvector of A2 with eigenvalue 2. This property may be generalized to all positive
integer powers of matrix A:
Amx = m x, m = 1,2,3,... Eq 13
We next show that any polynomial expression in square matrix A has the same eigenvectors as those
of A. Let
where I is the unit matrix, and the coefficients ci are constants. All matrices in this expression are
square of order n. Due to the property Eq 13, and the fact that multiplication of a matrix with a scalar
is commutative (i.e. Ai=Ai), one has
Bx = c0Ix + c1Ax + c2A2x + c3A3x + ... + cmAmx = (c0 + c1 + c22 + c33 + ... + cmm)x
Hence
Thus, an eigenvector x of matrix A with eigenvalue is also an eigenvector of matrix B in Eq 14; the
eigenvalue of matrix B is b, obtained from Eq 15.
e.g.6 Taking the matrix A in Eq 7, what are the eigenvalues of the matrix B below?
108 −136 −4
2
B = 2I - 3A + 5A = ( 58 −64 −14)
46 −68 14
A’ = c0I + A Eq 16
The matrix A’ differs from A only in the diagonal elements: all diagonal elements of A are shifted by
the same amount c0 in forming A’. From Eq 15, the eigenvalues of A’ are ’ = + c0. In other words,
when diagonal elements of a square matrix A are shifted by c0, (all of) its eigenvalues are shifted by
the same amount.
As an example of Eq 16, consider the matrices used in e.g.1 and e.g.2. The matrix A in Eq 7, and
matrix B in Eq 10 are reproduced below.
8 −8 −2 6 −8 −2
𝑨 = (4 −3 −2) , 𝑩 = (4 −5 −2 ) Eq 17
3 −4 1 3 −4 −1
Comparison of the two matrices reveals that they are related to each other as in Eq 16:
B = A ─ 2I Eq 18
’ = ─ 2 Eq 19
As remarked above, det B = 0, implying that one of the three vaues of ’ is 0, and hence solving the
characteristic equation of B is slightly easier than that of A.
Now consider a problem in which we are asked to find the eigenvalues of A, as in e.g.1. If by shifting
the diagonal elements of A we can find a matrix B as in Eq 18, such that det B = 0, our task becomes
somewhat simplified because then we may calculate the eigenvalues of B first, and then use Eq 19 to
obtain eigenvalues of A.
In e.g.2, we found ’1= -1, ’2 = 0, ’3 = 1 so that eigenvalues of A are =’ + 2; hence, 1=-1+2=1,
2=0+2=2, and 3=1+2=3. Note once again that eigenvectors of A and B are identical.
Symmetric matrices
In the majority of real-life applications, the square matrix A of order n in the eigenvalue-eigenvector
equation, Eq 5, is a symmetric matrix. A square matrix A is said to be symmetric if the transposed
matrix is equal to itself:
AT = A symmetric matrix Eq 20
Graphically, this relation means that in a symmetric matrix, the off-diagonal elements in the lower
triangle of a square matrix A are mirror images through the diagonal line, of those in the upper
triangular part. Or, in terms of elements, aji = aij for all i,j=1,2,...,n.
The significance of symmetric matrices lies in the following two properties:
These properties are not necessarily true for non-symmetric square matrices.
1 3 0
𝑨 = (3 1 3)
0 3 1
1 3 0 0 3 0 1 0 0
(3 1 3 ) = (3 0 3 ) + (0 1 0) A = A’ + I
0 3 1 0 3 0 0 0 1
Eigenvalues of A are related to those of A’ by: = ’ + 1. The first and third columns of A’ are
identical, hence det A’ = 0. One of the eigenvalues of A’ is zero, and therefore we prefer to work with
A’, and then switch to A. Eigenvectors of A’ and A are the same, and their eigenvalues are related as
shown above.
The characteristic equation (Eq 3) of matrix A’ gives: -’(’2─18) = 0. Hence, 𝜆1′ = −3√2, 𝜆′2 = 0,
𝜆′3 = 3√2. For each eigenvalue, the corresponding eigenvector is obtained from
−𝜆′𝑥1 + 3𝑥2 =0
3𝑥1 − 𝜆′𝑥2 + 3𝑥3 =0
3𝑥2 − 𝜆′𝑥3 =0
1 1
𝐹𝑜𝑟 𝜆1′ = −3√2, 𝒙1 = (−√2)
2
1
1 1
𝐹𝑜𝑟 𝜆′2 = 0, 𝒙2 = ( 0)
√2 −1
1 1
𝐹𝑜𝑟 𝜆′3 = 3√2, 𝒙3 = (√2)
2
1
Now we turn to our actual problem: eigenvalues and eigenvectors of A. Eigenvalues of A are: 𝜆1 =
1 − 3√2, 𝜆2 = 1, and 𝜆3 = 1 + 3√2. The corresponding eigenvectors are the same as those of A’,
given above.
Let us consider all scalar products of two eigenvectors 𝒙𝑇𝑖 𝒙𝑗 with ij. For the three eigenvectors
above, one finds
1
𝒙1𝑇 𝒙2 = [(1)(1) + (√2)(0) + (1)(−1)] = 0
2√2
1
𝒙1𝑇 𝒙3 = [(1)(1) + (−√2)(√2) + (1)(1)] = 0
4
1
𝒙𝑇2 𝒙3 = [(1)(1) + (0)(√2) + (−1)(1)] = 0
2√2
Thus the three eigenvectors form an orthogonal set. The property of orthogonality of the
eigenvectors originates from the fact that matrix A is symmetric, and is generally true for any
symmetric matrix of order n.