Professional Documents
Culture Documents
MATRIX ALGEBRA
Trimester 2, 2021 - 2022
P
n
cij = aik bkj
k=1
An = |AA {z A} (2)
Note that multiplication of two matrices is de…ned only n tim e s
Unlike real numbers, matrices are not communicative, The identity matrix of order n (denoted I n or just I) is
i.e. AB is in general not equal to BA: an n n matrix with 1’s down the diagonal and zeros
everywhere else:
Further,
1
1. If two rows (or two columns) of A are interchanged,
AI n = A = I m A (5) the determinant changes sign but its absolute value
remains unchanged.
A symmetric matrix A such that AA = A is said to be 2. If all the elements in a single row (or column) of
idempotent. A are multiplied by a number c, the determinant
is multiplied by c:
If A = (aij )m n , the transpose of A, denoted by A0 , is
de…ned as A0 = (aji )n m : 3. If two of the rows (or columns) of A are propor-
tional, then j Aj = 0:
Transposition follows the following rules: 4. The determinant remains unchanged if a multiple
of one row (or one column) is added to another row
1. (A0 )0 = A
(or column).
2. (A + B)0 = A0 + B 0
5. jA0 j = jAj
0 0
3. ( A) = A
6. jABj = jAjjBj where A and B are square matrices
4. (AB)0 = B 0 A0 of the same dimension
7. jA + Bj =
6 j Aj + jBj (usually)
A square matrix such that A = A0 is said to be sym-
metric. 1
The inverse A of an n n matrix A has the following
properties:
Determinants and Matrix In-
verses B=A 1
() AB = I n () BA = I n (8)
The determinant of an n n matrix A, denoted by jAj;
is de…ned as
1
A exists () jAj 6= 0 (9)
jAj = ai1 Ai1 + ai2 Ai2 + + ain Ain (6) If A = (aij )n n and jAj 6= 0, the unique inverse of A is
given by
where Aij is referred to as the (ij)th co-factor de…ned as
the ( 1)i+j times the determinant of the (n 1) (n 1) 1
1
sub-matrix obtained from deleting the ith row and jth A = adj(A) (10)
jAj
column of A :
where
Aij =
a11 a1;j 1 a1;j+1 a1n 0 10
A11 A12 A1n
.. .. .. .. .. .. B C
. . . . . . B A21 A22 A2n C
adj(A) = B
B .. .. .. .. C
C (11)
ai 1;1 ai 1;j 1 ai 1;j+1 ai 1;n @ . . . . A
( 1)i+j
ai+1;1 ai+1;j 1 ai+1;j+1 ai+1;n An1 An2 Ann
.. .. .. .. .. ..
. . . . . . and Aij is the (ij)th cofactor of A:
an1 an;j 1 an;j+1 an;n
For a 2 2 matrix, this result gives
By this de…nition, the determinant of a 2 2 matrix is
given by
! 1 !
a b 1 d b
= ; (12)
c d ad bc c a
a11 a12
jAj = = a11 a22 a12 a21 (7)
a21 a22 a b
provided = ad bc 6= 0
c d
Equation (6) is referred to as the cofactor expansion of
jAj along the ith row. The following rules for inverses hold, provided the rele-
vant inverses exist:
The following rules for manipulating determinants are
1 1
often useful: 1. (A ) =A
2
1 1 1
2. (AB) =B A Null vector, denoted by 0; contains zeros as its elements
3. (A ) 0 1
= (A 1 0
) and the sum vector, denoted by , contains ones as its
1 1 1
elements.
4. ( A) = A
Sum of two vectors of the same order:
A linear system of n equations and n unknowns,
0 1
x1 + y1
a11 x1 + a12 x2 + + a1n xn = b1 B .. C
x+y =B
@ .
C
A (17)
a21 x1 + a22 x2 + + a2n xn = b2
.. xm + ym
. (13)
Scalar multiplication of a vector:
an1 x1 + an2 x2 + + ann xn = bn
3
Two vectors are orthogonal, denoted by x ? y; i¤ x0 y = (b) dim(A) = 2
0 (c) There are in…nitely many solutions. We need to …nd
two vectors, say v 1 and v 2 such that they are not collinear
In addition, if kxk = kyk = 1, they are said to be ortho- with each other and with a1 and a2 : A simple solution is
normal. 0 0
v1 = 0 0 1 0 and v 2 = 0 0 0 1
For example, the unit vectors; e1 =
0 0
1 0 0 ; e2 = 0 1 0 : : : em =
0
Partitioned Matrices and Their
0 0 1 in m dimensional Euclidean space
Inverses
are orthonormal.
When dealing with high order matrices, it is sometimes
A linear combination of the vectors: 1 x1 + 2 x2 + + convenient to consider subdiving the matrices into sub-
n xn
matrices, called partitioning the matrices.
0 1
a11 a12 a13 a14 a15
Vector Space B
A = @ a21 a22 a23 a24
C
a25 A (22)
A vector space, V, is any set of vectors that is closed a31 a32 a33 a34 a35
under addition and scalar multiplication. Further,
It can be partitioned in a number of ways, for example
1. 1. for all x 2 V, there exists a vector, denoted by 0,
0 1
such that x + 0 = x; ..
B a11 a12 . a13 a14 a15 C
2. for any x 2 V, there exists a vector, denoted by B .. C !
B C
x; such that x + ( x) = 0 B a21 a22 . a23 a24 a25 C A11 A12
A=B C=
B .. C A21 A22
B . C
A linearly independent set of vectors that span a vector @ A
..
space is known as a basis of that vector space. a31 a32 . a33 a34 a35
(23)
The maximum number of linearly independent vectors in
or
a vector space is known as the dimension of the vector
space. 0 1
..
A subspace is a subset of a vector space and is itself B a11 . a12 a13 a14 a15 C
B .. C !
a vector space e.g. the vector space spanned by the B C
B a21 . a22 a23 a24 a25 C A11 A12
0 A=B C=
independent vectors x = x1 x2 0 and y = B .. C A21 A22
B . C
0 @ A
y1 y2 0 is a subspace of R : 3 ..
a31 . a32 a33 a34 a35
(24)
Example: Let A be the subspace of R4 spanned by the
columns of the matrix
One can perform standard matrix operations on parti-
0 1
1 2 3 tioned matrices, treating the submatrices as if they were
B 2 3 8 C ordinary matrix elements. This requires obeying rules
B C
A=B C for sums, di¤erences and products.
@ 5 1 3 A
3 4 5
For example,
(a) Find a basis of A.
(b) What is the dimension of A?
! !
(c) Extend the basis of A to a basis in R4 : A11 A12 B 11 B 12
A21 A22 B 21 B 22
Solution: !
A11 B 11 A12 B 12
(a) Let a1 ; a2 ; a3 denote the columns of A. It is clear = (25)
A21 B 21 A22 B 22
that a1 and a2 are independent (otherwise they would be
proportional). But a1 2a2 + a3 = 0 so that fa1 ; a2 ; a3 g provided the dimensions of the submatrices are con-
are linearly dependent. Hence fa1 ; a2 g is a basis of A. formable for matrix addition/subtraction.
4
0 1
Further, ..
B 2 3 . 0 0 0 C
B .. C
B C
! ! B 3 4 . 0 0 0 C
B C
A11 A12 A11 A12 B .. C
= (26) B . C
A21 A22 A21 A22 A=B
B
C
C
B 1 ..
B 1 . 1 0 0 C
C
If A and B are partitioned conformably, then B .. C
B C
B 1 1 . 0 1 0 C
@ A
! ..
5 7 . 0 0 1
A11 B 11 + A12 B 21 A11 B 12 + A12 B 22 !
AB = A11 A12
A21 B 11 + A22 B 21 A21 B 12 + A22 B 22 =
(27) A21 A22
1
Inverting large square matrices is often made easier using Solution: Using
! equation (34), e = A11 , and e =
partitioning. Consider an n n matrix A which has an 4 3
: Then
inverse and is partitioned as follows: 3 2
0 1
! 1 1 !
A11 A12 e 1 B C 4 3
A= (28) A221 A21 = I3 @ 1 1 A =
A21 A22 3 2
5 7
0 1
1 1
where A11 is a k k submatrix with an inverse. Then B C
@ 1 1 A : Hence,
it can be shown that
1 1
0 1
4 3 0 0 0
! B 3 2 0 0 0 C
B C
A111 + A111 A12 1
A21 A111 A111 A12 1 1 B C
A 1
= A =B 1 1 1 0 0 C
1
A21 A111 1 B C
@ 1 1 0 1 0 A
(29) 1 1 0 0 1
where = A22 A21 A111 A12
Example: If P and Q are invertible square matrices,
prove that
Similarly, if both A 1
and A221 exist, then
1 1
! ! 1 !
1
e eA12 A221 P R P 1 P 1
RQ 1
A = 1 1 = 1
A221 A21 e A221 + A221 A21 e A12 A221 0 Q 0 Q
(30)
Solution: Using equation (33) where A11 = P ; A12 = R;
where e = A11 A12 A221 A21 A21 = 0 and A22 = Q, we obtain the result.
5
0 1 0 1
(b) What is the angle between two orthogonal vectors? a11 a1n x1
B .. .. C B C
Let A = B .. C and X = B .. C where
@ . . . A @ . A
Question 3
an1 ann xn
If x; y and z are linearly independent vectors, are x + y,
jAj 6= 0: Show that
x + z and y + z also linearly independent?
1 x1 xn
x1 a11 a1n
Question 4 jA + XX 0 j = .. .... .. ..
Expand the matrix product . .. . .
xn an1 ann
X = ([AB + (CD)0 ][(EF) 1
+ GH])0
= jAj(1 + X 0 A 1
X)
Assume that all matrices are square and E and F are non-
singular.
Linear Independence
Question 5 The n vectors a1 ; a2 ; :::; an in Rm are said to be linearly
For the matrix dependent if there exist numbers, c1 ; c2 ; :::; cn ; not all
zeros, such that
" #
0 1 1 1 1
X =
4 2 3 5 c1 a1 + c2 a2 + + cn an = 0 (33)
compute P = X(X0 X) 1
X0 and M = (I P). Verify that If this equation holds only when c1 = c2 = = cn = 0,
MP = 0: Let then the vectors are said to be linearly independent.
" # So, a linearly dependent set of vectors is such that one
1 3
Q= vector in the set can be expressed as a linear combination
2 8
of the other vectors.
Compute P and M based on XQ instead of X Conversely, a linearly independent set of vectors is such
that no vector in the set can be expressed as a linear
Question 6 combination of the other vectors.
Let A be any square matrix whose columns are
Consider the general system of m equations in n un-
[a1 ; a2 ; :::; aM ] and let B be any rearrangement of the columns
knowns:
of the M M identity matrix. What operation is performed
by the multiplication AB? By the multiplication BA?
a11 x1 + + a1n xn = b1
Question 7 .. .. ..
. . . (34)
Let A = faij g represent an m n matrix and B = fbij g a am1 x1 + + amn xn = bm
p m matrix.
(a) Let x = fxi g represent an n-dimensional column vec- which can be expressed in vector form:
tor. Show that the ith element of the p-dimensional column
Pm P
n
vector BAx is bij ajk xk : x1 a1 + + xn an = b (35)
j=1 k=1
(b) Let x = fxi g represent an n-dimensional column vector Here, a1 ; : : : ; an are the column vectors of coe¢ cients, and
and C = fcij g a q p matrix. Express the ith element of the b is the column vector with components b1 ; : : : ; bm :
q-dimensional column vector CBAx in terms of the elements Suppose (39) has two solutions (u1 ; : : : ; un ) and
of A; B; C and x: (v1 ; : : : ; vn ): Then
Question 8
Show that the product AB of two n n symmetric matrices u1 a 1 + + un an = b and v1 a1 + + vn an = b
A and B is itself symmetric if and only if A and B commute. Subtracting the second equation from the …rst yields
Question 9
(u1 v1 )a1 + + (un vn )an = 0
Show that A0 and B0 commute if A and B commute.
c1 a1 + + cn an = 0
Question 10 where ci = ui vi :
6
The two solutions are di¤erent if and only if not all A is obtained by deleting all but k rows and k columns,
c1 ; : : : ; cn are zeros. Hence if (39) has more than one and then taking the determinant of the resulting k k
solution, then the column vectors a1 ; : : : ; an are lin- matrix.
early dependent. Equivalently, if the column vectors
a1 ; : : : ; an are linearly independent, then (39) has at
Example:
0 1 all the minors of the matrix A =
Describe
1 0 2 1
most one solution. Without saying more about the right- B C
hand side vector b, however, we cannot know if there are @ 0 2 4 2 A
0 2 2 1
any solutions at all, in general.
Solution: Because there are 3 rows, there are minors of
For an n n matrix A with column vectors a1 ; : : : ; an :
orders 1,2 and 3. There are
0 1 (a) 4 minors of order 3. These are obtained by deleting
a11 a12 a1n
B C any one of the 4 columns:
B a21 a22 a2n C
A = B .. .. .. C , (36) 1 0 2 1 0 1 1 2 1 0 2 1
B .. C
@ . . . . A 0 2 4 ; 0 2 2 ; 0 4 2 ; 2 4 2
an1 an2 ann 0 2 2 0 2 1 0 2 1 2 2 1
0 1
a1j (b) 18 minors of order 2. These are obtained by delet-
B a C
B 2j C ing one row and two columns in all possible ways. For
where aj = B . C
B . C example, four of these are
@ . A
anj 1 0 2 4 2 1 0 2
; ; ;
0 2 2 2 4 2 2 2
the n column vectors are linearly independent if and only
if jAj 6= 0 (c) 12 minors of order 1. These are all the 12 individual
elements of A:
Example: Suppose a; b and c are three linearly indepen-
dent vectors in Rn : Are a b; b c and a c linearly The rank r(A) of a matrix A is equal to the order of the
independent? largest minor of A that is non-zero.
7
Solution. Perform the following elementary row opera- Example. Determine if the system of equations:
tions to get the resulting matrix:
0 1
1 2 3 2
B C 2x1 x2 = 3 (39)
@ 2 3 5 1 A (R1 2) + R2 !
1 3 4 5 (R1 1) + R3 4x1 2x2 = 5 (40)
0 1
1 2 3 2
B C has a solution.
@ 0 1 1 3 A
0 1 1 3 !
0 1 2 1
Solution. Here A = and Ab =
1 2 3 2 4 2
B C !
@ 0 1 1 3 A !
2 1 3
0 1 1 3 R2 + R3 : Since jAj = 0, so r(A) < 2: Because
0 1 4 2 5
1 2 3 2 A is not the null matrix, it follows that r(A) = 1: But
B C
@ 0 1 1 3 A r(Ab ) = 2 because the minor obtained by deleting the
0 0 0 0 …rst column is non-zero. Thus, r(A) 6= r(Ab ), so the
The rank of the last matrix is obviously 2 because there system does not have a solution. To see this, multiply
are precisely two linearly independent rows. So the orig- the …rst equation by 2 to obtain 4x1 2x2 = 6, which is
inal matrix has rank 2. inconsistent with the second equation.
Consider the general linear system of m simultaneous Suppose that equation (41) has solutions and that
equations in n unknowns which can also be expressed in r(A) = r(Ab ) = k < m: Since the rank of A is less
matrix notation as: than the number of equations there are m k "super‡u-
ous" equations in the system. That is, these equations
Ax = b (37) are not required to …nd the solutions of the system. If
we choose any collection of k equations corresponding
where A is the m n coe¢ cient matrix. to k linearly independent rows of A, then the solution
to these equations will also satisfy the remaining m k
De…ne a new m (n+1) matrix Ab ; called the augmented equations.
matrix, that contains A in the …rst n columns and b in
column n + 1 :
Now, suppose that (41) has solutions and that r(A) =
0 1 r(Ab ) = k < n, that is the rank of A is less than the
a11 a12 a1n number of variables in the system. Then there exist n k
B a21 a22 a2n C of the variables that can be chosen freely, whereas the
B C
A=B
B .. .. .. .. C
C remaining k variables are uniquely determined by the
@ . . . . A
choice of these n k free variables. We say the system
am1 am2 amn
has n k degrees of freedom.
and
Example. Determine whether the following system of
0 1 equations has any solutions, and if it has, …nd the num-
a11 a12 a1n b1
B a a22 a2n b2 C ber of degrees of freedom.
B 21 C
Ab = B
B .. .. .. .. .. C
C
@ . . . . . A
am1 am2 amn bm x1 + x2 2x3 + x4 + 3x5 = 1
The relationship between the ranks of A and Ab deter- 2x1 x2 + 2x3 + 2x4 + 6x5 = 2
mines whether equation (41) has a solution. 3x1 + 5x2 10x3 3x4 9x5 = 3 (41)
3x1 + 2x2 4x3 3x4 9x5 = 3
A necessary and su¢ cient condition for a linear system
of equations to be consistent (that is, to have at least
Solution. Here
one solution) is that the rank of the coe¢ cient matrix is 0 1
equal to the rank of the augmented matrix, i.e. 1 1 2 1 3
B 2 1 2 2 6 C
B C
A=B C and
@ 3 5 10 3 9 A
Ax = b has a solution () r(A) = r(Ab ) (38) 3 2 4 3 9
8
0 1
1 1 2 1 3 1 1. 0 r(A) min(m; n): The rank of A cannot be
B C
B 2 1 2 2 6 2 C negative and it cannot exceed the smaller of the
Ab = B C
@ 3 5 10 3 9 3 A number of row or column.
3 2 4 3 9 3
2. r(A) = 0 () A = 0. If r(A) = 0; then there are
no linearly independent columns. Hence, A = 0.
We know that r(Ab ) r(A): All minors of order 4 in Ab Conversely, if A = 0, then there are no linearly
are equal to 0, so r(Ab ) 3: Now, there are minors of order independent columns. Hence, r(A) = 0.
3 in A that are di¤erent from 0. For example, the minor
3. r(I n ) = n: The identity matrix of order n has n lin-
formed by the …rst, third and fourth columns and the …rst,
early independent columns, since I n x = 0 implies
second and fourth rows, is di¤erent from 0, because
x = 0:
1 2 1
2 2 2 = 36: 4. r( A) = r(A) if 6= 0: Multiplying the columns
3 4 3 of A by a nonzero constant , does not change the
dependence of the columns. Hence, the maximum
number of independent columns of A is equal to
Hence, r(A) = 3: Because 3 r(Ab ) r(A) and r(A) =
that of A:
3;so r(Ab ) = 3:Since r(A) = r(Ab ) the system has solutions.
There is one super‡uous equation as m = 4 and k = 3: Also, A real m n matrix A can be viewed as a collection of
since n = 5; there are 2 degrees of freedom. n columns in Rm , or m rows in Rn :
Next, we …nd all the solutions to the system of equations.
Using the non-zero minor above, we can write the equation Two subspaces associated with A
system in terms of 3 independent equations in the form:
1. column space of A, denoted by col A =
fx 2 Rm : x = Ay for some y 2 Rn g
x1 2x3 + x4 + x2 + 3x5 = 1
2. kernel (or null space) of A; denoted by ker A =
2x1 + 2x3 + 2x4 x2 + 6x5 = 2 fy 2 Rn : Ay = 0g
3x1 4x3 3x4 + 2x2 9x5 = 3
Similarly, two subspaces associated with A0
or in matrix form, as
1. col A0 = fy 2 Rn : y = A0 x for some x 2 Rm g
0 10 1 0 1 0 1
!
1 2 1 x1 1 3 1 2. ker A0 = fx 2 Rm : A0 x = 0g
B CB C B C x2 B C
@ 2 2 2 A @ x3 A+@ 1 6 A =@ 2 A
x5 The kernels are commonly known as orthogonal comple-
3 4 3 x4 2 9 3
ments
0 1 0 1 1 80 1 0 1 9
x1 1 2 1 > !>
< 1 1 3
x2 = 1. col? ( A) = fx 2 Rm : x ? Ag = ker A0
B C B C B C B C
@ x3 A = @ 2 2 2 A @ 2 A @ 1 6 A
>
: 3 x5 >
; 2. col? ( A0 ) = fy 2 Rn : y ? A0 g = ker A
x4 3 4 3 2 9
An important result relating to the rank of a matrix A
It is easily veri…ed is
0 1 1 0 1
1 2 1 1 5 3
B C 1 B C r(A) = r(AA0 ) = r(A0 A) (42)
@ 2 2 2 A = @ 6 3 0 A
18
3 4 3 7 1 3 To prove this result, we …rst establish the following:
Hence,
(a) ker A0 = ker AA0
0 1 0 1 0 1 0 1 (b) col? A = col? AA0
x1 1 0 1
B C B C B 1 C B 1 C (c) col A = col AA0
@ x3 A = @ 0 A @ x
2 2 A
=@ x
2 2 A
x4 0 3x5 3x5
Proof:
Given x2 and x5 the values for x1 ; x3 and x4 are uniquely
determined. This con…rms that there are 2 degrees of free- (a) If x 2 ker A0 , then A0 x = 0. Then, AA0 x = 0, so
dom. that x 2 ker AA0 : Conversely, if x 2 ker AA0 , then
AA0 x = 0 and x0 AA0 x = 0 ) A0 x = 0, so that
The rank of a matrix A satis…es the following: x 2 ker A0 :
9
0 1
(b) Since ker A0 = col? A and ker AA0 = col? AA0 ; c11 c12 c1k
B c21 c22 c2k C
the result follows immediately. B C
where C = B
B .. .. .. .. C:
C
(c) col A = (col? A)? = (col? AA0 )? = col AA0 @ . . . . A
cn1 cn2 cnk
From (c) result above, it immediately follows that
r(A) = r(AA0 ). Further since r(A) = r(A0 ); (c) im- That C must have rank k can be shown as follows.
plies r(A) = r(A0 A):
r(C) = r(AB) min(r(A); r(B)) (44) Eigenvalues and eigenvectors are also called characteris-
tic roots and characteristic vectors, respectively.
Every m n matrix A of rank k can be written as A =
BC 0 , where B m k and C n k both have rank k: It should be noted that if x is an eigenvector with eigen-
value , then x is another eigenvector for every scalar
Since A has rank k, there are k independent columns, 6= 0: Thus, eigenvectors are not unique.
say b1 ; : : : ; bk such that each column ai of A is a linear
combination of b1 ; : : : ; bk , i.e. To obtain unique eigenvectors, it is often required that
the eigenvectors are of unit length, known as orthonor-
malizing the eigenvectors, i.e. x0 x = 1:
P
k
ai = cij bj ; i = 1; 2; ::; n Here are some general properties of eigenvalues and
j=1
eigenvectors:
Stacking the ai (i = 1; 2; ::; n) side by side, we get
I The eigenvalues of symmetric matrices are real.
10
the diagonal elements. Further, if ej denotes the jth
unit vector in Rn , having all components 0, except for 5:6056
x22 = x21 = 1:8685x21
the jth component which is 1, then any nonzero multiple 3
of ej is an eigenvector associated with the eigenvalue dj : The eigenvector
! corresponding to 2 is therefore x2 =
1
Example: Find the eigenvalues and eigenvectors of the t ; t 6= 0:
! 1:8685
2 3
matrix A = :
3 6 If orthonormalized eigenvectors are required, then x1
and x2 need to satisfy two further conditions x01 x1 = 1
Solution: The eigenvalue equation is and x02 x2 = 1: Using these conditions, we obtain:
( ! !) ! ! 1
2 3 7:6056 0 x21 0 P AP = D (52)
=
3 6 0 7:6056 x22 0
Since the diagonal elements of a diagonal matrix are its
5:6056x21 + 3x22 = 0 (50) eigenvalues, if A is diagonalizable then the diagonal ele-
3x21 1:6056x22 = 0 (51) ments of D are its eigenvalues.
11
An n n matrix A is diagonalizable i¤ it has a set of n A quadratic form Q(x) = x0 Ax; as well as its associ-
linearly independent eigenvectors x1 ; x2 ; :::; xn : In that ated symmetric matrix A, are said to be positive de…-
case, nite (p.d.), positive semide…nite (p.s.d.), negative de…-
nite (n.d.) or negative semide…nite (n.s.d.) according
as
1
P AP = diag( 1; 2; : : : ; n) (53)
where P is the matrix with x1 ; x2 ; :::; xn as its columns, Q(x) > 0; Q(x) 0; Q(x) < 0; Q(x) 0
and 1 ; 2 ; : : : ; n are the corresponding eigenvalues.
for all x 6= 0. The quadratic form Q(x) is inde…nite
if there exists vectors x and y such that Q(x ) < 0
Since for symmetric matrices their eigenvalues are real and Q(y ) > 0: Thus, an inde…nite quadratic form can
and their eigenvectors are orthogonal, symmetric matri- assume both negative and positive values.
ces are diagonalizable.
Let A = (aij )n n be an n n matrix. An arbitrary
principal minor of order k, denoted by k ; is the deter-
Quadratic Form minant of the matrix obtained by deleting n k rows and
A quadratic form in n variables is a function Q of the corresponding n k columns in A: The determinant jAj
form itself is a principal minor (no row or column is deleted).
12
3 0
3 0 3 Tutorial Exercise No. 2
D1 = 3; D2 = = 3; D3 = 0 1 2 =3 Question 1
0 1
3 2 8 Consider the following matrix M :
0 1
1 1 1
Since Dk > 0; for k = 1; 2; 3; Q is positive de…nite. 0
B 03 1
2 6
0 0 C
B C
M =B C
@ 0 0 1 0 A
(b) The associated symmetric matrix 0 0 0 1
By suitably partitioning M, determine M 100 =
100 tim e s
0 1 z }| {
1 3 0 M M M
B C
A=@ 3 9 0 A:
0 0 2 Question 2
0 1
2 3 0 0 0
B 3 4 0 0 0 C
The leading principal minors are D1 = 1; D2 = 0; B C
1 B C
D3 = 0: It follows that Q is not positive de…nite, neither is it Compute A for A = B 1 1 1 0 0 C
B C
positive semide…nite or negative de…nite. To determine if it @ 1 1 0 1 0 A
is negative semide…nite, we need to examine all the principal 5 7 0 0 1
minors of A:
(m) Question 3
Let k denote the mth principal minor of order k: Then,
Suppose that a; b; c 2 R3 , are all di¤erent from 0, and
for A we have
that a ? b; b ? c; b ? c: Prove that a; b and c are linearly
independent.
(1) (2) (3) (m)
1 = 1; 1 = 9; 1 = 2; so ( 1)1 1 > 0; for
m = 1; 2; 3 Question 4
1
Calculate jAj; tr(A) and A for
Question 6
1. Q is positive de…nite () 1 > 0; : : : ; n >0 Compute the characteristic roots of
2. Q is positive semide…nite () 1 0; : : : ; n 0 2 3
2 4 3
3. Q is negative de…nite () 1 < 0; : : : ; n <0 6 7
A=4 4 8 6 5
4. Q is negative semide…nite () 1 0; : : : ; n 0 3 6 5
13
Let A be a m n matrix and ai its ith column. Then
2 3
a b c vec(A) is de…ned as
6 7
A=4 b d e 5
c e f 1 0
a1
has the eigenvectors B . C
vec(A) = B C
@ .. A (58)
2 3 2 3 2 3
1 1 1 an
6 7 6 7 6 7
v1 = 4 0 5 ; v2 = 4 2 5 ; v3 = 4 1 5
1 1 1 Example. Let
Note that the Kronecker product is de…ned for any pairs 1. (A1 + A2 ) B = (A1 B) + (A2 B) (A1 and
of matrices A and B regardless of their dimensions. A2 of the same dimension)
14
2. A (B 1 + B 2 ) = (A B 1 ) + (A B 2 ) (B 1 and Solution. (A B)(A B) = (AA BB) = (A B)
B 2 of the same dimension)
Example. Show that (A B) 1 = A 1 B 1 when A
3. A B= (A B)
and B are both nonsingular, not necessarily of the same
4. (A B)(C D) = (AC BD) dimension.
5. A B 6= B A, in general
Solution. (A B)(A 1 B 1 ) = (AA 1 BB 1 ) =
6. A (B C) = (A B) C (I m I n ) = I mn : So, (A 1 B 1 ) = (A B) 1
7. diag(A B) = diag(A) diag(B) (A and B
Example. Let A be an m m symmetric matrix with
square)
eigenvalues 1 ; : : : ; m and let B be an n n symmetric
8. (A B)0 = A0 B0 matrix with eigenvalues 1 ; : : : ; n . Show that the mn
9. tr(A B) = (tr A)(tr B) eigenvalues of A B are i j (i = 1; : : : m; j = 1; : : : n):
10. A B is idempotent if A and B are idempotent.
Solution. There exist orthogonal matrices P and Q such
11. (A B) 1 = A 1 B 1 when A and B are both that P 0 AP = and Q0 BQ = where and are
nonsingular, not necessarily of the same dimension. diagonal matrices whose diagonal elements are the eigen-
!
A11 A12 values of A and B, respectively. This gives
12. If A is a partitioned matrix A =
A21 A22
!
A11 B A12 B
then A B = (P 0 Q0 )(A B)(P Q) = (59)
A21 B A22 B
0 1 0 1
13. Let A and B be square matrices of dimensions m Since P = P and Q = Q equation (63) can be
and n, respectively. Then jA Bj = jAjn jBjm written as
15
AA0 and f j g are the eigenvalues of BB 0 : The eigenvalues Exercise. For any two matrices of the same order, show
i j is nonzero if and only if both i and j are nonzero. that
Hence, the number of nonzero eigenvalues of AA0 BB 0
equals the product of the number of nonzero eigenvalues of
AA0 and the number of nonzero eigenvalues of BB 0 : This (vec A)0 vec B = tr A0 B (64)
implies r(A B) = r(AA0 )r(BB 0 ) = r(A)r(B)
Solution. Let A = (aij ) and B = (bij ) : Then
The vec operator has the following properties:
PP PP 0
(vec A)0 vec B = aij bij = (A )ji bij
1. vec(A + B) = vec A + vec B P 0 j i j i
bn a !
=b a 1 3
For example, A =
3 4
(b) Let B = b1 bn be an m n matrix and let
the n columns of I n be denoted by e1 ; : : : ; en : Then B can x0 Ax=x21 + 4x22 + 6x1 x2
Pn
be written as B = bi e0i : Hence, using (a) ! ! !
i=1
P
n P
n @x0 Ax 2x1 + 6x2 2 6 x1
vec ABC = vec Abi e0i C = vec((Abi )(C 0 ei )0 ) Then, @x
= =
6x1 + 8x2 6 8 x2
i=1 i=1
P
n P
n = 2Ax
= (C 0 ei ) (Abi ) = (C 0 A)(ei bi )
i=1 i=1
P
n P
n In general this result holds when A is symmetric. If A
= (C 0 A) (ei bi ) = (C 0 A) vec bi e0i
i=1 i=1 is not symmetric, then
0 P
n
= (C A)vec bi e0i = (C 0
A)vec B @x0 Ax
i=1 @x
= (A + A0 )x
16
Returning to the preceding double summation, we …nd An example.
that for each term aij the associated term is xi xj . There-
fore
maximizex R = a0 x x0 Ax
0
2 3
@x Ax 2 1 3
= xi xj . h i
@aij 6 7
where a0 = 5 4 2 and A=4 1 3 2 5
The square matrix xx has ij 0 th
element xi xj . Hence 3 2 5
@x0 Ax Solution
@A
= xx0 .
17
Form the Lagrangean function
L(x; ) = a0 x x0 Ax + 0
Cx (79)
X
J Di¤erentiating L(x; ) gives us the necessary conditions
0
L(x; ) = f (x) + j cj (x) = f (x) + c(x) (70)
j=1
a 2Ax + C 0 =0 (80)
First order conditions:
2. If the Lagrange multipliers are zero, then the constrained Constrained maximum: 2.25
solution will equal the unconstrained solution.
18
Then:
(a) Dk (x ) > 0; k = 1; : : : ; n ) x is a local minimum 1 1
f (x) = exp (x )0 1
(x ) (93)
point (2 )k=2 j j1=2 2
(b) ( 1)k Dk (x ) > 0; k = 1; : : : ; n ) x is a local maxi-
mum point and are, respectively, the mean and variance-
(c) Dn (x ) 6= 0 and neither (a) nor (b) is satis…ed ) x is covariance matrix of X.
a saddle point Each variable in X has a marginal distribution that
is univariate normal, that is, Xi N ( i ; 2i ) for i =
Consider the following constrained optimization problem 2 th
1; :::; k, where i and i are the i element of and the
ith diagonal element of , respectively.
P
m 1 1
L(x) = f (x) j (gj (x) bj ) (89) f (x) = 2 )k=2
exp 2
(x )0 (x )
j=1
(2 2
Qk 1 1 2
De…ne the bordered Hessian = p exp (xi i)
i=1 2 2 2
@g1 (x) @g1 (x) = f (x1 )f (x2 ) f (xk ) (94)
0 0
@x1 @xr
.. .. .. .. .. .. Thus, the multivariate density is the product of the
. . . . . . marginal densities, that is, the X 0 s are distributed in-
@gm (x) @gm (x) dependently of one another. Zero correlations between
0 0
Br (x) = @x1 @xr normally distributed variables imply statistical indepen-
@g1 (x) @gm (x)
L0011 (x) L001r (x) dence. This result does not necessarily hold for variables
@x1 @x1
.. .. .. .. .. .. that are not normally distributed.
. . . . . .
@g1 (x) @gm (x)
... L00r1 (x) L00rr (x)
@xr @xr
(90)
Distributions of Quadratic
Let x be an interior point that satis…es the …rst-order Forms
conditions: Suppose x N (0; I), that is the k variables in x havein-
dependent (standard) normal distributions, each with
@L(x) @f (x) P
m @gj (x) mean zero and unit variance.
= j = 0; i = 1; :::; n(91)
@xi @xi j=1 @xi
The sum of squares x0 x = ki=1 x2i 2
k: Note that this
gj (x) = bj ; j = 1; : : : ; m (92) a particular quadratic form x0 Ix:
Then 1
Suppose now that x N (0; 2
I): Then x0 2
I x=
x2
1 x2 x2 2
+ 2
+ + k
k:
(a) If ( 1)m Br (x ) > 0 for r = m + 1; :::; n; then x solves 2 2 2
the local minimization problem; Similarly, it can be shown that if x N (0; ) where
(b) If ( 1)r Br (x ) > 0 for r = m + 1; :::; n; then x solves is a positive de…nite matrix, x0 1
x 2
k
the local maximization problem.
Consider the quadratic form x0 Ax where x N (0; I)
and A is a symmetric, idempotent matrix of rank r k:
Multivariate Normal Distribu- Then it can be shown that x0 Ax 2
r:
19
Because the matrices are symmetric idempotent, Question 3
Let F = (fij ) and G = (gij ) represent p q and q r
matrices of continuous
0
di¤erentiable functions of a vector x =
x0 Ax = x0 A0 Ax x1 xm of m variables. Show that
= (Ax)0 Ax
@(aF + bG) @F @G
and similarly =a +b
@xk @xk @xk
where a and b are constants.
x0 Bx = (Bx)0 Bx
20
0
@f @g @H
= tr
@xj @Y @xj
Question 8
The function f (x1 ; x2 ; x3 ) = x21 +x22 +3x23 x1 x2 +2x1 x3 +
x2 x3 has one stationary point. Show that it is a local mini-
mum point.
Question 9
Solve the problem
max f (x; y; z) = x + 2z
subject to
g1 (x; y; z) = x+y+z =1
7
g2 (x; y; z) = x2 + y 2 + z =
4
Question 10
Let x N (0; ): Show that x0 1
x 2
k:
Question 11
Let x N (0; I) and A is a symmetric, idempotent matrix
of rank r k: Show that x0 Ax 2
r:
21