Professional Documents
Culture Documents
Appendex D
Appendex D
We use uppercase boldface letters to denote matrices. We can write an m x n matrix generically
as
where aij represents the element in the ith row and the jth column. For example, a25 stands for the
number in the second row and the fifth column of A. A specific example of a 2 X 3 matrix is
A square matrix has the same number of rows and columns. The dimension of a square matrix is
its number of rows and columns.
A 1 X m matrix is called a row vector (of dimension m) and can be written as x = [x1,x2,…, xm]
A square matrix A is a diagonal matrix when all of its non-main diagonal elements are zero, that
is, aij = 0 for all i≠j. We can always write a diagonal matrix as
1
The n x n identity matrix, denoted I, or sometimes I n to emphasize its dimension, is the diagonal
matrix with unity (one) in each diagonal position, and zero elsewhere:
The m x n zero matrix, denoted 0, is the m x n matrix with zero for all entries. This need not be a
square matrix
0=
Matrix Addition: Two matrices A and B, each having dimension m x n, can be added
element by element.
2
Scalar Multiplication: Given any real number γ (often called a scalar), scalar
multiplication is defined as γA= [γ aij], or
Matrix Multiplication: To multiply matrix A by matrix B to form the product AB, the
column dimension of A must equal the row dimension of B. Therefore, let A be an m x n matrix
and let B be an n xp matrix. Then matrix multiplication is defined as
In other words, the (i,j)th element of the new matrix AB is obtained by multiplying each element
in the ith row of A by the corresponding element in the jth column of B and adding these n
products together. A schematic may help make this process more transparent:
3
We can also multiply a matrix and a vector. If A is an n x m matrix and y is an m x 1 vector, then
Ay is an n x 1 vector. If x is a 1 x n vector, then xA is a 1 x m vector.
1.
2.
3.
4.
5.
6.
7.
8.
9.
10.
11.
12.
13.
14.
4
If A is m x n and B is n x m, then AB and BA are both defined, but they are not usually the
same; in fact, they have different dimensions, unless A and B are both square matrices. Even
when A and B are both square, AB ≠ BA, except under special circumstances.
Transpose
Properties of transpose
1.
2.
3.
4.
5.
6.
A square matrix A is a symmetric matrix if and only if A’=A. If X is any n x k matrix, then X’X
is always defined and is a symmetric matrix.
5
Partitioned Matrix Multiplication
Trace
The trace of a matrix is a very simple operation defined only for square matrices.
Properties of trace
1.
2.
6
3.
4.
5.
Determinant of a matrix
The determinant is a value that can be computed from the elements of a square matrix. The
determinant of a matrix A is denoted det(A), det A, or |A|. Geometrically, it can be viewed as the
scaling factor of the linear transformation described by the matrix.
Each determinant of a 2 × 2 matrix in this equation is called a "minor" of the matrix A. This
procedure can be extended to give a recursive definition for the determinant of an n × n matrix,
the minor expansion formula.
A matrix is invertible if and only if its determinant is non-zero, and correspondingly the matrix is
singular (non-invertible) if and only if its determinant is zero.
Properties of Determinants
The determinant of a product of matrices is equal to the product of determinants.
Special types of matrices have special determinants; for example, the determinant of an
orthogonal matrix is always plus or minus one, and the determinant of a complex
Hermitian matrix? is always real.
Orthogonal matrix:
7
A real square matrix A, where AA’=A’A=I, then matrix A is orthogonal. An orthogonal matrix is
always invertible, and A-1=A’
Example:
Inverse
Finding inverse
Properties of Inverse:
9
D.3 Linear Independence: Rank of a Matrix
Linear Independence
For a set of vectors having the same dimension, it is important to know whether one vector can
be expressed as a linear combination of the remaining vectors.
0 is a vector of 0’s
0 is a scalar 0
If holds for a set of scalars that are not all zero, then {x1,x2, …,
xr} is linearly dependent. The statement that {x1,x2, …, xr} is linearly dependent is equivalent to
saying that at least one vector in this set can be written as a linear combination of the others.
Solution:
rewriting
10
Rank of a Matrix
Simply the rank of a matrix is the number of linearly independent column of the matrix.
The matrix has a full column rank if its determent is different from zero. If its determinant is zero
it has less than the number of column.
1. Find determinant of the matrix. If determinant of the matrix is non-zero, its rank is the
number of the columns. If determinant is zero then go to step 2
2. By deleting one raw and column find determinant of all possible minors. If determinant
of one of the minor is non-zero then rank is equal with number of the column of the
minor. If all determinants of the column is zero, find the minor of the minor and
continues.
|A|=1(-4)-2(-4)+(-1)(6-2)=0 so the matrix has less than 3 rank. So we have to find the rank of
2x2 matrix by deleting one column and one row and finding the determinant of all possible 2x2
matrices.
11
So, since at least one of the 2x2 matrix determinant is non-zero, the rank of the matrix is 2.
1.
2.
3.
12
Properties of Positive Definite and Positive Semi-Definite Matrices:
1. A positive definite matrix has diagonal elements that are strictly positive, while a p.s.d.
matrix has nonnegative diagonal elements;
2. If A is p.d., then A-1 exists and is p.d.;
3. If X is n x k, then X’X and XX’ are p.s.d.;
4. If X is n x k and rank(X) = k, then X’X is p.d. (and therefore nonsingular).
Example 1: the matrix which is square and symmetric is given below. Find AA
13
Properties of idempotent matrix
1. Let A be an n x n idempotent matrix. rank(A) = tr(A)
2. A is positive semi-definite
We can construct idempotent matrices very generally. Let X be an n x k matrix with rank(X) = k.
Define
14
D.7 Moments and Distributions of Random Vectors
In order to derive the expected value and variance of the OLS estimators using matrices, we need
to define the expected value and variance of a random vector. As its name suggests, a random
vector is simply a vector of random variables. We also need to define the multivariate normal
distribution. These concepts are simply extensions of those covered in Appendix B.
Expected Value
Or
15
Variance-Covariance Matrix
Assuming
Properties of Variance:
17
Chi-Square Distribution
In Appendix B, we defined a chi-square random variable as the sum of squared independent
standard normal random variables. In vector notation, if u ~ Normal(0,In), then u’u ~ χn2.
t Distribution
F Distribution
An F random variable is obtained by taking two independent chi-square random variables and
finding the ratio of each standardized by degrees of freedom.
Properties of F distribution
18