Professional Documents
Culture Documents
March 8, 2024
This set of notes is to supplement the material presented in our classes of Linear Algebra
in the first half of MA110 in Spring 2024 at IIT Bombay. The primary content was
developed by using the reference:
Linear Algebra and its Applications by G. Strang, 4th Ed., Thomson.
Most of this content in these Appendices were not covered in class, but are the result
of various questions asked by curious students. These are being shared primarily
to give such students more content to mull over.
The topics covered in the class were:
2. V ECTOR S PACES
A PPENDIX I - D ETERMINANTS
N OTE : (i) The notation in these notes is the same as that discussed in class.
(ii) Work out as many examples as you can.
1
Appendix I - D ETERMINANTS
• det(AB) = det(A)det(B).
• det(A) = det(AT ).
• A is invertible ⇔ det(A) ̸= 0.
If this happens, then det(A−1 ) =
3. If A is n × n, and its row echelon form U is obtained without row exchanges, then
det(U ) = det(A).
Q: What happens if there are row exchanges? Exercise!
2
4. If A has a zero row, then det(A) = 0.
Proof: Let the ith row of A be zero, i.e., Ai∗ = 0.
Let B be obtained from A by Ri = Ri + Rj , i.e., B = Eij (1)A. Then Bi∗ = Bj∗ .
If at least one diagonal entry is zero, then elimination will produce a zero row ⇒
det(A) = 0.
a 0 0 b 0 b c 0 a b
= 0, = 0, =− = −bc ⇒ = ad − bc
c 0 0 d c 0 0 b c d
3
Cofactors: 3 × 3 Case
Cofactors: n × n Case
Let C1j be the coefficient of a1j in the expansion
X
det(A) = (a1α1 . . . anαn ) det(P )
all permutations P
where M1j is obtained from A by deleting the 1st row and j th column.
7.
det(AB) = det(A) det(B)
d(A) = det(AB)/det(B)
(a) d(I) = 1.
(b) If we interchange two rows of A, then d changes its sign.
(c) d is a linear function in each row of A.
Then d is the unique determinant function det and det(AB) = det(A) det(B). □
4
Determinants of Transposes (Proof)
8.
det(A) = det(AT )
5
Determinant: Geometric Interpretation
• Test for invertibility: An n × n matrix A is invertible ⇔ det(A) ̸= 0.
• n-dimensional volume: If A is n × n, then |det(A)| = the volume of the box
(in n-dimensional space Rn ) with edges as rows of A.
Examples: (1) The volume (area) of a line in R2 = 0.
1 2
(2) The determinant of the matrix A = is −4 .
1 −2
(3) Let’s compute the volume of the box (parallelogram) with edges as
rows of A or columns of A.
y
y
=4
x x
6
1
R EMARK . If A is invertible, then A−1 = CT .
det(A)
For n ≥ 4, this is not a good formula to find A−1 . Use elimination to find A−1 for n ≥ 4.
This formula is of theoretical importance.
Applications: 2. Solving Ax = b
C RAMER ’ S RULE : If A is invertible, the Ax = b has a unique solution.
C11 · · · Cn1 b1
1 1 .. .
x=A b= −1 T
C b= . .. ...
det(A) det(A)
C1n · · · Cnn bn
1 1
Hence xj = (b1 C1j + b2 C2j + · · · + bn Cnj )= det(Bj ), where Bj is obtained by
det(A) det(A)
replacing the j th column of A by b, and det(Bj ) is computed along the j th column.
R EMARK : For n ≥ 4, use elimination to solve Ax = b. Cramer’s rule is of theoretical
importance.
Since the edges of the box spanned by rows of A are at right angles, the volume of the
box
= the product of lengths of edges = |det(A)|
.
• det(A
e1 ) = a11 = d1
• det(A
e2 ) = d1 d2 = det(A1 ) d2
• det(A
e3 ) = d1 d2 d3 = det(A2 ) d3 etc.,
• If det(A
ek ) = 0, then we need a row exchange in elimination.
7
Appendix II - O R THOGONAL S UBSPACES
(1, −1, 0) y
x
y=x
8
Orthogonal Subspaces:Examples
1 0 1
1 1 −1
E XAMPLE 2: Let v1 = 0 , v2 = 1 and w = 1 .
0 0 0
If V = Span {v1 , v2 }, and W = Span{w}, then V ⊥ W .
Sketch of Proof. It is enough to see that v1 , v2 ∈ V are orthogonal to w ∈ W .
C OMPARE : In Example 1, dim (L) + dim (P ) = dim (R3 ).
In Example 2, dim (V ) + dim (W ) = 3 < dim (R4 ).
Q: Can we enlarge W to W ′ ⊆ R4 such that V ⊥ W ′ and dim (V ) + dim (W ′ ) = dim (R4 )?
A: Yes!
O BSERVE : (0, 0, 0, 1) is orthogonal to both V and W .
T
v 1 1 0 0
Let A = 1T = . Then C(AT ) = V .
v2 0 1 1 0
Note that x ∈ N (A) ⇔ v1T x = 0 = v2T x. ⇒ (1, −1, 1, 0)T ∈ N (A) ⇒ W ⊂ N (A).
By the Rank-Nullity Theorem, rank(A)+ dim(N (A)) = 4. Since rank(A) = 2,
we get dim(N (A)) = 2. Therefore if W ′ = N (A), then V ⊥ W ′ and dim (V ) + dim (W ′ ) = 4.
By inspection, W ′ = Span{w1 = (1, −1, 1, 0)T , w2 = (0, 0, 0, 1)T } is orthogonal to V .
The dimension of the two spaces sum up to dim (R4 ).
9
Orthogonal Complements
T HEOREM (Orthogonal Complement)
Given a subspace W ⊆ Rn , dim (W ) + dim (W ⊥ ) = n.
Proof. Let v1 , . . . , vr be a basis of W . Let A be a matrix with rows v1 , . . . , vr .
Then rank(A) = r and W = C(AT ). W ⊥ = N (A) is of dimension n − r.
This proves the theorem.
O BSERVE : Let V = Span v1 = (1, 1, 0, 0)T , v2 = (0, 1, 1, 0)T
Orthogonal Complements
T O SUMMARISE : If W is a subspace of Rn with basis B, and B ′ is a basis of W ⊥ ,
then B ∪ B ′ is a basis of Rn , W ∩ W ⊥ = 0, dim (W ) + dim (W ⊥ ) = n, W + W ⊥ = Rn ,
and for every x ∈ Rn , ∃ unique w1 ∈ W and w2 ∈ W ⊥ such that x = w1 + w2 .
S OME C ONSEQUENCES : Let A be m × n of rank r. Then
1. C(AT ) ∩ N (A) = 0 and Rn = C(AT ) + N (A).
Similarly C(A) ∩ N (AT ) = 0 and Rm = C(A) + N (AT ).
2. If x ∈ Rn , there is a unique expression x = xr + xn , where xr ∈ C(AT ), xn ∈ N (A).
Hence Ax = Axr ∈ C(A). Thus
The matrix A transforms its row space C(AT ) into its column space C(A).
10
Appendix III - A DDITIONAL T OPICS
Introduction
Topics included (in no particular order) are:
1. An Application of QR-factorization
2. Angles in Rn
Recall that QT Q = = Ir .
T T
qr q1 . . . qr qr
Let Ax = b be an inconsistent system.
We apply least squares method to get best approximate solution.
AT A = (QR)T QR = RT QT QR = RT R. Therefore we solve
AT A x̂ = RT R x̂ = AT b = RT QT b. Since RT is invertible,
Thus least square solution is given by Rx̂ = QT b .
III.2 Angles in Rn
Angles in R2
Note: w ∈ R2 can be written as (||w|| cos α, ||w|| sin α), where
y
v
w
θ
α
x
cos(θ + α)
v = ∥v∥
sin(θ + α)
cos α
w = ∥w∥
sin α
11
v T w = ∥v∥∥w∥(cos(θ + α) sin α + sin(θ + α) cos α)
= ∥v∥∥w∥ cos(θ + α − α) = ∥v∥∥w∥ cos θ
This gives a formula for the angle between two vectors in terms of the inner product and
norm (length) of vectors.
Angles in Rn
To generalise the angle between two vectors to Rn , we use the notion of inner product
and norm in Rn . But to do this, we need the Cauchy-Schwartz inequality for v, w ∈ Rn :
|v T w| ≤ ∥v∥∥w∥
2
vT w
To verify this expand the inequality w − v ≥ 0.
∥v∥∥w∥
Define the angle between any two vectors v, w ∈ Rn as
T
−1 v w
θ = cos .
∥v∥∥w∥
Inner Product
We have defined orthogonality only in Rn .
Can we do something similar for other vector spaces?
If V is n-dimensional, we can use coordinate vectors and the inner product in Rn .
P OINT TO P ONDER :
Can different bases and coordinate vectors be used to define new inner products on Rn ?
N OTE : On the infinite dimensional space R∞ , the space of all real sequences,
the dot product does not generalise.
If xn = √1n , then (xn ) · (xn ) = Σn n1 which does not converge, and hence is not finite.
However, this generalization works for sequences (xn ) and (yn ) if Σn (xn yn ) is convergent.
Define ℓ2 (R) = {(x1 , . . . , xn , . . .) ∈ R∞ | Σi x2i < ∞}. E.g., (1, 1/2, 1/3, . . . , 1/n, . . .) ∈ ℓ2 (R).
Which of the following are in ℓ2 (R)? (1, 0, . . . , 0, . . .), (1, 0, 10, 1, 0, . . .), (1, 1/2, 1/22 , 1/23 , . . .)
12
• (Symmetry) ⟨x , y⟩ = ⟨y , x⟩
13
III.4 Symmetric Matrices & SVD
λ1 (v1T v2 ) = (λ1 v1 )T v2 = (Av1 )T v2 = (v1T AT )v2 = v1T (Av2 ) = v1T (λ2 v2 ) = λ2 (v1T v2 ).
A = QΛQT .
Given that A can be diagonalised, then A = QΛQT follows from the previous slide ,
together with the Gram-Schmidt process applied to each eigenspace of A.
∼∼∼ f in ∼∼∼
14