Professional Documents
Culture Documents
Michael L. Baker
<mbaker@lambertw.com>
UNIVERSITY OF WATERLOO
Preface
The aim of this document is to serve as a reference of problems and solutions from the fourth edition
of Linear Algebra by Friedberg, Insel and Spence. Originally, I had intended the document to
be used only by a student who was well-acquainted with linear algebra. However, as the document
evolved, I found myself including an increasing number of problems. Therefore, I believe the
document should be quite comprehensive once it is complete.
I do these problems because I am interested in mathematics and consider this kind of thing to be
fun. I give no guarantee that any of my solutions are the best way to approach the corresponding
problems. If you find any errors (regardless of subtlety) in the document, or you have different
or more elegant ways to approach something, then I urge you to contact me at the e-mail address
supplied above.
This document was started on July 4, 2010. By the end of August, I expect to have covered up to
the end of Chapter 5, which corresponds to the end of MATH 146, Linear Algebra 1 (Advanced
Level) at the University of Waterloo. This document is currently a work in progress.
Contents
1 Vector Spaces
1.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . .
1.2 Vector Spaces . . . . . . . . . . . . . . . . . . . . . . .
1.3 Subspaces . . . . . . . . . . . . . . . . . . . . . . . . .
1.4 Linear Combinations and Systems of Linear Equations
1.5 Linear Dependence and Linear Independence . . . . .
1.6 Bases and Dimension . . . . . . . . . . . . . . . . . . .
1.7 Maximal Linearly Independent Subsets . . . . . . . . .
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
3
3
3
4
10
16
21
30
Vector Spaces
1.1
Introduction
Section 1.1 consists of an introductory, geometrically intuitive treatment of vectors (more specifically, Euclidean vectors). The solutions to the exercises from this section are very basic and as such
have not been included in this document.
1.2
Vector Spaces
Section 1.2 introduces an algebraic structure known as a vector space over a field, which is then used
to provide a more abstract notion of a vector (namely, as an element of such an algebraic structure).
Matrices and n-tuples are also introduced. Some elementary theorems are stated and proved, such
as the Cancellation Law for Vector Addition (Theorem 1.1), in addition to a few uniqueness results
concerning additive identities and additive inverses.
8. In any vector space V, show that (a + b)(x + y) = ax + ay + bx + by for any x, y V and any
a, b F .
Solution. Noting that (a + b) F , we have
(a + b)(x + y) = (a + b)x + (a + b)y
= ax + bx + ay + by
= ax + ay + bx + by
(by VS 7)
(by VS 8)
(by VS 1)
as required.
9. Prove Corollaries 1 and 2 [uniqueness of additive identities and additive inverses] of Theorem
1.1 and Theorem 1.2(c) [a0 = 0 for each a F ].
Solution. First, let 0, 00 V be additive identities, and let x V. We have, by VS 3,
x = x + 0 = x + 00
Whereupon VS 1 yields
0 + x = 00 + x
By Theorem 1.1 [Cancellation Law for Vector Addition], we obtain 0 = 00 , proving that the
additive identity is unique.
Second, let x V and y, y 0 V be such that x + y = x + y 0 = 0. Then we obtain, by VS 1,
y + x = y0 + x
By Theorem 1.1 [Cancellation Law for Vector Addition], we obtain y = y 0 , proving that each
additive inverse is unique.
Third, let a F . We have, by VS 3 and VS 7,
a0 = a(0 + 0) = a0 + a0
By Theorem 1.1 [Cancellation Law for Vector Addition], we obtain a0 = 0, as required.
3
10. Let V denote the set of all differentiable real-valued functions defined on the real line. Prove
that V is a vector space with the operations of addition and scalar multiplication defined in
Example 3.
Solution. Let f, g, h V and c, d, s R. Due to rules taught in elementary calculus, we
have f + g V as well as cf V for all f, g V and c R. Now, note that by the
commutativity of addition in R, we have f + g = f (s) + g(s) = g(s) + f (s) = g + f for
all s. Therefore, V satisfies VS 1. Next, by the associativity of addition in R, we have
(f + g) + h = (f (s) + g(s)) + h(s) = f (s) + (g(s) + h(s)) = f + (g + h) for all s. Therefore,
V satisfies VS 2. Next, let 0V (s) = 0F for all s, where 0F is the additive identity of R.
Then f + 0V = f (s) + 0F = f (s) = f for all s. Therefore V satisfies VS 3. Now, for each
s we have f (s) R and so we are guaranteed an additive inverse, say f (s), such that
f (s) + (f (s)) = 0F . Now, 0V (s) = 0F for all s also, so we note f + (f ) = 0V for all f .
Therefore V satisfies VS 4. Next, let 1F represent Rs multiplicative identity. Then we have
1F f = 1F f (s) = f (s) = f for all s. Therefore, 1F f = f for all f . So VS 5 is satisfied. Note
also that (cd)f = (cd) f (s) = c (df (s)) = c(df ) for all s. This is because multiplication in R
is associative. So VS 6 is satisfied. We also have c(f +g) = c(f (s)+g(s)) = cf (s)+cg(s) =
cf +cg. This is because multiplication in R is distributive (over addition). So VS 7 is satisfied.
Finally, note (c + d)f = (c + d) f (s) = c f (s) + d f (s) = cf + df . This is again because
multiplication in R is distributive (over addition). So VS 8 is satisfied. This completes the
proof that V, together with addition and scalar multiplication as defined in Example 3, is a
vector space.
22. How many matrices are there in the vector space Mmn (Z2 )?
Solution. There are 2mn vectors in this vector space.
1.3
Subspaces
Section 1.3 examines subsets of vector spaces, and defines a special type of subset, known as a
subspace, which is a subset of a vector space that can be considered a vector space in its own right
(only certain subsets satisfy this property). A certain theorem that is referred to as the subspace test
is stated and proven, which provides a fast way of checking whether a given subset is a subspace.
More concepts relating to matrices are introduced as well. The section closes with a proof that the
intersection of two subspaces is itself a subspace.
3. Prove that (aA + bB)t = aAt + bB t for any A, B Mmn (F ) and any a, b F .
Solution. We have (aA)ij = a Aij , (bB)ij = b Bij , so (aA + bB)ij = a Aij + b Bij . So
[(aA + bB)t ]ij = a Aji + b Bji
Now, (At )ij = Aji , (B t )ij = Bji , (aAt )ij = a Aji , (bB t )ij = b Bji . So
[aAt + bB t ]ij = a Aji + b Bji
Therefore (aA + bB)t = aAt + bB t as required.
4. Prove that (At )t = A for each A Mmn (F ).
4
Solution. We have (At )ij = Aji . Thus [(At )t ]ij = (At )ji = Aij , so that (At )t = A as
required.
5. Prove that A + At is symmetric for any square matrix A.
Solution. We have, from exercises 3 and 4:
(A + At ) = At + A
= At + (At )t
= (A + At )t
6. Prove that tr(aA + bB) = a tr(A) + b tr(B).
Solution. We have
tr(aA + bB) =
=
n
X
i=1
n
X
(aA + bB)ii
(aAii + bBii )
i=1
n
X
=a
i=1
Aii + b
n
X
Bii
i=1
= a tr(A) + b tr(B)
19. Let W1 and W2 be subspaces of a vector space V. Prove that W1 W2 is a subspace of V if
and only if W1 W2 or W2 W1 .
Solution. To prove the first direction, assume W1 W2 is a subspace of V. Assume also that
W1 6 W2 . Then we are guaranteed some a W1 such that a
/ W2 . Choose an arbitrary
b W2 , and consider the sum a + b. Clearly (a + b) W1 W2 , and so (a + b) W1 or
(a + b) W2 . Assume that (a + b) W1 . Then, since W1 and W2 are subspaces, we have
(a + b) a = b W1 . The other possibility is that (a + b) W2 , in which case we obtain
(a + b) b = a W2 . However, this is a contradiction, and so we conclude that b W1 ,
implying W2 W1 .
To prove the second direction, assume W1 W2 or W2 W1 . Then clearly the union W1 W2
will be the larger of these two subspaces, and therefore W1 W2 is a subspace.
20. Prove that if W is a subspace of a vector space V and w1 , w2 , . . . , wn are in W , then
a1 w1 + a2 w2 + . . . + an wn W for any scalars a1 , a2 , . . . , an .
Solution. For each i such that 1 i n, we have ai wi W due to Ws closure under scalar multiplication. We also know that ai wi + ai+1 wi+1 W for 1 i n 1,
since W is closed under addition. An inductive argument can then be used to show that
a1 w1 + a2 w2 + . . . + an wn W.
Definition. If S1 and S2 are nonempty subsets of a vector space V, then the sum of S1 and S2 ,
denoted S1 + S2 , is the set {x + y : x S1 and y S2 }.
Definition. A vector space V is called the direct sum of W1 and W2 if W1 and W2 are subspaces
of V such that W1 W2 = {0} and W1 + W2 = V . We denote that V is the direct sum of W1 and
W2 by writing V = W1 W2 .
23. Let W1 and W2 be subspaces of a vector space V.
(a) Prove that W1 + W2 is a subspace of V that contains both W1 and W2 .
Solution. We have that W1 + W2 = {x + y : x W1 and y W2 }. Since W1 , W2 are
subspaces, they both contain the additive identity 0 from V. Therefore, 0 + 0 = 0
(W1 + W2 ). Next, let a, b (W1 + W2 ). Then we have a = xa + ya and b = xb + yb for
xa , xb W1 , ya , yb W2 , and so a + b = (xa + ya ) + (xb + yb ) = (xa + xb ) + (ya + yb ).
Since W1 , W2 are subspaces, we have (xa + xb ) W1 and (ya + yb ) W2 , and therefore
a + b (W1 + W2 ). Now let c F and consider ca = c(xa + ya ) = (cxa ) + (cya ).
Since W1 , W2 are subspaces, we again have cxa W1 and cya W2 , and therefore
ca (W1 + W2 ). We conclude that W1 + W2 is a subspace, by the subspace test. To
show that W1 W1 + W2 , simply consider any element u W1 . Since 0 W2 , we can
write u = u + 0 W1 + W2 by the definition of W1 + W2 , and so W1 W1 + W2 . To
show that W2 W1 + W2 , note 0 W1 also, and argue similarly.
(b) Prove that any subspace that contains both W1 and W2 must also contain W1 + W2 .
Solution. Assume W is some subspace of V, and W1 W and W2 W. Let a
(W1 + W2 ). Then a = x + y for x W1 and y W2 . However, because of the above,
we obtain x W and y W, and since W is a subspace and closed under addition, we
obtain a W. This shows that W1 + W2 W.
24. Show that Fn is the direct sum of the subspaces
W1 = {(a1 , a2 , . . . , an ) Fn : an = 0}
and
W2 = {(a1 , a2 , . . . , an ) Fn : a1 = a2 = . . . = an1 = 0}.
Solution. First, lets examine W1 W2 . The definitions of W1 and W2 tell us that if
(a1 , a2 , . . . , an ) W1 W2 , then a1 = a2 = . . . = an = 0. This clearly only holds for the ntuple (0, 0, . . . , 0). Next, let x = (x1 , x2 , . . . , xn ) Fn . Then we can clearly write x = w1 + w2
for some w1 W1 and w2 W2 ; more specifically, where w1 = (x1 , x2 , . . . , xn1 , 0) and
w2 = (0, 0, . . . , 0, xn ). So Fn W1 + W2 . It is not hard to demonstrate that W1 + W2 Fn
to prove that Fn = W1 + W2 , proving that Fn is by definition the direct sum of W1 and W2 .
31. Let W be a subspace of a vector space V over a field F . For any v V the set {v} + W =
{v + w : w W} is called the coset of W containing v. It is customary to denote this coset
by v + W rather than {v} + W.
(a) Prove that v + W is a subspace of V if and only if v W.
Solution. To prove the first direction, assume v + W is a subspace of V. Clearly,
6
1.4
Section 1.4 discusses the notion of linear combination and also provides a method for solving systems
of linear equations which will later serve as the foundation for the row reduction of matrices. The
span of a subset is introduced, in addition to the notion of a generating subset. Some geometric
interpretations of the concepts are given.
2. Solve the following systems of linear equations by the method introduced in this section.
(a)
2x1 2x2 3x3
= 2
(1.4.1)
(1.4.2)
x1 x2 2x3 x4 = 3
(1.4.3)
Solution. Begin by replacing equation (1.4.2) with (1.4.2) 3 (1.4.3) to obtain 4x3 +
8x4 = 16. Divide equation (1.4.2) by 4 to obtain x3 + 2x4 = 4. Now replace equation
(1.4.1) with (1.4.1) 2 (1.4.3) to obtain x3 + 2x4 = 4 also. Switch equations (1.4.3) and
(1.4.1). We obtain
x1 x2 2x3 x4 = 3
x3 + 2x4 =
Which gives us
x3 = 4 2x4
x1 = x2 3x4 + 5
10
(1.4.4)
x1 2x2 + x3 = 3
(1.4.5)
2x1 x2 2x3 = 6
(1.4.6)
(1.4.7)
(1.4.8)
(1.4.9)
Solution. Begin by replacing equation (1.4.8) with (1.4.8)(1.4.7) to obtain 2x2 2x3
4x4 = 1. Next, replace equation (1.4.9) with (1.4.9) 2 (1.4.7) obtain x2 + x3 + 2x4 =
2. Now, replace equation (1.4.9) with (1.4.9) + 12 (1.4.8), to obtain 0 = 52 . This
shows that the system is inconsistent and therefore our solution set is the empty set,
S=
3. For each of the following lists of vectors in R3 , determine whether the first vector can be
expressed as a linear combination of the other two.
11
(1.4.10)
0
3
(1.4.11)
(1.4.12)
Applying the same procedure as previously, we obtain s = 4 and t = 3. The system therefore has solutions, and so we conclude (2, 0, 3) can be expressed as a linear
combination of (1, 3, 0) and (2, 4, 1).
4. For each list of polynomials in P3 (R), determine whether the first polynomial can be expressed
as a linear combination of the other two.
(b) 4x3 + 2x2 6, x3 2x2 + 4x + 1, 3x3 6x2 + x + 4
Solution. We need to verify that there exist s, t R such that
4x3 + 2x2 6 = s(x3 2x2 + 4x + 1) + t(3x3 6x2 + x + 4)
= (s + 3t)x3 + (2s 6t)x2 + (4s + t)x + (s + 4t)
This yields the following system of equations:
s + 3t =
(1.4.13)
2s 6t =
4s + t =
2
0
(1.4.14)
(1.4.15)
s + 4t = 6
(1.4.16)
Applying the same procedure as previously, we obtain 2s 6t = 8, which is inconsistent with (1.4.14). The system therefore has no solutions, and so we conclude that
4x3 + 2x2 6 cannot be expressed as a linear combination of x3 2x2 + 4x + 1 and
3x3 6x2 + x + 4.
10. Show that if
M1 =
1
0
0
,
0
M2 =
0
0
0
,
1
M3 =
0
1
1
,
0
12
1
x11
0
0
0
+n
0
1
0
1
+ x22
0
0
0
1
=
x11
n
n
x22
=a
and hence a span(S) where we define S = {M1 , M2 , M3 }. This shows that A span(S). A
similar argument demonstrates that span(S) A, and this completes the proof.
11. Prove that span({x}) = {ax : a F } for any vector x in a vector space. Interpret this result
geometrically in R3 .
Solution. Since {x} contains only one vector, then for all a span({x}) we can write, for
some c F ,
a = cx
which proves span({x}) {ax : a F }. Clearly, if b {ax : a F } then b is a linear
combination of the vectors in {x}, namely, a scalar multiple of x itself, so that {ax : a F }
span({x}). This proves span({x}) = {ax : a F }. In R3 this result tells us that the span of
any single vector will be a line passing through the origin.
12. Show that a subset W of a vector space V is a subspace of V if and only if span(W) = W.
Solution. Assume W V such that span(W) = W. 0 is in the span of any subset of V due
to the trivial representation. Now let a, b W. Since span(W) = W, we have a, b span(W),
so we can write
!
!
n
n
n
X
X
X
a+b=
(ai + bi )vi
ai vi +
bi vi =
i=1
i=1
i=1
i=1
which demonstrates that ca span(W) and hence ca W. This proves that W is closed under
scalar multiplication. By the subspace test, we have that W is a subspace.
Assume next that W V with W a subspace of V. Since W is a subspace of V, clearly
0 W, and W is closed under addition and scalar multiplication (in other words, linear
combination). Any vector a W can be trivially represented as a linear combination of
vectors in W (viz., itself multiplied by the scalar 1 from F ) and so we have a span(W),
proving that W span(W). Now, let a span(W). Then, by the definition of span(W), we see
that a is a linear combination of vectors in W. By noting first that W is closed under scalar
multiplication and then noting its closure under addition, we clearly have a W, proving
span(W) W, and hence span(W) = W.
13. Show that if S1 and S2 are subsets of a vector space V such that S1 S2 , then span(S1 )
span(S2 ). In particular, if S1 S2 and span(S1 ) = V, deduce that span(S2 ) = V.
13
However, we have S1 S2 , which implies that for each i, we have vi S2 also. By the
definition of the linear combination, then, we obtain that a span(S2 ), demonstrating that
span(S1 ) span(S2 ). Now assume S1 S2 and span(S1 ) = V. We obtain (from the
proposition we just proved) that span(S1 ) = V span(S2 ). However, each element in S2
is also in V. Since V is closed under vector addition and scalar multiplication, surely every
linear combination of vectors in S2 must be in V. Therefore, span(S2 ) V, which, since
V span(S2 ) also, shows that span(S2 ) = V in this case.
14. Show that if S1 and S2 are arbitrary subsets of a vector space V, then span(S1 S2 ) =
span(S1 ) + span(S2 ). (The sum of two subsets is defined in the exercises of Section 1.3.)
Solution. We will define
A = span(S1 ) + span(S2 ) = {(a + b) : a span(S1 ), b span(S2 )}
B = span(S1 S2 )
Assume a A. Then we have a = u + v for some u span(S1 ) and v span(S2 ). Therefore, we can write, for some a1 , a2 , . . . , am , b1 , b2 , . . . , bn F and x1 , x2 , . . . , xm S1 and
y 1 , y 2 , . . . , y n S2 :
!
!
m
n
X
X
a=
ai xi +
bi y i
i=1
i=1
Since, for each i, we have xi S1 and hence xi (S1 S2 ), and yi S2 and hence yi (S1 S2 ),
we clearly see that u, v span(S1 S2 ). Since by Theorem 1.5, the span of any subset of a
vector space V is a subspace of V, it is closed under vector addition, implying (u + v) = a B,
we have obtained A B.
Now, assume b B. Then we can write, for some a1 , a2 , . . . , an F and z1 , z2 , . . . , zn
(S1 S2 ), the following:
n
X
b=
ai zi
i=1
Therefore, for each i, we see that zi satisfies either zi S1 or zi S2 (or both), for otherwise
we have zi
/ (S1 S2 ). Now, we partition the set of summands as follows:
P = {ai zi : zi S1 , 0 i n}
Q = {ai zi : zi S2 , zi
/ S1 , 0 i n}
It can easily be proved that P Q yields all the summands in the expression for b and that
P Q = . Then we can write the decomposition
b=
n
X
ai zi =
i=1
X
xP
14
x+
X
xQ
P
P
We clearly have that ( xP x) span(S1 ) and ( xQ x) span(S2 ) which proves that
b A, whereupon we obtain B A. We have therefore proved that A = B, as required.
15. Let S1 and S2 be subsets of a vector space V. Prove that span(S1 S2 ) span(S1 )span(S2 ).
Give an example in which span(S1 S2 ) and span(S1 ) span(S2 ) are equal and one in which
they are unequal.
Solution. We will define
A = span(S1 S2 )
B = span(S1 ) span(S2 )
Assume a A. Then we can write, for some a1 , a2 , . . . , an F and v1 , v2 , . . . , vn ,
a=
n
X
ai vi
i=1
such that for each i, vi satisfies both vi S1 and vi S2 . Because of this, we see immediately
that a span(S1 ) and a span(S2 ), proving that a (span(S1 ) span(S2 )). Therefore,
A B, as required.
Now, let V = R2 and S1 = {(1, 1)}, S2 = {(1, 1)}. Then S1 S2 = , so that A =
span(S1 S2 ) = {0} and B = span(S1 ) span(S2 ) = {0}. This is a case in which A = B.
Next, let V = R2 and S1 = {(2, 0)} and S2 = {(1, 0)}. Then S1 S2 = , yielding A =
span(S1 S2 ) = span = {0}. Now,
span(S1 ) = {k(2, 0) : k R}
span(S2 ) = {k(1, 0) : k R}
It is not hard to show that span(S1 ) = span(S2 ), so we conclude B = span(S1 ) span(S2 ) =
span(S1 ). Clearly, span(S1 ) 6= , so this is a case in which A ( B.
16. Let V be a vector space and S a subset of V with the property that whenever v1 , v2 , . . . , vn S
and a1 v1 + a2 v2 + + an vn = 0, then a1 = a2 = = an = 0. Prove that every vector in
the span of S can be uniquely written as a linear combination of vectors of S.
Solution. Assume v span(S), and assume also that we can represent v in two ways, like
so, for some a1 , a2 , . . . , an , b1 , b2 , . . . , bn F and v1 , v2 , . . . , vn S:
v=
n
X
ai vi =
i=1
n
X
bi vi
i=1
n
X
!
ai vi
i=1
n
X
!
bi vi
i=1
n
X
(ai bi )vi
i=1
At this point, we use the given property of S, and this gives us that for each i, we have
ai bi = 0
15
which implies ai = bi . This clearly proves that our two representations of v cannot be distinct,
as required.
17. Let W be a subspace of a vector space V. Under what conditions are there only a finite
number of distinct subsets S of W such that S generates W?
Solution. I believe the following conditions would lead to there only being a finite number
of distinct generating sets for W:
The case where W = {0}, in which case there would only be the set {0} and that
generate W.
The case where F is a Galois field such as Z2 .
1.5
Section 1.5 discusses the notion of linear dependence, first through the question Can one of the
vectors in this set be represented as a linear combination of the others? and then reformulates
this notion in terms of being able to represent the zero vector as a nontrivial linear combination of
vectors in the set. Afterwards, some proofs concerning linear dependence and linear independence
are given.
2. Determine whether the following sets are linearly dependent or linearly independent.
1 3
2
6
,
(a)
in M22 (R)
2
4
4 8
Solution. We consider the equation
1 3
2
6
a1 2a2
a1
+ a2
=
2
4
4 8
2a1 + 4a2
3a1 + 6a2
4a1 8a2
=
0
0
0
0
(1.5.1)
3a1 + 6a2 = 0
(1.5.2)
2a1 + 4a2 = 0
(1.5.3)
4a1 8a2 = 0
(1.5.4)
Solving the system with the usual method, we obtain a1 = 2a2 . Therefore there are
infinitely many nontrivial solutions and so the set is linearly dependent in M22 (R).
1 2
1
1
(b)
,
in M22 (R)
1
4
2 4
Solution. We consider the equation
1 2
1
1
a1 a2
a1
+ a2
=
1
4
2 4
a1 + 2a2
16
2a1 + a2
4a1 4a2
=
0
0
0
0
(1.5.5)
2a1 + a2 = 0
(1.5.6)
a1 + 2a2 = 0
(1.5.7)
4a1 4a2 = 0
(1.5.8)
Solving the system with the usual method, we obtain a1 = 0 and a2 = 0 (which proves
that only trivial solutions exist) and so the set is linearly independent in M22 (R).
4. In Fn , let ej denote the vector whose jth coordinate is 1 and whose other coordinates are 0.
Prove that {e1 , e2 , . . . , en } is linearly independent.
Solution. Upon writing out the system of equations, we see that the system immediately
yields ai = 0 for all i satisfying 1 i n. Therefore, {e1 , e2 , . . . , en } is linearly independent.
7. Recall from Example 3 in Section 1.3 that the set of diagonal matrices in M22 (F ) is a
subspace. Find a linearly independent subset that generates this subspace.
Solution. Since all diagonal matrices are in the form
a 0
1 0
0
=a
+b
0 b
0 0
0
0
1
0
0
,
0
0
1
0
0
1
8. Let S = {(1, 1, 0), (1, 0, 1), (0, 1, 1)} be a subset of the vector space F3 .
(a) Prove that if F = R, then S is linearly independent.
Solution. Assume F = R. Then we must prove there exist no nontrivial solutions
a1 , a2 , a3 R to
a1 (1, 1, 0) + a2 (1, 0, 1) + a3 (0, 1, 1) = (0, 0, 0)
The equation above yields the following system:
a1 + a2 = 0
(1.5.9)
a1 + a3 = 0
(1.5.10)
a2 + a3 = 0
(1.5.11)
(a1 c + a2 )u = 0 by VS 8. Then the equation will hold for any choice of a1 , a2 F such that
a1 c + a2 = 0. This equation can clearly be satisfied by some choice of a1 , a2 F , not both
zero, because if we let a1 = 1 and a2 = c (these elements are guaranteed to be in F due
to the field axioms), we obtain a1 c + a2 = 1 c + (c) = c + (c) = 0 by the field axioms.
Therefore, {u, v} is linearly dependent.
Now, assume {u, v} is linearly dependent. Then there exist a1 , a2 F , not all zero, such that
a1 v + a2 u = 0. Since F is a field, division is possible, allowing us to write
v=
a2
u
a1
19
Solution. Assume S is linearly dependent. Then by exercise 14, either S = {0} or there
are v, u1 , . . . , un S such that v is a linear combination of S. If S = {0} then u1 = 0 as
required. Otherwise, we can write uk = v so that uk = a1 u1 + a2 u2 + . . . + an un for some
a1 , . . . , an F . Either k > n and the proof is complete, since the fact that we can write uk =
a1 u1 +a2 u2 +. . .+an un +0un+1 +. . .+0uk1 clearly implies uk span({u1 , . . . , un , . . . , uk1 })
as required. Otherwise we can simply isolate un in the equation and use the same reasoning.
Now, assume either S = {0} or there exist uk+1 S such that uk+1 span({u1 , u2 , . . . , uk })
for some k (1 k n). If S = {0}, then S is linearly dependent by exercise 14. Otherwise,
the fact that uk+1 span({u1 , u2 , . . . , uk }) for some k satisfying 1 k n implies that there
exists some vector v S (namely, uk+1 ) that is a linear combination of some other vectors in
S (namely the set {u1 , u2 , . . . , uk }). Exercise 14 then implies that S is linearly dependent.
16. Prove that a set S of vectors is linearly independent if and only if each finite subset of S is
linearly independent.
Solution. Assume S is linearly independent. For each finite subset R S, R must be linearly
independent by exercise 12, as required.
Next, assume that each finite subset R S is linearly independent. Assume, for the sake
of contradiction, that S is linearly dependent. Then there exist vectors v1 , . . . , vn S, and
scalars a1 , . . . , an F , not all zero, such that a1 v1 + . . . + an vn = 0. This implies that some
finite subset R S is linearly dependent, which is contradictory to what we assumed. This
completes the proof.
17. Let M be a square upper triangular matrix (as defined in Exercise 12 of Section 1.3) with
nonzero diagonal entries. Prove that the columns of M are linearly independent.
Solution. Clearly, Mi,j = 0 whenever i > j and Mi,j 6= 0 whenever i = j. We wish to prove
that the equation
0
M1,n
M1,1
M1,2
a1 + a2 + . . . + an =
Mn,n
0
Mn,2
Mn,1
only holds when a1 = . . . = an = 0. This, of course, yields the n equations:
a1 M1,1 + a2 M1,2 + . . . + an M1,n = 0
a1 M2,1 + a2 M2,2 + . . . + an M2,n = 0
..
.
a1 Mn,1 + a2 Mn,2 + . . . + an Mn,n = 0
We proceed by induction. Let P (k) be the statement that ank+1 = 0 (1 k n). Consider
P (1). We wish to show that an1+1 = 0, or in other words an = 0. Looking at the equation
n 1 + 1, or in other words equation n:
a1 Mn,1 + a2 Mn,2 + . . . + an Mn,n = 0
we note that each Mn,i = 0 unless i = n, since this is the only case when n > i does not hold
(the matrix is upper triangular). This allows us to remove the first n 1 summands from the
20
1.6
Section 1.6 presents, and gives examples of, the concept of a basis, that is, a linearly independent
subset of a vector space that generates the vector space. Theorem 1.8 says that vectors in the
vector space can be represented as unique linear combinations of vectors in a subset if and only
if the subset is a basis. Theorem 1.9 says that every finite generating set for a vector space can
be reduced to a basis for that vector space. Theorem 1.10 (the replacement theorem) gives a very
important result: that a linearly independent subset in a vector space has no more elements in it
than a set that generates that vector space, and that a subset of this generator can be unioned
with the independent subset to form a generating set. Corollaries to this theorem are the fact that
every basis in a finite-dimensional vector space contains the same number of elements, that any
finite generating set for a vector space V contain at least dim(V) vectors, and that a generating set
containing exactly dim(V) vectors is a basis. Additionally, any linearly independent subset of V
with exactly dim(V) vectors is a basis, and every linearly independent subset of V can be extended
to a basis for V.
21
1
0
1
0
,
0
0
0
1
1
0
1
0
,
0
1
0
0
7. The vectors u1 = (2, 3, 1), u2 = (1, 4, 2), u3 = (8, 12, 4), u4 = (1, 37, 17), and u5 =
(3, 5, 8) generate R3 . Find a subset of the set {u1 , u2 , u3 , u4 , u5 } that is a basis for R3 .
Solution. Start with the set {u1 } and note that u2 is not a multiple of u1 . Therefore
{u1 , u2 } will be linearly independent. Next, guess that u5 is not a linear combination of
the two vectors already in the set. To verify, solve the usual homogeneous system of linear
equations to determine whether {u1 , u2 , u5 } is linearly independent. As it turns out, it is, and
therefore by a corollary to the replacement theorem, the set generates R3 , and so is a basis
for R3 also.
9. The vectors u1 = (1, 1, 1, 1), u2 = (0, 1, 1, 1), u3 = (0, 0, 1, 1), and u4 = (0, 0, 0, 1) form a basis
for F4 . Find the unique representation of an arbitrary vector (a1 , a2 , a3 , a4 ) in F4 as a linear
combination of u1 , u2 , u3 , and u4 .
22
2
Y
x ck
k=0
k6=0
f1 (x) =
2
Y
x ck
k=0
k6=1
f2 (x) =
c0 ck
c1 ck
2
Y
x ck
k=0
k6=2
c2 ck
x+1
1
x+2
1
x+2
3
x1
3
x1
2
x+1
2
1 2
(x 1)
3
1
= (x2 + x 2)
2
1 2
(x + 3x + 2)
6
We then use these polynomials in a linear combination, with the coefficient of each fi
equal to bi :
g(x) =
2
X
bi fi (x)
i=0
1 2
1
1
(x 1) + 5 (x2 + x 2) + 3 (x2 + 3x + 2)
3
2
6
5
5
1 2 3
x + x+1
= 2x2 + 2 + x2 x + 5 +
2
2
2
2
= 6
= 4x2 x + 8
(c) (2, 3), (1, 6), (1, 0), (3, 2)
Solution. Using the notation from the book, we set c0 = 2, c1 = 1, c2 = 1, c3 = 3,
23
3
Y
x ck
k=0
k6=0
f1 (x) =
3
Y
x ck
k=0
k6=1
f2 (x) =
c1 ck
3
Y
x ck
k=0
k6=2
f3 (x) =
c0 ck
c2 ck
3
Y
x ck
k=0
k6=3
c3 ck
=
=
=
=
x+1
1
x+2
1
x+2
3
x+2
5
x1
3
x1
2
x+1
2
x+1
4
x3
5
x3
4
x3
2
x1
2
1 3
(x 3x2 x + 3)
15
1 3
(x 2x2 5x + 6)
8
1 3
(x 7x + 6)
12
1 3
(x + 2x2 x 2)
40
Note that there was no actual use in calculating f2 (x) since b2 = 0. We then use these
polynomials in a linear combination, with the coefficient of each fi equal to bi :
g(x) =
3
X
bi fi (x)
i=0
1
1
1 3
= 3 (x3 3x2 x + 3) 6 (x3 2x2 5x + 6) 2
(x + 2x2 x 2)
15
8
40
= x3 + 2x2 + 4x 5
11. Let u and v be distinct vectors of a vector space V. Show that if {u, v} is a basis for V and a
and b and nonzero scalars, then both {u + v, au} and {au, bv} are also bases for V.
Solution. Since {u, v} is a basis for V, each vector w V can be written uniquely as
w = a1 u + a2 v. Note
a2 (u + v) +
a1 a2
(au) = a2 u + a2 v + a1 u a2 u
a
= a1 + a2 v
=w
a1
a2
= (au) + (bv)
a
b
and therefore {u + v, au} and {au, bv} generate V. Now, since {u, v} is a basis for V, the
equation
a1 u + a2 v = 0
is satisfied only by a1 = a2 = 0. Therefore, the equation a1 (u + v) + a2 (au) = 0 can be
rewritten as (a1 + a2 a)u + a1 v = 0, whereby we obtain a1 = 0 and a1 + a2 a = 0 by the linear
independence of {u, v}. This last equation yields a2 a = 0, however a 6= 0 as stated by the
exercise, which implies that a2 = 0. This proves the linear independence of {u + v, au}. Next,
we examine {au, bv}, which gives the equation a1 (au) + a2 (bv) = 0. This can be rewritten as
24
(a1 a)u + (a2 b)v = 0. Again, a 6= 0 and b 6= 0, which gives a1 = a2 = 0, proving the linear
independence of {au, bv}. Therefore both sets are linearly independent and generate V, so
both are bases for V.
13. The set of solutions to the system of linear equations
x1 2x2 + x3 = 0
2x1 3x2 + x3 = 0
is a subspace of R3 . Find a basis for this subspace.
Solution. Solving this system, we obtain that our solution set is the set
S = {(x1 , x2 , x3 ) : x1 = x2 = x3 }
Therefore, a basis for this set could be {(1, 1, 1)}. The aforementioned set is linearly independent, since a1 (1, 1, 1) = (0, 0, 0) clearly implies a1 = 0. It can also be readily seen that any
vector in S can be expressed in the form a(1, 1, 1) where a R.
14. Find bases for the following subspaces of F5 :
W1 = {(a1 , a2 , a3 , a4 , a5 ) F5 : a1 a3 a4 = 0}
and
W2 = {(a1 , a2 , a3 , a4 , a5 ) F5 : a2 = a3 = a4 and a1 + a5 = 0}
What are the dimensions of W1 and W2 ?
Solution. For W1 , consider the set 1 = {(0, 1, 0, 0, 0), (0, 0, 0, 0, 1), (1, 0, 1, 0, 0), (1, 0, 0, 1, 0)}.
For W2 , consider the set 2 = {(0, 1, 1, 1, 0), (1, 0, 0, 0, 1)}. Let v1 W1 . Then v1 =
(a1 , a2 , a3 , a4 , a5 ) satisfying a1 a3 a4 = 0, that is, a1 = a3 + a4 . Now, note
a2 (0, 1, 0, 0, 0) + a5 (0, 0, 0, 0, 1) + a3 (1, 0, 1, 0, 0) + a4 (1, 0, 0, 1, 0) = (a3 + a4 , a2 , a3 , a4 , a5 )
= (a1 , a2 , a3 , a4 , a5 )
= v1
This proves that 1 spans W1 . Next, let v2 W2 . Then v2 = (b1 , b2 , b3 , b4 , b5 ) satisfying
b2 = b3 = b4 and a1 + a5 = 0. The latter equation implies a1 = a5 . Note
b2 (0, 1, 1, 1, 0) + b1 (1, 0, 0, 0, 1) = (b1 , b2 , b2 , b2 , b1 )
= (b1 , b2 , b3 , b4 , b5 )
= v2
This proves that 2 spans W2 . It is trivial to verify that both 1 and 2 are linearly independent. 1 and 2 therefore constitute bases for W1 and W2 , respectively. Furthermore,
because 1 contains 4 vectors and 2 contains 2 vectors, we conclude that W1 and W2 are of
dimension 4 and 2, respectively.
25
15. The set of all n n matrices having trace equal to zero is a subspace W of Mnn (F ) (see
Example 4 of Section 1.3). Find a basis for W. What is the dimension of W?
Solution. For each i, j satisfying 1 i n, 1 j n, we define S ij Mnn (F ) in the
following way:
(
1 if k = i and l = j
ij
Skl =
0 otherwise
Next, for each i satisfying 2 i n, define Di
1
i
Dkl =
1
Consider the set = {S ij : i 6= j} {Di : 2 i n}. We will prove this set is a basis for W.
First, we show is linearly independent. We need to show that the equation
n
X
ai Di +
i=2
bij S ij = 0
1in
1jn
i6=j
holds only when, for each i satisfying 2 i n, we have ai = 0, and for each i, j satisfying
1 i n, 1 j n, and i 6= j, we have bij = 0. First, note that the matrix 0 has 0 for all
of its entries. For each k satisfying 2 k n, we obtain
0=
n
X
i
ai Dkk
+
i=2
ij
bij Skk
=
n
X
i
ai Dkk
+
i=2
1in
1jn
i6=j
bij 0 =
n
X
i
ai Dkk
i=2
1in
1jn
i6=j
ij
To understand why this is, note that Skk
is only nonzero when k = i and k = j. However
ij
this can never be true in this case because the summation specifies i 6= j. Hence all the Skk
i
are zero. Furthermore note that for all i satisfying 2 i n, Dkk = 1 only when i = k. This
permits us to simplify the equation even more:
n
X
i
ai Dkk
= ak 1 +
i=2
ai 0 = ak
2in
i6=k
Therefore, ak = 0 (for 2 k n). We now wish to show that bij = 0 for all 1 i n,
1 j n, and i 6= j. For each k, l satisfying 1 k n, 1 l n, and k 6= l, we have
0=
n
X
i=2
i
ai Dkl
+
X
1in
1jn
i6=j
ij
bij Skl
=
n
X
i=2
26
ai 0 + bkl 1 +
X
1in
1jn
i6=j
()
bij 0 = bkl
where the (**) denotes the condition that the equations i = k and j = l do not hold simultai
neously. To understand why this is, note that Dkl
= 0 always (since k 6= l). Note also that
ij
kl
Skl = 0 unless i = k and j = l. We therefore pull out the unique term bkl Skl
from the
ij
kl
summation. Skl = 1 by definition, but all the other Skl = 0, and so we are left with bkl = 0
(for 1 k n, 1 l n, and k 6= l). Therefore we have shown that the only representation
of the zero matrix as a linear combination of the vectors within is the trivial one, proving
that is linearly independent.
Now we need to show that span() = W. Let M W and denote M s entry in row i, column
j by Mij . Then, since tr(M ) = 0, we have
M11 = M22 . . . Mnn
We will define the matrix M 0 by
M0 =
n
X
Mii Di +
i=2
Mij S ij
1in
1jn
i6=j
n
X
i
Mii D11
+
i=2
ij
Mij S11
=
n
X
i=2
1in
1jn
i6=j
Mii (1) +
Mij 0 =
n
X
Mii = M11
i=2
1in
1jn
i6=j
i
To make the above manipulations, we simply noted that D11
= 1 (for 2 i n), and
ij
that Skk = 0 (for 1 i n, 1 j n, and i 6= j). With the same kind of reasoning, for
2 k n, we obtain
0
Mkk
=
n
X
i
Mii Dkk
+
i=2
ij
Mij Skk
= Mkk 1 +
1in
1jn
i6=j
Mii 0 +
2in
i6=k
Mij 0 = Mkk
1in
1jn
i6=j
0
It now remains to verify that Mkl
= Mkl when 1 k n, 1 l n, and k 6= l. We obtain
0
Mkl
=
n
X
i=2
i
Mii Dkl
+
X
1in
1jn
i6=j
ij
Mij Skl
=
n
X
i=2
Mii 0 + Mkl 1 +
Mij 0 = Mkl
1in
1jn
i6=j
()
where the (**) denotes the condition that the equations i = k and j = l do not hold simultaneously. This completes our proof that the entries of M 0 are exactly those of M , and
so M 0 = M , which demonstrates W span(). Clearly we also have span() W, for
otherwise, noting that each vector in is also in W, W would not be closed under linear
combination, contradicting the fact that it is a subspace, which was given in the exercise.
Therefore span() = W. This, as well as the linear independence of , demonstrates that
27
is indeed a basis for W, completing the proof. There are n 1 vectors in {Di : 2 i n}
and n2 n vectors in {S ij : i 6= j}. Since is the union of these two sets (note that the
two sets are clearly disjoint as a result of the linear independence of ), contains exactly
(n 1) + (n2 n) = n2 1 vectors.
16. The set of all upper triangular n n matrices is a subspace W of Mnn (F ) (see Exercise 12
of Section 1.3). Find a basis for W. What is the dimension of W?
Solution. For each i, j satisfying 1 i n, 1 j n, we define
(
1 if k = i and l = j
ij
Skl =
0 otherwise
and form the set = {S ij : i j}. We will prove is a basis for W. To verify the linear
independence of , we have to show that the equation
X
aij S ij = 0
1in
1jn
ij
1in
1jn
ij
()
and thus akl = 0 for all 1 k n, 1 l n, and k l. This proves that is linearly
independent. We now demonstrate that generates W. Assume M W, that is, M is an
upper triangular matrix. We define the matrix M 0 Mnn (F ) by
X
M0 =
Mij S ij
1in
1jn
ij
1in
1jn
ij
()
where the (**) denotes the condition that the equations i = k and j = l do not hold siij
multaneously. This is true because Skl
is defined to be 0 unless i = k and j = l. We have
0
therefore demonstrated that M = M . This shows that W span(), and for the same
reasons as in exercise 15, that span() = W. Therefore is a basis for W. Since contains
n + (n 1) + (n 2) + . . . + (n (n 1)) = n2 (1 + 2 + . . . + (n 1)) = 12 (n2 + n) elements,
we conclude dim(W) = 12 (n2 + n).
28
17. The set of all skew-symmetric n n matrices is a subspace W of Mnn (F ) (see Exercise 28
of Section 1.3). Find a basis for W. What is the dimension of W?
Solution. A skew-symmetric matrix is a matrix A such that At = A. Since the diagonals
of a matrix are invariant under the matrix transpose operation, special attention must be
paid here, for in a field of characteristic 2, a = a does not imply a = 0, and every element is
its own additive inverse. The definition of skew-symmetry is equivalent to that of symmetry
for matrices with their entries in a field of characteristic 2 (A = A in such a field, so the
equation At = A becomes At = A which is precisely the definition of a symmetric matrix).
We will assume F s characteristic is not 2.
For each i, j satisfying 1 i n, 1 j n, we define
1 if k = i and l = j
ij
Skl
= 1 if k = j and l = i
0 otherwise
and form the set = {S ij : i < j}. The proof that is a basis is very similar to those given
in exercises 15 and 16, so I will not repeat it. It is trivial to see that dim(W) = 21 n(n 1).
19. Complete the proof of Theorem 1.8.
Solution. We wish to prove that if V is a vector space and = {u1 , u2 , . . . , un } is a subset
of V such that every v V can be uniquely expressed as a linear combination of vectors in ,
then is a basis for V.
Note first that is trivially a generating set for V, since we are told each v V is also in
span(). We know that if v = a1 u1 + . . . + an un = b1 u1 + . . . + bn un , then ai = bi for each
i such that 1 i n for otherwise each v would not be uniquely expressible as a linear
combination of vectors from . This implies that ai bi = 0 for each i in the equation,
proving that is linearly independent. is therefore a basis for W.
20. Let V be a vector space having dimension n, and let S be a subset of V that generates V.
(a) Prove that there is a subset of S that is a basis for V. (Be careful not to assume that S
is finite.)
Solution. Since V is finite-dimensional (with dimension n), then some set = {1 , . . . , n }
is a basis for V. Since V and S generates V, then for each i satisfying 1 i n,
there exists some positive integer ci such that we can write
i = ai1 si1 + . . . + aici sici
for some ai1 , . . . , aici F and si1 , . . . , sici S. For each i such that 1 i n, define
Si = {si1 , . . . , sici }
and let
=
n
[
i=1
29
Si
Clearly span() from above, and so by Theorem 1.5 we have span() = V span().
Clearly, span() V since V is a vector space, and therefore span() = V. Theorem 1.9
implies that there exists some 0 such that 0 is a basis for V. This set 0 is clearly
a subset of S, since itself is a subset of S.
(b) Prove that S contains at least n vectors.
Solution. Since 0 is a basis for V (which is of dimension n), 0 must contain n vectors,
by Corollary 1 to the Replacement Theorem. The fact that 0 S then gives us that S
contains at least n vectors.
1.7
Section 1.7 presents an alternative formulation of the concept of a basis. First, it defines the
concept of a chain and a maximal set (with respect to set inclusion). Then, a principle known
as the Maximal Principle (also known as the Hausdorff Maximal Principle) is presented, which is
equivalent to the Axiom of Choice. Maximal linearly independent subsets are defined as sets that
are linearly independent and are contained in no linearly independent subset with the exception of
themselves. The equivalence between maximal linearly independent subsets and bases is proved.
At the end of the section, it is shown that if S is a linearly independent subset of a vector space,
there exists a maximal linearly independent subset of that vector space containing S, giving the
obvious but important corollary that every vector space has a basis.
4. Let W be a subspace of a (not necessarily finite-dimensional) vector space V. Prove that any
basis for W is a subest of a basis for V.
Solution. Let be a basis for W. is clearly a linearly independent subset of W and hence
of V. Therefore by Theorem 1.12, V possesses a maximal linearly independent subset (basis)
containing .
2.1
In Section 2.1, linear transformations are introduced, as a special kind of function from one vector
space to another. Some properties of linear transformations are discussed. Two special subspaces
(of the domain and codomain, respectively), the nullspace and range, are presented, and their dimensions are called the nullity and rank, respectively. Furthermore, results are proven concerning
the injectivity and surjectivity of linear transformations satisfying certain criteria.
For Exercises 2 through 6, prove that T is a linear transformation, and find bases for both N(T)
and R(T). Then compute the nullity and rank of T, and verify the dimension theorem. Finally, use
the appropriate theorems in this section to determine whether T is one-to-one or onto.
2. T : R3 R2 defined by T(a1 , a2 , a3 ) = (a1 a2 , 2a3 ).
Solution. Let a = (a1 , a2 , a3 ) and b = (b1 , b2 , b3 ). Also let c R. Then T(ca + b) =
(ca1 +b1 ca2 b2 , 2ca3 +2b3 ) = (ca1 ca2 , 2ca3 )+(b1 b2 , 2b3 ) = c(a1 a2 , 2a3 )+(b1 b2 , 2b3 ) =
30
cT(a) + T(b). This proves that T is linear. Consider the set = {(1, 1, 0)}. Let v span().
Then v = (a, a, 0) for some a R, so that T(v) = T(a, a, 0) = (a a, 2 0) = (0, 0). This
proves v N(T) and hence span() N(T). We also clearly have N(T) span(), and
so spans the nullspace of T. is clearly linearly independent since it consists of a single
nonzero vector, so is a basis for Ts nullspace. Let 0 = {(1, 0, 0), (0, 1, 0), (0, 0, 1)}. By
Theorem 2.2, R3 = span(T( 0 )) = span({(1, 0), (1, 0), (0, 2)}) = span({(1, 0), (0, 1)}). Since
the set {(1, 0), (0, 1)} is linearly independent, it constitutes a basis for the range of T. Thus,
nullity(T) = 1 and rank(T) = 2. Therefore 1 + 2 = 3, verifying the dimension theorem. Since
nullity(T) 6= 0, we see that N(T) contains some nonzero vector. By Theorem 2.4, then, T
is not one-to-one. Note that rank(T) = dim(R2 ), proving by Theorem 1.11 that R(T) = W.
Therefore T is onto.
6. T : Mnn (F ) F defined by T(A) = tr(A). Recall (Example 4, Section 1.3) that
tr(A) =
n
X
Aii .
i=1
i=1
i=1
Now, assume T(cx + y) = cT(x) + T(y) for all x, y V and c F . Let x, y V. Then we
obtain T(x + y) = T(1x + y) = 1 T(x) + T(y) = T(x) + T(y). Next, let x V and c F .
Then we obtain T(cx) = T(cx + 0) = c T(x) + T(0) = c T(x). This proves T is linear. The
same type of reasoning can be used to show that if T satisfies
!
n
n
X
X
ai xi =
ai T(xi )
T
i=1
i=1
10. Suppose that T : R2 R2 is linear, T(1, 0) = (1, 4), and T(1, 1) = (2, 5). What is T(2, 3)? Is
T one-to-one?
Solution. By Theorem 2.6, T is defined by
T(ae1 + be2 ) = aw1 + bw2
So that we obtain w1 = (1, 4) and w2 = (2, 5) (1, 4) = (1, 1). Therefore we conclude
T(2, 3) = 2w1 + 3w2 = 2(1, 4) + 3(1, 1) = (2, 8) + (3, 3) = (5, 11). Now, consider x N(T).
Clearly, x = ae1 + be2 , so T(x) = aT(e1 ) + bT(e2 ) = aw1 + bw2 for some scalars a, b. However,
{w1 , w2 } is linearly independent, so that a = b = 0. This implies that N(T) = {0}, implying
T is one-to-one by Theorem 2.4.
11. Prove that there exists a linear transformation T : R2 R3 such that T(1, 1) = (1, 0, 2) and
T(2, 3) = (1, 1, 4). What is T(8, 11)?
Solution. Since {(1, 1), (2, 3)} is a basis for R2 , by Theorem 2.6, there exists a unique linear
transformation such that T(1, 1) = (1, 0, 2) and T(2, 3) = (1, 1, 4). This transformation is
defined by
T(a(1, 1) + b(2, 3)) = a(1, 0, 2) + b(1, 1, 4)
To determine T(8, 11), we write (8, 11) = 2(1, 1) + 3(2, 3). This gives T(8, 11) = 2(1, 0, 2) +
3(1, 1, 4) = (2, 0, 4) + (3, 3, 12) = (5, 3, 16).
12. Is there a linear transformation T : R3 R2 such that T(1, 0, 3) = (1, 1) and T(2, 0, 6) =
(2, 1)?
Solution. There clearly exists no such linear transformation. Assume T(1, 0, 3) = (1, 1).
Then by the linearity of T we must have T(2, 0, 6) = 2T(1, 0, 3) = 2(1, 1) = (2, 2) 6=
(2, 1).
13. Let V and W be vector spaces, let T : V W be linear, and let {w1 , w2 , . . . , wk } be a linearly
independent subset of R(T). Prove that if S = {v1 , v2 , . . . , vk } is chosen so that T(vi ) = wi
for i = 1, 2, . . . , k, then S is linearly independent.
Solution. Assume, for the sake of contradiction, that S is linearly dependent. Then there
exist a1 , . . . , ak F , not all zero, such that
a1 v1 + a2 v2 + . . . + ak vk = 0
Then by the linearity of T, such scalars also satisfy
a1 T(v1 ) + a2 T(v2 ) + . . . + ak T(vk ) = 0
Now recall that T(vi ) = wi for 1 i k. This implies that {w1 , w2 , . . . , wk } is a linearly
dependent subset of R(T), a contradiction. Therefore S must be linearly independent.
14. Let V and W be vector spaces and T : V W be linear.
(a) Prove that T is one-to-one if and only if T carries linearly independent subsets of V onto
linearly independent subsets of W.
32
Proof. Assume T(S) is linearly independent, and for the sake of contradiction, that
S is linearly dependent. Then there exist scalars a1 , . . . , an , not all zero, and vectors
s1 , . . . , sn S such that
a1 s1 + . . . + an sn = 0
However, T(0) = 0, so by linearity we would obtain that T(S) is linearly dependent, a
contradiction. Therefore S is linearly independent.
Solution. Let dim(V) < dim(W) and assume, for the sake of contradiction, that T is
onto. Let be a basis for W. Then contains dim(W) elements. Since T is onto,
= T(S) for some subset S V. However, is clearly linearly independent in W, so
by the lemma we see S is linearly independent in V. Therefore V contains a linearly
independent subset with dim(W) elements. Since dim(W) > dim(V), this contradicts
the replacement theorem. Therefore T cannot be onto.
(b) Prove that if dim(V) > dim(W), then T cannot be one-to-one.
Solution. If T is one-to-one, then T carries linearly independent subsets of V onto
linearly independent subsets of W. Let be a basis for V. Then contains dim(V)
elements, and is linearly independent. So T() will be a linearly independent subset of
W containing dim(V) elements, by (b) of Exercise 14. However, dim(W) < dim(V), so
we obtain the same contradiction to the replacement theorem as in (a). Thus, T cannot
be one-to-one.
2.2
Section 2.2 discusses a correspondence between linear transformations and matrices. It defines the
matrix representation of a linear transformation in such a way that the columns of said matrix can
be deduced simply by examining the image of each vector in the domains basis, represented as a
coordinate vector with respect to the basis of the codomain. The vector space of linear transformations from V to W, L(V, W), is also defined. We also see that sums and scalar multiples of linear
transformations have matrix representations as we would expect, i.e. the matrix representation of a
sum of transformations is the sum of the matrix representations of the individual transformations.
2. Let and be the standard ordered bases for Rn and Rm , respectively. For each linear
transformation T : Rn Rm , compute [T] .
(a) T : R2 R3 defined by T(a1 , a2 ) = (2a1 a2 , 3a1 + 4a2 , a1 ).
Solution. Recall that the j th column of [T] is the column matrix [T(j )] . Since
were working with the standard ordered bases here, this simplifies to [T(ej )] . Notice
T(e1 ) = T(1, 0) = (2 0, 3 + 0, 1) = (2, 3, 1) = 2e1 + 3e2 + 1e3 , and T(e2 ) = T(0, 1) =
(0 1, 0 + 4, 0) = (1, 4, 0) = 1e1 + 4e2 + 0e3 . Therefore
2 1
4
[T] = 3
1
0
(c) T : R3 R defined by T(a1 , a2 , a3 ) = 2a1 + a2 3a3 .
34
3 1
1
[T] = 0
2
0
3
Next, we see that T(1, 2) = (1, 1, 4) = 73 (1, 1, 0) + 2(0, 1, 1) + 23 (2, 2, 3) and T(2, 3) =
4
(1, 2, 7) = 11
3 (1, 1, 0) + 3(0, 1, 1) + 3 (2, 2, 3). Therefore, the coordinate vectors for T(1, 2)
7
2
11
and T(2, 3) are ( 3 , 2, 3 ) and ( 3 , 3, 43 ) respectively. So we obtain
7
3 11
3
3
[T] = 2
2
3
4
3
Note that in both cases, a system of linear equations had to be solved to express each image
vector as a combination of the vectors in the codomains basis, although this was not shown.
8. Let V be an n-dimensional vector space with an ordered basis . Define T : V Fn by
T(x) = [x] . Prove that T is linear.
Solution. Let x, y V and c F . Assume that = {v1 , . . . , vn } and that
x=
n
X
ai vi
and
i=1
y=
n
X
bi vi
i=1
ca1 + b1
a1
b1
ca2 + b2
a2 b2
T(cx + y) = [cx + y] =
= c .. + .. = c[x] + [y] = cT(x) + T(y)
..
. .
.
can + bn
an
bn
2.3
Section 2.3 uses composition of linear transformations to motivate the usual definition of matrix
multiplication. Some useful theorems are proved about compositions of transformations and matrix
multiplication. An important relationship between matrices and linear transformations is demonstrated through the so-called left-multiplication transformation, LA .
13. Let A and B be n n matrices. Recall that the trace of A is defined by
tr(A) =
n
X
Aii
i=1
n
X
(BA)ii =
i=1
n
n
X
X
i=1
!
Bik Aki
k=1
n
n
X
X
!
Aki Bik
i=1
k=1
n
X
(AB)kk = tr(AB)
k=1
which proves the first proposition. To prove the second, simply note
tr(At ) =
n
X
(At )ii =
i=1
n
X
Aii = tr(A).
i=1
15. Let M and A be matrices for which the product matrix M A is defined. If the jth column of A
is a linear combination of a set of columns of A, prove that the jth column of M A is a linear
combination of the corresponding columns of M A with the same corresponding coefficients.
Solution. Assume M is an mn matrix, and A is an np matrix. Then for each k satisfying
1 k p and k 6= j, there exist ak , with
Aij =
p
X
ak Aik
k=1
k6=j
p
p
n
n
X
X
X
X
ak
Miz
ak Azk =
(M A)ij =
Miz Azj =
z=1
z=1
k=1
k6=j
k=1
k6=j
n
X
!
Miz Azk
z=1
p
X
ak (M A)ik
k=1
k6=j
z=1
y=1
y=1
36
z=1