Professional Documents
Culture Documents
Vector Spaces
3.1 Vectors in Rn
3.2 Vector Spaces
3.3 Subspaces of Vector Spaces
3.4 Spanning Sets & Linear Independence
3.5 Basis and Dimension
3.6 Rank of a matrix
3.7 Inner Product Spaces
3.8 Linear Transformation
n
3.1 Vectors in R
◼ An ordered n-tuple :
a sequence of n real numbers ( x1 , x2 , , xn )
n
▪ R -space :
the set of all ordered n-tuples
n=1 R1-space = set of all real numbers
(R1-space can be represented geometrically by the x-axis)
3.2
◼ Notes:
(1) An n-tuple ( x1 , x2 , , xn ) can be viewed as a point in Rn
with the xi’s as its coordinates
(2) An n-tuple ( x1 , x2 , , xn ) also can be viewed as a vector
x = ( x1 , x2 , , xn ) in Rn with the xi’s as its components
▪ Ex:
(x1 , x2 ) (x1 , x2 )
or
(0,0)
a point a vector
※ A vector on the plane is expressed geometrically by a directed line segment
whose initial point is the origin and whose terminal point is the point (x1, x2)
3.3
u = ( u1 , u2 , , un ) , v = ( v1 , v2 , , vn ) (two vectors in Rn)
▪ Equality:
u = v if and only if u1 = v1 , u2 = v2 , , u n = vn
▪ Notes:
The sum of two vectors and the scalar multiple of a vector
in Rn are called the standard operations in Rn
3.4
▪ Difference between u and v:
u − v u + (−1) v = (u1 − v1 , u2 − v2 , u3 − v3 ,..., un − vn )
▪ Zero vector:
0 = (0, 0, ..., 0)
3.5
▪ Notes:
A vector u = (u1 , u2 , , un ) in R n can be viewed as:
Use comma to separate components
3.8
◼ Ex 5: Practice standard vector operations in R4
Let u = (2, –1, 5, 0), v = (4, 3, 1, –1), and w = (–6, 2, 0, 3) be
vectors in R4. Solve x in each of the following cases.
(a) x = 2u – (v + 3w)
(b) 3(x+w) = 2u – v+x
Sol: (a) x = 2u − ( v + 3w )
= 2u + (−1)( v + 3w )
= 2u − v − 3w (distributive property of scalar multiplication over vector addition)
= (4, − 2, 10, 0) − (4, 3, 1, − 1) − ( −18, 6, 0, 9)
= (4 − 4 + 18, − 2 − 3 − 6, 10 − 1 − 0, 0 + 1 − 9)
= (18, − 11, 9, − 8)
3.9
(b) 3(x + w ) = 2u − v + x
3x + 3w = 2u − v + x (distributive property of scalar multiplication over vector addition)
3.10
▪ Notes:
(1) The zero vector 0 in Rn is called the additive identity in Rn (see
Property 4)
(2) The vector –u is called the additive inverse of u (see
Property 5)
▪ Ex 6:
Given x = (–1, –2, –2), u = (0,1,4), v = (– 1,1,2), and
w = (3,1,2) in R3, find a, b, and c such that x = au+bv+cw.
Sol:
− b + 3c = − 1
a + b + c = −2
4a + 2b + 2c = − 2
a = 1, b = −2, c = −1
Thus x = u − 2 v − w 3.12
3.2 Vector Spaces
V ≠ ∅: Addition 𝐮 + 𝐯 & Scalar multiplication 𝑐𝐮
If these 10 axioms are satisfied, then V is called a vector space, and
the elements in V are called vectors
(1) 𝐮 + 𝐯 ∈ 𝑉, ∀𝐮, 𝐯 ∈ 𝑉 (closure under vector addition)
3.14
◼ Examples. (It is straightforward to show that these vector spaces satisfy the
above ten axioms)
n
(1) n-tuple space: R
(u1 , u2 ,un ) + (v1 , v2 , v2 ) = (u1 + v1 , u2 + v2 ,un + vn ) (standard vector addition)
k (u1 , u2 ,un ) = (ku1 , ku2 , kun ) (standard scalar multiplication for vectors)
※ By the fact that the sum of two continuous function is continuous and the
product of a scalar and a continuous function is still a continuous
function, C (−, ) is a vector space
3.16
▪ Summary of important vector spaces
R = set of all real numbers
R 2 = set of all ordered pairs
R 3 = set of all ordered triples
R n = set of all n-tuples
C (−, ) = set of all continuous functions defined on the real number line
C[a, b] = set of all continuous functions defined on a closed interval [a, b]
P = set of all polynomials
Pn = set of all polynomials of degree n
M m ,n = set of m n matrices
M n ,n = set of n n square matrices
3.17
◼ Notes: To show that a set is not a vector space, you need
only find one axiom that is not satisfied
◼ Ex 6: The set of all integers is not a vector space
Pf: 1 V , and 12 is a real-number scalar
( 12 )(1) = 12 V (it is not closed under scalar multiplication)
noninteger
scalar
integer
Pf:
(8) (9) (5)
(1) 0v = (c + (−c)) v = cv + (−c) v = cv + (−(cv)) = 0
3.20
(4) (7)
(2) c0 = c(0 + 0) = c0 + c0
c0 + (−c0) = (c0 + c0) + (−c0) (add (–c0) to both sides)
(3)
c0 + ( −c0) = c0 + [c0 + ( −c0)]
(5) (4)
0 = c0 + 0 0 = c0
(3) Prove by contradiction: Suppose that cv = 0, but c 0 and v 0
1 (9) 1
(10)
1
v = 1v = c v = ( cv ) = ( 0 ) = 0 (By the second property, c0 = 0)
c c c
→ if cv =0, either c = 0 or v = 0
(8)
(4) 0 v = (1 + (−1)) v = 1v + (−1) v
0 = v + (−1) v (By the first property, 0v = 0)
(5)
(−1) v = − v (By comparing with Axiom (5), (–1)v is the additive inverse of v)
3.21
Exercise 1
Let (d) be a straight line in xy-plane. Find the condition such
that (d) together with the usual + and scalar multiplication
form a vector space
y (d’)
x
O
(d)
3.22
Exercise 2
Decide whether W, equipped with the usual addition “+” and
usual scalar multiplication “.” on R3, is a vector space
(a) W = ( x1 , x2 , 2 x1 + 3 x2 ) x1 , x2
(b) W = ( x1 , x2 ,3 x1 − 4 x2 + 1) x1 , x2
3.23
3.3 Subspaces of Vector Spaces
◼ Subspace:
(V , +, ) : a vector space
W
: a nonempty subset of V
W V
(W , +, ) : The nonempty subset W is called a subspace if W is
a vector space under the operations of vector
addition and scalar multiplication defined on V
▪ Trivial subspace :
Every vector space V has at least two subspaces
(1) Zero vector space {0} is a subspace of V (It satisfies the ten axioms)
(2) V is a subspace of V
※ Any subspaces other than these two are called proper (or nontrivial) subspaces
3.24
Test a subset to be a subspace
Theorem 3.4:
A nonempty subset W of a vector space V is a subspace of V if and
only if W is closed under addition and scalar multiplication, that is:
3.25
Pf:
1. Note that if u, v, and w are in W, then they are also in V.
Furthermore, W and V share the same operations.
Consequently, vector space axioms 2, 3, 7, 8, 9, and 10 are
satisfied automatically
2. Suppose that the closure conditions hold in Theorem 3.4, i.e.,
the axioms 1 and 6 for vector spaces are satisfied
3. Since the axiom 6 is satisfied (i.e., cu is in W if u is in W), we
can obtain
3.1. for a scalar c = 0, cu = 0 W zero vector in W
axiom 4 is satisfied
3.2. for a scalar c = −1, ( − 1)u W − u ( − 1)u
st. u+( − u) = u+( − 1)u = 0
axiom 5 is satisfied
3.26
▪ Ex 2: A subspace of M2×2
Let W be the set of all 2×2 symmetric matrices. Show that
W is a subspace of the vector space M2×2, with the
standard
Sol: operations of matrix addition and scalar multiplication
First, we knon that W , the set of all 2 2 symmetric matrices,
is an nonempty subset of the vector space M 22
Second,
A1 W, A2 W ( A1 + A2 )T = A1T + A2T = A1 + A2 (A1 + A2 W )
c R, A W (cA)T = cAT = cA (cA W )
The definition of a symmetric matrix A is that AT = A
Sol:
1 0 0 0
A= W , B = W
0 0 0 1
1 0
A+ B = = I W (W is not closed under vector addition)
0 1
W is not a subspace of M 22
3.28
▪ Ex 4: The set of first-quadrant vectors is not a subspace of R2
Show that W = {( x1 , x2 ) : x1 0 and x2 0} , with the standard
operations, is not a subspace of R2
Sol:
Let u = (1, 1) W
(− 1)u = (− 1)(1, 1) = (− 1, − 1) W
(W is not closed under scalar multiplication)
W is not a subspace of R 2
3.29
▪ Ex 5: Identify subspaces of R3
Which of the following subsets is a subspace of R 3?
(a) W = ( x1 , x2 ,1) x1 , x2 R (Note: the zero vector is not in W)
(b) W = ( x1 , x1 + x3 , x3 ) x1 , x3 R (Note: the zero vector is in W)
Sol:
(a)
3.30
(b)
3.32
▪ Theorem 3.5: The intersection of two subspaces is a subspace
If V and W are both subspaces of a vector space U ,
then the intersection of V and W (denoted by V W )
is also a subspace of U
Pf:
(1) For v1 and v 2 in V W , since v1 and v 2 are in V (and W ),
v1 + v 2 is in V (and W ). Therefore, v1 + v 2 is in V W
(2) For v1 in V W , since v1 is in V (and W ),
cv1 is in V (and W ). Therefore, cv1 is in V W
▪ Linear combination :
A vector u in a vector space V is called a linear combination of
the vectors v1 , v 2 , , v k in V if u can be written in the form
u = c1 v1 + c2 v 2 + + ck v k ,
3.35
▪ Ex 1: Finding a linear combination
v1 = (1,2,3) v 2 = (0,1,2) v 3 = ( − 1,0,1)
Prove (a) w = (1,1,1) is a linear combination of v1 , v 2 , v 3
(b) w = (1, − 2,2) is not a linear combination of v1 , v 2 , v 3
Sol:
(a) w = c1 v1 + c2 v 2 + c3 v 3
c1 − c3 =1
2c1 + c2 =1
3c1 + 2c2 + c3 =1
3.36
1 0 − 1 1 1 0 − 1 1
2 1 0 1
⎯⎯⎯⎯⎯⎯
Gauss Elimination
→0 1 2 − 1
3 2 1 1 0 0 0 0
c1 = 1 + t , c2 = −1 − 2t , c3 = t
3.37
(b)
w = c1 v1 + c2 v 2 + c3 v 3
1 0 − 1 1 1 0 − 1 1
2 1 0 − 2 ⎯⎯⎯
G.-J. E.
→ 0 1 2 − 4
3 2 1 2 0 0 0 7
3.38
▪ The span of a set: span(S)
If S = {v1, v2,…, vk} is a set of vectors in a vector space V,
then the span of S is the set of all linear combinations of
the vectors in S,
span(S ) = c1v1 + c2 v 2 + + ck v k ci R
(the set of all linear combinations of vectors in S )
3.39
▪ Notes: The above statement can be expressed as follows
span( S ) = V
S spans (generates) V
V is spanned (generated) by S
S is a spanning set of V
▪ Ex 2:
(a) The set S = {(1, 0, 0), (0,1, 0), (0, 0,1)} spans R3 because any vector
u = (u1 , u2 , u3 ) in R 3 can be written as
u = u1 (1, 0, 0) + u2 (0,1, 0) + u3 (0, 0,1)
span( S ) = R 3
3.42
▪ Theorem 3.6: span(S) is a subspace of V
If S = {v1, v2,…, vk} is a set of vectors in a vector space V, then
(a) span(S) is a subspace of V
(b) span(S) is the smallest subspace of V that contains S, i.e.,
every other subspace of V containing S must contain span(S)
any subspace
containing S
v1
vk
v3 v2
span(S)
V
3.43
Pf:
(a)
Consider any two vectors u and v in span(S ), that is,
u =c1 v1 + c2 v 2 + + ck v k and v =d1 v1 + d 2 v 2 + + d k v k
3.44
(b)
Let U be another subspace of V that contains S , and we want to
show span(S ) U
k
Consider any u span(S ), i.e., u = ci v i , where v i S
i =1
(because U is closed under
U is a subspace k vector addition and scalar
U contains S v i U u = ci v i U multiplication, and any linear
combination can be evaluated
i =1 with finite vector additions
and scalar multiplications)
3.45
Exercises
and r ( x) = 2 x + 3 in P2
3.46
Linear Independence and Linear Dependence
Definition
S = v1 , v 2 , , v k : a set of vectors in a vector space V
For c1 v1 + c2 v 2 + + ck v k = 0
3.47
▪ Ex 4: Testing for linear independence
Determine whether the following set of vectors in R3 is L.I. or L.D.
S = v1 , v 2 , v 3 = (1, 2, 3) , ( 0, 1, 2 ) , ( −2, 0, 1)
Sol: c1 − 2c3 = 0
c1 v1 + c2 v 2 + c3 v 3 = 0 2c1 + c2 + =0
3c1 + 2c2 + c3 = 0
1 0 − 2 0 1 0 0 0
2 1 0 0 ⎯⎯⎯ G.-J. E.
→ 0 1 0 0
3 2 1 0 0 0 1 0
c1 = c2 = c3 = 0 (only the trivial solution)
(or det(A) = −1 0, so there is only the trivial solution)
S is (or v1 , v 2 , v 3 are) linearly independent 3.48
◼ Ex 5: Testing for linear independence
Determine whether the following set of vectors in P2 is L.I. or L.D.
S = v1 , v 2 , v 3 = 1 + x − 2 x 2 ,2 + 5 x − x 2 ,x + x 2
Sol:
c1v1+c2v2+c3v3 = 0
i.e., c1(1+x – 2x2) + c2(2+5x – x2) + c3(x+x2) = 0+0x+0x2
c1+2c2 =0 1 2 0 0 1 2 0 0
1 5 1 0 ⎯⎯⎯
c1+5c2+c3 = 0
G. E.
→ 0 1 3 0
1
–2c1– c2+c3 = 0 −2 −1 1 0
0 0 0 0
This system has infinitely many solutions
(i.e., this system has nontrivial solutions, e.g., c1=2, c2= – 1, c3=3)
S is (or v1, v2, v3 are) linearly dependent
3.49
◼ Ex 6: Testing for linear independence
Determine whether the following set of vectors in the 2×2
matrix space is L.I. or L.D.
2 1 3 0 1 0
S = v1 , v 2 , v 3 = , ,
0 1 2 1 2 0
Sol:
c1v1+c2v2+c3v3 = 0
2 1 3 0 1 0 0 0
c1 + c2 + c3 =
0 1 2 1 2 0 0 0
3.50
2c1+3c2+ c3 = 0
c1 =0
2c2+2c3 = 0
c1 + c 2 =0
2 3 1 0 1 0 0 0
1 0 0 0 G.-J. E. 0 1 0 0
⎯⎯⎯→
0 2 2 0 0 0 1 0
1 1 0 0 0 0 0 0
S is linearly independent
3.51
◼ Theorem 3.7: A property of linearly dependent sets
A set S = {v1,v2,…,vk}, for k 2, is linearly dependent if
and only if at least one of the vectors vi in S can be written
as a linear combination of the other vectors in S
Pf:
()
c1v1+c2v2+…+civi+…+ckvk = 0
ci 0 for some i
c1 ci −1 ci +1 ck
v i = − v1 + +− v i −1 + − v i +1 + +− vk
ci ci ci ci 3.52
()
Let vi = d1v1+…+di-1vi-1+di+1vi+1+…+dkvk
d1v1+…+di-1vi-1+(–1)vi +di+1vi+1+…+dkvk = 0
c1 = d1 , c2 = d2 ,…, ci = –1 ,…, ck = dk (there exits at least this
nontrivial solution)
S is linearly dependent
3.53
Exercises
Let S={A, B, C, D} in V=M2x2
0 −3 0 2
A= , B= ,
2 −1 −5 1
1 0 0 1
C= ,D = in M 22
0 −3 3 6
3.54
Exercises
3.55
3.4 Basis and Dimension
◼ Basis: Linearly
Spanning Bases
V: a vector space Sets
Independent
Sets
S = {v1, v2, …, vn}V
V
(1) S spans V (i.e., span(S) = V)
(2) S is linearly independent
3.56
▪ Notes:
(1) the standard basis for R3:
{i, j, k}, for i = (1, 0, 0), j = (0, 1, 0), k = (0, 0, 1)
3.57
(3) the standard basis for mn matrix space:
aij = 1
{ Eij | 1im , 1jn }, and in Eij
other entries are zero
Ex: 2 2 matrix space:
1 0 0 1 0 0 0 0
, , ,
0 0 0 0 1 0 0 1
3.58
▪ Ex 2: The nonstandard basis for R2
Show that S = {v1 ,v 2 } = {(1,1), (1, −1)} is a basis for R2
c1 + c2 = u1
(1) For any u = (u1 , u2 ) R , c1 v1 + c2 v 2 = u
2
c1 − c2 = u2
c1 + c2 = 0
(2) For c1 v1 + c2 v 2 = 0
c1 − c2 = 0
3.63
◼ Dimension:
The dimension of a vector space V is defined to be the
number of vectors in a basis for V
V: a vector space S: a basis for V
◼ Finite dimensional:
A vector space V is finite dimensional if it has a basis
consisting of a finite number of elements
◼ Infinite dimensional:
If a vector space V is not finite dimensional, then it is called
infinite dimensional
3.64
▪ Notes: dim(V) = n
Linearly
(1) dim({0}) = 0 Spanning
Bases Independent
Sets Sets
S: a basis #(S) = n
3.65
◼ Ex 3: Find the dimension of a vector space according to the
standard basis
dim(V) = n
Linearly
Spanning
Bases Independent
Sets
Sets
#(S) n #(S) n
#(S) = n 3.69
Exercises
Determine whether the vectors form a basis for the given vector
space
3.70
3.6 Rank of a Matrix
◼ Definition. The rank of a matrix A, denoted by rk(A), is the
maximum number of linearly independent row vectors of A.
◼ We say that a matrix A is row-equivalent to a matrix B if A
can be obtained from B by (finitely many) elementary row
operations
3.71
Rank of a Matrix
◼ The maximum number of linearly independent row vectors of
a matrix does not change if we change the order of rows or
multiply a row by a nonzero constant c, or take a linear
combination by adding a multiple of a row to another row.
This proves that rank is invariant under elementary row
operations
3.72
Rank of a Matrix
3 0 2 2
A = −6 42 24 54
21 −21 0 15
3.73
Example
◼ Determine the rank of the matrix
3 0 2 2 3 0 2 2
R 2+ 2 R1
A = −6 42 24 54 ⎯⎯⎯ ⎯
R 3− 7 R1
→ 0 42 28 58
21 −21 0 15 0 −21 −14 1
3 0 2 2 Row vectors
R 3+ R 2/2 containing
⎯⎯⎯⎯R 2/2
→ 0 21 14 29 = U pivots are
0 0 0 30 linearly
independent
The 3 row vectors of U are linearly independent.
So, rk(U)=3=rk(A)
3.74
Example 2
◼ Determine rk(A), where
1 1 0 2
−2 −2 1 −5
A=
1 1 −1 3
4 4 −1 9
Solution:
1 1 0 2 1 1 0 2
−2 −2 1 −5 0 0 1 −1
A=
1 1 −1 3 →⋯→ U =
0 0 0 0
4 4 −1 9
0 0 0 0
rk(A)=2
3.75
A Criterion for Linearly Independent Vectors
3.76
Rank in terms of Column Vectors
3.77
Rank in terms of Column Vectors
◼ Proof…
3.79
Proof…
◼ We collect components in columns:
a1k c11 c12 c1r
a c c c
2k = v 21 + v 22 + + vrk 2 r
1k 2k
amk cm1 cm 2 cmr
◼ where k=1,…,n. The vector on left is the kth column of A.
◼ Each of these n columns is a linear combination of the same r
columns on the right
◼ Hence A cannot have more linearly independent columns
than rows, whose number is rank(A)=r
3.80
Proof…
◼ Rows of A are columns of AT
◼ For AT our conclusion is that AT cannot have more linearly
independent columns than rows
◼ So that A cannot have more linearly independent rows than
columns
◼ Together, the number of linearly independent columns of A
=rank(A)=r
3.81
Exercises
2. Determine rk(A)
2 2 −1 3
0 6 5 11
A=
1 −2 −3 −4
3 3 2 2
3.82
Null space
Let A be an m x n matrix and consider the
linear system Ax = 0 where x Rn. Define
W = {x Rn | Ax = 0 }.
W is a subspace of Rn called the null space of A, denoted by
NS(A).
Let x and y be solutions of the linear system,
i.e. Ax = 0 and Ay = 0.
Then A( x + y ) = Ax + Ay = 0 + 0 = 0.
So, x + y W. Let c be a scalar, then
A(c x) = c ( Ax ) = c 0 = 0. So, c x W and W is a subspace
of Rn. The null space is also called the kernel of A
3.83
Example
◼ Find a basis and dimension of NS(A), where
1 1 0 2
−2 −2 1 −5
A=
1 1 −1 3
Solution: 4 4 −1 9
1 1 0 2 0 1 1 0 2 0
−2 −2 1 −5 0 0 0 1 −1 0
A= → ... →
1 1 −1 3 0 0 0 0 0 0
4 4 −1 −9 0 0 0 0 0 0
3.84
1 1 0 2 0 x1 + x2 + 2 x4 = 0
0 0 1 −1 0 x3 − x4 = 0
0 0 0 0 0
Set x4 = s
0 0 0 0 0
→ x3 = s
Set x2 = r
Solution of Ax=0: → x1 = − r − 2s
−r − 2s −r −2s −1 −2
r r 0 = r + s 0
1
x= = +
s 0 s 0 1
s 0 s
0 1
−1 −2
1 0
u = , v = span NS ( A)
0 1
0 1
3.85
u , v are linearly independent, since
au + bv = 0
(− a − 2b, a, b, b)T = (0,0,0,0)T
a=b=0
3.86
Row Space and Column Space
◼ Given an mxn matrix A. We call the span of the row vectors
the row space of A, denoted by RS(A)
◼ and the span of the column vectors the column space of A,
denoted by CS(A)
3.87
Important Remarks
1. RS(U) = span{nonzero vectors of U}, where U is a matrix
in row echelon form
2. If B is obtained from A by an elementary row operation,
then RS(B)=RS(A)
3. If A is reduced to a row echelon form matrix U by using
elementary row operations, then RS(A)=RS(U)
4. Thus, to find RS(A), we reduce it to U, and
RS(A)=span{nonzero vectors of U}, where U is a matrix
in row echelon form
3.88
Example
◼ Determine rk(A), find a basis for RS(A), where
1 1 0 2
−2 −2 1 −5
A=
1 1 −1 3
4 4 −1 9
Solution:
1 1 0 2 1 1 0 2
−2 −2 1 −5 0 0 1 −1
A=
1 1 −1 3 U =
0 0 0 0
4 4 −1 9
0 0 0 0
rk(A)=2
3.89
Quiz
1. Determine rk(A), find a basis for RS(A), a basis for
CS(A), a basis for NS(A), and verify
dim(RS(A))+dim(NS(A))=n
1 2 0 −2 3 1
−3 −2 −4 5 3 −3
A= ,
−4 −2 3 6 5 2
−3 0 3 4 8 3
CS(A)=RS(AT)
3.90
Exercises
1. Show that Ax=b has a solution if and only if b is in CS(A)
2. Determine if b lies in the column space of the given
matrix. If it does, express b as a linear combination of the
columns
1 −3 2
A= , b =
2 5 15
3.91
3.7. Inner Product Spaces
◼ Inner product of two vectors in n-space
I. Linearity: , , u, v, w n
III. Positive-definiteness: u n
u.u 0, and u.u = 0 if and only if u = 0
3.92
Inner Product
◼ A vector space V is called an inner product space
if we can define an inner product on V. That is, any
pair u, v in V, is associated with a real number,
denoted by (u, v), called the inner product of u and v
such that
I. Linearity: , , u, v, w V
(u + v, w) = (u, w) + (v, w),
II. Symmetry: (u, v) = (v, u ), u, v V
III. Positive-definiteness: u V
(u, u ) 0, and (u, u ) = 0 if and only if u = 0
3.93
Inner Product Spaces
◼ Vectors whose inner product is zero are called orthogonal.
That is (u,v)=0, then u is orthogonal to v
◼ The length or norm of a vector u in V is defined by
|| u ||= (u, u )
◼ A vector of norm 1 is called a unit vector
3.94
Properties of norms
3.96
3.8 Linear Transformations: Concept of functions
◼ A function f consists of sets D and R, and a rule f that assigns
to each element x of D a unique element y=f(x) of R
◼ The set D is called the domain of f
◼ We write “f: D→ R”
3.97
Matrix induces a function
◼ If A is an mxn matrix, its induced function
fA : n
→ m
, defined by f A ( x) = Ax m
f I ( x) = Ix = x, so f I : n
→ n
is just the identity function
3.98
Example 2
◼ Let
−1 0
A= . Then
0 2
x −1 0 x − x
fA = =
y 0 2 y 2 y
−x x 0 0
f A (u ) = , u = f A (v ) = , v =
0 0 2y y
f A reflects the x-axis f A expands the y-axis
3.99
Example 2
◼ Let
−1 0
A= . Then
0 2
x −1 0 x − x
fA = =
y 0 2 y 2 y
−x x
f A ( w) = , w =
2y y
f A combines the two cases
for a general vector
3.100
Linear Transformations
◼ Definition
3.101
Matrix induced by linear transformation
Let T : n
→ m
be a linear transformation.
Then there exists an m n matrix A such that
T = fA
We call this A the matrix induced by T and
denote it by A = [T ]
◼ A is the matrix whose columns are just the image of elements
of the standard basis
3.102
Quiz
◼ Determine whether the given formula defines an inner
product
1) u = (a1 , a2 , a3 ), v = (b1 , b2 , b3 ) 3
3.104