Professional Documents
Culture Documents
2.1 Vectors
To begin with, we appeal to the intuitive sense of a vector, i.e., as a magnitude
and direction, as in the two-dimensional case shown below.
w
Figure 2.1 The concept of a vector with magnitude and direction. Such vectors can
be easily pictured in two and three dimensions only.
In Figure 2.1, vectors v and w are depicted as arrows, the length of which indicates
their magnitude, with directions implied by the existence of the reference arrows
lying horizontally and vertically. It is clear that the two vectors have an angle
between them (), but in order to uniquely fix them in “space,” they should also
be oriented with respect to these references. The discussion of these reference
directions is as important as the vectors themselves.
dot(x,y) Inner (dot) productM: The inner product of two vectors, v and
w, is denoted as v , w . Although there may exist many
definitions for the computation of an inner product, when
vectors are interpreted as columns of n real numbers, the inner
product is customarily computed as:
n
v , w v T w w T v v i wi (2.1)
i 1
Chapter 2. Vectors and Vector Spaces 47
v, w v w cos (2.2)
1 n
v v, v 2 v i2 (2.3)
i 1
v w vw wv (2.4)
2.2.1 Fields
First, we cannot introduce the idea of a vector space without first giving the
definition of a field. Conceptually, a field is the set from which we select elements
that we call scalars. As the name implies, scalars will be used to scale vectors.
Field: A field consists of a set of two or more elements, or
members, which must include the following:
1. There must exist in the field a unique element called 0
(zero). For 0 F and any other element a F , 0( a ) 0
and a 0 a .
2. There must exist in the field another unique element called
1 (one). For 1 F and a F , 1( a ) a (1) ( a / 1) a .
3. For every a F , there is a unique element called its
negative, a F , such that a ( a ) 0 .
Solutions:
1. No. This cannot be a field, because 1 1 2 , and the element 2
is not a member of the set.
2. Yes. This set has the identity matrix and the zero matrix, and
inverse matrices of the set have the same form. However, this set
has the special property of being commutative under
multiplication, which is not true of matrices in general. Therefore,
the set of all 2 2 matrices is not a field.
3. No. The inverse of a polynomial is not usually a polynomial.
4. Yes. This is the most common field that we encounter.
5. No. Like polynomials, the inverse of an integer is not usually an
integer.
Now we can proceed to define vector spaces. The definition of a vector space
is dependent on the field over which we specify it. We therefore must refer to
vector spaces over fields.
* Such arithmetic operations need not generally follow our familiar usage. For instance,
in Example 1, the answer will vary if the set is interpreted as binary digits with binary
operators rather than the two real numbers 0 and 1.
50 Part I. Mathematical Introduction to State Space
( a b) x ax bx
(2.7)
a( x y ) ax ay
1. The set of all n-tuples of scalars from any field F, defined over F.
For example, the set of n-tuples n over the field of reals , or
the set of complex n-tuples C n over the field of complex numbers
C.
2. The set of complex numbers over the reals.
Chapter 2. Vectors and Vector Spaces 51
Solutions:
1. Yes. These are the so-called “euclidean spaces,” which are most
familiar. The intuitive sense of vectors as necessary for simple
mechanics fits within this category.
2. Yes. Note that a complex number, when multiplied by a real
number, is still a complex number.
3. No. A real number, when multiplied by a complex number, is not
generally real. This violates the condition of closure under scalar
multiplication.
4. Yes in both cases. Remember that such matrices are not generally
fields, as they have no unique inverse, but they do, as a collection,
form a space.
5. Yes.
6. Yes. Again, such a set would not form a field, but it does form a
vector space. There is no requirement for division in a vector space.
7. Yes. The sum of symmetric matrices is again a symmetric matrix.
8. No. One can easily find two nonsingular matrices that add to form
a singular matrix.
9. Yes. This is an important property of solutions to differential
equations.
10. No. If one were to add the particular solution of a nonhomogeneous
differential equation to itself, the result would not be a solution to
the same differential equation, and thus the set would not be closed
under addition. Such a set is not closed under scalar multiplication
either.
52 Part I. Mathematical Introduction to State Space
With the above definitions of linear vector spaces and their vectors, we are
ready for some definitions that relate vectors to each other. An important concept
is the independence of vectors.
n
a1 x1 a 2 x 2 a n x n a i x i 0 (2.8)
i 1
l q
If we consider the vectors x i to be columns of numbers, we can use a more
compact notation. Stacking such vectors side-by-side, we can form a matrix X as
follows:
Chapter 2. Vectors and Vector Spaces 53
X x1 x 2 x 3
Now considering the set of scalars a i to be similarly stacked into a column called
a (this does not mean they constitute a vector), then the condition for linear
l q
independence of the vectors x i is that the equation Xa 0 has only the trivial
solution a 0 .
a1 x1 ( t ) a n x n ( t ) 0
l q
for all t [t 0 , t1 ] , then the functions x1 ( t ), , x n ( t ) are linearly dependent in
that interval. Otherwise they are independent. Independence outside that interval,
or for that matter, in a subinterval, is not guaranteed. Furthermore, if
X ( t ) [ x1 ( t ) x n ( t )] is an n 1 vector of functions, then we can define
the Gram matrix (or grammian) of X ( t ) as the n n matrix:
z
t2
G X ( t1 , t 2 ) X ( t ) X * ( t ) dt
t1
It can be shown (see Problem 2.26) that the functions x i ( t ) are linearly
independent if and only if the Gram determinant is nonzero, i.e., G X ( t1 , t 2 ) 0
.
LM 2 OP LM1OP LM0OP
x1 M1P x2 M3P x3 M7P
MN 0 PQ MN4PQ MN8PQ
This is a linearly dependent set because we can choose the set of a-coefficients as
54 Part I. Mathematical Introduction to State Space
3
a i x i x1 2 x 2 x3 0
i 1
Some readers may be familiar with the test from matrix theory that allows
one to determine the linear independence of a collection of vectors from the
det(A) determinantM formed by the matrix X as we constructed it above. That is, the
dependence of the vectors in the previous example could be ascertained by the
test
2 1 0
det( X ) 1 3 7 0
0 4 8
If this determinant were nonzero, then the vectors would have been found linearly
independent.
However, some caution should be used when relying on this test, because the
number of components in the vector may not be equal to the number of vectors in
the set under test, so that a nonsquare matrix would result. Nonsquare matrices
have no determinant defined for them.* We will also be examining vector spaces
that are not such simple n-tuples defined over the real numbers. They therefore
do not form such a neat determinant. Furthermore, this determinant test does not
reveal the underlying geometry that is revealed by applying the definition of linear
dependence. The following example illustrates the concept of linear independence
with a different kind of vector space.
* In such a situation, one can examine all the square submatrices that can be formed from
subsets of the rows (or columns) of the nonsquare matrix. See the problems at the end
of the chapter for examples.
Chapter 2. Vectors and Vector Spaces 55
LM 1 OP LM s 2 OP
x1 M
M s 1 PP x2 M
M ( s 1)( s 3) PP
MM 1 PP MM 1 PP
Ns 2Q N s3 Q
If the chosen field is the set of all real numbers, then this set of two vectors is
found to be linearly independent. One can verify that if a1 and a2 are real numbers,
then setting
LM 1 OP LM s 2 OP
0
LM0OP a x a2 x2 a M
M s 1 PP a MM ( s 1)( s 3) PP
N0Q 1 1
MM 1 PP MM 1 PP
1 2
Ns 2Q N s 3 Q
will imply that a1 a 2 0 . See Problem 2.3 at the end of this chapter.
If instead the field is chosen as the set of rational polynomials, then we have
an entirely different result. Through careful inspection of the vectors, it can be
found that if the scalars are chosen as
s3
a1 1 and a2
s2
then
LM 1 OP LM 1 OP
2
ai xi M
M s 1 PP MM s 1 PP LM0OP
i 1 MM 1 PP MM 1 PP N0Q
Ns 2Q N s 2Q
The vectors are now seen to be linearly dependent. Note that this is not a matter
of the same two vectors being dependent where they once were independent. We
prefer to think of them as entirely different vectors, because they were selected
from entirely different spaces, as specified over two different fields.
This particular example also raises another issue. When testing for linear
independence, we should check that equality to zero of the linear combinations is
true identically. The vectors shown are themselves functions of the variable s.
Given a set of linearly independent vectors, there may therefore be isolated values
56 Part I. Mathematical Introduction to State Space
for s that might produce the condition ai xi 0 even though the vectors are
independent. We should not hastily mistake this for linear dependence, as this
condition is not true identically, i.e., for all s.
Geometrically, the concept of linear dependence is a familiar one. If we think
of two-dimensional euclidean spaces that we can picture as a plane, then any two
collinear vectors will be linearly dependent. To show this, we can always scale
one vector to have the negative magnitude of the other. This scaled sum would of
course equal zero. Two linearly independent vectors will have to be noncollinear.
Scaling with nonzero scalars and adding them will never produce zero, because
scaling them cannot change their directions.
In three dimensions, what we think of as a volume, we can have three linearly
independent vectors if they are noncoplanar. This is because if we add two scaled
vectors, then the result will of course lie in the plane that is formed by the vectors.
If the third vector is not in that same plane, then no amount of nonzero scaling
will put it there. Thus there is no possibility of scaling it so that it is the negative
sum of the first two. This geometric interpretation results in a pair of lemmas that
are intuitively obvious and can be easily proven.
This is obvious from the geometric description we have given. If the original
set of vectors is dependent, it adds to zero with at least one nonzero scaling
coefficient. We can then include the new vector into the linear combination with
a zero coefficient, as in Figure 2.2. At least one of the other scalars is already
known to be nonzero, and the new sum will remain zero. Thus, the augmented set
is still dependent.
0 a1 x1 a 2 x 2 a n x n a n 1 x n 1
n
a x
i 1 i i
0 with at least one a coefficient not equal to zero.
Suppose that a j 0 . Then without risking a divide-by-zero,
we can perform the following operation:
a1 x1 a 2 x 2 a j 1 x j 1 a j 1 x j 1 a n x n
xj
aj
a1 a a j 1 a j 1 a
x1 2 x 2 x j 1 x j 1 n x n
aj aj aj aj aj
(2.12)
By defining bi a i a j ,
x j b1 x1 b2 x 2 bn x n (2.13)
Already, our intuitive sense is that we can have (at most) as many
independent vectors in a set as we have “dimensions” in the space. This is strictly
true, although we will need to rigorously show this after discussion of the
mathematical notion of dimension.
2.2.3 Bases
If we are working within a particular vector space and we select the maximum
number of linearly independent vectors, the set we will have created is known as
a basis. Officially, we have the definition:
58 Part I. Mathematical Introduction to State Space
One must be careful to note that in an n-dimensional vector space, there must
be exactly n vectors in any basis set, but there are an infinite number of such sets
that qualify as a basis. The main qualification on these vectors is that they be
linearly independent and that there be a maximal number of them, equal to the
dimension of the space. The uniqueness condition in the definition above results
from the fact that if the basis set is linearly independent, then each vector contains
some unique “direction” that none of the others contain. In geometric terms, we
usually think of the basis as being the set of coordinate axes. Although coordinate
axes, for our convenience, are usually orthonormal, i.e., mutually orthogonal and
of unit length, basis vectors need not be.
le q le , e ,, e q
i 1 2 n
0 x 1e1 n e n 0 (2.18)
l q is, as
Suppose 0 0 . If this were the case and if the set e i
specified, linearly independent, then we must have
Chapter 2. Vectors and Vector Spaces 59
1
x e1 2 e2 n e n
0 0 0
1e1 2 e2 n e n (2.19)
n
i ei
i 1
n n
x x i e i i e i
i 1 i 1
n
e j
i i e i
i 1
(2.20)
0
vector space is isomorphic* to the space of n-tuples. This allows us to always use
the n-tuple operations such as dot-products and the induced norm on arbitrary
spaces. It also allows us to rely on the intuitive picture of a vector space as a set
of coordinate axes. For these reasons, it is often much more convenient to use the
representation of a vector than to use the vector itself. However, it should be
remembered that the same vector will have different representations in different
bases.
There is weaker terminology to describe the expansion of a vector in terms
of vectors from a particular set.
Note that in this terminology, the set x i l q is not necessarily a basis. It may
be linearly dependent. For example, five noncollinear vectors in two dimensions
suffice to span the two dimensional euclidean space, but as a set they are not a
basis. We sometimes use this definition when defining subspaces, which will be
discussed later.
We will see after the following examples how one can take a single vector
and convert its representation from one given basis to another.
e1 1 e2 s e3 s 2 e4 s3
* Spaces are said to be isomorphic if there exists a bijective mapping (i.e., one-to-one and
onto) from one to the other.
Chapter 2. Vectors and Vector Spaces 61
x 3s 3 2 s 2 2 s 10
x 3e 4 2e 3 ( 2)e 2 10e1
l f q l f , f , f , f q os
i 1 2 3 4
3
t
s 2 , s 2 s, s 1, 1
x 3( s 3 s 2 ) 5( s 2 s ) 3( s 1) 13
or
LM 3 OP
xM P
5
MM 3 PP
N13Q
This can be verified by direct expansion of this expression.
Solution:
The first function space can be described with power series expansions because
of the use of the key word analytic. Therefore, for example, we can write a Taylor
series expansion
62 Part I. Mathematical Introduction to State Space
1 d n f (t )
f (t ) n! dt n
( t t0 )n (2.22)
n0 t t 0
a0
f (t ) a n cos
2 n 1
d t i b sind t i
2 n
T n
2 n
T
(2.23)
or
1
f (t ) c n e j( 2n T )t
2 n
(2.24)
o
e i 1, cos d i, sind i, cosd i, sind i,t
2
T
2
T
4
T
4
T (2.25)
or
o
f i 1, e j ( 2 T )t , e j ( 4 T )t , t (2.26)
n n
x x j v j x i v i (2.27)
j 1 i 1
Chapter 2. Vectors and Vector Spaces 63
n
v j bij v i (2.28)
i 1
By gathering the vectors on the right side of this equation side-by-side into a
matrix, this expression can also be expressed as
LM b OP
1j
v j v1 v 2 v n MMb PP
2j
MM b PP
N Q
nj
v1 v 2 v n B
where the matrix B so defined in this equation will hereafter be referred to as the
“change of basis matrix.”
Substituting the expression (2.28) into Equation (2.27),
n
xj
LM b v OP x v
n n
(2.30)
j 1 N Q
i 1
ij i
i 1
i i
By changing the order of summation on the left side of this equality, moving
both terms to the left of the equal sign, and factoring,
n F n I
G bij x j x i J vi 0 (2.31)
i 1 H j 1 K
64 Part I. Mathematical Introduction to State Space
n
x i bij x j (2.32)
j 1
This is how we get the components of a vector in a new basis from the
components of the old basis. The coefficients bij in the expansion come from our
knowledge of how the two bases are related, as determined from Equation (2.29).
Notice that this relationship can be written in vector-matrix form by expressing
the two vectors as n-tuples, x and x . Denoting by B the n n matrix with
coefficient bij in the ith row and jth column, we can write
x Bx (2.33)
le , e q RSLMN10OPQ, LMN10OPQUV
le , e q RSLMN12OPQ, LMN01OPQUV
and
1 2
T W T 1
W 2
representation in the le q basis, we must first write down the relationship
i
between the two basis sets. We do this by expressing the vectors e j in terms of
vectors e i :
N Q
2 1
2
B
LM 0 OP 1
2
N1 Q 1
2
Chapter 2. Vectors and Vector Spaces 65
l q basis,
giving, in the e i
x Bx
LM 0 OP LM2OP LM 1 OP
1
2
REMARK: The reader might verify that the expansion of the e i vectors in terms
of the e j vectors is considerably easier to write by inspection:
e1 1e1 2e 2 e1 e2
LM1OP
N2Q
e 2 ( 1)e1 0e 2 e1 e2
LM1OP
N0Q
Therefore, the inverse of the change-of-basis matrix is more apparent:
B 1
LM1 1OP
N2 0 Q
Note that the columns of this matrix are simply equal to the columns of
lq n s
coefficients of the basis vectors e i because the e j basis happens to be the
same as the standard basis, but this is not always the case. We can display the
relationship between the two basis sets and the vector x as in Figure 2.3.
ê1
x
e2
ê2 e1
Figure 2.3 Relative positions of vectors in Example 2.7. It can be seen that
x 2e1 2e 2 and that x 1 e1 1 e 2 .
66 Part I. Mathematical Introduction to State Space
LM x 11 x12 x1n OP
M PP
x x 22 x 2n
A x1
MM
21
x2 xn (2.34)
Nx m1 x m2 x mn
PQ
With this interpretation of a matrix, the concept of rank can be reconsidered.
Note that this definition for rank is consistent with the test for linear
independence of a collection of vectors. Therefore, it is apparent also that the rank
is the dimension of the space formed by the columns of the matrix (the range
space, see Chapter 3). If the collection of columns forms a matrix with a nonzero
determinant, then the set of vectors is linearly independent. Of course, this test
only works with a number of vectors equal to the number of coefficients in the
representation (which is equal to the dimension of the space). If the number of
vectors is different from the dimension of their space, then all square subsets of
the rows and columns of the matrix must be separately tested with the determinant
test. If matrix A is of dimension m n , then of course, r ( A) min( m, n ) .
Furthermore, if an m n matrix has r ( A) min( m, n ) , it is said to be of “full
rank.” If an m n matrix has r ( A) min( m, n ) , it is said to be “rank deficient,”
or “degenerate.” If the number of vectors considered is the same as the dimension
of the space, so that the matrix A is square, then the matrix is also called
“singular.” A square matrix that is of full rank is “nonsingular.”
q( A ) n r ( A ) (2.36)
Chapter 2. Vectors and Vector Spaces 67
In addition, there are some useful properties for the rank and nullity of
products of matrices, given by the following theorem.
r ( A) + r ( B ) n r ( C ) min( r ( A), r ( B ))
(2.37)
q( C ) q( A) q( B )
[The second line can be derived from the first, using the
definition of nullity (2.36)]. Furthermore, if D is an m m
nonsingular matrix and E is an n n nonsingular matrix, then
4. x , y x , y . (2.42)
5. x , y x, y . (2.43)
Vector spaces with inner products defined for them are referred to as inner
product spaces. One can verify that the inner products given at the beginning of
this chapter for n are valid inner products as defined above. Note, however,
that if the vectors are defined over a complex field, then the conjugate operator as
shown in the conditions must be observed. For example, the inner product of two
complex-valued vectors can be written x, y x * y , where x * is the complex-
conjugate transpose.
z
b
x, y x *( t ) y( t ) dt (2.44)
a
2.2.7 Norms
A norm generalizes our concept of “length” or “magnitude” of a vector.
Vector spaces with such norms defined for them are referred to as normed linear
spaces.
Norms are sometimes induced from the given inner product. For example,
the euclidean norm given earlier in this chapter is induced by whatever inner
product is defined:
1
x x, x 2 (2.49)
A unit vector is one that has unit length, i.e., whose norm is 1. Of course, any
vector can be converted to a unit vector (“normalized”) by dividing it by its own
norm. This will not change its direction, only scale its magnitude.
n
x k
k xi k (2.50)
i 1
n
xi
2
2. l2 norm: x 2
(2.52)
i 1
(i.e., the euclidean norm). If the vector is a signal vector, x( t ) , the l2
norm becomes:
z
x( t ) 2 x *( t ) x( t ) dt (2.53)
3. l norm: x
max x i (2.54)
i
70 Part I. Mathematical Introduction to State Space
( x , y ) x y (2.55)
x, y x y cos (2.56)
All euclidean spaces are complete, and we will not generally be concerned
with the completeness of the spaces we discuss in this book. We will therefore
not elaborate on this definition further.
l q is orthogonal if
A set of vectors x i
x i , x j 0 if i j and
(2.60)
x i , x j 0 if i = j
x i , x j 0 if i j and
(2.61)
x i , x j 1 if i = j
spans ny s ,
j j 1,, m , for every m n . We build the orthonormal set by
starting with one vector from the original nonorthonormal set, and successively
include additional vectors, one at a time, while subtracting from each new vector
any component that it shares with the previously collected vectors. Therefore, as
we add more vectors to the orthonormal set, each will contain only components
along directions that were not included in any previously found vectors.
The technique will be presented as the following algorithm:
Let v1 y1 .
Choose v 2 to be y 2 with all components along the previous vector,
subtracted out, i.e.,
v 2 y 2 av1
v1 , v 2 0 v1 , y 2 a v1 , v1
so
v1 , y2
a
v1 , v1
Choose v 3 as y 3 with the components along both previous v i vectors
subtracted out:
v 3 y3 a1v1 a 2 v 2 (2.62)
0
v1, v3 0 v1, y3 a1 v1, v1 a2 v1, v2
v2 , v3 0 v 2 , y3 a1 v2 , v1 a2 v2 , v2
0 (2.63)
These two equations can be solved independently of one another to give
v1 , y3 v 2 , y3
a1 and a2
v1 , v1 v2 , v2
i 1 i 1 v k , yi
v i yi a k v k yi vk (2.64)
k 1 k 1 vk , vk
vi
v i (2.65)
vi
It should be noted that there are a number of variations on this procedure. For
example, the normalization step (step 5) could be executed each time a new vector
v is computed. This will affect the numerical results in subsequent steps, and make
the end result look somewhat different, though still entirely correct. It has the
benefit, though, that all terms v i , v i 1 for subsequent iterations. Also, the
results will depend on which basis vector is chosen at the start. There is no special
reason that we chose y1 to start the process; another vector would suffice.
If we always use orthonormal basis sets, we can always find the component
of vector x along the jth basis vector by performing a simple inner product
lq
operation with that jth basis vector. That is, if e i , i 1,, n is an orthonormal
n
x j e j , x xi e j , ei
i 1
0 0 0
x1 e j , e1 x 2 e j , e 2 x j e j , e j x n e j , e n
1
(2.66)
x
L O
M c P
n
2
1
2
(2.67)
N Q i 1
i
1
x x, x 2
1
n n 2
c j e j , ci e i
j 1 i 1
L F IO 1
2
M c G c e , e J P
n n
MN H
i 1
i
K PQ j 1
j j i
L O L O
M c b c gP M b c g P
n
1
2 n
2
1
2
N i 1 Q N
i i
Q i 1
i
x
L O
n
M x P 2
1
2
(2.68)
N Q
i 1
i
Note that this is the formula often given for the norm of any vector, but it may not
be true if the vector is not expressed in orthonormal bases. See Example 2.7
(Change of Basis) on page 64. Parseval’s theorem is particularly useful when
finding the norm of a function that has been expanded into an orthonormal set of
basis functions (see Example 2.11 on page 79); the theorem applies to infinite-
dimensional spaces as well.
2.4.1 Subspaces
Strictly, this definition specifies only that the subspace is closed under scalar
multiplication and vector addition. Some of the other properties that must be
satisfied for the subspace to be considered a space unto itself are automatically
satisfied given the fact that vectors in subspace M are gathered from the space X.
Other properties can be inferred from closure. For example, that the subspace by
necessity contains a zero vector is implied by the above definition if one chooses
0.
Note that space X can be considered a subspace of itself. In fact, if the
dimension of M, dim( M ) , is the same as the dimension of X, then X M . If
dim( M ) dim( X ) , then M is called a proper subspace of X. Because of the
closure requirement and the fact that each subspace must contain the zero vector,
we can make the following observations on euclidean spaces:
1. All proper subspaces of 2 are straight lines that pass through the
origin. (In addition, one could consider the trivial set consisting of only
the zero vector to be a subspace.)
76 Part I. Mathematical Introduction to State Space
2. All proper subspaces of 3 are either straight lines that pass through the
origin, or planes that pass through the origin.
3. All proper subspaces of n are surfaces of dimension n 1 or less, and
all must pass through the origin. Such subspaces that we refer to as
“surfaces” but that are not two-dimensional, as we usually think of
surfaces, are called hypersurfaces.
This situation is depicted in Figure 2.4. In the figure, vector x is in the three-
dimensional space defined by the coordinate axes shown.
x
xu
X 3
u
U X
Figure 2.4 Illustration of the orthogonal projection theorem. Vectors u and y lie in
the shaded plane. Vector x extends off the plane, and x u is orthogonal
to u and every other vector in the plane.
Chapter 2. Vectors and Vector Spaces 77
The shaded plane that passes through the origin is the subspace U. Vector x
does not lie in the plane of U. The vector x u referred to in the theorem is
orthogonal to the plane itself; i.e., it is orthogonal to every vector in the plane,
such as the vector y. It is this orthogonality condition that enables us to find the
projection.
In the following example, we will practice using the projection theorem by
considering subspaces that are the orthogonal complements of one another. For
this we give the definition:
Solution:
We will use two approaches to solve this problem. In the first, we use the
projection theorem without first orthonormalizing the vectors x1 and x 2 . To
directly apply the projection theorem, we propose that the projection of f onto W
be written in the following form, with as yet undetermined coefficients i :
Pf 1 x1 2 x 2
f , x1 Pf , x1 1 x1 , x1 2 x 2 , x1
Similarly,
f , x 2 Pf , x 2 1 x1 , x 2 2 x 2 , x 2
78 Part I. Mathematical Introduction to State Space
7 110 2 3
1 1 3 2 2
LM1OP
M P
2
g Pf x1 ( 1) x 2
MM1PP
N0Q
This being the desired vector in W, it is a simple matter to compute:
LM 3 OP
h f g M P W
2
MM 1 PP
N 1Q
As an alternative path to the solution, which avoids the simultaneous equation
solving at the expense of the Gram-Schmidt operations, we could first
lq
orthonormalize the x i vectors. Denote the orthonormalized version as v i and
compute them as described:
LM1OP
x M P
2
v1 1
MM2PP
N1Q
v 2 x 2 av1
where
v1 , x 2
a 0.3
v1 , v1
giving
Chapter 2. Vectors and Vector Spaces 79
LM 0.3OP
M
0.6P
v2
MM 0.4PP
N 0.7Q
Normalizing the two new basis vectors:
LM0.3162OP LM0.2860OP
M
0.6325P
M
v1 v2 0.5721P
v1
v1 MM0.6325PP and v 2
v2 MM 0.3814PP
N0.3162Q N 0.6674Q
Then to find the components of f along these directions,
f1 f , v1 2.2136
f 2 f , v 2 10488
.
giving
LM1OP
M P
2
g Pf f1v1 f 2 v 2
MM1PP
N 0Q
exactly as before.
z
1
f ( t ), g ( t ) f ( t ) g ( t ) dt (2.72)
l
W span 1, cos t , sin t , cos 2t , sin 2t . q
1. From the inner product given, induce a norm on this space.
2. If it is known that the vectors that span W are linearly independent and
therefore form a basis for this 5-dimensional subspace, show that the set
l q
of vectors 1, cos t , sin t , cos 2t , sin 2t is an orthogonal basis for W.
Normalize this set to make it an orthonormal basis.
3. Find the orthogonal projection of the following function, f ( t ) , defined
in Figure 2.5, onto the space W.
f (t )
t
0
2 2
Solution:
To induce a norm, we use
f ( t ) f ( t ), f ( t )
1
2
L1
M z f
2
O
( t ) dt P
1
2
(2.73)
MN
PQ
Use the notation:
If this set is linearly independent (can you show this?), then it automatically
qualifies as a basis. To show that the set is orthogonal, we must demonstrate that
condition (2.60) holds. Finding all these inner products,
Chapter 2. Vectors and Vector Spaces 81
z
1
e1 , e1 1 dt 2
z
1
e2 , e2 cos 2 t dt 1
z
1
e3 , e3 sin 2 t dt 1
z
1
e4 , e4 cos 2 2t dt 1
z
1
e5 , e5 sin 2 2t dt 1
e1 , e 2 e 2 , e1 0
e i , e j 0 for all i j
z z
1
1 1 1 2 1 1
C1 f ( t ), e1 f ( t ), f ( t ) dt dt
2 2
2 2
2
z z
1 1 2 2
C2 f ( t ), e 2 cos t f ( t ) dt cos t dt
2
z z
1 1 2
C3 f ( t ), e3 sin t f ( t ) dt sin t dt 0
2
z z
1 1 2
C4 f ( t ), e 4 cos 2t f ( t ) dt cos 2t dt 0
2
z z
1 1 2
C5 f ( t ), e5 sin 2t f ( t ) dt sin 2t dt 0
2
1 2
g( t ) cos t (2.74)
2
Note that this is simply the five-term Fourier series approximation for the
given signal f ( t ) . The original function and its projection are shown in Figure
2.6 below.
It is useful at this point to also remember Parseval’s theorem. Because the
projection g( t ) is written in an orthonormal basis, its norm can easily be found
from the coefficients in this basis as:
L O
5
g( t ) M C P 2
1
2
LM 1 4 OP 1
2
0.95
N Q
i 1
i
N2 Q 2
(2.75)
Chapter 2. Vectors and Vector Spaces 83
g (t )
f (t )
0
t
2 2
Figure 2.6 Original function f ( t ) and its projection g( t ) onto the subspace W.
The euclidean vectors that we have been using as examples throughout this
chapter are in fact not linear algebras because we have defined no multiplication
for them. Nevertheless, some common vector spaces that we will discuss will
indeed be linear algebras. An example is the space of n n matrices, which we
will introduce as representations of linear operators in the next chapter. The
familiar procedure for matrix multiplication will be the multiplication operator
that makes the space of matrices into an algebra of matrices. The interpretation
of these matrices as operators on vector spaces is the topic of the next chapter.
2.6 Summary
In this chapter we have reviewed the basics of vector spaces and linear algebra.
These fundamental concepts are rightfully the domain of mathematics texts, but
it is useful to present them here in order that we can stress engineering relevance
and the terminology we will find important later. The most important concepts
include the following:
• Vectors need not be thought of as “magnitude and direction” pairs, but
this is a good way to imagine them in low dimensions.
• All linear finite dimensional vector spaces, such as time functions or
rational polynomials, may be treated as vectors of n-tuples because they
are isomorphic to n-tuples. We then treat their representations as we
would n-tuples, unifying the vector-matrix operations we must perform
into one computational set.
• The basis of a space is an important generalization of the “coordinate
system” that we are familiar with from basic mechanics and physics.
Basis expansions and change-of-basis operations can be performed with
vector-matrix arithmetic.
• The rank of a matrix represents the dimension of the space formed by its
columns, i.e., it is the number of linearly independent columns the matrix
possesses.
• Inner products are used to relate the differences in direction of two
vectors, and norms are used to gauge their “size.”
• Orthonormal bases are the most convenient kind of basis for
computational purposes and can be constructed with the Gram-Schmidt
procedure.
• There exists a simple theorem that enables us to find the vector closest
to a subspace but not actually in it. This process will be important in
engineering approximations wherein we must find the best solution to a
problem when an exact solution is unavailable or impractical.
Chapter 2. Vectors and Vector Spaces 85
In the next chapter, we will consider linear operators, which will help us solve
simultaneous equations, map vectors from one space into another or into itself,
and compute other functions performed on vectors.
2.7 Problems
2.1 The set W of n n matrices with real entries is known to be a linear vector
space. Determine which of the following sets are subspaces of W:
a) The set of n n skew-symmetric matrices?
b) The set of n n diagonal matrices?
c) The set of n n upper-diagonal matrices?
d) The set of n n singular matrices?
2.2 Consider the set of all polynomials in s of degree less than or equal to four,
with real coefficients, with the conditions:
p'( 0) p( 0) 0
p(1) 0
Show that this set is a linear vector space, and find its dimension. Construct
a basis for this space.
2.3 For Example 2.4 on page 54 where we tested the linear independence of
vectors of rational polynomials, show that indeed, if a1 and a 2 are real,
then they must both be zero.
2.4 Are the following vectors in 4 linearly independent over the reals?
field of reals.
c) o2s 2
t
2s 1, 2s 2 2s 1, s 2 s 5 in the space of polynomials
over the field of reals.
2.7 Determine a basis for the space W 4 spanned by the four vectors
R|L1O L 2 O L 3 O L 5 OU|
|M 1 P M 1 P M 2 PP, MM 5 PP|V
{ y , y , y , y } S M P , M P, M
1 2 3 4
||MM5PP MM 7 PP MM 10 PP MM 15 PP||
TN 7 Q N8Q N11Q N15QW
LM 3 OP LM 4 OP LM1OP
Is z M Is u M P W ? Is v M P W ?
1 P 9 1
MM 13 PP W ?
MM12 PP MM3PP
N Q
17 N Q
8 N3Q
2.8 Use the matrix form of the change of basis operation, i.e., Equation (2.28),
to perform the change of basis on the vector x 10 2 2 3 given
in Example 2.5 on page 60, arriving at the new representation shown in
that example.
Chapter 2. Vectors and Vector Spaces 87
2.9 Prove that the number of linearly independent rows in a matrix is equal to
the number of linearly independent columns.
2.10 For the following matrices, find the ranks and degeneracies.
LM 1 2 3 OP LM0 2 0 OP LM1 6 3 2 2 OP
a) MM2 2 2P b) MM2 0 0P c) MM2 12 4 6 10P
N4 1 4PQ N0 2 0PQ N3 18 0 15 15PQ
2.12 If it is known that a valid basis for the space of polynomials of degree less
than or equal to three is {1, t , t 2 , t 3 } , show that every such polynomial has
a unique representation as
p( t ) a 3t 3 a 2 t 2 (1 t ) a1t (1 t ) 2 a 0 (1 t ) 3
2.13 Construct an orthogonal basis of the space spanned by the set of vectors
R|L2O L 1 O L 5 OU|
|SMM2PP, MM 0 PP, MM3PP|V
||MM3PP MM 1 PP MM 4 PP||
TN 1 Q N1Q N 8 QW
Repeat for
2.14 Using the Gram-Schmidt technique, construct an orthonormal basis for the
subspace of 4 defined by the vectors
LM1OP LM 4 OP LM 3 OP
M P, M P, M P
4 3 1
x1
MM 1 PP x2
MM 1 PP and x3
MM2PP
N1Q N1Q N2Q
2.15 Let C 4 denote the space of all complex 4-tuples, over the complex
numbers. Use the “standard” inner product and the Gram-Schmidt method
to orthogonalize the vectors
2.17 Let
R| 4x if 0 x 0.25
f ( x ) S2 4 x if 0.25 x 0.5
1
|T 0 otherwise
f,g z0
1
f ( x ) g ( x ) dx
and let W be the subspace spanned by the three functions f 1 , f 2 , and f3.
Find the dimension of the subspace W.
Chapter 2. Vectors and Vector Spaces 89
Now let
g( x )
RS1 if 0.5 x 1
T0 if 0 x 0.5
2.18 Let V be the vector space of all polynomials, with inner product
f,g z
0
f ( x ) g ( x ) e x dx
z
1
f,g f ( t ) g ( t ) dt
1
f,g z 3
0
f ( x ) g ( x ) dx
2.22 Use the complex vector x 1 2i 2 i to demonstrate that, in a
complex vector space, x, y x y is not a valid inner product.
J z
1
1
p 2 ( t ) dt
2.24 Let V be the space of all polynomials with real coefficients, with the inner
product
f,g z
2
f ( x ) g ( x ) e x dx
f,g z
1
1
f ( x ) g ( x ) dx
[1] Brogan, William L., Modern Control Theory, 3rd edition, Prentice-Hall, 1991.
[2] Golub, Gene H., and Charles F. Van Loan, Matrix Computations, Johns Hopkins
University Press.
[3] Halmos, Paul R., Finite Dimensional Vector Spaces, 2nd edition, D. Van Nostrand,
1958.
[4] Kolmogorov, A. N., and S. V. Fomin, Introductory Real Analysis, Dover, 1970.
[5] Naylor, A. W., and G. R. Sell, Linear Operator Theory in Engineering and
Science, Springer-Verlag, 1982.
[6] Shilov, Georgi E., Linear Algebra, Dover, 1977.
[7] Strang, Gilbert, Linear Algebra and Its Applications, Academic Press, 1980.
[8] Zadeh, Lotfi, A., and Charles A. Desoer, Linear System Theory: The State Space
Approach, McGraw-Hill, 1963.