Professional Documents
Culture Documents
Electrical Engineering
Consulting Editor: John B. Thomas
Springer Texts in Electrical Engineering
Linear Programming
M. Sakarovitch
Introduction to Probability
lB. Thomas
Charles A. Desoer
Department of Electrical Engineering
and Computer Sciences
University of California
Berkeley, CA 94720
USA
ISBN 978-1-4612-6961-8
PREFACE
This book is the result of our teaching over the years an undergraduate course on
Linear Optimal Systems to applied mathematicians and a first-year graduate course on
Linear Systems to engineers. The contents of the book bear the strong influence of the
great advances in the field and of its enormous literature. However, we made no
attempt to have a complete coverage.
Our motivation was to write a book on linear systems that covers finite-
dimensional linear systems, always keeping in mind the main purpose of engineering
and applied science, which is to analyze, design, and improve the performance of phy-
sical systems. Hence we discuss the effect of small nonlinearities, and of perturbations
on the data; we face robustness issues and discuss the properties of feedback. It is our
hope that the book will be a useful reference for a first-year graduate student.
We assume that a typical reader with an engineering background will have gone
through the conventional undergraduate single-input single-output linear systems
course; an elementary course in control is not indispensable but may be useful for
motivation. For readers from a mathematical curriculum we require only familiarity
with techniques of linear algebra and of ordinary differential equations.
The purpose of this book is to provide a systematic and rigorous access to a) the
main topics of linear state-space system theory in both the continuous-time case and
the discrete-time case, and b) the I/O description of linear systems. The main thrusts
of the book are: analysis of system descriptions and derivation of their properties, LQ-
optimal control, state-feedback and state-estimation, and a short study of MIMO
unity-feedback systems. We cover both continuous-time and discrete-time systems; in
most cases, the discrete-time case is covered in an isomorphic chapter, e.g. Chapter 2d
is the discrete-time coverage of Chapter 2.
The contents of the book can be described roughly as follows (for a topic by
topic description, see the table of contents). As an introduction, Chapter 1 discusses
the relation between physical systems, their models and their mathematical representa-
tions. It also raises the issues of sensitivity and robustness. Chapters 2 through 4
describe and structure the trajectories of a generic class of linear system representa-
tions in both the time-varying and time-invariant cases. A treatment of finite horizon
LQ-optimization is included. Based on the results of these chapters, Chapter 5
discusses general system concepts and the four-way classification· of systems: linear
versus nonlinear, time-invariant versus time-varying. Chapter 6 derives the relation
between a continuous-time system and its related discrete-time (approximate)
equivalent obtained as a result of the AID and D/A conversions. Chapter 7 covers sta-
bility: the three main types of stability for linear systems, namely, the I/O stability and
the two state-related stability concepts i.e., asymptotic stability, and above all,
exponential stability. Chapter 8 focuses on the coupling of the input to the state, i.e.
controllability, and that of the Slate to the output, i.e. observability. Related concepts
such as stabilizability and detectability are also covered. Chapter 9 covers (time-
invariant) minimal realizations and the McMillan degree; the
vi
Special thanks are due to our wives Nicole and Jackie for their strong support
over the years. We thank Oswaldo Garcia for drawing the figures. Finally, Bettye
Fuller deserves special thanks for handling a difficult manuscript.
Reference numbers are used to number items such as results. definitions. state-
ments. and formulas. These reference numbers are started at the beginning of each
section listed with two symbols. e.g. section 2.1 (Fundamental Properties of R 0) or
Appendix A.3 (Linear Spaces). When referring to an item inside a section we use bare
reference numbers, e.g. (55). When referring to an item outside a section. we use
three symbols. e.g. (2.2.44) indicates item (44) of Section 2.2.
The index at the end of the book is preceded by a) a list of Mathematical Sym-
bols (followed by a brief definition). and b) a list of often-used Abbreviations.
CONTENTS
PREFACE ............................................................................................................................ v
INTRODUCTION
Thus engineers invent new devices. invent ways to manufacture them cheaply and
reliably. and interconnect them into useful systems. It is for this last step that it is
important to study the properties of engineering systems. and delineate their capabili-
ties and their limitations. As a result of these studies. engineers can 1) specify the
features the system must have in order to accomplish the required tasks. 2) use the
analysis methods to design the system.
Let us now tum to the following question: on the one hand. engineers design.
build. and operate physical systems and. on the other hand. the main tool of system
theory is mathematics: The relation between these two activities is the subject of the
next section.
t For example. a very large flexible satellite may for certain problems be usefully
modeled as a particle. Or. an integrated circuit. part of a digital signal processor. may
be modeled as a set of binary operations performed on streams of "zeros" and "ones."
3
for electrical circuits we may use node equations or modified node equations or tableau
equations.
Having picked a mathematical representation for the chosen model, the next step
is to analyze the model, that is, find its properties, its capabilities and its limitations;
these three goals are crucial to engineering design. The study of these questions is the
task of system theory.
System theory is part of engineering in the same way that theoretical physics is
part of physics. In both cases, system theory and theoretical physics use mathematical
tools to study the main models of engineering and physics, respectively. These
mathematical tools give predictive power to engineering, thus the merits of the various
design alternatives may be sorted out before investing labor and materials in building
anyone of them.
Of course in the study of specific engineering systems one uses computer simula-
tion. The test of the validity of the whole procedure is the agreement between the
simulation results and the measurements performed on the physical system. This
whole process is summarized in Fig. 1.1.
_______MEASUREMENTS
.>----....,..---....... --..-.. . . . - ,
' ......
,
I I
......
AGREE?
,
\
\
J
i
_/
Fig. 1.1. A physical system and its relation to its models, their mathematical
representations, and the measurements.
1.3. Robustness
Robustness is an important concern in system design, we briefly discuss two
aspects of the robustness problem.
The first one arises because every physical system is designed on the basis of a
necessarily approximate model. Then the question arises: will the physical system
have a performance that is sufficiently close to the performance predicted by the
4
(idealized) model? For example, if the linear model is exponentially stable, or con-
trollable, etc ... will the physical system, which, say, is in fact a small (possibly non-
linear) perturbation from the linear model, still be exponentially stable or controllable,
etc .. ?
The second aspect of robustness appears in the following manner: the design is
based on the (idealized) model and it specifies a number of nominal properties of the
system to be built, for example, physical dimensions, geometric configuration, compo-
sition of materials, etc. Once produced and operating in the field, the physical system
differs from the nominal design mainly because of environmental conditions -- tem-
perature, humidity, wear, radiation, etc. --, and manufacturing deviations -- the
manufactured physical components are only approximations of the nominal in terms of
physical dimensions, configuration, composition, etc. So the question is: what is the
effect of all such deviations from nominal on the performance of the physical system
in the field? Such questions can be studied by calculating the effect on system perfor-
mance due to changes in design parameters and exogeneous disturbances. It is a fact
that some designs, which perform nominally perfectly, are totally inadequate in the
field because they are too sensitive to small perturbations in some of their parameters.
In the course of these notes, we'll discuss from time to time these questions of robust-
ness.
CHAPTER 2
2.1.1. Definitions
We study here dynamical systems represented by the following equations
3 (as is usual), the state x(t) eRn, the input u(t) eRn" and the output yet) e JRI1o;
y) Since A('), BO, CO, DO are functions of class PC, they are bounded on
bounded intervals.
E) equation (1) is called the state differential equation (d.e.) and equation (2) is
called the read-out equation.
Later on we shall see that at any time t the state x(t) and output yet) of represen-
tation R (.) are functions of t (actual time), to (initial time), Xo (initial state), and the
input u(-) e U. t Hence for all Ie 1Rr we shall have the functional relalions
7 x(l) = s(I,Io,xo,u)
8 y(l) = p(l,to,xo,u)
which are called the state transition map and the response (map) of R (.), respec-
tively. Their struclural propenies will be studied next.
t Later we shall see that the dependence on u(-) (for t to) is actually on u[Io. I ]
(the values of u(·) during the lime-interval [Io,t] c R+).
7
and (8.1.4) of existence and uniqueness theorem (8.1.6). Indeed if DclR+ denotes
the union of sets of discontinuity points of A('), B(·) and u(·), then D has at most a
finite number of points per unit interval. Moreover:
a) For all fixed x ERn, the function t E \D -+ p(x,t) E RD is continuous and for
every 'tE D the left-hand and right-hand limits p(x,'t+) and p(x,'t-) resp., are finite vec-
tors in RD.
14 Theorem [Existence of the state transition map]. Under the assumptions and
notation above, for every triple (to,xo,u) E R+ x R n xU, the state transition map
is a continuous map well defined as the unique solution of the state differential equa-
tion 0), more precisely,
\>' t E \D x(t) = A(t)x(t) + 8(t)u(t)
where
where X(t) E R nxn and A10, A20 and FO are of class PC (R+,Rnxn), (A.3.6).
Show that, for all (to,XO)E R+xRnxn, (18) has a unique solution X(')E
(A.3.7), s.t. X(to) = Xo.
(Hint: by "stacking columns" (18) is convertible into a vector d.e.
8
22 Theorem [The structure of R (')]. Let C (R+,Rn ) and PC (1R+,RIlo) denote the
linear spaces of continuous functions from lRt into R n and resp. piecewise continuous
functions from R+ into RIlo. Consider any representation R (-) described by (1)-(5)
where for every input U(')E U, DcR+ denotes the union of sets of discontinuity
points of A('), BO, and u(·).
V.t.c. the state transition map s, given by (7), and the response map p, given by (8),
have the following structure:
a) For every triple (to,xo,u) E R+ x R n xU
b) [Linearity in (xo,u)]:
c) [additive property]:
For every quadruple (t,to,xo,u) E x x R n xU
30 y{-}=C(·)s(',1(j,xo,u)+D(')u('),
where all functions on the RHS are at least piecewise continuous on R+: the latter
clearly defines the response map p(',to,xo,u) e PC (1R+,IRIlo).
y) the partial maps s(t,1(j,xo,Su) and p(t,1(j,xo,Su) are called the zero-input (z-i) state
transition map, resp. the z-i response-, because of property b) they have the property
that, for fixed (t,1(j) e R+ x R+ the maps
and
are linear; hence by matrix representation Theorem (A5.3) they are representable by
matrices. Therefore there exists a matrix <I>(t,1(j) e R nxn S.t.
31
32
the matrix function (t,1(j)e R+xR+ <I>(t,1(j)e R nxn is caHed the state transition
matrix (map) and will be studied below.
S) The partial maps s(t,1(j,S,u) and p(t,to,S,u) are called the zero-state (z-s) state
transition map, the z-s response, resp. Because of propeny b) they are linear in
u(·) e U, the benefits of this will be discussed later.
e) The additive property c) means that both the state transition map and the response
can be calculated by adding their z-i and z-s contributions.
33 Proof of Theorem (22). By comments (29.0.) and (29.y) property a) holds and
property c) follows from property b). So we have to show property b), i.e. linearity
relations (25) and (26).
(25): Call the LHS of (25) x(t) and the RHS o.IXt(t) + o.2xZ(t) with xi(t) := s(t,to,xOi,Ui),
10
i = 1,2. Observe that by Theorem (14), xO and xk) for i = 1,2 are the unique solu-
tions of the d.e. (I) for given (to,alxOI + ul + (to,xOi,Uj) (i=I,2), resp.
Therefore the latter xl) can also be combined s.t. is a solution of (1)
for the fonner given triple. Since the solution is unique, (25) follows.
(26) is obtained by combining (30) and (25). (Note that the read-out equation (2) is
linear in (x,u).) •
34 Exercise [Linear matrix d.e.'s, cfr. Exercise (17)]. Denote by X(·,to,Xo,F) the
unique continuous solution of matrix d.e. (18) for a given triple
(to,XO,F)E s.t. X(to)=X o.
<1>(... ) is called the state transition matrix. Therefore by setting Xo = Ej, (the ith stan-
dard unit vector), we get (using Theorem (14» that, for all to E for all i E n.
x(·)= <1>(·,to)Ej is. for x(to) = Ej' the unique continuous solution of the homogeneous
linear d.e.
Hence, since <I>(·,to)ej is the ith column of the matrix <1>(' ,to) , we have
37 Fact. 'V toE <I>(',to): R+ R nxn is uniquely defined as the unique con-
tinuous solution XO : R+ R nxn of the homogeneous linear matrix d.e. t
38 X(t)=A(t)X(t) a.a.tE R+
s.1.
39 X(to)=I,
or equivalently,
40
a <I>(t,to) =
at A(t)<I>(t,lo) a.a.1. E R+,
s.t.
41
•
42 Comment. If X(·) is the solution of (38) S.t. det X(t) 0 'V t E R+, then it is
called a fundamental matrix of (36). The following shows that 'V to E R+ <1>(' ,to) is
a fundamental matrix of (36).
43 Property. Consider the d.e. (38). If there exits a to S.t. detX(to) 0, then
detX(t) 0 'V t E R+.
t a.a.1. means "almost all I." This is to remind ourselves that the LHS of (36) is
not defined for those t where AO is discontinuous.
12
[0 ro(t) ]
(5) A(t)= -ro(t) 0
f
where O(t)= ro(t)dt»
o
[
Verify that <1>(1,0)=
[
e(a-I)I cos 1
-e
(a-I)I'
sin t
e- I sin t
e- I cos t II
(Example due to Marcus and Yamabe, 1960).
46 Exercise. Let MO, NO, PO be Cl-functions, (A.3.8), into JRnxn . Using stan-
dard calculus and linear algebra (note carefully the order of the factors), show that
d . .
- [M(t)N(t)] = M(t)N(t) + M(t)N(t)
dt
d . . .
- [M(t)N(t)P(t)] = M(t)N(t)P(t) + M(t)N(t)P(t) + M(t)N(t)P(t).
dt
Assuming that det M(t) yt. 0 V t and noting that M(t)· M(I)-I = I, verify that
13
47 Exercise. Let t X(t) E R nxn be any fundamental matrix of (36). Show that
o
010 <lJ(t,to)=-eI>(t,to)A(t).
<lJ(t,to)= I + J A('t)<lJ('t,to)d't
10
Xo(t) = I
t
r r
commute for all t E 1R+. Show that:
r [!
a) For all k = 0.1.2....
I Hl I
exp M =
k=O k!
,) :t [<xP [ A(')d' l l
A(t) . <xp [ A(,)d, <XP [ A(,)d, A(t)
50 Exercise. Show that for all t E A(t) and f A('t)d't commute if one of the
10
following conditions holds:
.
X(t) = Al (t) . X(t) + X(t) . A2 (t) * X(to)=X O
is
5. [Splitting]. If X(t) is any fundamental matrix of x(t) = A(t)x(t), then for all
t,toE R+,
6. [Determinant].
I
and
Hence, in (56), LO and RO satisfy the same d.e. and the same initial condition:
hence L(t) = R(t) 'V t E IRt.
4. By comment (42), for all to' <1>(',to) is a fundamental matrix and therefore for all
t,to <1>(t,lo) is nonsingular. Now set t = to in (56):
16
5. For (58) note that both sides satisfy the same d.e. and the same initial condition.
For (59) note that in (58) both factors are continuous. (exercise:
X('t)-l-X('or l =X('t)-I . [X('o)-X('t)] . X(,O)-I •... ).
6. Let .1(t) := det <1>(t.lo) and write for simplicity <1>(t) for <1>(t.'o). Observe that
Now. by calculation.
. 1
.1(t) = lim -h [.1(t+h)-.1(t)] = tr A(t) . .1(t) •
where
.1('0) = 1.
Therefore we obtain
•
l
I
.1(t) =det <1>(t.lo) =exp tr A('t)d't.
10
63 Exercise. Suppose that you have established (58). Show that (58) implies (56)
and (57).
R+ if and only if t --+ I tr[A('t)]d't is bounded from below on R+. or equivalently there
o
exists a constant k E R S.t.
17
!tr[A (t)] dt
t
k V tE 1R+ .
6S Exercise. Consider
x(t) = [A(t) + B(t)]x(t) , (a)
where A(') and B(') are of class Let <l>A(t,l(,) be the transition matrix
of x(t)=A(t)x(t). Let
66 Exercise. Let the constant matrix ME R nxn commute with A(t), Vte Rt;
then M commutes with <I>(t,to), V t,to E R+.
(Hint: consider [McI>(t,to) - <I>(t,to)M].)
t
= <I>(t,to)xo + f cI>(t,t)B(t)u(t)dt
10
s(t,to.e,u)
18
72 yet) = p(t,to,xo,u) =
I
f
C(t)q,(t,to)xo + C(t) q,(t,'t)B('t)u('t)dnD(t)u(t)
10
p(t,to,9,u)
For fixed (t,tO)' S.t. t to, sand p depend on U[to.l) t rather than u(·) (for t :s; to
upon u[I,Io)'
y) (72) is a consequence of the substitution of (71) in the read-out map, see (30);
moreover in (71) the first term on the RHS is known, see (31) where q, is defined by
Fact (37). Hence we are reduced to the
By Theorem (14) it is sufficient to prove that the RHS above denoted by z(t) satisfies
d.e. (1) S.t. x(to) =9, or equivalently z(') must satisfy the integral equation
t t
x(t) = f A('t)x('t)d't + f B('t)u('t)d't V' t E R+.
10 10
Now by Fact (37)
t
f
q,(t,'t)=I + A(cr)q,(cr,'t)da,
1
whence
19
77 z(t) := f
10
10 10 10
t t
= f B(t)u(t)dt + f A(t)z(t)dt .
10 10
Q.E.D.
Note that the conversion of the double integral has been obtained by changing the
order of integration and subsequently replacing a by t and t by a. •
Comments. a) The proof above uses the modern integral equation point of view
with an application of Fubini's theorem ([Rud.l,p.l50]) for changing the order of
integration. (Its application is justified here by the fact that the integrand of the double
integral is PC on a bounded triangle of JR2; hence it is bounded there and the double
integral converges absolutely.) Notice that nowhere derivatives have been used: a pol-
icy that is important in distributed systems (semi-group systems [Cur.1] and stochastic
d.e.'s driven by white noise [Kwa.l]).
/3) A more classical proof uses differentiation t under the integral sign, e.g.
[Rud.3,pp.236-237] viz.
d af
f f(t,t)dt=f(t,t) + Iof :;-
t t
-d (t,t)dt.
tlo Qt
t More precisely one should read "at almost all t e R+" (except at points of discon-
tinuity of (t,t».
20
d
f10 -:;-
t
z(t) = B(t)u(t) + <1>(t,t)B('t)uCt)dt
ut
= B(t)(t)u(t) + A(t)z(t) .
Since z(to)=e, obviously z(') satisfies d.e. (1) for x(tn)=e. Hence by the uniqueness
of the solution s(t,to,e,u) = z(t) for all t E 1R+. •
81 x(t)=<1>(t,to)x(to) = 9.
During the interval [t, t+dt) we obtain from the d.e. (1) and (81)
::::< B(t)u(t)dt,
where the approximately equal sign " ::::< " reminds us that higher order terms in dt
have been neglected. Since u(t) = e on [1: + d1:,t] we finally obtain using (82)
:::: <1>(t,Hdt)B(t)u(t)dt.
____ ________________ __
I I
I I
T T+dT t
Fig. 2.1. An elementary input.
21
Since for fixed (t,to), x(t) is linear in u(·), we are allowed to sum over all elementary
inputs of the previous form when u(·) is a step junction, creating the (integral) sum
I
x(t) = s(t,to,e,u)= cl>(t,t)B(t)u(t)dt.
10
Notice that each elementary contribution B(t)u(t)dt at time t is mapped by the transi-
tion matrix q,(t,t) into an elementary (state) contribution q,(t,t)B(t)u(t)dt at time t.
83 Exercise. [Variation of constants approach for solving (1)]. Consider the d.e. (1)
for x(to)=xo where AO, BO, and uO are assumed to be continuous. Write
where IRn is an unknown function. Substitute in (1) and
obtain that Hence rederive the solution given by (71).
84 Exercise. [Linear matrix d.e.'s cf. (17),(34),(51)]. Consider the matrix d.e.
18 .
X(t)=A 1(t)X(t)+X(t)A2(t) +F(t)* tE R+
86 Fact. [State composition property). Let s denote the state transition map (7), of
any system representation R (-) given by (1)-(5).
V.l.c.
For every fixed u(·) E U,
88 Comments. a) For every fixed u(·) E U, the state x(t2) can also be obtained
by proceeding from any other state x(tl) := s(tl'to,xO,u) on the trajectory through
x(lo)=xo·
For fixed u(·) E U and x(lo) = xo, s(t,lo,x(lo),u) defines a two-parameter family of
maps
With "0" denoting composition of maps [viz. (fog)(x)=f(g(x» for all x. see (Al.7)].
(87) reads
i.e.
y) Of course, if u(-) = Su. then (87) is the composition property of state transition
matrices (56).
91 Proof. Set t2=t. xI := s(tl.tO.xO,u) and call1(t) (r(t»; the LHS (RHS resp.) of
the second equality of (87).
Now 10 satisfies d.e. (1) S.t. x(t\)=xl and so does r(·). Hence, by the uniqueness of
the solution of (1) for a fixed uO (see (14», we have I(t)=r(t) for all tE IR+. •
zero. U.l.c. by (71) the state t will display a jump at t such that x(t+) *- x(t-)=e
and Vt t
94 = J <l>(t,a)B(a)Ej' a(a-'t)da
't-
=<l>(t,t)B(t)Ej'
= [C(t)<l>(t,t)B(t) + D(t)8(t-'t)]Ej'
Hence, if j=I,2, ... ,ni (i.e. "the unit impulse is successively applied at all inputs"), then
we fonn the ni columns of an no x ni real valued matrix H(t,t). The matrix function
defined by
V t t
96 H(t,t) := { 0 V t< t
C(t)<l>(I,t)B(t)+D(t)a(t-t) V t t
97 H(t,t) = { 0 'ltt<t.
p(t,t-,B,a( . -t»
98 h(t,t) := { 0 Vt<t.
t Taken to be right-continuous.
24
From the linearity in uO E U of the z-s response of R (.), (72), it follows by (97) that
the z-s response at time t due to the input uO applied at to reads
100 Comment [I/O point of view). In certain applications for all initial times to
under consideration x(to) = e and R (-) becomes an I/O map described by the super-
position law (99). Note that it reads approximately
i.e. each elementary input (Fig. 2.1), seen as a vector of impulses [u/t)dt 0(' -t)
at time t, is mapped by the impulse response matrix into an output contribution
H(t,'t)u(t)dt at time t t (Fig. 2.2). The output at t is the sum of the latter.
101 Exercise. For the time-invariant case R = [A,8,C,DJ. i.e. A,8,C,D are con-
stant matrices and <1>(t,lo) =exp[A(t-to)] (50).
a) Show that the impulse response matrix depends only on the elapsed time t-t. (It is
then customary to write H(t-t) instead of H(t,t); also w.l.g. t=O because H(t-t)
H(t,t) = H(t-t,O); therefore specification can be done by H(t) = H(t,O).)
u(TldT
to T
Fig. 2.2. z-s response: contribution of u(t)dt to y(t).
t See Appendix C.
25
c) The z-s response, (99), where w.l.g. to =0 (see later) is the convolution of HO by
u('), i.e.
•
A A A
108 Adjoint linear homogeneous d.e. 'so To the linear homogeneous d.e ..
where x(t) e R n and A(') e PC (R.,Rnxn), we associate the adjoint differential equa-
tion
where A * is the Hermitian transpose of A (for A real this is simply the transpose of
A) and x(t) e lRn. Note that (Rn,(','» is a Hilbert space with inner product
{x,y)=x y *
Hence for any matrix A e R nxn
is the defining relatio!! for A* e R nxn , the adjoint of A as a linear map (A.7.32).
Now, let «I>(t,to) and cl>(t,to) be transition matrices of (109) and (110) resp.; therefore
for all t,to e R.
112
-
x(to)=cI{to,t)x(t).
26
Proof. At a.a.t E 1R+, the derivative of the LHS of (114) is zero. To wit, by (109)-
(110),
v x(to),x(tl) E IRfi
(x(tl),lJ'x(to» = (o/*X(tl)'X(to».
Hence is the adjoint of lJ'=<1>(tl,to) (see (A.7.32». Note that 0/' "sends
the future x(tl) into the past x(to) ."
117 Fact. Under the assumptions and notations above, V t,toE lR+
118
For the last equality we used (112) and observe that at the outset x(t) was arbitrary.
Hence
* -
cP (t,to) = CP(to,t) .
where
128 Proof. for simplicity we suppress in our notation all the time dependences.
Now from (121)-(122) we obtain, using (1)-(2),
28
==
== :t (x,x)+(u,Y)-(Y,u).
where CO) denotes the usual vector inner product (v('t),w('t» = v ('t)* w('t), (A.7.1O); for
simplicity we shall also write L2 [t o ,t 1j instead of L 2([Io,ttl, JRIlo) or L2([Io,td, JR"').
Now by the density above and continuous extension, (see e.g. Comment we
may w.l.g. replace the input- and output-spaces (of class PC) of R (-) and R(-) by
L 2[Io,td. Hence, on [1o,tIl, R (.) induces a well-defined linear map
sending an (initial slate, input) pair into a (final state, output) pair.
Similarly on [lo,ttl RO induces a well-defined map
...
sending a (final state, input) pair into an (initial state, output) pair, (" P sends the
future into the past").
...
Note here that the domain and codomain of P and Pare JR n xL2[Io,td: this is a
product Hilbert space having the inner product (·,)lRn+ C·h. Hence from (127) (with
t=t 1) we have:
'v' (x(Io),iiO) E JRn xL2[Io,td = Domain of P ,
134 Corollary [Pairing for y=x]. Under the conditions of Lemma (126) let C(')=I
and DO = O. Then one has
t t
135 (x(t),x(t» + f (ii('t),x('t» d't= (x(to),x(to» + f (B('ttx('t),u('t» d't
where
x(t) = A(t)x(t) + B(t)u(t)
where
and 11·11 denotes the Euclidean nonn, (A.6.1). It follows by (71) that for fixed Xo the
cost J is a functional of uO, more precisely it reads
J(·):PC([to,tl],1Rnl) R:u(·) J(uC·».
Standard LQ.problem: minimize J(.) over all possible inputs u(·) of class pC,tt
thus solve
(3) In the expression of the cost J(.), (138), C(·)x(·) is the weighted state-trajectory
to be penalized and S is the final-state penalty-matrix; ( II C(·)xO 112 reads also
IICOx(·)1I 2=(x(·),QOx(·» with QO=C(·)*C(·) CO and S depend on the
problem at hand.
y) In many cases one uses a weighting matrix RO to penalize the input uO in the
cost J(.), namely, a tenn J(u(t),R(t)u(t) )dt is present (instead of JII u(t) " 2dt), where
10 10
R(·) = R(.)* > 0 is a matrix-valued function of class PC (which is positive definite
'It). However by introducing U(·)=R I12 (·)u(·) and (where R II2 (.)
is any square root of class PC of R(·» we have
145 Analysis. For simplicity we omit mostly all time-dependences in the notations:
e.g. x for x(t), C for C(t), etc.
where o(e) --+ 0 as I e I --+ 0 and oJ(8u) is the directional derivative (a functional
e
linear in ou). Equation (146) shows that for u to be a minimizer of J, (Le.
V Ou E PC, VeE R, J(u+eou) ?!. J(u», it is necessary that
147 PC.
Hence if e approaches zero from above then in the limit BJ(ou) ?!.O; similarly if the
approach is from below then 15J(ou) O. Hence condition (147) is necessary.
Now condition (147) is also sufficient. Indeed, by the linearity in (xO'u) of the
state transition map (71) of R 0, any perturbed input u+eou (with Xo fixed) generates a
perturbed state trajectory x+eox (where u and ou map into x and ox, resp., and
oxo=9). Hence, by calculation, using (138) and Exercise (141), VUEPC,
V OUE PC VeE R
where (a) BJ(Bu) is specified in (149)-(151) below and (b) by (138)-(139). J(Bu)
for allliu(')e PC. t Hence if ue PC is S.t. &(Bu)=O Vliue PC. then by (148)
with £= 1.
149 f
SJ(Su)= (u,Su)dt + f (c"'Cx.Sx)dt + (SxI.SxI)
10 10
151 Sxo=9.
Note that SJ(Su) is a linear functional in Su which we want to make explicit in Su for
expressing (147).
Now for (150)-(151) the pairing corollary (134) gives
Viie PC
where
153 + ii.
t When calculating J(u). Eq. (1) is integrated with x(to)=xo. whereas J(Su) is ob-
tained by integrating ()x = ASx + BSu, with Sx(to) = 9.
33
With this choice, (152) converts the directional derivative (149) into the explicit func-
tional
II
Recall from Step 1 that for u e PC to be optimal it is necessary and sufficient that
oJ(ou) = 0 for all OU e PC: choosing OU = u+B*x shows that this will happen if and
only if
The substitution of (156) and the choices (154) in (136)-(137) and (153) resp., finally
result in the following optimality characterization.
157 Theorem [Solution of the standard LQ-problem]. The: solution of the standard
LQ-problem (136)-(140) is given by the optimal input
u(t) =-B(t)*x(t)
where xO E C ([to,til, 1Rn) is defined as the partial state trajectory of the 2n-
dimensional two-point boundary value problem or [to,t 1] given by:
158 . *
x(t) = A(t)x(t) - B(t)B(t) x(t) ,
160
•
161 Comments. a) The homogeneous 2n-dimensional linear d.e. (158)-(159) is
usually called a Hamiltonian system and
A(t) -B(t)B(t)* ]
162 [
H(t):= -C(t)*C(t) -A(t)*
163 Optimal cost. We shall work on R+ with an arbitrary pair lo,t 1 E 1R+ s.t.
10 :s; tl·
Under these conditions, let us recall our substitutions, i.e.
+ li(t)
such that
Now using this interpretation of (158)-(160) in the pairing corollary (134) (with Xo
arbitrary), we obtain successively:
tl tl
= 2Jo(uC·»,
where (a) for the second equality we used (154) and (156) and (b) the last RHS IS
twice the optimal cost 10 of the standard LQ-problem on [lo,t1l starting at x(to)=xo
(indeed the pairing is perfonned, with to :s; tl' under the conditions of Theorem
(157». Hence we have obtained an optimal cost equation. •
35
The optimal cost equation (164) is useful for converting the solution of the
standard LQ-problems into a linear feedback law dictated by either the Hamiltonian
system or the matrix Riccati d.e. This result is given in the next two theorems, which
we prove immediately.
168
-
°
where XC-> and XC-> are nxn real matrix-valued functions defined on [O,td, with
det X(t) "# for all t E [O,tl]' as the unique solution of the backwards Hamiltonian
linear matrix d.e.
169 :t
[ X(t) ]
X(t) =
[A(t) -B(t)B(t)*
-C(t)*C(t) -A(t)*
1 [X(t)]
X(t)
= H(t) [ X(t) ]
X(t) , te [O,td
with
-
170 X(tl)=I and X(tl)=S,
b) On any interval [to,til cR+, the LQ-problem has the optimal cost 10 given by the
quadratic form
1 -
171 10="2 (X(to)X(to)-lxO,xo)
and generates the optimal closed-loop system dynamics described by the linear homo-
geneous d.e.
172 x(t) = [A(t)-B(t)B(t)*X(t)X(t)-1 ]x(t) t e [to,td
173 x(to)=xo,
Let
175 F(t) ;= B(t)*X(t)X(tr l E R nixn
then (i) F(') is independent of xo; (ii) (172) can be interpreted as describing the
dynamics (136) modified by a linear time-varying state feedback law u(t) =-F(t)x(t),
(see Fig. 2.3 below).
176 Proof of Theorem (167). The proof is done in two steps. The first step shows
that the partial solution X(t)E Rnxn of (169)-(170) is nonsingular for all tE lO,td: it is
based on the optimal cost equation (164) and may be skipped on a first reading. The
second step shows that the theorem holds.
that f II u(t) 11 2dt=0 and therefore u(t)=O for all tE [to,ttl. Therefore, since (158)
't
reads x=Ax+Bu (with u=-B*x), there results on ['t,t l ] x=Ax with x('t) = 9. Hence
x(t)=X(t)k=9 'litE ['t,td. In particular at t1 we have (using also (170»:
9=x(tt)=X(tt)k=k::l- -H-. e:
Step 2. The theorem holds.
Let (X(·),X(·» be the backwards solution of the matrix d.e. (169)-(170). By Step
1 we know that detX(t) 0, '*
'litE [O,ttl. Therefore for the given initial condition
x(to) = xo, (137), there exists a unique vector k E lRn s.,t. at to
177 xo=X(to)k.
is the unique solution of the Hamiltonian system (158)-(160) on [to,t 1]. Note here
that, since detX(t) 0 '*
'litE [O,ttl, by (177) and (178)
37
Hence by Theorem (157) for any [to,td c the solution of the LQ-problem will be
given by:
u(t) =-B(t)* x(t) =-B(t)*-X(t)k=-B(t)*-X(t)X(t)-lx(t),
where we used successively (158), (178) and (179). Note especially that the feedback
matrix function multiplying x(t) is a fixed function on [0,t 1] which does not depend
on the specified initial state Xo E 1Rn and on to E s.t. to :::; II. Hence conclusion (a)
holds.
Finally for conclusion (b) we note that by (164), (178) and (179) we have for the
optimal cost
2Jo= (x(to),xo) = (X(to)k,xo)
= (X(to)X(to)-lxO'xO) .
180 Theorem [Optimal LQ state feedback by the Riccati d.e.]. Consider the stan-
dard LQ-problem, (136)-(141), where the horizon tl is fixed and toE [O,tl) is arbi-
trary.
U.t.c. a) On any interval [to,td c R+ the LQ-problem is solved by the fixed linear
state feedback law
181 u(t) = -B* (t)P(t)x(t) t E [O,tl]'
where PO=P(.)* °
is the nxn real matrix-valued function defined on [0,t 1] as the
unique solution of the backwards matrix d.e.
182 . * * *
-P(t) = A(t) P(t)+ P(t)A(t) - P(t)B(t)B(t) P(t) + C(t) C(t) t E [O,td
with
183
b) On any interval [to,tt1 c R+ the LQ-problem has the optimal cost JO given by the
quadratic form
184
38
and generates the optimal closed-loop system dynamics described by the linear homo-
geneous d.e.
185 x(t) = [A(t)-B(t)B(t)*P(t))x(t) t E [to,td
186 x(to)=xo,
190 Note that, for any given Hermitian pet), the RHS of (182) gives a Her-
mitian -pet). Therefore, since (by (183» P(t1) is Hermitian, P(·) will be Hermitian.
Moreover, since the cost is nonnegative and cost formula (184) will be shown to hold
for all Xo E R n and all to E [O,td, PO will be positive semidefinite. Therefore on com-
paring the statements of Theorems (167) and (180) we are done if the solution of the
d.e. (182)-(183) reads
191 pet) = X(t)· X(t)-l for t E [O,td,
where (X(·),X(·» is the backwards solution of the Hamiltonian matrix d.e. (169)-(170).
Now note that (191) and (169)-(170) imply
=-[-CCX-A*XJX-l + P[AX-BB*XJX-l
Hence (191) defines a solution of the d.e. (182)-(183). Conversely if PC') solves
(182)-(183), then the transition at t J of the closed loop_d.e. (185) is well
defined. Call the latter XO and set XO := P(·)X(·). Then (X(·),X(·» is the unique
solution of (169)-(170) where (191) holds, (exercise). Hence (191) defines a bijection
between the solutions of (182)-(183) and (169)-(170) resp., and we are done.
e x
! }L..___
- 8·....----....
U_ _ "_'"_' " ; - - - - '
Fig. 2.3. State-feedback realization of standard LQ-problem: F(t) is the optimal state
feedback matrix.
39
•
192 Comments. ex) Theorems (167) and (180) show that, for a fixed horizon t l ,
the standard LQ problem on any interval [to,td c is solved by computing the fixed
state feedback matrix
193 F(t)=B(t)*P(t) Vte [O,td .
The latter, by closing the loop as in Fig. 2.3, will generate the optimal input
194 u(t)=-F(t)x(t)
which steers the system on its optimal state trajectory through x(to) = xo.
The nxn matrix P(·) in the feedback gain (193) can either be computed from the
linear Hamiltonian matrix d.e. (169)-(170) through
196 Concluding remarks. Five features of our solution of the LQ-problem are
worth emphasizing.
1. A local optimality analysis, (146), leading to the annihilation of the directional
cost derivative along any direction, (147). A generalization of this technique
involving a Lagrangian cost derivative leads to the maximum principle of optimal
control e.g. [Var.I,Fle.I,Alex.l]. The two-point boundary value problem of
Theorem (157) is typical for expressing that necessary principle of optimality.
2. A global optimality analysis, (148) et seq., leading to the sufficiency of condition
(147) for global optimality. This combined with point 1 results in the necessary
and sufficient condition of Theorem (157); the interested reader will easily derive
from (148) that
Vue PC Voue PC Vee R
this shows the convexity of the quadratic cost in u (that is, the tangent space is
always below the cost, e.g. [Var.I]): in optimal control, convexity is the driving
force for global optimality.
3. A systematic study of the optimal cost, (164), to obtain the solution in feedback
form: this borrows ideas from dynamic programming, e.g., [Var.I]. That theory
derives the Riccati d.e. by the Hamilton-Bellman-Jacobi equation (" backwards
40
2.2. Applications
This section contains various applications using a system representation
R (.) = [AO,B('),C('),D(')]: the variational equation encountered in linearization, exam-
ples of nonlinear control, dynamic optimization and periodically varying differential
equations.
x(t) = f(x(t),u(t),t)
where x(t)e R n and u(t)e R n; and f:RnxRn;xR+ Rn.
Suppose that for a given to' Xo and a given input u(-) we have calculated the
corresponding state trajectory, say xO. What happens if for given to'x o ' we now have
Uc')+ouO as input, where ou(t) is small for all times of interest? The new input will
give a new (perturbed) state trajectory xO+axO and we would expect that oxO will
be small, say of the same "order" as ou(')' This is usually the case; for precise condi-
tions see e_g. lDes.2J, [In. 1].
Proceeding formally, we obtain
2 *(t)+ox(t) = f(x(t)+ox(t),u(t)+ou(t),t) .
Since t x(t) is known, this is a differential equation in ax('); and ox(to)=9 since the
perturbed trajectory starts from Xo at to'
For each fixed t, let us expand f in a Taylor series about the point (x(t),u(t),t):
f(x(t)+ox(t),u(t)+ou(t),t) = f(X(t),U<:t),t)
41
+ higher-order tenns .
4 X(t) = f(X(t),u(t),t) ,
8x(to)=9
where A(') and B(') are the state and control Jacobian matrices defined in (3). Equa-
tion (5) is called the variational equation about the trajectory X(.) generated by
(to,Xo.l(». Note that since we dropped higher-order tenns, (5) gives an approximation
to the difference between the perturbed trajectory and the reference trajectory X(.).
This approximation is, however, extremely useful in science and engineering (e.g.
design by optimization).
*7 Exercise. Assume that in (2) auo reads au(t)=ev(t) where v(') is a specified
function bounded on R+ and e is a real parameter S.t. ee (-£o,eo) for some small
eo> O. Assume also that xo ' the initial condition at to' has been changed into xO+8xO
(thus in (2), 8x(to) = axo 9). Note that under these conditions, (a) Equation (1)
reads
8 x(t) = f(x(t),t,e) := f(x(t),u(t)+ev(t),t),
indeed u(·) and v(·) are specified, hence the RHS of (8) depends on t, x(t) and e, and
(b) if 8xa..= e and e=O, then X(t)=x(t) where X(.) satisfies (4) and X(to)=xo. Assume
now that f in (7) satisfies the condition of Theorem (8.2.6), whence it can be applied.
Show that at any fixed time t> to 8x(t):= x(t)-x(t) satisfies
42
+E J
10
10 + J
10
where
a) <1>(',') is the transition matrix of the state Jacobian matrix A(t)=fx(X(t),u(t),t) of (3)
and B(t) = fu(X(t),\(t),t) is the control Jacobian matrix of (3).
b) [ / II II ] 0 as (9,0).
11 Comment. The solution of (5) with gives (10) without the error
term o(Sxo,e): (10) asserts, in addition to (5), that at any fixed time t > to the error is
*12 Exercise.
Let Eo> 0 be small. Let v E IRn, be a fixed control perturbation
value. Let t be a fixed time S.t. t > to' For all £ E [0'£0] define in (2) (where
Sx( to) = 9) as
'r;f t E [t,t+e)
13 elsewhere on IRt .
v 'r;f t E [t,t+e)
14
\(t) elsewhere
[" short pulse" perturbation]. Assume that 'r;f e E [O,Eo] (1) has a continuous solu-
tion due to (to,xo,u E('», satisfying
16
Call this solution x£(-) and note that (a) for e=O xO(t) = X(t) (the solution due to
and (b) for any fixed e e [O,eo] the composition property holds, i.e.
Now assume that the RHS of 0), namely, f(x,u,t) and its Jacobian fx(x,u,t) are con-
tinuous on R n x R n, Then a careful analysis, [Alex. 1,pp.334 et seq.], gives:
where (o(e)/e) -7 0 as e -7 O.
•
[Hints: Take into account the explicit relation on lR+: for t using (17),
44
Hence if tl = to' (thus x£(t\) = xo), and t E [to,t], then (20) follows.
Now at t = t+E, by (14), (15), (18), and the continuity of xO,
= Ef(X(t),v,t) +O(E)
and
X(t+E) - X(t) = Ef(x(t),u(t),t) + O(e) .
Hence
23 xE(t+e)-X(t+e)=e[f(x(t),v,t)-f(X(t),U(t),t)] +O(e) .
Finally, for t>'t+e (for e sufficiently small, this is any time t>t), observe that on
['t+e,t] uE(·) = U('), hence (see (22»
xE(t) = 'I'(t,'t+E,XE('t+e),O) .
Moreover,
x(t) = 'I'(t,t+e,X(t+e),O) .
xE(t) = v
X(t) + dx (t,t+E,X(t+E),O)' ["t;(t+E)-X(t+E)]
v v
+ o(lIx£(t+e)-X(t+e)II) .
Use now finally (19) (with tl = 't+e) and (23) to obtain (21).]
*24 Comment. The control perturbations ouO in Exercises (7) and (12) are small
LOO-perturbations and small L I_perturbations resp. Their effect on the trajectory is
different. Compare (10) and (21) where Ox(t) = xE(t)-X(t). In (10) a uniform effect is
observed. In (21), at time t, the "short pulse" causes a sudden change in velocity, and
for t>t, its effects are propagated along the trajectory by <l>(t,t).
Problem. The movement of a satellite is described by the evolution of its state (posi-
tion, velocity, ... ) on a fixed time-interval (to,td c R+ and described by the nonlinear
31 x(to)=xo,
where x(t) e R nt, Xo e R n is a specified initial state and f: R n x R nj x R.t Rn. The
function fin (1) is obtained by applying the laws of Mechanics to the chosen model of
the satellite.
We are also given a fixed final state Xl e R n at tl and we want to find a control
u(')e PC S.t. we reach the state Xl at tl' i.e.
find u(')e PC S.t.
32 x(t1) = s(1,t1'to,u) = Xl ,
produces a small state error XI-X(t l ) at time t i . The corresponding variational equa-
tion (5)-(6), i.e.
5 i)x(t) = A(t)i)x(t) + B(t)i)u(t)
6 Sx(to)=O
Neglecting the error and noting that control corrections i)U[Io.I,] e PC produce state
corrections ax(tl) given by
I,
33 ox(t,) = J<l>(t;t)B(t)ou('t)dt ,
to
This map is called the reachability map with properties described by the following
nxn real-valued symmetric p.s.d. matrix
I,
37 Wr(lo,t,) := J<l>(tl,t)B(t)B(t)*<l>(t1,t)*dt;
to
the latter matrix is called the reachability grammian. Now, by theorem (8.2.12), Lr
and W r(Io,II) are shown to have the same range, i.e.,
Hence
Note that (33), (35) and (36) suggest to refonnulate problem (34) as:
Solve for ou(-) the linear equation
+This is not automatic, e.g. for A = diag[ -1,-2] and b * = [1,0], del W /1 0 ,1 I) = o.
47
I,
42 = J<1>(tl ,'t)B('t)ou('t)d't ,
10
for some xI E lRn : note especially that the first and second equality follow from (41),
(38) and (37), while the last equality follows from (41) and (36) defined by (33). Now
(42) indicates a solution to problem (34), viz.
Indeed this follows by back substitution. Hence a solution to problem (34) is given by
the following
W(to,to)=0
il=-A(t)*x(t) x(tl)=xI
where
48
11
Comments a) No control correction is needed if and only if the guess U(.) is exact.
i.e. xl =S(tI.to.XO,lJ).
b) It can be argued from this that. under certain conditions. a convergent iterative
scheme based on successive control corrections (43). i.e. ui+'(·)=ui(·)+/)uj (·) j=I.2....
will converge to a control u(') which solves the original problem (32). i.e.
xl = s(t,.to.xo.u). (cf. convergence of Newton-Raphson's iteration scheme.)
31 x(Io)=Xo
where
51 <I> : R n -+ R : x -+ $(x)
and
Comment. The cost (50) depends on the final state, which itself is detennined by
the control you use: e.g. the state may be the (position, velocity) vector of the center
of a mass of a rocket, -cp(x) may be the height and u(') the thrust you want to apply
on [to,t 1] to maximize height.
Now, usually this problem has no closed-fonn solution. However, from a practi-
cal point of view, we may start from a reasonable guess and try to improve upon it.
Thus we assume that we have a guess, i.e. a control and its corresponding state
trajectory i.e. the function defined by
X<to)=xo·
Suppose that in our search for a "better" control we change u into UrBu; the trajectory
changes fonn x to x+Bx where, after dropping higher-order tenns,
6 Bx(to)=9 .
The first equation is the variational equation about the trajectory X('), (see Sec. 2.2.1).
Let cl>(t;t) be the state transition matrix corresponding to (5); then, by (6),
11
54 J
Bx(tl)= cl>(tl,'t)B('t)Bu('t)d't.
to
Moreover, standard analysis reveals that the directional derivative BJ(Bu) at U('),
defined by
55 J(UreBu)=J(u)+eBJ(Bu)+o(e) ee IR
is given by
56 BJ(Bu) = ,Bx(tl)}'
i.e. <!lxl is the row gradient of cp at Xctl)' Using (54) in (56), we obtain successively
50
t,
51
58 YCt) := B(t)*x(t)= .
Therefore, by (51), the directional derivative oJ(ou) (a linear functional in ou) has the
explicit form
59 oJ(ou) = J(y(t),ou(t» dt .
to
Remember that we want to minimize our cost J(u) having at our disposal J(u). Note
also that, by (55) where (0(10)/10) -) 0 as E -) 0, there exists an EO>O such that for all
10 E (0,100]
J(UtEOU) < J(u)
if ou(·) satisfies
60 oJ(ou) < 0,
i.e. ouO is a direction of descent. Indeed by (55) and (60) with E > 0 small
since aJ(au) dominates (o(£)/e). Hence the biggest decrease in the cost from Umay be
expected in a direction of steepest descent au('), which is the solution of the problem
Note that since we are interested in a direction, length must be kept constant; we
r= I II
I,
choose as length the L2-norm, (A.7.1O), given by lIauOII Bu(t)11 2dt. Note,
10
also that the constant a is arbitrary and we may set
62 a=IIY(')112
I,
where y(.) is given by (58). Now using the L 2 -inner-product (f,gh= (f(t),g(t»dt, it f
10
follows immediately that the solution of (61)-(62), i.e a direction of steepest descent,
reads
63 Bu(t) =-y(t) =-B(t)*<1>(tl,t)* cjl,tl t e [to,td.
BJ(Bu)=(Y('),8u('»2 -lly(-)lIi=-(Y(-);Y('»'
Finally, having our direction of steepest descent cSu(·) our search for an improved
guess reduces to try to find an 00 such that, at the new control u+-ecSu, J(U+-£cSu) is
significantly smaller than J(u). The determination of this £>0 is called a line search,
e.g. [Oro.1], and belongs to the art of computing.
where the matrix A(')e PC [R+,Rnxn] is a periodic function with period T (T-periodic,
for short), i.e.
A(t+T) = A(t) 'tte
From the periodicity of A('), direct substitution into (65) shows that if <1>(t,to) is the
transition matrix of (65) then the map t <1>(t+T,to) is a fundamental matrix of (65).
52
68 <I>(T,O)= C,
a) PO is T-periodic,
75 tE
Proof. a) From (72) and then using (67), (70) and the properties of the exponential.
we see that for all t E
P(t+T)=<1>(t+T,O) exp[-B(t+T)]
=<1>(t.O) exp(-Bt)
=P(t) ,
=P(t) exp[B(t-to)]p(to)-l.
d) If x(t) is any solution of (1) then, for some IoE R+ and XOE ern, x(t) = «1>(t,tO)xO'
whence by the state transformation (74), with defined by
= p(tr1x(t)
= p(t)-l<1>(t.to)xo =
where the last step follows by (73). Hence f9r every to E for every corresponding
state := p(tor1xo. defined by (74), satisfies the d.e. (75). •
54
76 Exercise. Consider the periodically varying d.e. of Exercise (2.1.45) item (7)
and choose the period T = 2x. Compute the matrix B, (69), and the transition matrix
cI>(t,tO)' [Hint: use (73).]
77 Exercise. Consider the periodically varying d.e. (65) with BE fCnxn given by
(69). Show that,
where (1) for all to the matrix-valued function t P(t,to) is T-periodic and (2) P(t,tO)
is nonsingular for all t,to;
79 tE R+,
80 Remark. By (78) the transition matrix of a periodically varying d.e. is the pro-
duct of a periodically modulated amplitude P(t,tO) by an exponential exp[B(t-tO)]'
81 Remark. The matrix-valued functions pet) and P(t,tO) defined by (72) and (78),
resp. are related by
82 pet) = P(t,O) .
•
CHAPTER 2d
Introduction
This chapter starts by discussing how to obtain a discrete-time linear system
representation R d(') from a continuous-time system. The state and output trajectories
of Rd(') are then derived and structured. The dual-system representation Rd(') is next
defined and related to R d(') via a Pairing lemma. We then handle finite horizon linear
quadratic optimization and end with coverage of periodically varying recursion equa-
tions.
Physical Setting
For most engineering problems, the basic laws of Physics are those of Mechanics
and Electromagnetism: Newton's laws, Lagrange's equations, the Navier-Stokes equa-
tions, Maxwell's equations, Kirchhoff laws, etc. Each of these laws describe
continuous-time phenomena. At present it is cost-effective to manipulate signals in
digital fonn. For this purpose, the continuous-time signals are periodically sampled by
an AID converter (analog-to-digital converter) and transformed to digital fonn: the out-
put of the AID converter can be thought as a sequence of numbers. This sequence of
numbers may be manipulated by a digital computer (controller) and the resulting
sequence of numbers must be restored to analog fonn by a D/A converter; indeed, the
analog fonn is required to actuate the physical devices.
In control problems (robots, measuring instruments, airplanes, satellites, process
control problems, etc.), sensors measure physical variables (e.g. position, velocity,
acceleration, temperature, pressure, voltage, etc.) and the AID converter transfonns it
to digital form. After treatment by a digital computer (controller), the digital signal is
restored to analog fonn in order to operate the actuators (motors, valves, reaction
wheels, ailerons, ... ).
In communications systems, the signal from a microphone or picture tube is sam-
pled and transmitted in digital fonn. At the receiving end, the signals are restored to
analog fonn to actuate loudspeakers or TV tubes.
Throughout this chapter we neglect the quantization error: that is the error occur-
ring in the process of transforming a sample of an analog signal into a finite sequence
of binary digits. We assume that this error is a smalI noise that will barely affect the
performance of the system.
Throughout this chapter, we assume that the sampling period h>O is given;
roughly speaking, the sampling frequency 1/h must be definitely larger than twice the
highest frequency present in the signals being sampled.
where ueO and YeO are the continuous-time input and output, resp., A(') and BO are
piecewise continuous and, for simplicity, we assume CO and DO continuous, with
A(·) and B(') continuous at the sampling points: A(kh+) = A(kh-),B(kh+) = B(kh-),
'ike N.
Suppose we want to drive S by a digital input specified by the sequence
ud('):= (Ud(kh) J;: we apply this sequence to the D/A converter which, by assump-
tion, produces a piecewise-constant continuous-time input uc(') to S. (The D/A con-
verter is assumed to behave as a zero-order hold circuit.) Thus
The input ucO of S produces the continuous-time output YeO, which is fed into an
AID converter to product the output sequence Yd(')::::: (Yd(kh) J;. (See Fig. 2d.1.)
(k+l)h
6 Ad(k) :=: Cl>«k+l)h,kh); Bik ):= J Cl>«k+l)h,t)B('t)dt
kh
7 ; Dd(k):= D(kh)
where x(kh) is the state of S when it is driven by ud ' as shown in Fig. 2d.1. Finally,
let us write ud(k) for ud(kh), then
ke N.
The linear time-varying equations (10) and (11) relate the output sequence (Yd(k)];
and the state sequence (Xd(k) ]; to the input sequence (Ud(k) ];.
Remark. For the physical set up of Fig. 2d.l -- a continuous-time system preceded
by a D/A and followed by a synchronized AID converter -- the matrix Ad is always
nonsingular by (6). The same holds for series and/or parallel connections of such sys-
tems. However this is not necessarily the case for feedback interconnections, (e.g.
deadbeat controllers).
Comments. IX) Since in the system of Fig. 2d.l, S is driven by the piecewise-
constant input ucO, S has a state trajectory xO : R+ --t R n which, at time kh, coin-
cides with xd(k), the state of the discrete-time representation R d(') defined by (10) and
(11). In fact, x(·) is discontinuous at the sampling points as well as at the points of
discontinuity of A(·) and B(·).
1 A(k),B(k),C(k),D(k) ke N
Remarks I. It is understood that the recursion equation (2) is run forward, calculat-
ing the later state x(k+l) in terms of the input u(k) and the previous state x(k) is
always possible.
II. In most cases, A(k) e Rnxn is nonsingular for all k (see eq. (0.6) above).
However, some engineering systems are "deadbeat," and A(k) is singular: in that case,
x(k+l) and u(k) in (2) do not define x(k) uniquely, and we cannot solve (2) backward
in time.
III. To emphasize the analogy between the continuous-time representation R (-)
treated in Chapter 2 and the discrete-time representation R dO we number the main
properties of R dO using the same numbers as those labeling those of R ('). For brev-
ity. we do not state obvious definitions. (Most of the time. the proofs are by inspec-
tion. If not. a hint is given.)
Remark. It is only when A(k) is nonsingular V kEN that (40) can be solved for
<1>(k,ko) in tenns of <1>(k+l,ko). In that case, <1>(',') is defined by (40) and (41) on all of
NxN.
k-l
y(k) = C(k)<1>(k,ko)xo + C(k) L <1>(k.k'+1)B(k')u(k') + D(k)u(k) .
k'=ko
96 Impulse response.
The zero-state response (Y(k»)k;; due to the input (U(k) )k;; is
k
y(k) = L H(k,k')u(k')
k'=k o
where the impulse response is the matrix-valued sequence (k,k') -4 H(k,k') given by
114 Consider the equation together with x(k+ 1) = A(k)x(k); for all solutions,
k (x(k),x(k» is constant.
The state transition matrix of the adjoint equation is given by
Note that since A(k) is not necessarily invertible, the dual system runs naturally back-
wards in time.
Proof. Compute: use first (121) and (122), then (2) and (3):
(x(l),x(l» + (Y(l+l),u(l)
= (A(l )*x(l + 1)+C(1 )*u(l + I),x(l» + (C(l )*x(l + l)+D(l )*1I(l + 1),u(l»
=(x(l+I),x(l+I» + (u(/+l),y(l».
*129 Comment: adjoints. It rums out that it is easier to consider pointwise maps:
Fd: (IR n xIR n,) (IRn x 1R""), (IRnxlRllo) (IRnxlR ll ,) where
61
F d : (x(l),u(/» (x(l+l),y(l)
F d* : (x(l+l),ii(l+l) (x(l),Y(i+l»
the maps being, respectively, specified by (2) and (3), and (121), (122)0
= (x(ko),x(ko)} +
k-l
L *
(B(l) x(1 + l),u(/)} 0
leo
148 As in the continuous-time case, it is easy to check that the control u(-) is a glo-
bal minimizer of J if and only if
BJ05u)=O V Bu
where
where Bx is related to by
62
we obtain
k,-I
Bl(Bu) = L (u(k) + B(ktx(k+l),Bu(k)}.
leo
161 Comments. ex) Naturally, the recursion equation (Le.) of the system goes
forward and the r.e. of the adjoint goes backward. In (158), we used the nonsingular-
ity of A(k) to make the system r.e. go backward.
\3) Let us denote the Hamiltonian by H(k)
A(kr l A(k)-I B(k)B(k)* ]
162 H(k):= [
C(k)* C(k)A(k)-l A(d +C(ktC(k)A(k)-1 B(k)B(k)* .
Optimal cost.
Given Rd(') and the cost J(u), (136)-(138), by substituting the optimal control
(156) and using the pairing corollary (134),t we get \i (ko,kl) E (N X N)+,
\ixOE R n ,
= 2Jo(u)
167 Theorem [Optimal state-feedback by the Hamiltonian system]. For the stan-
dard LQ-problem (136)-(138), let the horizon kl be fixed :and koe [O,kl-l] be arbi-
trary.
U.th.c., on any [kO,k 1],
a) there exists a unique linear state-Jeedback law
168 u(k) =-B(k)*5qk+ I)X(k)-lx(k) for k E [0,k1-1]
where X(·) and X (-) are n x n real matrix--::alued sequences defined on [0,k 1], with
detX(k) "" 0, \ikE [O,kd, and XO and X(·) are the unique solution of the back-
wards Hamiltonian matrix r.e.
169
170
173 x(ko)=xo .
•
t Where the substitutions (154) are applicable.
64
174 Comment. The control law (168) is of the fonn u(k)=-F(k)x(k) where F(k)
does not depend on xO'
]= [X x(k l )
180 Theorem [Optimal state-feedback by the Riccati r.e.]. For the standard LQ
problem (136)-(138), let the horizon kl be fixed and ko E [O,k l ] be arbitrary. U.l.c.,
on any interval [kO,k}],
a) the standard LQ problem is solved by the linear state-feedback law
where PO=P(·)* is the nxn real matrix-valued sequence defined on [O,k l ] as the
unique solution of the matrix Riccati r.e.
182 P(k-l) = C(k)*C(k) + A(k)*P(k)[I+B(k)B(k)*P(k)r1 A(k)
with
183
186 x(!co)=xo.
187 Comments. a.) In (181) and (185), the inversion is legitimate; indeed, as we
shall see P(k) = P(k)* 0 and
det[I+B(k)B(k*)P(k+ 1)] =det[I+B(k*)P(k+ I)B(k)] > 0 Vk
y) If we put h:k) := [I+B(k)B(ktp(kW l A(k), then simple calculations show that the
Riccati equation (182) becomes
189 r
P(k-l) = N:k)*[P(k)+P(k)B(k)B(k)*P(k) 1N:k) + C(k)* C(k)
Since P(k 1) = S = S* 0, (189) shows that P(k) = P(k)* 0 for all k < k 1.
S) Using standard matrix manipulations, the matrix Riccati Le. (182) can also be
written as (we drop the dependence on k in the RHS),
190 P(k-l)= C"C+ A*PA- A*PB(I+B*PB)-lB*PA
We are going to prove that if (192) holds in the RHS of the Riccati equation (182),
then P(k-l)=X(k-l)X(k-l)-I. Consider Eq. (182), a) multiply it on the right by
A(k)-I[I+B(k)B(ktp(k)] and b) multiply the result by X(k) OIl Ihe right to get
Using the Hamiltonian equations (169)-(170) and (192), equation (194) becomes
P(k-l)X(k-l)=X (k-l)
equivalently
195 P(k-l)=X (k-l)X(k-l)-I.
Thus we have shown that if P(k) is given by (192) and if (X(,),X (.» are solutions of
(169)-(170) then P(k-l), given by (195), satisfies the Riccati r.c.
The calculations above can be performed in reverse order: assume (192) and
(195), write (194) and obtain from it the Riccati r.e. (182).
66
We conclude that, given the relation (192) between P(k) and X (k) and X(k),
(X('),X('» is a solution of (169)-(170) if and only if P(·) is a solution of the Riccati
r.e. (182).
II. The cost formula (184) follows immediately form (171) and (192).
III. The optimal control (181) follows from (168) indeed from (168)
196 u(k) = -B(k)*X (k+ I)X(kr 1x(k) =-B(k>*P(k+ I)' X(k+ I)X(k)-lx(k) .
equivalently,
198 X(k+ l)X(k)-l = (I+B(k)B(k)*P(k+l))-l A(k)
65 x(k+l) = A(k)x(k) ke N,
From the periodicity of A('), direct substitution into (65) shows that if ct>(k,ko) is the
transition matrix of (65), then k is a fundamental matrix of (65), hence
there exists a constant nonsingular matrix C s.1.
67 cl>(kf-p,ko) = <'l>(k,ko)C ,
68 ct>(p,O) = c.
Now, for any nonsingular matrix Me a: nxn, the pth root of M, denoted by M1/P is
67
We have then
71 Theorem. Let k A(k) be p-periodic, where (66) holds. Consider the matrix-
valued function PO, defined on N by
72 P(k) := <Il(k,O) . B-k ,
75 /;(k+l)=B/;(k) ke N,
3.1. Preliminaries
1 Notations [Rings and fields]. R[s], ( ((: [s)), denotes the ring of polynomials in s
with real, (complex, resp.), coefficients; 8[p(s)] denotes the degree of the polynomial
p(s). lR(s), ( ((:(s», denotes the field of rational functions in s with real, (complex
resp.), coefficients; thus f(·)e ((:(s) iff f(s) = n(s)d(sr 1 where nO and d(·)e ((:[s].
Rp(s), «((:p(s», denotes the ring of proper rational functions with real, (complex
resp.), coefficients; thus f(·) e ((: p(s) iff fO e ((: (s) and f(oo) exists as a finite complex
number. Rp,o(s), ( ((:p,o(s», is the ring of strictly proper rational functions with real,
(complex resp.), coefficients; thus f(')e ((:p,o(s) iff f(')e ((:(s) and f(oo)=O. Mat [ ... ]
denotes the set of matrices having entries in the ring between the brackets, e.g. Mat
[ ((: [s]] is the set of matrices having complex polynomial entries (also called complex
polynomial matrices). If the size of these matrices remains constant then this is indi-
cated by superscripts: e.g. Rp,o(s)n><Jn denotes the linear space of nxm real strictly
proper rational matrices, For more about these rings, fields, and matrices we refer to
section A2 of the Appendix A. Note that most polynomials and rational functions
have real coefficients; however, complex coefficients may creep in when computing
roots or partial fraction expansions. Hence we prefer to work with complex
coefficients.
3 Fact. Let Ae ((:nxn and let be a complex number. The following statements
are equivalent:
i)
4 ii) There exists a nonzero vector e e ((:n S.t. Ae=Ae;
5 iii) There exists a nonzero vector '11 e ((:n S.t. *
'11 *;
iv) XA*("X)=O .t
•
t denotes the complex conjugate of
Proof. Exercise: apply Theorem (AA.13) to the matrix A - AI and its Hermitian
*-
transpose A -A.I. I
7 Any nonzero vector eE crn S.t. (4) holds is called a (right) eigenvector of A
associated with the eigenvalue A.
8 Any nonzero vector 11 E crn S.t. (5) holds is called a left eigenvector of A associ-
ated with the eigenvalue A.
The eigenvalue-eigenvector structure is important to explain the action of a
matrix. The following definitions single out interesting cases.
11 Let A E cr nxn . We call spectrum of A the point set of the (necessarily pairwise
distinct) eigenvalues of A we denote it by o[A].
13 Exercise [Real matrices]. Let A E Rnxn. Show that eigenvalues and eigenvec-
tors have the property of complex conjugate symmetry, i.e. if I.E c: is an eigenvalue
of A and (e,11) is a corresponding pair of right- and left-eigenvectors, then is an
eigenvalue and (e,Tf) is a corresponding pair of right- and left-eigenvectors. Hence
3.2.1. Definition
In this chapter we consider system representations of the form
1 x(t) = Ax(t) + Bu(t)
tE R
l yet) = Cx(t) + Du(t)
wheret
3 the state x(t) E G: n, the input u(t) E G: n, and the output yet) E G: n" ;
4 A,B,C,D are constant complex matrices of dimension n x n, nxni, noxn, noxni' resp. ;
8 <1>(0,0) = I .
Let us solve the matrix differential equation by using Picard iterates, (cf. section
(B. 1.3)), and call <1>0,<1>( ,<1>2' . .. the successive iterates; we obtain for all t E IR
<1>2(t) = 1 + At +
By the proof of the fundamental Theorem (B.1.6), we know that the sequence of
iterates converges on any bounded interval to the state transition matrix <1>(t,O). There-
fore it is given by the convergent Taylor series expansion
9 <1>(t,O)= L
00
1=0
I
-It [At]1 =1
•
1
+ ,[At] +
1.
+ -h [Alii +
11 Theorem. For any t,lo E 1R the state transition matrix of (1) is given by
12 <1>(t,lo) = exp[A(t-to)] .
A= then exp[At]= 1;
A. 1 0 0 0
0 A. 0 0
A=
0 0 0 I.. 1
0 0 0 0 I..
show that
0 eAt
eAt =
0 eAt ie'\!
0 0 eAt
(Hints: A=A.I+N, where AI and N commute; exp (At) = exp (At)exp[Nt]; compute
exp[Nt] using (9): for k Nk=O)
18 (sI-A)<1>(s)=I.
Hence
19 <l>(s)=L[exp(At)]=(sI-Ar l .
The elements of the matrix (sl-A) are polynomials with coefficients in tr; therefore
they are elements of the field of rational functions with coefficients in tr;
20
21
and
n .
21b XA(s)=det(sl-A)=sn + L XiSn-l
i=1
with BiE (fnxn for all i = O.I ....... n-l. and XiE cr for all iE n.
22 Theorem. Assume that the polynomial XA(S) is known. then the coefficient
matrices Bi can be successively calculated by the formulas:
Bo=1
Proof. Multiply both sides of (21) on the right by (sl-A)' XA(S): we get
Proof. If we successively eliminate B n_1, B n_2 ,... in the equations (23) we obtain
0=Bn_2A2+Xn_lA+XnI
The Cayley-Hamilton theorem implies that for any n x n matrix with elements in a
field F, An is a linear combination of I,A,A2,... ,An- 1: i.e. in c: nxn, the n+l "vectors"
I,A, ... ,A n are linearly dependent.
the spectral radius of A (note that PAis the maximum absolute value of the eigen-
values of A); PAis important for the following result.
32 Comments. a) Fonnula (31) shows that (sl-Af 1 is strictly proper and ana-
lytic in BC(O;PA):= ( s E (C I I s I > PA ). t
13) In [Des.3,p.27] one shows that for any matrix A E (Cnxn and for any £>0, there
exists an induced matrix nonn 11'11 such that IIAII-£ ::;; PA ::;; IIAII (i.e. the spectral
radius can be approximated arbitrarily close from above by IIAII for some induced
nonn 11·11). Hence if I s I > PA' then w.l.g. I s I > IIAII for some induced nonn IHI.
33 Proof of Theorem 30. Because of comment 13) w.l.g. we may assume that
I s I > IIAII. Now call S(s) the RHS of (31), then S(s) converges; indeed
00
IIS(s)1I Is I-I L (IIAIII S I-I )n= (I s I -IIAID- I < 00. Furthennore, (exercise),
n=O
S(s)(sI-A) == I . •
36 Fact. Consider the square matrices A and B in (Cnxn. Let XAB and XBA denote
the characteristic polynomials of AB and BA in (Cnxn. Then
37 XAB(S)= XBA(s)
or equivalently
37a det(sI-AB) = det(sI-BA) .
38 Comment. The proof below displays the advantage of a density argument based
on theorem (A.6.84) (i.e. the continuous extension of identities).
and so (37a) follows by taking the detenninant of the equation above. Now let
n .
XAB(s)=sn + L (Xi' sn-1,
i=1
n .
XBA(S) = sn + L Pi' sn-t,
i=l
hence:
Observe also that for fixed BE (Cnxn, the coefficients <Xi and depend continuously
on the entries of A, Le. for all i E n the maps
are continuous, in fact they are polynomials in the a.k's. Hence the equality of (37)
when A is nonsingular is equivalent to J
VAEM j :== (AE (Cnxn:detA ¢. O)
40 Comment. The proof above shows that for discovering a matrix property
involving equalities it pays to investigate first a simple case (here det A '¢ 0). The
same philosophy could also be used to prove the Cayley-Hamilton Theorem. Indeed
in this case it follows easily that XA(A)=O for a semi-simple matrix A (cfr. section 3
below).
45 x(1) = s(t,to,xo,u) == e
A(t-Io)
Xo + J
10
77
J
t
46 + C + Du(t) .
to
+ p(t,to,e,u)
47 Comment. Let T't be the shift operator, i.e. for every function f(·) defined on
R, T'tf is the function f(·) delayed by 't seconds according to
i.e. by delaying the application of Xo and of the input by 't seconds, the output is
delayed 't seconds. Note also that the same applies to the state transition map (45),
viz.
Hence under the shifting conditions above, the behavior of R is independent of the
initial time to: so w.l.g. to = O. For this reason the system representation R is called
time-invariant.
52 Exercise. Consider the nth order scalar linear d.e. with constant coefficients
Z(n)+ClIZ(n-I)+Cl2Z(n-2)+ ... + Cln_lz(l) + Clnz=u(t) .
Show that such an equation can always be put into the form (1) by setting
z =: Xl, z(l) =: X2, z(2) =: X3, ... ,z(n-l) =: xn and obtaining:
78
o 1 o 0 xl
0
o o X2 0
+ u(t)
1 0
0 0
-(X2 -(Xl Xn
ceA1B + D/)(t) Vt L 0
{
58 H(t) := H(t,O) = 0 'V t < 0
where we remember that t represents here the elapsed time t-t since the application
of impulses is at 't, i.e. H(t-'t) = H(t,t) = H(t-t,O). In a similar fashion we introduce a
normalized state impulse response matrix
60
65 = C(sI-A)-lxo +
(Hint: use (63t and (31); the matrices CAiB for i=O.I,2.... are called the Markov
parameters of H(s».
2 Comments a) If ei satisfies (I) then so does !lej for any ae (!:: hence w.l.g.
we shall assume that
or equivalently,
13
Since T-! is nonsingular, (13) yields (5) and (6). Moreover by (6) and (8)
det(sI-A)=det(sI-A)='XA(s); hence (Ai J.n is a list of n roots of 'XA counting multi-
1=1
plicities.
Sufficiency. By assumption (6) holds where T-! and A are given by (7) and (8).
Therefore multiplication of (6) on the right by T-! gives (13). Finally equation (13)
yields columnwise the eigenvalue-eigenvector relations (1). So A is semisimple. •
16 Fact [Distinct eigenvalues]. Let A e (Cnxn and {Ai 1.1 c (C be any point set of
L-1
(necessarily pairwise distinct) eigenvalues. Then any fami!; of eigenvectors (e i ].11=1
such that
is linearly independent.
).1
19 Proof of Fact (16). By contradiction. Assume that the family (ei
1-' is l.d .. t
Hence there exists an I-tuple (ai ).1 of scalars not all zero, say a, ¢ 0, that
1='
1
20 L aiei=9.
i='
Multiply now (20) from the left by (A-A2I)(A-A3I) ... (A-All). Then, with (17), we
get
where the factor multiplying e l is nonzero because a, ¢ 0 and the eigenvalues are by
assumption distinct. Hence e 1 is a zero eigenvector: --+ Therefore the family
( ei ).1 is l.i. . I
1='
An important special case of semisimplicity is the following.
25 Proof of Theorem (23). The first assertion follows from Fact (16) with I=n:
indeed, any family rei ).n1='
of eigenvectors associated with n pairwise distinct eigen-
values A. according to (1) is l.i.. Now let e be any eigenvector associated. say with
AI' We1show that e=a,e, for some a, e a:. Indeed since there exists an eigenvec-
tor basis (ei J.n there exists an-tuple (ai J.n of scalars not all zero such that
1=1 1=1
n
e= L aiei. i.e.
i='
n
26 (-e+ale,) + L aiej=9.
i=2
30 Fact [Dyadic Expansion]. Let A.B and C be respectively mxn nxp and mxp
matrices over a commutative ring. For k=I ..... n. let a'k and denote the successive
columns of A and rows of B. resp. Then C := AB has the expansion
31
•
32 Comment. Each term in the sum of (31) is a matrix product of a column by a
row: such matrix is called a dyad and (31) is called a dyadic expansion.
n
33 Proof of Fact (30). Observe that the (ij)th element of C reads Cij= L !ljkbkj'
k=1
Hence with i.j free we get
n n
C=[Cij]i.j= L L a·kbk ..
k=1 k=1
•
We have now the main result of this section. It uses the concepts of right- and
left eigenvectors and spectral list (see section 3.1). In (42) below
ii) A has a corresponding basis of left eigenvectors (Tlj )j:l such that
41a
iii) The bases (ej );1 and (Tli )i:1 of right- and left-eigenvectors (of ern) are
84
43 1= L e;TJ;
n *
;=1
44 X
n
=L
*
e;TJ; x =
n
L (TJ;,x) e; =: L
n
xiei,
;=1 i=1 i=1
45 A=Ln A;e;TJ;*
;=1
and, for every x E ern, y=Ax reads w.r.1. to the eigenvector basis (e.] n
1 i=1
n * n n
46 y=Ax=L AieiTJ; x=L A; (TJi,x)e; =: L A;xiei
bl bl bl
i.e. for all i each ith component of x has been multiplied by the eigenvalue Ai' •
By (4Ia) the eigenvector dyads eiTJt satisfy (e;TJt)(ejTJt)=Bij(e;TJt) for all i,j:
they are projection matrices e.g. [Kat.l,p.21]; their sum associated with one distinct
eigenvalue is called an eigen projection, [Kat.l,p.42].
'Y. Dyadic expansions are paramount to display a basis for the solution space of the
d.e. x=Ax and to characterize the poles of a transfer function matrix: see below.
8. Note that the spectral list of eigenvalues (Ai ).n1=1
= 01 [A] and the eigenvector
bases (e; ).n1=1 and (TJ; ).n1=1 define matrices
85
49
111*
112*
50 N* := E trnxn,
11n*
where E and N* are nonsingular. Therefore equations (40)-(43) and (45) are
equivalent to respectively
40a AE=EA
41a N*A=AN* ,
42a N*E=I
43a EN*=I
45a A=EAN* ,
(for (43a) and (45a) use Fact (30». So, if we observe that equations (44) and (46) are
straightforward in view of (43) and (45) resp., then five fonnulas condense the infor-
mation of the theorem namely (48)-(50) and (45a) with
51
•
54 Exercise. Let the n x n complex semisimple matrix A have the eigenvector
dyadic expansion (45). Show that:
i) for A nonsinguiar,
55 A-I = t
i=!
(1-)-1 e{Tlt ,
56 A
k
= Ln k *
(I"i) eir\j ,
i=1
86
57
= Ln exp(l.,it)eiTli
*
i=1
58 (sI-A)-1 = ±
i=1
(s-A)-l ei11 t
where IE JR, x(l) E c: n , and A E {f nxn is semisimple. By (3 .. 2.56) the state trajectory
xC') is given by
67 X(I) = s(t,0,xo,8 u)= exp[At]xo
as a linear function of Xo. Hence the set {x(·): XOE c: n } is a linear space. Also,
using the eigenvector basis [ek ) n of Theorem (40), we get successively from (67)
k=l
by using (44) and linearity in Xo
87
n
x(t) = s( t.O. L (llk.Xo) ek.au)
k=l
n
= L (11k.xo) s(t,O,ek,au)
k=1
n
68 = L <11k.xo}exp[At]ek
k=l
n
=L (11k,xo) expO"kt)ek ,
k=l
where for the last equality we used (57) and the mutual orthonorrnality condition (42).
Hence, by equation (68). if we think of the initial state as a linear combination of
eigenvectors. then the resulting motion of the state is a linear combination of very sim-
ple motions s(t,O.ek,au) = exp(Akt)ek. These motions are called modes and form a
basis of the linear space of solutions of it = Ax. Observe that in (68). i)
(11k'xo) =: Xok E .r is the kth coordinate of the initial state w.r.1. to the eigenvector
basis and measures the excitation of the kth-mode by the initial state Xo and ii) if
xO=ek then x(t)=exp(Akt)ek for all t meaning that. for all t. the state remains on
Sp(ek) = (cxek : cx E (C). i.e. the axis supported by ek .
*70 Real algebra. Consider now the case where A E R nxn is semisimpie and
n
XE lR is a real vector. By exercise (3.1.13) eigenvalues and corresponding eigenvec-
tors have the property of complex conjugate symmetry. 111erefore in Theorem (40)
we may assume that the eigenvalue-eigenvector triples (Ak,ek,11k) i) occur in complex
conjugate pairs (Ak,ek.llk).(Ak.ek,11k) for k=I,. ..• m and ii) are real for k=2m+ 1,...• n. As
a consequence for any x E lRn, for any k=I, ... ,m, complfx conjugate contributions in
(44) and (46) generate real contributions by the equations
while. for k=2m+ 1,...• n. the contributions (11k.x}ek and A.k (11k.x}ek resp. are real.
Therefore for k=l •...• m expressions (71)-(72) will have real parameters if we introduce
real and imaginary parts according to:
A.k = CXk + jOlJ< Ak = Uk - jOlk
F1
t rn z= (zl.q •...• Zn) E .r n• Re(z) denotes the vector in R n with components
( Rezi .
i=l
88
73
where, for k=l, ... ,m, xkr=2{l1kr,x) and Xki=2(l1ki,X) are real and, for k=2m+l, ... ,n,
xk = (l1k'X) is real;
ii) For any x E IRn
m n
75 y = Ax = L [Ykrekr + Ykieki] + L Ykek
k=1 k=2m+1
Note especially that for k=l •...• m the l.i. pairs of nonreal eigenvectors (ek,ek) and
(l1k,l1k) have been replaced by the l.i. real pairs (ekr,eki) and (l1kr.11ki) resp., where, by
considering real spans of the form ckek + c;;- ek and dk11 k + d k11k for complex
coefficients c k and d k ,
= Sp(ekr,eki)'
and
Recall now that for ken the ek and the 11k are l.i. Therefore IRn has now two bases
of n real vectors namely
and (ii) are mutually orthogonal for distinct subscripts. Hence (74) and (75) are the
unique representations of XE R n and y=AXE IRn w.r.1. to the real basis (76); more-
over for k=l,oo.,m with Ak=uk + jrok for k m, we see that
for x = ekr we have y = Aekr = Ukekr - rokeki '
and
we have
So to any complex conjugate eigenvalue pair (Ak,Ak) there exists a subspace neces-
sarily oj dimension two, viz. Sp(ekr,eki) such that its image under A is contained in it;
moreover the same will happen to its image under AI for any I and under exp[At]
for any t E R. Hence, for Xo E IRn and k E ill, the explicit writing of the complex con-
jugate modes of (68), viz.
with
90
Moreover, to any nonreal eigenvalue pair O"k'X;;) there exists a two-dimensional sub-
space Sp(ekr,eki) such that if the initial state Xo lies in this subspace, then so does the
z-i state trajectory x(·): this trajectory is an exponential spiral; see Fig. 3.1 for the
case <Xk<O.
*79 Special case: XOE Sp[ek,ek)' As before let AE R nxn , be semisimple and let the
notations (73) hold. As above let XOE 1Rn but now let Xo E Sp[ek,ek)'
Since XOE Rnc ten. xo=xo. and since in addition XOE Sp[ek.ek) we have
80
x(t) = exp(At) . Xo
Consider the special case where Xo = ekr. the real part of the eigenvector ek ; then (80)
shows that ck is real and. in fact, ck =0.5. Then (81) reduces to
x(t) = eUkl cos (Ilk! ekr"'" eUkl eki .
83 Poles of the transfer function. The transfer function matrix R(s) of any system
with poles necessarily in the spectrum of A (see (3.2.26». Assume now that A is
semi simple with spectral list cr/[A)= (Ai J.n1=1 .
Then, by Theorem (40) and (58), (sI-
85
•
H(s) = Ln *
(S-Ajr 1 Cej'11 jB + D.
j=l
Note that Cej . 11*jB is a dyad with column Cei and row 11*jB. Hence a pole contribu-
tion at A; disappears iff the corresponding dyad is zero or equivalently Cej = e or
* . .
11*jB = e . Hence we have with P [H(s)]. the set of poles of H(s), the theorem.
. .
86 Theorem [poles of H(s)]. Let H(s) be the transfer function matrix of a system
representation R == [A,B,C,D] with AE (tnxn semisimple and spectral list
( I..; J.n1=1 =cr/[A].
V.t.c.
pE P[H(s)]
if and only if
i) there exists an eigenvalue Aj E 0/ [A] S.t. p== \ '
ii) Cej *' e,
...
111
) *B
11j e*
*' ,
where e i and 11i are right- and left-eigenvectors of A corresponding to \ in the dyadic
expansion (85). I
yet) = p(t,O,xo,u)
90
91 Comment. Clearly in (85) and (90) the nonzero vectors Ce i and 11*jB represent
the strength of the coupling of the ith-mode with the output and the input resp. Note
that Ce i depends on the location and the sensitivity of the sensors while l,*jB depends
92
93a ±
j=1
ej exp(l"jt)· [<Tli,X O) + J e -A,'t <B* TJj,u(t» d't
0
1
(b) Let xo=8n at t=O- and, for some P=(Pl,P2, ... Pn i)E CI: n" let u(t) = p6(t) (i.e. for
k E llj, the kth-scalar input is an impulse of area Pk applied at t=O): show that
93b
Remarks. a.) In (93b), (B*TJj,p) measures the coupling between the impulsive
vector input p6(t) and the ith-mode; in particular, if (B*TJj,p)=O, then the ith-mode is
not excited by that particular input.
*
If B TJj = 8 ni , then, by Eq. (93a), we see that no input can excite the ith-
mode, i.e. the actuators are not coupled to the ith-mode.
r) If Cej = 8110 , then Eq. (90) shows that the jth-mode does not contribute to the
output, i.e. the sensors are not coupled to the ith-mode.
cos Ijl]
sin Ijl .
Note that
[o -cotg Ijl ]
(sin <1»-1
93
II
e1
Fig. 3.2 The angle between the normalized eigenvectors el and e2 is III.
(a) Show that the t volume of the parallelepiped based on e l and e2 , is given by
97 OS; Idet.11 S; I,
where for c!> variable the following minimum and maximum occur
99 Comment: The results in (a) and (b) generalize in (CD. In fact if det. 1 ::: I,
then T- 1 is almost unitary, the e· are almost orthonormal and Vi IIllill::: I; if
det. 1 ::: 0, then T- 1 is almost the ei are almost linearly dependent and :3
i s.t. Illlili is very large. In order to calculate precisely, in truncated arithmetic, the
similarity transformation A = TA.1, it is required that T- 1 be almost unitary; if not,
t Of course for n=2 volume should read "area" and parallelepiped "parallelogram,"
resp.
94
many significant figures may be lost, (see e.g. [Gol.l ,p. 197-198]). A similar waming
applies also to the solution of the equation Ax=b.
(c) In order to illustrate Comment (99) above, consider the matrix A defined above
and let cp=1t-£, with £ small; more precisely £2 is negligible.
i) Calculate Iderrll and the condition number of A; sketch the vectors el' e 2,
11 1> 112'
ii) Consider the equation Ax=b and a perturbation in b: hence the resulting
perturbation in x is = A-I
Let Bb=/}·th/ll11dl, with /}>O small. Show that Bx is very large, in fact
100 Comment. Though 118bll = /} > 0 is small, 118xll is huge because of the almost
linear dependence of e 1 and e 2 and the disparity of the eigenvalues: A is not well con-
ditioned.
101 Comment. The fast mode exp(-lOOt) disappears almost immediately and for t
not large, IIx(t)ll > > Ilxoll = I; so here IIx(t)ll increases first: this shows that even for
real negative eigenvalues IIx(t)II may increase considerably above IIxoil before even-
tually decreasing exponentially to zero.
CHAPTER 3d
3d.1 Preliminaries
In the sequel the notations of Section 3.1 apply with the exception that the
transform variable s has to be replaced by z; e.g. R[z], ( tr [z]), denotes the ring of
polynomials in z with real, (complex, resp.), coefficients, etc.
Remarks. I. It is understood that the recursion equation (l) is run forward: since
A may be singular (e.g. "deadbeat" case), x(k+ 1) and u(k) do not always uniquely
define x(k). However, if A is nonsingular then (1) can also be run backwards and it is
better to define Rd on all integers i.e. on the set Z, (instead of the "half-line" of
integers N).
II. The system representation R d = [A,B,C,D] is a special case of the representation
R d(') = [A('),B('),C('),D(-)] where A('),B(-),C(-),D(-) are constant sequences. For this
reason R d = [A,B,C,D] is called a time-invariant representation: see (47) below.
where V k,l e N
In contrast to the continuous-time case one has easily the following theorem.
12 Theorem. Consider the recursion equation x(k+l) = Ax(k) and let No=N[A Illa]
be the algebraic eigenspace of A e cr nxn associated with its eigenvalue A=O, [see
(4.3.3) below]. Then for all x(ko)e No we have for kl := ko+mo where mo :s; n
x(k t )=6
At 0
A= [ 0
1, then Ak=
[Ar
0
0
AJ< 1
;
k!
where Cmk := --..;;;;.;--
m!(k-m)! .
A k by the z-transform
_ Taking the z-transform of <I> (k+ 1,0) = A<I> (k,O) with <I> (0,0) = I and denoting
<l>(z) := Z[<I> (k,O)] we have by (D2.S)
IS
- -
z [<I>(z)-I] = A<I>(z) .
Hence
19 q)(z) = Z [A k] = z(zl - Af! E G::p(z)" X" •
Comment. The only yifference between (19) and (3.2.19) is the matrix factor zl,
which multiplies (zI-Af . Hence using similar results of Section 3.2, we have the fol-
lowing.
20
-
Exercise [Computing <l>(z)J:
-
Show that <l>(z) given by (19) reads
97
rrbij(Z)]. .
I.JE D
=: B(z) E (Cnxn[z] is an n x n polynomial matrix reading
s.t. with
n .
2Ib XA(Z)=Zn + L XjZn-l,
j;l
Bo=I
_ [Z(Z-A)-I 0
27a if then cl>(z) = 0 l;
27b
OIl -
if A= [ 0 0 ' then cl>(z)=
[1 1
0
z-I
1 .
29 Remark [see Chapter 4]. Note in particular that rnO is the size of the largest Jor-
dan block associated with the eigenvalue 1..=0. lienee, if A is semisimple (i.e.
diagonable), then 1110 ::;; I and 0 is not a pole of <I>(z), [compare (pa) and (27b»).
Moreover if A is nonsingular then mo=O and 0 is not a pole, indeed <1>(0)=0.
k-l
45 x(k) = s(k,ko,xo,u) = Ak-koxO + L Ak-k'-IBu(k')
k';ko
and
k-ko k-l ,
46 y(k)=p(k,ko,xo,u)=CA xo+ L CAk-k-1Bu(k')+Du(k).
k'=ko
47 Comment. Let lEN and define the delay t operator T/ S.t. for every sequence
fO:= )0', T/ f is the sequence f(-) delayed by I sampling periods, more precisely
f(k-l) for k I
{
48 [T/f](k)= e for 0::;; k < I .
i.e. by delaying the application of Xo and the input I sampling periods, the state and
the output of R d are delayed by the same amount.
Therefore, under the shift operations above the behavior of R d is independent of
°
ko' so w.l.g. k = 0 . For this reason the system representation R d is called time-
invariant and the state transition- and response-maps are normalized to ko=O.
CAk_1B for k 0
58 H(k) := H(k,O) = { DO for k=0
for k < 0
60
-
H(z) := Z[H(k)] ,
64 x(z) = Z[s(k,D,xo,u)]
= Cz(zI-A)-lxO+ I{z)u(z)
= Cz(zI-A)-lxo + [C(zI-ArlB+D]ii(z)
-
where H(z) is the transfer function of Rd'
•
3d.3 Properties of Rd when A has a Basis of Eigenvectors
As in section 3.3 we assume that A E er nxn has a basis of eigenvectors, i.e. is
semisimple. Hence as in Theorem (3.3.40) A has a dyadic expansion
45
71 x(k+ 1) = Ax(k) kE N,
where x(k) E R n for a real initial state x E IRn at k=O in the span of two (nonreal) com-
plex conjugate eigenvectors e/ and e;
of A corresponding to a pair of (nonreal) com-
plex conjugate eigenvalue-eigenvector triples (A/,e/,'Il/) and (A/,e;,Tl/)' Introducing for
AI E fr a polar representation
101
72
x(k) = 2 Re (11/'x)'A/x)
where the pair of real coefficients xr(k) and xi(k) are given by
[
Xr(k) ] [ cos kcll, sin kcll, ] [2(rll r'xO) ]
xj(k) = P/k -sin kcll, cos kcll, 2(11/i.xo)
with
Note that V k. x(k) e Sp(e/pe/i). a two-dimensional real subspace in which its tra-
jectory is an exponential spiral.
83 Poles of the transfer function. Note that. with H(z) the transfer function of R d
-
given by (61). then by the dyadic expansion (45)
85 H(z) = C(zI-A)-IB+D = L
- n
(z-'A.j)-ICej '11j B+D *
i-l
y(k) = p(k,O,xo,u)
n n k-l
90 =1: Cej(TJj,xo))"jk + 1: Cej' 1: Ajk-k'-I(B*TJj,u(k'»+Du(k).
j=1 j=l](=O
•
J:,
92 Exercise [State transition]. Given R d = [A,B,C,D] where A is semisimple with
dyadic expansion (45). Show that for any XOE ffn and any u(·)= (U(k) the state
for k > 0 is given by
x(k) = s(k,O,xo,u)
93
•
CHAPTER 4
This chapter develops the main properties of the linear time-invariant representa-
tion R = [A,B,C,D] when the matrix A is general.
The main topic is the representation of a linear map A and its consequences. Our
approach will be mainly geometric and certain details will be omitted. A key tool here
is matrix representation theory and especially the effects of a change of basis: see sec-
tion A.5. of Appendix A. Some key references are [Kat.l,I.3 and 1.5] for theory,
[Nob.l,chs.8 and 10] for algebraic formulations and [001.1], [001.2] for numerical
aspects. The chapter concludes by discussing the function of a matrix, the spectral
mapping theorem and a discussion of the linear map X AX + XB. The discrete-time
results are implicitly present throughout the chapter.
We start by giving some preliminaries.
4.1. Preliminaries
We introduce the notions of A-invariant subspace, direct sum of subspaces and
nilpotent matrix.
1 Invariant subspaces. Let (V,F) be a linear space over the field F s.t. dim V = n.
Let M be a subspace of (V,F) and let A : (V ,F) (V,F) be a linear map. The sub-
space M is said to be A-invariant iff
2 xeM:>AxeM.
A:M
n
5 A bj = L aijb i for all j En,
i='
(see Section A.S.I). The vectors in M are then described in terms of vectors E P
and M is A-invariant if and only if
6 EM=:> EM.
7 Example. (V,F) is represented by (1:2 by the basis r£i)2 . Let the linear map A
1='
be represented by the matrix A = [6 g]. then the subspace M := Sp[ed is A-
invariant.
representation of x as
k
13 +Xk=L Xi,
i=1
lOS
14 Example. Let (V.R) be the linear space of all periodic functions from R into 1R
of the fonn
f(t) =
N 'k
1: t
k=-N
We now use the concepts of direct sum and A-invariance to prove an extremely
useful theorem.
Fig. 4.1. These illustrations show that the direct sum decomposition of V =]B2 is not
unique.
106
A
20
n-k
ii) If both M 1 and M 2 are A-invariant then V has a basis such that A has a matrix
representation A e F" x n of the form
k n-k
A= tAi'-
k
21
n-k.
and rbj].n
Proof of Theorem (19).
respectively.
r i)
r M
Let (bj].k
Now by Exercise (16). since V = M
J=I J=k+l
1U 2
be bases for M 1 and M 2'
and MIn M 2 = { e} we
have V = M 1 Ea M 2; hence (bj);1 is a basis for V and any x in V has the unique
representation (4); moreover A has a matrix representation A = (aij) dictated by (5).
Now T/j=I.2.....k. bj M which is A-invariant; hence A b e M with basis (b i1:1'
E 1 j 1
Therefore. by (5). T/ j E k
k
Abj=L Ilij bj
i=l
•
i.e. Ilij=O T/ i = k+ I ..... n. T/ j = I.2 ..... k. Hence one has the matrix representation
(20).
22a Remarks. a) In (20). All is the matrix representation of the part of A in M I'
In (21). All and A22 are matrix representations of the parts of A in M I and in M 2.
107
resp.
13) The representation (21) will be blockdiagonal -- irrespective of what
bases are chosen in M 1 and M 2.
24 Comment. It will turn out that any square matrix A e c: nxn is uniquely decom-
posable as the sum of a semisimple matrix and a nilpotent one. (See Theorem (4.4.14)
below.)
25 Exercise. Let Me c: nxn be nilpotent. Show that all eigenvalues of M are zero.
(Hint: First. prove that if A. is an eigenvalue of M. then A.m is an eigenvalue of Mm.)
M\= 0 0 I
10] and M2=
000 0 0
i) A\ = ,XA(s) = (s-A.\)2(s-'-2).
o 0 '-2
108
where
5 dkE N S.t. dk 1 for k= 1,2,,,.,0
and
6 d\+d2 + ... +do=n,
i.e. the positive integer dk is the multiplicity of the zero Ak as a zero of the characteris-
tic polynomial lA('): if A has repeated eigenvalues, then for some k,dk > 1 and o<n.
If on the other hand, dk= I for all k, the 0 = n and A is simple.
Theorem (14) below shows that the minimal polynomial "'A(s) has the factored
form:
11 Nk := N ([A-Ak I ]m-J.
According to Exercise (4.1.8), Nk is an A-invariant subspace of (C".
109
The notations dk, mk, XA' 'I' A, 'l'k' and N k will be used without comment in the follow-
ing developments.
where the ci are the complex coefficients of 'I' A(·)' Therefore for all k, 'I' A(Ak) = O.
2) We claim that for all k the multiplicity mk of the zero Ak satisfies mk :s; dk; i.e. the
polynomial 'l'A is a divisor of the polynomial XA-
Indeed the minimal polynomial 'I'A has necessarily a degree 0['1'A1 which does not
exceed the degree 15[XA] of XA, (an annihilating polynomial). Hence using Euclidean
division, there exist polynomials q and r in c: [sl such that
where, if res) == 0, XA is a multiple of 'I'A and mk :s; dk for all k, and, if res) =1= 0,
then its degree satisfies 15[r] < 15['1'A]' The latter alternative leads to contradiction.
Indeed by (15), with s replaced by A, and the Cayley-Hamilton Theorem (3.2.24):
18 Remark. Part 1) of the proof shows that any annihilating polynomial has zeros
at Ak for k = 1,2, ... ,a. Setting XA = p in part (ii) we conclude that any annihilating
polynomial, having zeros solely at the Ak' has the form
110
CJ
19 pes) = n (S-Akl
k=l
k
where
20 PkeN S.t. for k=1,2, ... ,0'.
24 ii) There exists a unique exponent v 1 S.t. in (23) above, for Jl=O,I, ... ,v-l the
inclusion is a strict t inclusion of subspaces, and, for all Jl v, the inclusion is an
equality of subspaces.
iii) Establish the corresponding property for the successive ranges and ranks of Til.
[Hint: N[TIl+l] = (xe c;n:TxeN[TIl] J.]
25 Remark. The unique exponent v, mentioned in (ii) above, is called the ascent of
T, [Tay.l,p.290]. It will turn out that the multiplicity mk of the zero Ak of the minimal
polynomial 'I'A(') in (7) above is the ascent of A-AkI.
26 Exercise. Let 'I'A(s) and XA(S) denote the minimal- and characteristic polynomi-
als resp. of Ae (Cnxn, given by (7) and (4). For any polynomial p(s)e C;[s] denote
by pes) the polynomial obtained from pes) by replacing its coefficients by their com-
plex conjugates. Consider A* e (Cnxn (the Hermitian transpose of A). Show that its
minimal- and characteristic polynomial read, resp.
and
[Hints: for (27) show that p(.) is an annihilating polynomial of A iff p(.) is an annihi-
lating polynomial of A*; for (28) consider XA(s)=det[sI-A].]
characteristic- and minimal polynomials XAO and 'l' AO, resp. given by (4.2.4)-(4.2.6)
and (4.2.7)-(4.2.9). Consider the A-invariant subspaces N k as given by (4.2.11).
U.th.c.
o
2 (Cn= $ Nk=N, $N 2 $ ... $N o '
k='
3 Comments. a.) Nk := N[(A-AkI)mk] and N[A-AkI] are called, resp., the
algebraic- and geometric eigenspace of Ae c: nxn at its eigenvalue Ak'
4 while Ok := dim N k
5 and rk := dimN[A-AkI] are called resp. the algebraic- and geometric multiplicity
of the eigenvalue Ak' [KaLl,pp.42,37]. Note that N[A-AkI] is the subspace of eigen-
vectors of A at Ak; moreover in Corollary (25) below it will tum out that if mk> 1 then
N [A-AkI] is a proper subspace of N k: in this case N k cannot be spanned by eigenvec-
tors:
P) Note that for ke Q, Nk=N[(A-AkI)mk] is A-invariant. Hence by
Theorem (4.1.19), the decomposition (2) implies that if we pick as basis for <en, the
union of cr arbitrary bases of, respectively, N ,,N 2,''',N 0' we obtain for A a blockdiag-
onal representation of the form
A, 0 0
0 A2 0
-
6 A=block diag[A"A 2 , ... ,A cr l=
0 0 0
0 0 Ao
In Corollary (28) below it will turn out that Ok := dim N k = dim Ak = dk, i.e. the alge-
braic multiplicity of the eigenvalue Ak is the exponent of (S-Ak) in the characteristic
polynomial XAo, (see (3.2.7»; moreover, det(sI-A k) = (S-Ak)d k •
where nk e c: [s], a polynomial Multiplying (8) by 'l' A(s) and using the definition
112
The RHS of (9) is a polynomial in s such that the coefficients of all positive powers of
s are zero. Therefore, if we replace s by A in the RHS of (9), all coefficients of posi-
tive powers of A will be zero and
a
10 1= L nk(A)'I'k(A).
k=l
with
We claim that
13 xkENk for k=I,2, ... ,a.
I
= [nk(s)'I'A(s)] I x
IA
t Identity (9) expresses the fact the polynomials 'l'kO are coprime. It is called the
Bezout condition.
113
with xkE Nk for k=1.2 .... 0". W.l.g. we may assume that xI '" e. (otherwise reorder
the eigenvalues such that to Al corresponds XI '" e). From (14)
now by the definitions of 'l'1(s) (see (4.2.10» and the Nk for k > 2, premultiplication
of (15) by '1'1 (A) gives
17 'l'1(A)xl =8.
because 'l'1(A)xk=8, 'v'k L2. Now the polynomials (S-A2)m1 and 'l'1(s) are
coprime, hence by the Bezout condition we can find polynomials hl(s) and h2(s) S.t.
'v'SE (t.
Therefore
where, in view of (16) and (17), each term on the LHS of (18) is 8. Hence xI =8.
which contradicts (15). Therefore the decomposition (11)-(13) is a unique decomposi-
tion of x as a sum of vectors in the N/s. This means that
V =$ (V (1 N k) .
k=1
Comment. The corollary asserts that "every A-invariant subspace is spanned by gen-
eralized eigenvectors."
Proof. Let XE V. Then, using the unique decomposition (11)-(12) of the proof of
114
Theorem (1), x = xl + X2 + ... + xa ' where, for all k= 1, ... , 0', xk = nk(A)Pk(A)x
belongs to N k. Now, since xe V and V is A-invariant, for each k, xk belongs also to
V. Hence
a
•
Vee (V n N k) c V .
k=l
[Hints: pick 0' = 2 with each N k spanned by one real eigenvector. Let M be any
line through the origin not parallel with the Nk'S .... J
*21 Remark. The proof of Theorem 1 uses only the fact that 'VA(s) is an annihilat-
ing polynomial of A with zeros solely at the Ak' Therefore, recalling the general
expression (4.3.19)-(4.3.20) of such polynomial, there results also
(i.e. strict inclusion). Hence by (4.3.24) and (23)-(24) we have established the follow-
ing
and
29 Ok := dimNk=d k
and
30 XAN.(S)=
34 XA(s)=TI
k=1 •
We claim that AN. has only one eigenvalue: Ak' Indeed, suppose Jl were another
eigenvalue. Then there exists a nonzero vector ZEN k C ([ n s. t.
35 AZ=Jlz
and
36 [A-AkI tkZ=9n.
38 Exercise. Let A E tr nxn and let x be a nonzero vector of (Cn. Let 'VA be the
minimal polynomial of A. Let p denote a polynomial. call o[p] its degree. Show that
i) there exists an eigenvalue Ak of A and a polynomial pes). with 0 s; orp] < 0['1'A],
s.t. p(A)x is a ( nonzero) eigenvector of A at Ak.
t·
ii) If x E N k then, in i), Ak is the eigenvalue corresponding to N k and p is a divisor
of [S-Ak with 0 O[p] < mk·
[Hint: there exists a polynomial 'V of minimum degree such that 'V(A)x = e.]
40 Nk[A*]
Show that
a) Nk[A*] is A*-invariant.
b) Nk[A*) is the algebraic eigenspace of A* at X"k of dimension dk,
c) tr n is decomposed into algebraic eigenspaces of A* according to
cr
41 (Cn= $ N [A*].
k=! k
[Hints: (a) Use Exercise (4.1.8); (b) use (4.2.27) and (4.2.28); (c) in Theorem (4.3.1)
replace A by A*.]
42 Comment. The spaces N k[A] and N k[A*]' defined above. are called resp. the
right- and left algebraic eigenspace of A E tr nxn at its eigenvalue Ak. (Note that
117
Nk[A) contains the right eigenvectors of A at A.k' and Nk[A*] contains the left eigen-
vectors of A at A.k (i.e. T\* A=A.kT\*).
2 AEk=EkAk=Ek [A.kIdk+Rk]'
where Ak E (t d.xd. is the kth diagonal block of (4.3.6) and the dkxd k matrices Ak1d.
and Rk := Ak-AkId. satisfy
3 AkId.Rk = RkAk1d.
Hence defining
where
118
11 AR=RA
and
12 R is nilpotent of index m.
Therefore defining
13 D:=lIAT, M:=l IRT,
15 A=D+M
18 Theorem [Partial Fraction expansion]. Let Ae cr nxn and let (Co be decom-
posed into algebraic eigenspaces N k according to (4.3.2). Pick any basis of cr n as the
union of cr bases, respectively, N IN 2, . . . ,N ("J' Hence (1)-(15) applies.
U.th.c
m-I
19a (sl-Ar l = (sI-D)-1 + :E (sl-D)-I-IM 1
1=1
20 Comments. <X.) (19b) shows that (SI-A)-I has a pole at each eigenvalue of A;
moreover the pole at Ak is of order mk the ascent of A-AkI. Note that if A is semisim-
pIe then, by Theorem (14) and (13), R=O, and all poles of (sI-A)-1 are simple.
P) Equations (19) give the partial fraction expansion of (SI-A)-I: (19a)
is the "coordinate-free" form of that expansion.
00 m-I
= (sl-D)-I MI = (81-0)-1 M' ;
1=0 1=0
in (25). equality holds for lsi sufficiently large and. by analytic continuation. for all
s *- Ak for k=1.2 .....0'. (19a) follows by combining (24) and (25). •
28 Remark (Uniqueness of M and 0). Formula (l9c) is the partial fraction expan-
sion of (sI-Ar l at its poles A.k' Hence the coefficient matrices of (S-Ak)-l and
(S-Akr2. namely
29 r l . blockdiag [O .....O.ldk.O.....O] . T
and
30 r 1 • blockdiag [O.....O.Rk.O.....O] . T •
32 Remark. Let jl E (l:nxn be the nonsingular matrix given by (5). Define now
33a T:= N*
(1
34 1= L EkNk"
k=!
36 A= f Ek rAk · I
k=! [
d_ + Rk]N k*
38 V't.
44
•
121
48 *
Nk[A] := N - )m.]
[[A*-AkI
Show that
*
a) For every ke 0:. Nk[A) and Nk[A) have bases that are mutually orthonormal.
i.e. there exist n x d k matrices Ek and Nk• both ofrank d k• s. t.
51
and
52 Nk*Ek = Id••
b) For every k E 0:
a
53 N [A).1
k
= 1=1
Efl Nk[A*).
I .. k
54 N_[A] .1 =N+[A). *
122
55 Note on terminology. Consider the spaces N _[A] and N +[A*] defined in (49)-
(50). In Theorem (7.2.33) below we prove that the solutions of the d.e. x=Ax (viz.
x(t)=exp[At]xo) are exponentially decaying iff XOE N_[A] (Le. the direct sum of the
algebraic eigenspaces of A associated with its eigenvalues with negative real parts).
For this reason we call N _[A] the (right) stable subspace of A. Similarly N +[A*] is
called the (right) unstable subspace of A* (or the left unstable subspace of A). Of
course the same terminology applies to N _[A*] and N +[A] (obtained by exchanging A
and A* in (50». Recall now that the algebraic eigenspace Nk[A*], (48), is called the
left algebraic eigenspace of A at its eigenvalue Ak' (see (4.3.42».
with dimR[Nk]=dk=dimNdA*].
b) (53): use (35) and (51).
c) (54): by (53) and Exercise (A.7.30b)
L
[keK Nk[A]).L = n
keK
[f
1=1
1 "k
N1*[AJ].]
is nilpotent of index mk' The question hence is to find a basis of N k such that Rk• the
nilpotent part of A k• is as simple as possible. The answer to this question is yes and is
documented by Fact (13) below.
An important notion is the following.
1 Let Ae (rnxn have the eigenvalue Ak with algebraic eigenspace N k. We call Jor-
dan chain of length fl at the eigenvalue Ak of A any I.i. family of vectors ];1 in
Nk S.t.
where := e.
3 Comment. The nonzero vectors of N k are called generalized eigenvectors; hence
(2) defines a chain of generalized eigenvectors. Note that any chain of generalized
eigenvectors contains one eigenvector, namely. E N [A-AkI].
Jordan chains are useful because of the following.
5 Ef= e(
i.e. R =Span ];1' Show that. w.r.t. R A has the matrix representation
dictated by
6
where
Ir
7 If = AkII1 + Rt
with
11-1
8 Rf= [-:-
o I
t In the symbols It and Rt. fl is a superscript and not an exponent; III is the
flxfl unit matrix as usual.
124
10 Comments. Jt:, defined by (7)-(8), is called a Jordan block at Ak of size 11; Rt:
is the nilpotent part of Jt: of index 11 (see (4.4.15»; a Jordan chain of length 11 gen-
erates a Jordan block of size 11 by (5)-(6).
The comments suggest the following affirmative answer to the question posed at
the start of this section.
13 Fact [Jordan basis of N k]' Let A E (l: n x n have the eigenvalue Ak with algebraic
eigenspace N k' Let d k ;= dimN k and rk ;= dimN [A-AkI] be resp. the algebraic- and
geometric multiplicity of Ak and let mk be the ascent of A-AkI.
V.th.c.
N k has as basis a union of rk Jordan chains of length for 't= 1,2, ... ,rk such that
15 ii) there exists at least one 't E Ik such that Ilk1 = mk '
16 iii)
20 Comments. a) The hard pan of the proof of the Fact above is the proof of
existence of the Jordan basis of rk chains; see [Kat.l,p.22] and [Nob.l,Th.1O.3]. The
other parts of Fact (13) are reasonably straightforward; in particular for (15) the
interested reader is referred to Exercise (30) below.
Note that, by (15) and Exercise (4), Ak in (17) contains always a Jordan
block of maximal size mk' Hence it makes sense to order the blocks in (17) according
to decreasing size, i.e.
125
Ak 0 0 0
0 Ak 0 0
3
0 0 Ak 0 0
Ak=
0 0 0 Ak 2
0 0 0 0 Ak
Ak 0 0 0
0 Ak 0 0 3
0 0 Ak 0 0
Ak=
0 0 0 Ak 0
0 0 0 0 Ak
if and only if
26 A and A. have the same Jordan form.
•
28 Comments. (X) [cf. (A. 5.22)]. In other words. since (25) reflects a change of
basis. any linear map A : (V. (C) (V. (C) where the linear space (V. (C) has dimen-
sion n. has a unique matrix representation in Jordan form.
The numerical computation of the Jordan form is a formidable task and
should be avoided except in simple cases: see [GoI.2] for an extensive review of the
numerical difficulties and [Kag.l] for an algorithm: imponant numerical considerations
are (i) the difference in magnitude of the eigenvalues. (ii) whether or not the algebraic
eigenspaces N k are almost orthogonal. etc.
y) The structural information of the Jordan form. however. is useful in theoret-
ical considerations: for instance in (4.4.15) its semisimple pan D equals A. where A is
the diagonal matrix of eigenvalues (4.4.6) and its nilpotent part M has zero entries
everywhere except for some entries of the superdiagonal, which are 1. For structural
infonnation on the Jordan fonn of A* see [KatJ,I.5.5.] and [Nob.1.,Th.10.4].
We conclude this section by two exercises: they may be skipped.
**30 Exercise. Let Ae (Cnxn have the eigenvalue Ak with algebraic eigenspace N k •
Let dk=dimN k and let mk be the ascent of A-AkI with 1 mk d k. Recall
definition (1). Show that A has a Jordan chain ];1 in Nk of length Il=mk'
More precisely, set T = A-AkI; hence (2) reads with Il = mk
:_1 for
for
j=I,2, ... ,/
j=l+l .... ,I1·
127
iv) By (i) and (iii) the family tl' with ef given by (i). is l.i. In particular. if
for complex coefficients Uj.
f
j=1
...... 31 Exercise. Let A E tr nx n and let N k be its algebraic eigenspace at the eigen-
x
value Ak' Consider = Ax. x(O) = xo. Show that 'v' Xo '" en. Xo E N k' the solution
contains a term expO"kt). where is a nonzero vector of tr n.
(Hint: use the Jordan form).
then. by definition.
The purpose of this section is to develop a systematic method for defining the func-
tion of a matrix. The most important cases are
Ak. ke N
Throughout this section. we use the following notations. Let A e trnxn. let
{ A\.A2 ....• Ao} be the point set of its (distinct) eigenvalues. also called the spectrum of
A and denoted by o(A). We write its minimal polynomial as in (4.2.7)
128
2 Lemma. Given two polynomials PI.P2 e c: [s]. let us divide them by 'I'A: call qi
and ri the respective quotients and remainders. Thus we have
6 VAe c:
7
•
8 Remarks. ex) If we evaluate Eq. (3) at A and recall that 'l'A(A)=9 we get
PI(A)=rl(A). So whenever we want to calculate any polynomial pI(A) we never
need to calculate powers of A larger than d\v A-I.
Conditions (7) are called the spectral interpolation conditions.
•
One can define a function of a matrix A, say f(A), in tenns of the power series of
f. However, in view of Lemma (2), it turns out that it is much more efficient and
much more insightful to do it in the following way.
By definition,
12 f(A) := p(A).
•
(J
15 Exercise. Let f satisfy the conditions above. Show that (a) Af(A) = f(A)A; (b)
f(AT)=f(A)T; (c) let Te (Cnxn be nonsingular, then f(TArl) = Tf(A)r 1.
16 Exercise. Let f satisfy the conditions above and, as is usual in applications, let
fez) = fOO. V z e!J.. Show that the polynomial p can be chosen with real
coefficients and hence
f(A*)=f(At .
Expression (I), which defines f(A), seems at first somewhat arbitrary: let us verify
that if g and h are, respectively, the sum and the product of the analytic functions f 1'[2'
i.e.
then we do have
and
19 Theorem [Properties of f(A)]. Given AE (Cnxn. Let I'l be a domain that con-
tains all the eigenvalues of A. If f) ,f2 and f are analytic functions mapping I'l inw (C,
then
20
where the last step follows by direct calculation. We are going to show that the poly-
nomial S --4 P,(s)P2(s) is an interpolating polynomial for f,(s)f2(s). Indeed
VAkE a(A)
where the second step follows from the definition of p, and pz and the last one from
Leibnitz's rule. If mk > 2, the same type of calculations establishes the required equali-
ties. So (21) is established.
It remains to establish the equivalence (22).
(¢). Since the right hand side of (22) holds, we may choose p(A)=O, VA, as an
interpolating polynomial; hence peA) := f(A) = e.
°
By contradiction: assume that at least one interpolating condition (22) is not
satisfied, say, t</)(Ak) ¢. for some k a and I mk-l. Since I mk-1, any
such interpolating polynomial, say p, cannot have [S-Ak ]m. as a factor, hence p is not
a multiple of 'JfA hence, by definition of the minimal polynomial, peA) := f(A) ¢. 8,
which contradicts the left-hand side of (22). •
30 Theorem [General formula for f(A)]. Let A E c: nxn have. a minimal polynomial
'Jf A given by (1). Let the domain contain a(A), then for any analytic function
f : --4 c: we have
132
o mk- I
31 f(A) = L L t</)(Ak)Pk/(A) ,
k=1 1=0
32 Comments. The importance of formula (31) lies in its form: a) the polynomials
Pki are independent of f and depend exclusively on A. So the matrices Pki (A) are
independent off.
b) The weighting factors of the PkI's, namely, t<I)O"k) depend only on
the values of f (and some of its derivatives) on the spectrum of
A : a(A) = {"'I'''-2, ... ,1.. 0 } .
o mk- I
b) '<;f V E N, AV = L L v(v-l) ...
k=1 1=1
o mk- I dl
c) for A nonsingular, InA=L L -I (lnA)'IA"Pkl(A)
k=1 1=0 dA
Note that (33d) shows that (sI-A)-l has a pole of order mk at S=Ak' (Indeed
Pk(mk-1)(A) 9: if not, we could drop one of the interpolating conditions (22); but in
Theorem (19) we showed that all the m conditions were necessary.) The fact that
(SI-A)-I has a pole of order mk at Ak can also be seen by taking a Laplace transform
of (33a).
and
133
37 { Ak -t 8 as k -t 00 1 {v Ai e a(A), I Ai I < 1 }
and
V Ai E a(A), I Ai I ::; 1 }
{ } {
38 k -t Ak is bounded on IN+ and mi = 1 whenever I Ai I = 1 .
39 Calculating the Pkl (A)'s. Before proving Theorem (30), l<:t us show how to cal-
culate the matrices Pkl (A) without obtaining an interpolating polynomial. The trick is
to use Comment (32a).
Suppose that, for the given A, a=2, mt =2, m2= 1; thus
'JIA(A) = (A-Al)2(A-A2)' Then formula (31) gives for any analytic -t (t
f 2(A) = A-A} , VA
f3 (A) = (A-Al)2, VA
1 2
P2o(A) = 2 (A-A} I)
41 Exercise. Let J be a Jordan block of size 3 with eigenvalue AI' Then, for any f
analytic in a neighborhood of AI,
AI 1 0 f(AI) J...
2!
fUCA )
I
J= [ 0 Al 1 and f(1)= 0 nA I )
o 0 o f(AI)
42 Proof of Theorem (30). To alleviate the notations let us consider a special case
where
44 p(A)=A3ao+A2al+Aa2+a3
where the coefficients ao, ... ,a3 must satisfy the interpolating requirements (11):
This is a system of four linear algebraic equations in the coefficients ao,al,a2,a3' If \jf A
had four simple zeros, the matrix of the coefficients would be the classical nonsingular
Vandermonde matrix. Here, A.3, is a double zero of \jfA' hence the matrix is a modified
Vandermonde matrix [AiLl,p.ll9]; it is known that it is nonsingular. Consequently
the system (45) has a unique solution and by Cramer's theorem each coefficient ak of
p(A.) is a linear combination of f(A"I),f(AI),f(A3) and f'("-:3).
Therefore, in general we have
135
where each coefficient is a linear combination of the rt)(Ak)' k=I,2, ... ,0 and
I=O,I, ... ,mk-l. Therefore if we rearrange the tenns of the sum in (46) we obtain an
expression of the fonn
o m.-I
47 f(A) = peA) = L L rl)(Ak)Pkl (A) .
k=1 1=0
where the matrices Pkl (A) are the result of the rearrangement of the tenns. This com-
pletes the proof of fonnula (31). •
2 o[f(A)] = }.
r, r,
no guarantee that f(A) has 0 distinct eigenvalues. (Take f(A) = 1, V AE cr). So in
(2) (f(A1).£(A2), ... f(A o) l denotes the point set contained in the listed elements.
r,
b) By (2) the spectra of exp(At), log A and Av are {exP(\t) {lOg Ai and
{A;V respectively.
where the last step follows from the interpolation condition of the definition of f(A),
136
see (4.6.11).
•
8 Application to the numerical integration of x=Ax. Given x=Ax, x(O)=xo,
AE tr nxn , XE trn. Call t -t x(t) the exact solution x(t)=exp(At)·xo. Note
that t -t x(t) is analytic in t. Call the sequence of computed values.
°
9 Forward Euler method.
For h > and small, we have, for any tk E JR+,
11 m=O,1.2 ....
From the spectral mapping theorem (1) and (11). we have the following.
o
12 Fact. Suppose a(A) c tr_ (equivalently, the origin is exponentially stable). Let
ho be the largest positive h such that
max 11 +hAj I = 1 .
o
Conclusion. Thus even if a(A) c tr_ (and hence the exact solution x(t) -t en
exponentially). for h > ho, for almost all xo. the sequence of computed vectors
blows up. It is for this reason that engineers prefer the backward Euler
method.
(C, devise a graphical test for obtaining the ho defined in Fact (12). [Hint: consider a
unit circle centered on (-1,0»)
°
15 Backward Euler Method.
For h > and small we have, for any tk E JR,
16 m=D,I,2, ....
Note that if Re Aj < 0, then Il-hAj I > 1, since h> 0. Thus we have proven the follow-
ing.
o
17 Fact. Ifa(A)c (C_,then, jorallh>O, 'VXOE (Cn, the computed sequence
(Sm ); obtained by the backward Euler formula (16) goes to en exponentially. •
(The idea behind method (c) is to do one step of backward Euler with step size h/2
followed by one step of forward Euler with step size h/2.)
Select the method that is most appropriate for this problem and justify your choice.
(Hint: What can you say about A and cr(A); consider as k --7 00.)
1 Theorem. Let A and BE (Cnxn. Let L: (Cnxn --7 (Cnxn be the linear map
defined by
2 L :X --7 AX + XB .
Let [Ai).n and r/lj),n be the list of eigenvalues of A and B, respectively. V.l.c.
1-\ J=\
6 L [VjWjoJ := AVjwj" ,
+viWj B=(lI.j+llj)VjWj . 0
8 v*Pw=al1(v*VI)(WI *W)=O
139
iction. Hence the n2 eigenvectors viwt are linearly independent and (3) and (4) hold.
Note especially that (3) and (4) hold for A and B simple, (3.3.23).
II. Consider the case where either A or B or both are not simple. Let us reorder the
elements of X E tr nxn in lexicographic order, say column by column, so that we
obtain an n2 -vector SE tr n 2. Then the eigenvalue equation L (X) = AX is rewritten as
where E (C n2, as defined above, and M is an n2xn 2 matrix whose elements are the
ai/s, bi/s and zeros; M depends continuously on the 1ti/s and bi/S. The eigenvalue
equation
9 det(AI-M) = 0
shows that the eigenvalues A of L are the zeros of a polynomial of degree n2 whose
coefficients are continuous functions of the 1ti{s and bi{s. Hence these eigenvalues are
continuous functions of the 1ti{s and bi/s. Let <l>A' (<I>B), denote the discriminant t of the
polynomial det(AI-A), (det(AI-B), resp.): these are polynomials in the 1ti/s and bi{s
resp. Now for all ai/s and b;{s such that <l>A ¢ 0 and <l>B ¢ 0 we have shown above
that A= for some (i,j); more precisely, we have shown that a) both sides of this
equation are continuous functions of the ails and bi/S, and b) that the equations holds
V Ae tr nxn and VBe tr nxn such that <l>A ¢ 0 and <PB ¢ 0, i.e. the equation holds
on an open dense set of (Cnxnx (Cnxn, (A.6.84a). Hence by continuity it holds for all
Ae (CnxnandBe (Cnxn. (SeeThmA.6.84)
L :X -t X-AXB .
t det{!"I-A) has a multiple zero !.. iff det(AI-A) = 0 and 'h(!..) = O. The discrim-
inant <PAC!..) is a polynomial, which is the derivative of det{AI-A).
CHAPTER 5
Note on Terminology. Since in System Theory we deal mostly with models and
their mathematical representations, there will be no confusion if we follow common
usage and use (1) the word system to refer to the model under consideration and (2)
the expression system representation to refer to any of its mathematical representa-
tions. At a later stage we also use the word system for system representation as
is common in mathematics.
1 Models and their representations: examples. Let the physical system of interest
be an electrical circuit.
Model I. Suppose we model the passive elements as linear time-invariant elements
as shown in Fig. 5.1. A mathematical representation of this model may be obtained by
using the fluxes <1>1' <1>2 and the capacitor charge q as state variables, noting that the
voltage of the voltage source is u(t) and writing the state equations dictated by
Kirchoffs current- and voltage-laws:
q=-L1I<l>1 - Lil<l>2
2 $1 =C-Iq-RIL1I<l>I-U(t)
From the fundamental Theorem of differential equations (B.l.6), we know that given
t Except for the fact that in mathematics systems have usually no inputs.
• •
¢1 ¢2
+ +
L1 ¢2 /L 2
v1
+ q• .+
R1 y
q/C
+
C
Fig. 5.1. This linear time-invariant circuit is described by e:quations (2) and (3).
any to E R, a set of initial conditions q(to) , $\ (to) and $2(to), and any piecewise con-
tinuous input u(') defined from to on, the differential equation (2) defines q(t), $\ (t) and
$2(t) for all t to, and hence by (3) a unique output yet).
Note that (2)-(3) is standard linear time-invariant differential system representation R
= [A,B,C,OJ, (3.2.1)-(3.2.5), with state x= (q,<i>\,<i>2) E R3 and A,B,C,O constant
matrices (check this).
Model II. Suppose that in the previous circuit we take into account the nonlinearities
of the RLC elements (see Fig. 5.2), which are specified by the element characteristics
q=-fL,($l) -fL,.($2)
5 [fL,($\),t]-U(t)
$2 = fc(q) - R2fL,.«h)
and
142
iR1 i2
+ q•
+
vR1 R1 R2 Y
+
Fig. 5.2. If the capacitor charge q and the inductor fluxes $1 and $2 are used as state
variables, this nonlinear time-varying circuit is described by the equations
(5) and (6).
8 x(t) = f(x(t),u(t),t)
9 y(t) = r(x(t),u(t),t)
13 x(k+ 1) = f(x(k),u(k),k) kE N,
14 y(k) = r(x(k),u(k),k)
16 Exercise [Model II, (5)-(6)]. a) With x = (q,«P\ ,cl>2) E 1R3 , u(t) E R, and
yet) E R, obtain from (5)-(6) explicit expressions for the functions f and r in (8)-(9).
From the assumptions on the element characteristics (4), show that. for any given
U(·)E PC(R+,R) and with XE R 3 , the function p(x,t):= f(x,u(t),t) satisfies conditions
(B.1.3) and (8.1.4) of the fundamental theorem of differential equations (B.1.6.).
y) Show that Model II has a differential system representation (8)-(9).
17 Exercise. Model a satellite S as a rigid body with center of mass 0b. Let
(Ob,b l ,b2 ,b3) be a dextral body frame centered at 0b with the b(s along the principal
axes of S. Suppose that gas jets apply to S a resultant force F and a torque (measured
with respect to 0b) t with components ('t\.'tZ,t3) in the body frame. Choose an inertial
frame and write the equations of motion of 0b and of S about 0b.
and a state space (Le. the set of all states). The inputs and outputs are functions of
time defined, typically, on (-00,00) or [0,00) in the continuous-time case or on
( nT: n e Z ) or (nT: n eN) in the discrete-time case. To tackle all possibilities
the time domain of the inputs and outputs will be denoted by T, where T is a subset
of R. The fundamental property of a dynamical system is that given any "initial" time
toe T, any "initial" state xoe and any input u(·)e U, both x(t) and yet) (the result-
ing state and output at some later time t) are uniquely defined. Also, given Xo and to,
the state x(t) and the output yet) depend only on the values of the input u(·) in the
elapsed interval [to,t). We are ready for a formal definition.
i.e. it delivers the state at time t1 reached from the state Xo at an earlier time
to as a result of the input u.
r is called the read-out function: the function r is defined for all times t E T , for
all states at time t, namely, x(t) E and for all inputs u(t) e U at time t, such
that
24 yet) := r(x(t),u(t),t) ,
i.e. it gives the output at time t given the time t, the state at time t and the
input at time t. Thus the read-out function r is "memoryless," (all arguments
25 State transition axiom. For all pairs (tl.to)E (TxT)+. for all XOE L.
if u.u E U such that u(t) = u(t) for all t E [to.ttl n T. then
26 x(t l ) = s(tl.to.xO.u[lo,td)'
where u[to.tll denotes the restriction of the input u to the "interval" [to.ttl n T. i.e.
i.e. to obtain the state x(t2) at time t2 we may first calculate x(tl)=s(tl.to.xO,u[lo,td) at
any intermediate time tt and then use this result to calculate the state at time t2 from
•
30 Any quintuple D = (U 'L'y ,S,T) associated with a dynamical system having the
properties given above is called a dynamical system representation. (Brevity will often
force us to speak about the dynamical system D). •
33 Comment [Dependence of the present state upon the input). Consider the state
transition axiom (26).
a) Given Xo and to, the present state x(tl) depends only on the values of the input u
during the elapsed past [to.ttl n T; the state Xo at time to summarizes the effect on D
of the inputs prior to time to.
Given Xo and to, x(t\) does not depend on the "future" values of the input after
the time tl; by (24) the same applies to the output y(tl) at time t l . This property is
referred to by saying that a dynamical system is
146
40 Example III [Delay systems]. Consider the system of Fig. 5.3. Since the output
at time t is yet), the input to the subsystem with transfer function (s+ 1)-1 is y(t)+y(t);
1 y
u
5+1
t F (T ,U) = {f: f: T
U ); thus F (N,lRn,) is the space of sequences [U(k»)O'
where, V' kEN, u(k) eRn,.
147
also the output of the delay line is y(t-I) since the delay is I se(:ond. Thus
y(t)=-y(t}-y(t-l}+u(t) .
How can we choose a state for this system at to? Reasoning intuitively, we see that
y(to+h) =-y(to+h}-y(to+h-l}+u(to+h)
so
y(to+2h) :::; y(to+h)+h[-y(to+h}-y(to+h-I}+u(to+h)],
and so on. Therefore to calculate the slope of y for any t E (to.to+ I) we need to know
Y[Io-I,101 as well as u[Io,ll' Hence our choice for the state at time to is
x(to) = Y[Io-1.101 .
The state at time to is an element of a function space (more precisely. the space of C l
real-valued functions defined on a unit interval). The system shown is thus a simple
example of a dynamical system whose state space is infinite-dimensional. For more on
this see e.g. [Cur.1] [HaI.2].
into
24 yet) = r(x(t).u(t).t)
we obtain
44 yet) = r(s(t.to.xo.u).u(t).t) .
Thus the output yet) is uniquely defined in tenns of t. to.xo. and u. Let us give the
name of p to the function in the RHS of (44). The function p is defined for all
(tl.to) E (TxT)+. for all Xo E L. for all u E U, and the output is given by
148
We call p the response function: it produces the output at time t in tenns of the initial
time to, the initial state Xo and the input U[Io,I)'
Observe now by (45) that with to and Xo fixed to standard values, (agreed upon in
advance; typically to=O and xo=9), then the output yet) is only a function of t and
U[Io.I)' i.e. for a given input u(·) e U the output yet) depends on t: this results in D hav-
ing an liD representation
46 u(')e U y(-)=F[u](')e Y .
Note in particular that u F [u] maps an input (a function of time) into an output
(another function of time): F is therefore called the I/O map of D.
52 f
y(t)=F [u](t)= H(t-1:)u('t)dt= (H*u)(t) Vt
o
with H(t) = H(t,O) the nonnalized impulse response (3.2.58). Note that in the linear
time-invariant case F is the convolution of the impulse response H (-) by the input u(·).
The discrete-time analogs for Rd(')=[A('),B('),C(')D('») and Rd=[A.B,C,D). are,
respectively, given by
k
53 y(k)=F[u](k)= L H(k,k,)u(k')
k'=ko
where <1>: R R is a nonlinear C 1 map s.t. <1>(0) =0, h(t) is the impulse response of
a linear time-invariant SISO system and u(') e F (lRt,lR).
57 Exercise. Consider the nonlinear feedback system of Fig. 5.4, where <)I: 1R 1R
is a nonlinear C 1 map S.t. <)1(0)=0, h(t) is the impulse response of a linear time-
invariant SISO system. Let the inputs u('), the errors e(') and the outputs yO belong to
an appropriate identical space L = E = U = Y of R-valued functions on 1R+.
Assume that the open-loop I/O map
58 e(')e L yO=G[e]O=[h*<1>(e)]Oe L
is invertible, (I denotes the identity map), more precisely (I+Gf i maps L into L.
u y
J
Fig. 5.4 Nonlinear feedback system.
u(')e L Y(')=F[u](')
60 = G[CI+G)-I[u]]C') e L .
•
61 Exercise. Show that the I/O map of the system of Fig. 5.3, where Y[O,-IJ - 0,
reads
F [u](t) = (h*u)(t)
where
150
Explain physically this series expansion. Sketch out h(t) for t 2:. O.
Clearly Ttu is the result of shifting u(·) to the right by 't seconds, i.e .• for't > 0, Ttu
is the input u delayed by 't seconds. U c F (T .U) is said to be closed under transla-
tions iff. for every 't E T, u E U ::::> TtU E U.
Observe that the value of the new input Ttu at the new initial time lQ+t is just u(to).
which is the value of the old input u(·) at the old initial time to. Thus (3) expresses
the fact that by delaying the application of Xo and the input u by t seconds. the output
y is delayed by the same amount. •
7 x(t) = f(x(t).u(t»
tE R
8 yet) = r(x(t).u(t»
151
2 p(t,to,a,x, + +
•
3 A dynamical system is said to be nonlinear iff it is not linear.
•
3a Remark. We purposefully state that to check whether a dynamical system is
linear we have to exhibit a dynamical system representation D that has linearity pro-
perties (a) and (b) above. Some representations may hide the fact that the system is in
fact linear. Consider the following simple scalar example:
x=-x+u
y=x.
+ (sech u
L y= sinh
In the the system appears to be nonlinear!
152
= +
6
response z-i response z-s response .
•
7 Linearity of the zero-input response. If D is a linear dynamical system
representation, then
V (t,to)E (T xT)+, VXI,x2E L' V(Xl,(X2E F,
A little thought shows that the linearity property (10) is the basis for the superpo-
sition integral (5.1.51).
5.4. Equivalence
1 Equivalent States. Let D and jj be two (not necessarily distinct) dynamical sys-
tem with the same input and output spaces U and Y. States Xo D
and Xo of D are said to be equivalent at time to iff D in the state Xo at to and D in
state Xo at to. when driven by any input u"o.ro) E U. yield the same output for all t to;
thus, with p and p the response functions of D and jj resp., VUE U, Vt to,
2
•
153
3 Exercise. Consider the linear time-varying system of Fig, 5.5, where the switch is
closed prior to t=1 and open for t 1. Pick the state x= E R2. Consider the
states xo=(O,I) and "0=(1,1) at times to=2 and to=O. Show that these states are
equivalent at to=2 but not at to=O.
i.e. they have the same set of I/O-pairs, where it is understood that Y[to. oo ) is the output
of D produced by [Xo,u[lo.oo)] and Y[to. oo) is the output of jj produced by ["o,u[to.OO) J.
(note the same input u).
9 yet) =r(x(t),u(t),t)
Let
10 ",:]R.n R n:
be a bijection S.t. '" and its inverse ",-1 are Cion ]R.n. Note that, for any ERn, the
Fig. 5.5
154
12 x= .
and by (8). (9). (11) and (12). there results a new differential system representation D
of the form
13 = . =:
14 yet) =
(prove this: note that by (11). t -+ x(t) is C l iff t -+ is Cl). Observe moreover. by
(11), that
V to E R+, the state Xo of D is equivalent to the state of jj at to. and
conversely,
V to E R+, the state of jj is equivalent to the state Xo := of D at to.
Hence D and fj are equivalent, having V to E T and V u[to. oo)' the same output
•
The nonlinear time-invariant state transformation (11) can be extended to the
time-varying case. For simplicity, we suppress the time dependence of x(t) and u(t)
where there is no ambiguity.
",(.,.) : R n x -+ lRn : -+
= [D 1 =: tE Rt.
=:
[Hints: x= Dl . + t x(t) is C 1 iff t is C1; exhibit equivalent
states.]
23 Remark. The definition of zero-state equivalence does not imply that the state
space 1: and thave the same dimension: see e.g. the Kalman decomposition Theorem
(8.6.16).
29 = [T(t)A(t)+T(t)]T(t)-lx(t) + T(t)B(t)u(t)
33 C(t)=C(t)T(t)-1 D(t)=D(t)
Furthermore u('), x(·), yO satisfy R 0 if and only if u('), T(')x('), y(.) satisfy RO.
36 Comments. a) In the study of stability, (see Chapter 7), it is important that
T(') as well as TO-I are bounded on x
whence the exponential stability of = A(t)x
becomes equivalent to that of i = A(t)", [Bro.1,p.188].
Algebraic equivalence is a special case of the dynamical system equivalence of
Exercise (15).
T (.) T20
2) if R -I-RIO and RIO --RIO, then
R 0 T2(')Tt<") R 2('),
41 a) ¢(t,"C)=T(t)¢(t,"C)T("C)-1 Vt,"CE ,
43 Proof. a) Note that the function t T(t)<l>(t,'t)T(t)-1 jis equal to I for t = 't and
satisfies the d.e. X= A(t)X fort a.a.t E 1R+ as can be easily checked using (32).
Hence, by the fundamental theorem of d.e., (41) follows.
b) Equation (42) follows immediately when (41), (32) and (33) are substituted in
= p(t,to,xo,u).
Hence 'V to E R+, the states Xo and Xo are equivalent at to iff Xo = T(to)xo. Therefore
R (-) and RO are equivalent. •
56
57 x(k) = T(k)x(k) .
Furthermore u('), x('i, yO satisfy R d(') if and only if uC'), T(·)x(·), y(.) satisfy RdO.
Therefore R dO and RdO are equivalent system representations with
61 Exercise. Let R = [A,B,C,D] and R= [A,B,C,D] have (1) the same input- and
output-spaces and (2) state spaces of the same dimension, say n.
Let TE R nxn be any nonsinguiar matrix. Show that
iff
63 A=TAr 1, B=TB, c = c r 1, D=D.
This concludes our brief overview of general system concepts; the point is that
these concepts are far more general than the special case of linear finite-dimensional
differential systems.
CHAPTER 6
Consider the basic laws of Physics that describe the behavior of physical objects:
Newton's laws, Lagrange's equations, the Navier-Stokes equations, Maxwell's equa-
tions, Kirchhoff's laws etc. Each of these laws describe continuous-time phenomena.
At present it is cost effective to manipulate signals in digital form: for this purpose,
signals are sampled by an AID (analog-to-digital) converter, the resulting sequence of
numbers is operated on by a digital computer (a controller) and the resulting sequence
of numbers must be restored to analog form (Le. to a continuous-time form) by a D/A
(digital-to-analog) converter. Indeed, the analog form is required to actuate the physi-
cal devices that are to achieve the engineering goals.
In communication systems, at the sending end the signals from a microphone or a
picture tube are sampled and transmitted in digital form. At the receiving end the sig-
nals are restored to analog form to actuate microphones and/or picture tubes.
In control applications, the continuous-time plant-output yc(t) is sampled by the
AID converter to produce the digital output Yo' The digital inputs Uo and Yo are fed
into the digital controller to produce the digital signal vo, which is then transformed to
a continuous-time signal vc' The latter is then used to drive the plant.
The aim of this chapter is to develop tools to study such problems, namely, linear
systems with digital and analog signals.
All the derivations are carried out in the SISO case, however they apply to the
MIMO case, as explained in Remark (6.4.4) below.
1 f has a Laplace transform with abscissa of absolute convergence O'f. Thus £(s) is
DIGITAL
,--- CONTROLLER
I
I
analytic for Re s > O'r and, for all 0' > O'r, f(s) -) 0 uniformly as I s I -) 00 in c:: (H' by
the Riemann-Lebesgue lemma (C.2.3).
As shown on the figure by the operator S, f: 1R+ -) IR is periodically sampled
(with period T) by an AID converter. Its output is the sequence of samples
Now it often happens that f is identically zero for t < 0, but f(t) "# 0 for t > 0
near zero: thus f(O+) "# O. By convention fo is taken to be f(O+); this suggests that
the sampling process commanded at time t=O had a minute delay in the execution of
the sampling. For these reasons, we assume that
t If, for some M and some c < 00 and Vt 0, I f(t) I ::;; M exp(ct) then
Pr ::;; exp(cT).
162
CONTINUOUS-TIME
FUNCTIONS SAMPLED-
FUNCTIONS
f: IR+... IR S
f(O+), flT),f(2Tl, ...
O"f < IX)
tL
LAPLACE TRANSFORMS
JZ
Z-TRANSFORMS
1\
f:C ... C
.....
T f(·)
O"f
00
then it can be shown that L is an injective map defined on the resulting equivalence
class. Now the inversion integral (C.2.11) recovers f, a representative of the
equivalence class.
8 Even with assumption (2), S is not injective. Indeed, the function f1(t):= 1,
f2 (t) := 1 + sin and f3(t) = 1 + e sin
t have the same sampled sequence. Note that
if we doubled the sampling rate, the sampled sequences would be different in the three
cases: hence it is important to sample fast enough. These examples illustrate the
well-known phenomenon of aliasing [Kwa.2] There is one case where S is injective.
00 sin [; (t-nT) ]
f(t) = L f(nT) --"----"-
-00 (t-nT)
T
163
Since by (8) and the examples above, two different time-functions f(') and g(')
may have the same sequence of samples, it follows that different f and gmay produce
f and g that are equal: in other words
10 T is not injective.
11 Theorem. Assume that f: --+ IR satisfies assumptions (1) and (2); let f be
defined as usual and let f be defined by (3); choose c E R and constrain s E € and
ze € by
12 Of < c < Re[s] and z=esT ,
14 (ii) -f(z)=f(e
- ST Il
)=- f(O+)+- +00.f 21tk 1;
fs+j-
2 T T
k=-oo
• n rk
15 (iii) if, in addition, f(s) = - - ; then
k=1 S-Pk
- n
fez) =
Z
16
•
rk -----'.:.--
k=1 z-eXP(PkT)
The map s -+ exp sT. Equation (3) shows that for sampled data systems, the
variables z and s are related by z = esT: this map takes the point s = O+jOl into the
point z = eoT(cos coT +j sin coT).
(d) Add to the above s-,Iane rectangle the triangle specified by the vertices
[oo,j ; ]. (0,0) and [0 -j; ;obtain the z-plane image of the resulting polygon.
0,
For example, left half-plane poles of ( map into poles of f that are in D(O; 1), the
unit disc centered on the origin.
23 Zeros of '(s) and f(z). Equation (14) shows that there is no simple relation
between the zeros of f(z) and those of (s). There is, however, an interesting limiting
case:
24 Asymptotic property of the zeros. t::t (s) be rational and let f(s) = O( 1/s2) for
I s I -+ 00; call ... ,Zm the zeros of f(s), then, as T -+ 0, (i.e. as the sampling
rate increases), lim f(exp(zjT» = O.
- 1 1
f(z) = "2 f(O+) + - [f(O-)+f(O+)] + L00
27 f(kT)z-k.
2 k=1
165
28
For each k and for fixed z. the integral converges absolutely since of < c; furthennore
since I z I > exp cT. in the integrand I z-k exp(pkT) I < 1 along the integration path;
therefore we may interchange the order of summation and integration and we may sum
the geometric series. and thus obtain
c+jm
I
A
f(z)=f(O+) + r.
k=1
f(kT)z-k
29 'VUE Z.
These poles lie to the right of the vertical integration path of (13a).
Note that if f(p) itself has poles. say, PI.P2' .... then for all such poles.
30 Re Pk Of < C
and they to the left of the integration path. Furthennore. f(p) is analytic in
Re s > Of: f is analytic on the integration path and to its right.
Note that the two observations above also hold for the integrand of (l3b). By
assumption (1) and the Lebesgue lemma (C.2.3), for all 00 > Of. as
Ipl 00. with Rep> 00. f(p) 0 uniformly. In order to calculate (13a). let us
use the residue theorem: so we first close the integration path to the right by a half
circle of radius R and then let R -+ 00. Since on the half circle p = R exp j6.
Since f(p) 0 uniformly on the right half circle. by the Jordan lemma (C.2.26) the
166
contribution of the half circle goes to zero as R -+ 00. Hence, in the limit of R -+ 00,
all the integrand poles to the right of the vertical path are included and, by calculating
the residues, we obtain (14).
" rl
33 Exercise. Let f(s)= (1-e-sT) - - . a) Show that we cannot close the
S-PI
integration path on the left to calculate fez) by (13a). b) Use (l3b) to calculate fez).
c) Use time-domain calculations to get
- z-1
f(z)=rl (T) .
z-exp PI
01 A It-_uc__
Fig.6.3 D/A converter.
The D/A converter shown in Fig. 6.3 maps input sequences into piecewise con-
stant functions; more precisely if the input is the sequence Uo = [uo(nT»); then the
output is the piecewise-constant function uc: t -+ uc(t) given by, 'V t 0,
1 uc<t)= l:
n=O
[1(t-nT)-I(t-(n+l)T)] uo(nT) .
T I -e-sT
3 uC<s)=uo(e S
A
) --- •
s
o T 2T 3T 4T 5T 6T
1 yo(nT) := yC<nT)
Note that (5) is even valid when Ys(s) = O( lis) as I s I -+ 00 - equivalently when
'* 0 - '*
m
Lrk in that case, yc<Ot-) yC<G-) = 0 and for (5) to be valid the first sample
1
must be yC<O+).
168
Yc YO
AID 1-----
--'
Fig. 6.5 NO converter: Ye is the continuous input.
1 a) the D/A and the AID converters have the same period T and are synchron-
ized.
Let Pl.P2.....Pm denote the poles of gthen the abscissa of absolute convergence of
g is Og := m!lX I
(RePi)'
3 Yo(z) = geq(z)uo(z)
and
I
,,-----4 CLOCK 1-----------7
I
_ u_
D__
11DIA
I I Uc
..
LINEAR SYSTEM
9(8)
Yc
EJ-'
AID ---
Yo
Fig. 6.6. Overall system under study: geq(z) is the pulse transfer function
from the input sequence Uo to the output sequence YD'
169
n
4 yo(n) = 1: geq(n-k)uo(k) .
k=O
Note that yc<s) is a rational function of s that is at most 0(1/s2) as 1s 1 -+ 00. Since
1.s g(s) is the input to the AID converter. its output is specified by (see (6.3.2»:
6 yo(z)=T [! g(S)]'
Finally
7 geq(z)=(l-z-I)T [+ g(S)]
and by (6.1.13a)
c+joo
8 ge I glat(p) dp
q 21tJ c-joo p[1-z-1exp pT]
I.
c+joo
10 - (z)=O-z-l) _1_ exp(pT) iiPl dp
geq 21tj z-exp(pT) p
C-jOO
11 Theorem. Let assumptions (1) and (2) hold, (i.e. synchronization and gO
rational and strictly proper), then the pulse transfer function geq(') relating the input
sequence uD to the output sequence Yo (see Fig. 6.6) is given by
170
where geqO is a strictly proper rational function in z, which is given by (8) and (10).
12 Exercise. Let assumptions (1) and (2) hold. Call Pl,P2, ... , Pm the poles of
g(s)/s, (they may have order 1).
a) Use (1) to show that
13 geq(z) = (l-z-l) f
i=l
Residue [ exp(pT) .
Pi z-exp(pT)
ili ] .
p
b) If, in addition all the poles of g(s)/s are simple, show that
14 geq(z) = (l-z-I)
m
E exp(PiT [g(S) ]
Residue.
i=1 z-eXP(PiT) Pi s
m' r
c) If, in addition, g(s)= E _k_, (1tkE 0: is the kth pole of gO), then
k=1 s-1tk
m' rk exp(1tkT)-l
15 g (z)-1: -
eq - k=1 1tk z-exp(1tkT)
16 Remark on Zeros. Examples show that even if gO is minimum phase, i.e. has
no €+-zeros, then geqO may have zeros outside the unit disk. [Ast.l]. For example,
g(s)= 1/s3
19u g(z)=T[g(s)] .
Now geq(z) is the pulse transfer function of the system shown in Fig. 6.5: its use is to
calculate Yo from lIO by (3): from Fig. 6.5 we see that
20 geq(z) = T [-7- g
1 -sT ]
(s) .
7 geq(z)=(l-z-I)T ].
g
Equation (7) is convenient because (s)/s is at most 0(1/s2) as I s I 00, hence we
may close the integration path of (8) and (10) to the left without changing the value of
the integral.
6.5. Example
Consider the control system shown in Fig. 6.7: the digital input is uo and the
continuous-time output is called Ye. The digital controller is specified by its pulse
transfer function c(z): hence vo(z)=c(z)co(z). The continuous-time plant is specified
by its transfer function g(s) which is assumed to be strictly proper and rational. The
NO and D/A converters are assumed to be synchronized. Hence geq(z) is the pulse
transfer function relating \to to Yo:
t More than one author uses the same notation for two different functions!
172
1 YD(Z)=Seq(Z)VD(Z) .
Note that y(s) is the product of five factors: the last three are periodic in s with period
. 2x
JT'-
STABILITY
This chapter describes the stability of a linear system from different points of
view: bounded-input bounded-output stability (I/O representation, external stability)
and the stability of the zero solution of x= A(t)x (state representation, internal stabil-
ity): for the latter case the notions of asymptotic- and exponential-stability are
developed. More specific topics conclude the chapter: bounded trajectories and regula-
tion, response of a linear stable system to T-periodic inputs, equilibrium solution of a
driven T-periodic differential system and slightly nonlinear systems.
where
H(','):RxR RDoXn; and 'Vte R, t H(t,t) is piecewise continuous,
u(')e PC (R,Rn,) S.t. the integral in (I) makes sense, (e.g. u(t)=9 'Vt < 0),
y(-):R Rllo.
4 IIxli := max I xi I
i
6 lIull
-
:= sup lIu(t)1I = sup {max Uj(t)
JElli
I I},
7 /lyll ::;;: sup IIy(t)ll = sup {'!lax I Yi(t) I } .
00 IE. lEn.
For a discussion of linear nonned Loo-spaces, see (A.6.S4) et seq.. Remember that a
function belongs to L 00 iff its L--nonn is finite or equivalently the function is
bounded. t
y) An attempt will be made to view the I/O map F : u y of (I) as a linear
transfonnation from L;.' into
8 I/O stability. The concept of input-output stability, (I/O stability), roughly asserts
that "any bounded input uO produces a bounded output yO using a finite gain." More
precisely:
We say that the system described by (1) is I/O stable iff
11 F : L;' ;u F [ul =; y ,
12
Hence by (9), the system described by (I) is I/O stable if and only if the I/O map F
maps L.;:' into such that there exists a k < 00 for which IIF II < k < 00.
t Mathematical purists should read "essentially bounded" and replace the sup in
(6)-(7) by the essential supremum.
175
II/II 00 > I .
•
14 Ilulll 00 = 1 and
17 Theorem [I/O stability]. The linear system described by (1) is I/O stable iff
The proof below shows that k, defined by (18), is an upperbound for the induced
norm of the I/O map F given by (12); hence it may be used as the fixed constant k of
I/O stability definition (9).
y) In (18) above, any matrix norm may be used since matrix norms are equivalent,
(A.6.45).
24 Proof of Theorem (17). 1) Sufficiency. We prove that (18) implies (9). There-
fore let uO be any input that is bounded on 1R, i.e. whose norm lIuli oo given by (6) is
finite. Then from (1), taking vector norms, we have, successively, 'It t E IR
s; f IIH(t,t)lIllu(t)lldt
-00
(we used successively (A.6.20), (A.6.17), (6) and (18». Hence, by (7), taking the
supremum w.r.t. t we obtain (9) with k as defined in (18).
2) Necessity. We use contraposition, i.e. the negation of (18) must imply the nega-
tion of (9) (i.e. system (1) is not UD stable). Therefore, let us negate (18); then,
negating equivalence (20), we obtain: there exists E !10 x OJ S.t.
2S
26 f IhJl(tl,t)ldt > I.
-00
with
sign[h 11 (tl,t)1 V t E (-co,td
{
27b ut(t):= 0 elsewhere
28 VI E N
I,
29 y/(tl)= f Ihll(tl,'t)ld't,
Hence by (28) and (30) the system described by (1) is not I/O stable by Exercise (13) .•
34 = JCCt)<ll(t,'t)B('t)u('t)dH D(t)u(t)
o
where uC·) e PC and y(.) e PC ClR+,lRl1o ).
35 We say that Cthe system described by) R (.) is I/O stable iff condition (9) holds
for the I/O map (34).
Note in particular that only the first tenn of the last expression of (34) is a superposi-
tion integral. However if DO is bounded on lR+ then R 0 = [AC'),BC'),CC'),DC')] is I/O
stable iff RO=[A(-),Bo,CC'),O] is I/O stable, (prove this). Therefore by Theorem
(17) and Comment (19. y), we have the following.
36 Corollary. Consider a linear system representation R (-) = [A(' ),B(· ),C(· ),D(-)l
where DO is bounded Then R (-) is I/O stable iff
then H(t,'t) depends only on the elapsed time t-'t. Thus J IIH(t,'t)lld't= JIIH('t)lIdt,
o
178
i.e. the first expression is independent of t e IR. Hence, by Theorem (17) we have.
41 Corollary. Let the system described by (1) be time-invariant. Then the system
described by (1) is I/O stable iff
00
R is 110 stable
00
44
A 0
45 P [H(s)] c (1:_ .
46 Comment. (45) means that B(s) has no poles in (1:+, or equivalently is analytic
on (I: +: a well known characterization of "external stability."
47 Proof of (44) <::> (45). Set G(t) = C eAtB 'v'te R+, hence B(s)=G(s)+D.
Observe now that (44) <::> (45) is equivalent to 'v' (i,j) e lloXfij
00
o
P [gij(s)] C c:_ .
I
48 gij(O = I: 1tk(t)exp[Akt] \;f t E R+ ,
k=1
o
where \;f k, Ak is a pole in Ci:_ and 1tk(t) is a polynomial in t.
Observe now that as t 00, any polynomial in t will grow at a slower rate than any
growing exponential exp[et] with e > 0, whence for any given polynomial 1tk(t) ,
\;f e > 0, \;f k E L
3 mk(e) E Rt s.t.
49 11tk(t) I mk(e) exp[et] \;f t E R+ '
00
f I gij(t) I dt
•
with Il-e > O. Hence m(ll-e)-1 < 00.
o
51 Comment. The proof above shows that for any proper transfer matrix
H(s) E Ci:p(S)n.><n1 with 0(8) := H(s)-H(oo) E Ci:p.O(S)n.><nI,
A 0
52 P [R(s)]!:: Ci:_
¢>
In other words the impulse response (corresponding to the strictly proper part of H(s»
is exponentially decaying with a positive decay rate arbitrarily close to J.1 > O. Hence,
the following result,
o
P [R(s)]!:;; (L.
•
7.2. State Related Stability Concepts and Applications
In this section we obtain some results that are partial answers to the following
questions. Given a linear differential system R (-) represented by
Under what conditions on A('), BO, C('), DO does a bounded input produce for all
XOE IRn and all toE R+ bounded state- and output-trajectories on [to,oo). In the first
subsection we consider only the state trajectory with u=elj, (i.e. the zero-input case)
and define concepts that guarantee that the state tends to zero as t 00. In the next
subsection we consider R (.), i.e. (I) and (2), and insure bounded trajectories for both
the state and the output; in addition we guarantee regulation, i.e. when the input tends
to zero as t 00, then so does the state and the output. The section concludes with
some specific applications.
3 x=A(t)x Vte
where x(t)e R n and A(')e PC(R+,Rnxn). We now take the differential equations
point of view, e.g. [Mil.1]. For obvious physical reasons, we restrict ourselves to the
forward motion of the state, by (2.1.31) the solution of (3) reads:
4 x(t) = cl>(t,to)xo
5 x=f(x,t)
where f('''): Rn x R R n is sufficiently smooth S.t. V (xo,to) E IRn x IR+ (5) has a
unique solution x(') = $(' ,to,xo) that is continuous on IR+. On comparing (5) and (3)-
(4), we observe that, in the linear case with L=lRn , in order to obtain all state trajec-
lSI
tories starting from to we need only know the solutions of x=A(t)x fort XO=Ei for all
i E n; moreover generically there is only one equilibrium point. viz. x ==
e. This is.
in general. not true in the nonlinear case: equation (5) may have more than one equili-
brium point (some stable or unstable); furthermore. one may encounter so-called
periodic closed orbits. some of which are stable and some unstable. For example,
consider the motion of the point (r,e) in the plane described by
r.=-sin r
{
e= r.
There are infinitely many solutions corresponding to periodic closed orbits, to wit: for
kEN, the equations
rk(t)= klt
{
Ok(t)=km
define state trajectories that are circles, say Ck , of radius klt, where for k=O Ck is a
stable equilibrium point, for k=2,4 .... , Ck is a stable closed orbit. and for
k= 1.3.5 ..... Ck is an unstable closed orbit. Indeed. consider the following initial con-
ditions:
1) if 0 < ro < 1t and 00 is arbitrary. the state trajectory spirals into the origin (i.e.
Co);
2) if 1t < ro < 21t and 80 is arbitrary, the state trajectory spirals outwards and
asymptotically approaches C 2;
3) if 211: < ro < 411: .... etc. (The reader should draw a sketch of these orbits and of a
few state trajectories.)
Thus. depending on how far away from the origin the initial state (ro,Oo) is. the nature
of the solution is qualitatively very different.
For more on this, see e.g. [Mil.lJ. In this section we will take full advantage of the
linearity of the d.e. (3).
Asymptotic stability. This property roughly asserts that every solution of x= A(t)x
tends to zero as t 00. More precisely we have the following.
6 Definition. We say that the zero solution x(t) == 8 of x=A(t)x (on t ;?.O) is
asymptotically stable (asy. stable) iff. for all Xo E IRn, for all to E 1R+
a) t x(t) = <l>(t,to)xo is bounded on t ;?. to,
b) t x(t) = <l>(t,to)xo tends to zero as t 00. •
9 0 as t 00 •
equivalent to
This interpretation is important for the definition of uniform asy. stability in Comments
(19) below.
\3) t can be calculated by solving successively x = A(t)x from Xo = Ej at
10=0 for all ie n.
11<1>(t,O)llII<1>(O,to)llllxoll
(where we used e.g. the sup vector norm, (A.6.17), (2.1.56), and (A. 6.1 8». Now
is a constant independent of t and equivalence (11) holds_ Hence by
(9) we satisfy the conditions of definition (6).
Necessity. By contraposition. So assume that t <1>(t,O) does not tend to zero as
t 00. This must hold for at least one entry of <1>(·,0), say w.l.g. <l>1l(·'0). Now
t <l>ll(t,O) is the first component of the state trajectory x(·) due to XO=EI (the first
standard unit vector) at to = O. Thus, this solution of x = A(t)x does not tend to zero
as t 00. Hence the zero solution of x=A(t)x is not asy. stable. •
Exponential stability. This property roughly asserts that every solution of x= A(t)x
is bounded by a decaying exponential depending on the elapsed time !-to- This
technically most desirable property is not always guaranteed by asymptotic stability.
16 Example. The linear time-varying circuit of Fig. 7.1 is described by the state d.e.
q=-q/(l+t)
and the zero solution is asy. stable. However for every Cl > 0 and for every solution
q(.), Iq(t)lexp[Cl(t-to)) 00 as t-to tends as 00. Hence Iq(t)1 cannot be bounded by
a decaying exponential.
We are now ready for a formal definition.
19 Comments. Cl) The positive constants Cl and m are fixed, i.e. independent of
toE R+; in particular Cl is afixed decay rate.
Using an induced matrix norm i.e.
11<1>(t,to)11 := sup (11<1>(t,to)xoll/llxoll : Ilxoll :t 0) we see that. with x(t) = <1>(t,to)xo.
exponential stability is equivalent to:
x
In other words every solution of = A(t)x is bounded by a decaying exponential in
t-to. whose defining constants mllxoll and Cl are such that m and Cl are fixed.
y) By abuse of language, the expression "the zero solution of x A(t)x is expo =
stable" is often replaced by "x = A(t)x is expo stable" or "the equilibrium x = e is expo
stable" or "A(·) is expo stable."
B) [Mil.l,p.188). A stability concept equivalent to exponential stability is the fol-
lowing:
x
The zero solution of = A(t)x on t 0 is said to be
[It is crucial to note that the constants k in (21) and T(e) in (22) are independent of
toE compare with the conditions (11) for asymptotic stability.]
e) In the time-invariant case x= Ax, (with A constant), asymptotic stability is
equivalent to exponential stability. Indeed <l>(t,to) = exp[A(t-to)] depends only on the
x
elapsed time t-to; hence the zero solution of = A(t)x is asy. stable, (i.e. (11) holds),
iff the zero solution is uniformly asy. stable, iff the zero solution is exponentially
stable.
<=:
=>: Observe that given any T > 0,
°
A(') is uniformly asy. stable ¢:> A(-) is expo stable, is as follows:
(21) and (22) follow from (18) with k=m and T(e) > S.t. exp[-aT] :5 em-I.
26 t-to=nT+s.
Let toE IRt be arbitrary but fixed and pick in (22) some T(e) > for e= 112. Then, °
by (22), 11<l>(s+to+T,to)11 :5 112 V s O. Hence, using composition and induced
matrix norms as in e.g.
<I>(s+to+2T,to) = <l>(x+to+2T,to+T)<l>(to+T,to) ,
36 exp[At] = f
k=!
1tk(t)e""I, where {Ak}O = o(A)
!
and
where the Pk(t) are polynomials S.t. Pk(t) II7tk(t)ll, p(t) := L° Pk(t) 0 and
k=l
o
41 Proof of Theorem (33). If a(A) c 4L, then, by (38), > O. Hence picking
EE we have that (18) holds with > O. Therefore sufficiency holds.
o
On the other hand if a(A) is not included in (1:_, then by (36) exp(At) does not tend
to the zero matrix as t 00 and the zero solution is not expo stable. •
46 A*P+PA=-Q
where A E (1:nxn, Q E (]:nxn S.t. Q= Q* > 0 are given t and a unique solution
p = p* > 0 is to be found. Equation (46) is called the Lyapunov equation. Its solva-
bility relates directly to the expo stability of x = Ax.
47 Lemma. Assume that given Q=Q* > 0, equation (46) has a unique solution
p=p* > 0, then the zero solution of x=Ax is expo stable, i.e. o(A)c
Indeed, P is Hennitian positive definite and P/'Pu can be chosen to be its least and
largest eigenvalues, resp.. Taking the derivative w.r.t. t of vO along any trajectory of
x=Ax, we have by (48) and (46)
Consider now the converse of Lemma (47). As a first step consider the following.
o
52 Exercise. With A E (Cnxn, let a(A) c (L. Consider the map
53 F : c;nxn -7 C;nxn: X -7 A*X+XA
o
S4 Lemma. Consider the Lyapunov equation (46) and let o(A) c (L. Then
'ltQ=Q* > 0 equation (46) has a unique solution p=p* > 0 given by
f eA*'QeA'dt .
00
55 P=
o
X=A*X+XA+Q, X(O)=O t
Now since A is expo stable it follows that the integral converges, as t -t 00, to the
limit
,
P := X(oo) = f eA*'tQeA'tdt ,
o
s.t. Q= R*R, if x is S.t. /Px = 0, then f IIR eA'x1l 2dt = 0, i.e. R eA'x === e, whence
o
x= e because ReA' is nonsingular 'It t E IR and V A E R nxn and all nonsingular
R.
•
From Lemmas (47) and (54) we have now our main result.
S6 Theorem. Consider the Lyapunov equation (46). Then the following statements
are equivalent:
a) 'ltQ=Q* > 0 the Lyapunov equation (46) has a unique solution p=p* > 0
given by (55);
o
b) o(A) C (L;
c) the zero solution of x=Ax is expo stable.
•
o
S7 Exercise. With o(A) c (L and P given by (55), show that
189
fo -dtd
00
*
A P+PA= [eA*'QeAtjdt=-Q .
60 x=A(t)x tE R;.
61 A(t+T)=A(t)
63 <b(t,to) = P(t)exp[B(t-to)]p(to)-1
where P(t)E cr nxn is nonsingular for all tE and t --t pet) is T-periodic, and b)
under the coordinate change x(t) = P(t)!;(t), (60) becomes a time-invariant d.e.
64
(where B is the constant matrix (62». Our objective is to show that the zero solution
of (60) is expo stable iff the zero solution of (64) is expo stable. Note that, by
o
Theorem (33), the latter holds iff O'(B) = cr_.
65 Exercise. Consider equation (62) and let D(O; 1) denote the open unit disc, i.e.
D(O;I):= {AE cr: 11..1 < 1). Show that
o
O'(B) c cr_
if and only if
O'[<I>(T,O)] c 0(0; 1) .
66 Theorem [Exp. Stability). Consider the T-periodically varying d.e. (60). Then
the zero solution of x = A(t)x is expo stable if and only if
190
67 a[<D(T,O)]cD(O;I).
68 Short proof (Exercise). Consider equation (63). Since pet) is nonsingular for all
te and PO is T-periodic and continuous, P(·) as well as PO-I are bounded on
equivalently, there exists positive constants M and N such that, for all t e 1R+,
IIP(t)11 S M and 1!P(t)-lli S; N. Therefore, by (63), for all toe R+, for all t
(MN)-llIexp[B(t-to)lII S; 11<1>(t,to)11 S; (MN)llexp[B(t-to)]II.
Hence, (18), the zero solution of x = A(t)x is expo stable if and only if the zero
solution of = is expo stable. The latter condition is equivalent to condition (67)
by Exercise (65). •
x
69 Comments. a.) Clearly the periodically varying d.e. = A(t)x is expo stable iff
the time-invariant d.e. = BS is expo stable.
Condition (67) can be tested easily; compute <1>(T,O) by integrating x = A(t)x over
[O,T] starting from XO=Ei for all i en; compute the eigenvalues Ai of <1>(T,O) and
check if I Ai I < 1 for all i e n.
70 Exercise. Consider the periodically varying d.e. (60). Assume that o[<1>(T,O)] is
not included in the disc D(O; 1) and let e E cr n be an eigenvector of <1>(T,O)
corresponding to an eigenvalue AE cr with I A I 1. Show that the solution
x(t)=<1>(t,O)e does not tend to zero as t -7 00.
[Hint: use <D(t+T,to+T) = <1>(t,to).]
where any vector nonn may be used for IIf(t)lI. Of course, IIflloo.o=lIfll oo ' i.e. the
usual Loo-nonn. For matrix functions M('): R+ lRnxm we denote by IIMII 00 the
Loo-nonn, i.e.
74 II M 1100 IIM(t)1I
where the matrix norm IIM(t)!I is the one induced by the chosen vector nonn.
Hence, if both u(t) and x(t) tend to zero as t 00, then so does yet). Hence, we are
reduced to show that
192
J
x(t)=<l>(t,to)xo+ <l>(t,t)B(t)u(t)dt =: X\(t)+X2(t)
10
tends to zero as t 00. This is immediate for xl(t) since x=A(t)x is expo stable. So
we are left to prove that lim X2(t)= 8. By taking vector norms and using (18),
1--> 00
I
and we are done if the integral on the RHS converges to zero as t 00. For this
e
purpose, first set u(t) = for t < to and then observe that
10 0
[in the last expression Ilu(t-t)1I = 0 for t > t-tol. Take now any sequence
[tk);C[to,OO) s.t tk 00 as k 00, and set 'ik, fk(t):= exp[-at] IIU(tk-t ) II
for t e R+. Thus we are done if we prove that
00
lim Jfk(t)dt = 0 .
k-->oo 0
the sequence (fkO ]; is dominated by the fixed absolutely integrable function gO. •
86 Exercise. Consider Exercise (84). Show that if A has some j<o-axis eigenvalues
then the bounded-input bounded-state stability of R is not robust w.r.t. small pertur-
bations in B.
87 Important comment. From Exercises (83), (84) and (86), we see that (in the
time-invariant case), for reasons of robustness, it is desirable that A be expo stable (Le.
o
a(A) c ([ J for obtaining bounded state and output trajectories and regulation. Now
in many cases it will be known that R is I/O stable. So an important question is: can
• 0
we decide that A is expo stable from the I/O stability of R, (Le. P [H(s)] c C_,
(43»1 The answer is affirmative iff A has no unstable hidden modes: see Corollary
(9.1.80) below. Similar facts can be developed for the time-varying case, but they are
quite technical, e.g. [Eng.1]. •
90
i.e. Vp(·)E PC[lR, .r"il is the unique T-periodic function on R s.t. vp(t)=up(t) for all
tE R+.
yet) = p(t,O,xo,u p)
t
95 = C exp[Atlxo + JH(t-'C)up(t)dt
o
=Cexp[At]xo + f H(t-t)vp(t)dt
o
(since the extension vp(·) agrees with upO on Hence, by (93), (95) and (3.2.58),
o
96 'VtE R+ y(t)-Yp(t)=CeAt[xo- J exp[-AtlBvp(t)dtl
where the last integral converges because A is expo stable, (indeed, using norms, (18)
and the notation I for the integral, 11111 (mJa)llBllllvplloo < 00). Therefore in the
RHS of (96) the expression in the brackets is a constant vector. Therefore since
eAt 0 for t 00, 'V Xo E.r" lim [Y-Yp](t) = 0. From this and claim (94) we
98 lim [Y-Yp](t) = 0 .
•
99 Comments. a) For to ¢ 0, the same result is obtained by straightforward
modifications (exercise).
The T-periodic output Yp(·) is often called the steady-state response to the T-
periodic input 1Ip(.).
102 Exercise. Using (18), show that, by (96), for all t E R+:
103 Exercise. Consider Theorem (97). Let the T -periodic input be of the form
m
104 1Ip(t):= Uk exp[j(ko>ot)] ,
where 0>0 := T21t ' and, for all Ik I E m, Uk E c: n;. Show that
m
105 yp(t) = I: H(jkooo) Uk expU(kooot)] "V t E R+ .
k=-m
00
Note. If in (104) m=oo, with IIUkll < 00, then (105) holds with m=oo, pro-
vided
00
IIH(jkO>o)ukll <
•
00 •
k=-oo
106 Exercise. Show that, under the conditions of Theorem (97), as t -+ 00, the state
x(t) tends to a T-periodic steady-state trajectory xp(t) given by
t
where 1) vpO is the T-periodic extension on R of upO given by (91), and 2) K(t) is
the state impulse response (3.1.59). •
196
where a) the state x(t) ERn, the input u(t) E IRn and A(t) E R nxn , b) A(-) and u(-) are
given piecewise continuous T-periodic functions on R+, whence A(t+T)=A(T) and
u(t+T)=u(t) for all tE 1R+, and c) the d.e. x=A(t)x is expo stable, or equivalently,
condition (18) holds for fixed positive constants m and a, where a is the exponential
decay rate.
Intuitively, based on our experience with the time-invariant case, we expect that
for all XOE IRn, for all to the solution of (110) tends as t -+ 00 to a unique T-
periodic solution xp(·).
To obtain this periodic solution xpO, the first idea is to, say, start from the zero
state at time 0 and integrate the d.e. (110) until a periodic solution is reached. Unfor-
tunately this may be very expensive due to a slow transition to the steady state (. .. the
exponent a in (18) may be very small ... ).
A better idea is as follows: if we knew the vector xp(O) then the T-periodic solu-
tion xp(·) would be defined for all tE 1R+ by integrating the d.e. (110) over [O,T]. Now
by T-periodicity
T
111 f
xp(O) = xp(T) = <lJ(T,O)xp(O) + <lJ(T;t)u(t)dt ,
o
and equivalently,
T
112 [I-<lJ(T,O)]xp(O) = f <lJ(T,t)u('t)dt .
o
Thus xp(O) is on a T-periodic solution if and only if (112) holds. Furthennore (110)
will have a unique T-periodic solution for any T-periodic input u(·) if and only if the
linear equation (112) has a unique solution.
Now the assumption that x(t) = A(t)x(t) is expo stable is (by Theorem (2.2.71);
see also (60)-(61) above) equivalent to the condition that all eigenvalues of <lJ(T,O)
satisfy I Ai I < 1. Hence all those of I-<lJ(T,O) are different from zero: hence I-<lJ(T,O)
is nonsingular. Consequently equation (112) has a unique solution for any T-periodic
input u(·).
For computation we proceed as follows:
1) The numerical matrix <lJ(T,O) is obtained by integrating x(t) = A(t)x(t) as in Com-
ment (69. 13).
2) The RHS of (112) is the z-s response of (110) at time T: so it is obtained by
integrating the differential equation (110) over [O,T] starting from (0,0).
3) xp(O) is the unique solution of the system of linear algebraic equations (112).
197
4) Integrating (110) over [O,T) starting from (xp(O),O) gives xp(t) for all t E [O,T] and
by periodicity for all t O.
113 Exercise.. Show that the solution of (11 0) starting at x(to) = Xo is given by
116 Comment. In the stability literature, the T -periodic solution xpO is said to be
"globally exponentially stable in the large."
119 Exercise. Consider the scalar d.e. x=-x+ex 2 for e > O. Show that, a) for all
Xo > e- 1, and all to, I x(t) I 00 as (t-to) In(exof(exo-l)),
b) for all Xo < e-1, and all to, x(t) 0 as t 00.
Clearly, in the exercise above if we neglect the nonlinear term (i.e. EX2) the zero solu-
tion of x= -x is expo stable; this is no longer true for the nonlinear equation if I Xo I is
large.
121 Theorem [Small nonlinearities]. Consider the d.e. (120) and assume,
x
a) the zero solution of the linear equation = A(t)x is expo stable, or equivalently (18)
holds with fixed positive constants m and a;
b) the nonlinear term h(x,t) is S.t.
198
and
123 P< (aIm) .
D.th.c.
for all (Xo,to)E every solution x(·) of (120) satisfies
124
where 'Y := a-pm > 0 and m > 0 are fixed constants independent of Xo and to, i.e. the
zero solution of (120) is expo stable.
125 Comments. a) The constant pin (122) must be independent of x and t; (123)
requires this constant to be sufficiently small.
(3) If the function h(',,) in (120) is of the form
h(x,t) = M(t) . x
Thus a sufficiently small L00_ perturbation of A(') will not upset its expo stability. •
o
127 Exercise. Let A E lRnxn S.t. alA] c C_. Consider the time-varying d.e.
x=A(t)x (on Rt), where A(')E has the property that A(t) 4 A as
t 4 00. Show that the zero solution of x=A(t)x is expo stable on [T,oo) for some
sufficiently large T> O.
129 Remark. Exercise (128) generalizes Exercise (83) to the time-varying case.
199
130 Exercise. Consider the scalar d.e. x=-x+£x 2, where £>0. Show that condi-
tion (122) is not satisfied.
131 Proof of Theorem (121). Conclusion (124) follows at once fonn the Bellman-
Gronwall inequality (B.1.15). From (120) we have V' (xo,to) e R n x R+ V' t to
I
S mllxollexp[-a(t-to)] + f exp[-a(t-t)]llx(t)lldt,
10
positive constants m and a. Let BA(') e PC [IRr,1Rnxn] S.t. f IIBA(t)!!dt < 00. Show
o
that the zero solution of the perturbed d.e. x= [A(t) + 6A(t)]x is expo stable, (more
precisely (18) holds with positive constants m'exp[m I lIoA(t)lIdt] and a).
o
[Hint: use the same technique as in the proof of Theorem (121)].
exponentially stable. We consider now a local result. We start from the nonlinear
differential equation
where A(t) is Dlf evaluated at (8,t), i.e. the Jacobian matrix of f evaluated at (8,t),
h(x,t) represents the remainder: intuitively as Ilxll 0, Ilh(x,t)ll 0 faster than IIxll.
The linear differential equation
is called the linearized equation: more precisely, (137) is the linearization of (135)
about its equilibrium point 8. Intuitively, if (137) is exponentially stable, we expect
that, for initial states Xo small enough, any solution of (135) starting from (xo,to),
with to arbitrary, would decay to 8 exponentially. The fact that this expectation is true
is very important: it allows us to predict the local exponential stability of the zero
solution of the nonlinear d.e. (135) once we know that its linearized equation
x(t) = A(t)x(t) has an exponentially stable zero solution. For brevity's sake, when that
is the case, we say that "equation (137) is expo stable" or "the function t A(t) is
exponentially stable." We now state precisely the result.
141 a) t A(t) be piecewise continuous and bounded on 1Rr, thus, for some a > 0, 'it t
143 b) ·
I1m sup Ilh(x,t)11 =0 .
IIxll->O t Ilxli
Under these conditions, if the linearized equation x(t) = A(t)x(t) is expo stable, then
there is an e > 0 S.t. all solutions of the nonlinear equation starting from any
(xo,to) E B(6;e) x goes to 6 exponentially. •
201
145 Comments. a) Assumption (143) is the technical way of expressing the idea
that, for all t, h(x,t) 4 e faster than x 4 e, uniformly in t.
In the proof of the theorem we use the Lyapunov function technique, which is
extremely useful especially for nonlinear problems.
We start with a lemma.
146 Lemma [ Properties of pet)]. Let AO satisfy (141), let x=A(t)x be expo
stable, define
then
148 *Z :0:::; *
v(z,t) := Z P(t)z :0:::; *Z .
•
Proof. By expo stability, :::I m,a > 0 S.t.
f z* <1>(t',t)*<1>(t',t)zdt'
00
v(z,t) := z*P(t)z=
t
=f Ils(t',t,z,O)11 2dt'
I
1.
a
j IIA(t')II'lls(t',t,z,O)ll2dt'
I
by (142)
1.
a
7t
I s(t',t,z,O)* A(t')s(t',t,z,O) I dt'
a
7I
I s(t',t,z,O)*
dt
s(t',t,z,O) I dt'
and replacing in the RHS the integral of the absolute value by the absolute value of
the integral, we finally get
202
To obtain the last line we used the expo stability of (137). Thus PI can be taken to be
lI2a. •
Thus the positive definite matrix pet) is such that Amin(P(t» and Amax(P(t» ::; PU'
Vt
154 · *
pet) =-A(t) pet) - P(t)A(t) - I
and
P(t)* = pet).
155 · *
Q(t)=Q(t)A(t) + A(t)Q(t)+ Q(t)Q(t) .
Proof of Theorem (140). Consider the solution of the nonlinear d.e. (135) starting
from (xo,to); for brevity call it x(t). Consider the composite function t v(x(t),t)
which maps R+ (C where v(z,t) is given by (147)-(148). Its derivative is called the
derivative o/v along the solution x(·) 0/(135). By the chain rule we have
d ()
156 - v(x(t),t)= D1v(x(t),t) . f(x(t),t) + : \ v(x(t),t)
dt vt
where the first term is the product of the gradient of the scalar function v, evaluated at
(x(t),t), with f(x(t),t). For brevity write x for x(t) and v(x,t) for the LHS of (156), use
(136) and (147) to obtain from (156) - dropping temporarily the dependence on x and
t-
hence
157 · * *
v(x,t)=-x x+x P(t)h(x,t)+h(x,t) P(t)x .*
View (157) as an equality between functions of x and t. By (143), given 1I4Pu'
=:Ie > 0 s.t. V IIxll < e,
159 * *
I x P(t)h(x,t) I < x x/4 .
160 v(x,t) rl .
V(X,t)
v(x(t),t) v(xo,to)
Hence, by (148),
Thus, for any to 0, any solution of (135) that starts in the ball B(B;e) tends to S
exponentially. •
We emphasize again: theorem (140) is a local result: it is only the solutions that
start sufficiently close to 9 that are guaranteed to tend to 9 exponentially as t 00.
CHAPTER 7d
where
4 II xII := max I Xi I
i
b) Vector sequences will be nonned by their (OO-norm, thus for the sequences u(')
and y(.) in (1)
7 II y II"" .- ke
sup II y(k) II = sup (max I Yi(k) I } .
Z ke Z ie Do
'Y) We view the I/O map F : u y of (1) as a linear transfonnation from I;; into I;;:',
(A.6,49).
8 I/O Stability. The concept of I/O stability roughly asserts that "any bounded
input u(-) produces a bounded output y(.) using a finite gain." More precisely: we say
that the discrete-time system described by (1) is I/O stable iff
9 3 'Y < 00 s.t., for all bounded u(') , lIylloo:S; 'Yllulloo '
10 Comment. The discrete-time system described by (1) is I/O stable iff the I/O
map (1) considered as a linear transformation
13 Exercise. Show that the discrete-time system described by (1) is not I/O stable
if there exists a sequence of inputs (uP(.) ];0 S.t. for all peN
17 Theorem [I/O stability]. The linear discrete-time system described by (1) is I/O
stable iff
k
the map k L II H(k,k') II is bounded on Z ,
k'=-oo
or equivalently,
k
20 'V (ij) e flo x llj , the map k L I hjj(k,k') I is bounded on Z .
k'=-oo
In (18) above, any matrix norm may be used, since all matrix norms are equivalent,
(A.6,45).
206
24 Proof of Theorem (17). 1) Sufficiency. We prove that (18) implies (9). There-
fore let u(') be any bounded input sequence, i.e. whose norm II u 11 00 , given by (6), is a
finite number. Then from (1), taking vector-norms, we have successively 'Ii k E Z
k k
II y(k) II = II L H(k,k')u(k') II:S; L II H(k,k')u(k') 1/
(we used successively an inductive application of the triangle inequality, (A.6.17), (6)
and (18». Hence, using (7), taking the supremum w.r.t. k, we obtain (9) with yas
defined in (18). •
2) Necessity. We use contraposition, i.e. the negation of (18) must imply the negation
of (9) (Le. system (1) is not I/O stable). Therefore let (18) not be true; then negating
equivalence (20) there exists (a,p) E !10 x Dj S.t.
,up I "",(k,k') I } =
kp
26 L I I > p.
k'=-""
with
sign [h ll (kp,k)]
{ O
27b ui(k) :=
elsewhere
28 V'pe N
Hence by (28) and (30) the system described by (1) is not I/O stable by Exercise (13) .•
35 We say that the system R dO is I/O stable iff condition (9) holds for the I/O
map (34).
Similarly as in the continuous-time case, we have the following corollary.
37 sup
k
{2:
k'=O
II C(k)<I>(k,k'+1)B(k') II } < 00
00
L IIH(k) II < 00
k=O
R d is 110 stable
45 P[H(z)]cD(O,l).
46 Comment. (45) means that H(z) has no poles outside the open unit disc, or
equivalently, is analytic in I z I 1 : a well-known characterization for "external sta-
bility."
00
L I gjik) I < 00
k=O
P[gij(Z)]cD(O,I) .
209
We prove this equivalence
00
00
<=: gij(Z) is a strictly proper rational function, hence by Exercise (51) below (partial
fraction expansion) and A E cr nxn (Comment (60)),
m
Therefore picking p := max I AI I < 1, eE (O,p-I-l), and m(e) := L m/(E). we have
I 1=\
by (48) and (49)
00
where (1+E)p < 1. Hence L I gij(k) I < 00.
k=O QED.
52 Z-l«z-A)-r)(k)= [ k-l]
r-l A.k- r l(k-r) Vk
where
[ k-l ]
r-l =«k-l)(k-2)··· (k-r+l»/(r-l)!
and
210
I(k)=
{ oI for k < 0
for k
where
0 for k"# 0
{
O(k)= 1 for k=O.
cr m,
57 g(z)= L L a/r<z-A/rr
1=1 r=1
leads to
a-I k
=: L A/ Tt/ (k).
/=1
211
'v'kE N
2 y(k) = C(k)x(k) + D(k)u(k)
and treat the same problems as in the continuous-time case: asymptotic and exponen-
tial stability of the zero solution of x(k+ 1) = A(k)x(k); condition for obtaining
bounded state- and output-trajectories as well as regulation, and the response to
periodic inputs.
3 x(k+ I) = A(k)x(k) kE N
7 Comment. Observe that any solution x(·), on any [ko,k] , is a finite set, hence
Moreover by the linearity of the solution we expect
8 Theorem [Asymptotic stability]. Let det A(k) 0 'v' kEN. The zero solution of
x(k+l)= A(k)x(k) on k is asy. stable if and only if
9 <1>(k,O) 0 as k 00 .
10 Exercise. Prove Theorem (8). [Hint: see the proof of Theorem (7.2.8); observe
that, because detA(k) * 0 V' ke N, V' k <I>(k,ko)='(I>(k,O)<I>(O,ko)].
19 Comments. a) The constants p E [0.1) and m > 0 are fixed, i.e. independent of
ko EN; the constant a 0 s. t. p = exp(--a) is the exponential decay rate.
I}) Using an induced matrix norm for II <l>(k.ko) 1/ we see that the zero solution of
x(k+l)= A(k)x(k) on k is expo stable if and only if :3 pE [0,1) and m > 0 s.t.
V' (xo.ko) E RRX N
22 V' e > 0 :3 an integer K(e) > 0 S.t. V' ko EN" <l>(k.ko II S e V' k
[It is crucial to note that the constants I in (21) and K(e) in (22) are independent of
ko EN; compare with the conditions (9) of asymptotic stability J.
e) In the time-invariant case x(k+I)=Ax(k). (with A constant), asymptotic stability is
equivalent to exponential stability. Indeed <l>(k,ko)=A k- ko depends only on the
elapsed time k-ko; hence the zero solution of x(k+ 1) = Ax(k) is asy. stable, (i.e. (9)
holds), iff the zero solution is uniformly asy. stable.
and (22) with 2E = I, and pick p e [O,I) s.t. 2pK > 1; hence II <I>(k,ko) II 21 pk-ko
for all k
o(A) eD(O,I)
35 Analysis. <I>(k,ko) = Ak- ko . Now using (4.4.36), (see Exercise (42) below),
(J
where (Ad = o(A)\ ( 0) and VI, TIL (k) is a matrix polynomial in k. Hence by
taking matrix norms V k n
(J
(J
(J
where the 1t1 (k) are polynomials S.t. 1t1 (k) II TIl (k)ll. 1t(k) := L. 1t1 (k) 0 and
1=1
42 Exercise. Let Ae cr nxn and let (AI Using the notations of Sec-
tion 4.4, esp. (4.4.36), show that
214
'v'k
46 P=A*PA + Q
where Ae cc nxn, Qe (Cnxn S.t. Q=ct' > 0 is given, and a unique solution
P=P'" > 0 is to be found. Equation (46) is called the (discrete-time) Lyapunov equa-
tion. It's solvability relates directly to the expo stability of x(k+l) =Ax(k).
47 Lemma. Assume that given Q=ct' > 0, Eq. (46) has a unique solution
P=P'" > 0, then the zero solution of x(k+l)=Ax(k) is expo stable, i.e.
o(A) c D(O,I).
Taking differences. of vex) along any trajectory of x(k+ 1) = Ax(k) we have by (48) and
(46)
=-x(k)*Qx(k) .
54 Lemma. Consider the Lyapunov equation (46) and let a(A) c D(O,l). Then
V'Q=Q* > 0 Eq. (46) has a unique solution p=P* > 0 given by
55 p= (A*)kQAk.
k=O
1) By an elementary induction
I
55b X(l+I)= L .
k=O
Now since A is expo stable it follows that, as I 00, the series converges to the limit
P := X(oo) = l: (A*)kQAk ,
k=O
216
which solves (46). (To see this let k 00 in (55 a) and use (55b).)
2) The solution defined by (55) is unique, since, by Exercise (52), the operator A,
given by (53), is injective.
3) Obviously P=P'" moreover P > 0, indeed, a) the first term of (55b) is posi-
tive definite by assumption and b) all the remaining terms are positive semi-definite by
inspection. •
From Lemmas (47) and (54) we have now our main result
56 Theorem. Consider the Lyapunov equation (46). Then the following statements
are equivalent:
a) VQ=ct > 0 the Lyapunov equation (46) has a unique solution P=P'" > 0 given
by (55),
b) a(A) c D(O,l),
c) the zero solution of x(k+l)=Ax(k) is expo stable.
•
Periodically varying r.e.'s. We study the expo stability of the Le.
60 x(k+l)=A(k)x(k) ke N ,
where the matrix sequence A(-) : N R nxn is p-periodic and nonsingular i.e.
61 A(k+p)=A(k) Vke N ,
and
61a det A(k) '" 0 V k e [O,p-I]
64 S(k+l)=BS(k)
(where B is the constant nonsingular matrix given by (62». Our objective is to show
that the zero solution of (60) is expo stable iff the zero solution of (64) is expo stable.
Note that, by Theorem (33) or (56), the latter holds iff a(B) eD(O,I).
217
a(B) cD(O,l)
if and only if
a( c D(O, 1) .
66 Theorem [Exp. stability]. Consider the periodically varying r.e. (60), where
detA(k) '" 0 'Vke N. Then the zero solution of x(k+l)=A(k)x(k) is expo stable if
and only if
67 a[ c D(O, 1) .
68 Short proof (exercise). Consider Eq. (63). Since P(k) is nonsingular for all
ke Nand P(·) is p-periodic, the sequences P(·) as well as p(·r 1 are bounded on N;
equivalently there exist positive constants M and N S.t. for all ke N, II P(k) II s: M
and II p(k)-lll s: N. Therefore by (63), for all ko e N and for all k
(MN)-IIlBk-koll s: s: (MN)IIBk-koli.
Hence, using (18), the zero solution of x(k+l)=A(k)x(k) is expo stable if and only if
the zero solution of 1) = B is expo stable. The latter condition is equivalent to
condition (67) by Exercise (65). •
where any nonn may be used for II f(k) II. Of course, II f 11 00 ,0 = II f 1100 ' i.e. the usual
loo-nonn of f(·). For matrix functions we denote by 11M 1100 the 100 _
nonn, i.e.
where the matrix nonn II M(k) II is the one induced by the chosen vector norm.
1) for every (xo,ko) E JR.n x N, for every bounded input u[k",oo)' then the state
k =
x(k) s(k,ko,xo,u) and the output k y(k) = p(k,ko,xo,u) are bounded on k ko,
more precisely with the constants given above,
2) Under the same conditions as in I), if in addition u(k) 8 as k 00, then x(k)
and as
k'=k" 1=0
Hence, if both u(k) and x(k) tend to zero as k 00, then so does y(k). Hence we arc
reduced to show that
k-I
x(k)=<l>(k,ko)xo + L <l>(k,k'+I)B(k')u(k') := xl(k)+x2(k)
k'=ko
tends to zero as k 00. Now this is immediate for XI (k) since x(k+ 1) = A(k)x(k) is
expo stable. So we are left to prove that lim x2(k) = 8. Now by taking vector norms
k->oo
and using (18),
k-l
o Ilx2(k)1I mllBlloo L p(k-k'-l)llu(k')II Vk> ko
k'=ko
219
and we are done if the series on the RHS converges to zero as k 00. For this pur-
pose, first set u(k) = e for k < leo and then observe that
k-J
L L
00
p(k-k'-l)lIu(k')11= pk'lIu(k-k'-I)1I
k=ko k'=O
[in the last expression lIu(k-k'-I)II=O for k' > k-IJ. Take now any sequence
[kd; S.t. kl > \:1/ fl(k'):=pk'llu(kl-k'-I)llfork'eN.
Thus we are done if we prove that
l:
counting measure on N). This theorem allows to pennute the operations of limit tak-
ing and summation because the sequence [flO is dominated by the fixed abso-
lutely summable series gO on N where g(k') := pk'il u Ilk ; indeed i) \:I I and k' e N
0.-
I fl (k') I :,; g(k'), and ii) g(k') = (l-p )-111 U Ilk < 00. Hence by the theorem,
k'=O 0._
00 00
91 We shall denote by k --) vp(k) the q-periodic extension of k --) up(k) on all of Z ,
i.e. vp(') is the unique q-periodic sequence on Z S.t. vp(k) = up(k) V kEN.
k
95 y(k) = p(k,O,xo,up) = CA kxo + L H(k-k')up(k')
k'=O
k Vke N
=CAkxo + L H(k-k')vp(k')
k'=O
where the last series converges because A is expo stable (indeed, using norms and (18)
and the notation i for the series, Iii II :s; (mI(l-p» II B IIII vp 1100 < (0). Therefore in
the RHS of (96) the expression between the brackets is a constant vector. Therefore
since Ak--)O as k--)oo, VXOE tr n lim [y-yp](k)=O. From this and claim (94)
k->oo
we have
221
103 Exercise. Consider the representation R d defined in Theorem (97). Let the q-
periodic input be of the form
m
104 up(k) := L ul exp[j(l 90)k]
1=1
Introduction
This chapter treats the coupling of the input to the state, i.e. controllability, and
that of the state to the output, i.e. observability. This is done for general dynamical
systems which are then specialized to the linear system representation
R (.) = [A(-),BC'},C(-),D(')]: first in the time-varying case and then in the time-invariant
case. For the latter systems this leads to Kalman decomposition and a discussion of
the absence of unstable hidden modes, viz. stabilizability and detectability. This
chapter ends with a brief study of 1) balanced representations (based upon normalized
controllability and observability grammians) and 2) the robustness of controllability
(for perturbed nonlinear systems).
We also say that the input u[to.t.] steers the phase (xo.to) of D to the phase (x!>t})
In some cases, one does not want to prespecify the time t}; in such a case we say
D is called controllable at to iff 'it xO,xI E L, there exists some tl > to such that some
u[lo. tIl E U steers (xo,to) to (xI,II)'
In most applications. if some input u[l().ttl steers (xo.to) to (x1.t1). there are
infinitely many other controls that do it also.
II. Controllability depends only on U. and the state transition function s. (r
and p have nothing to do with controllability).
For convenience. we will follow common usage and say "the dynamical system
D " rather than the more exact "the dynamical system representation."
7 D is controllable on [to.ttl
v
+.<f- U
y
YFS('II-Yx
Fig. 8.1 System D s with memoryless state feedback.
v + y
•
8 D. is controllable on [to,ttl
9 Do is controllable on [to,tIl .
By the uniqueness assumption (5). this exogeneous input v(-) will precisely produce
the input u that will steer (xo.to) to (xI.tl)'
(<:) By controllability of Ds. VXO,xIE L,
:::I v[100111 that steers the phase (xo.to) of
D s to the phase (xI.tl) of D s' Since D sand D have the same phase and. by (5), v
will produce a unique input U. the input u of D will produce the required transfer of
•
(xo.to) to (xI.tl)·
The second equivalence is proved in a similar way.
that is, the panial response map is a one-to-one map from L to Y(Io,ld'
225
21 rt(t)=y(t) + FrCu(t» .
For Do we make the uniqueness assumption (5) above.
22 Theorem. For the system Dr and the system Do satisfying (5), we have
23 D is observable on lto,td
Proof. Exercise.
Remark. Memoryless state feedback may affect observability. For example, for a
linear time-invariant system representation R = [A,B,C,D1, there may exist a linear
state feedback Fs such that for some states Xo and for some inputs uO, the state trajec-
tory remains, for all t, in the nullspace of C.
v y
v u y
Verify that for fs =[-I,O], the system Ds has some nonzero states that produce an out-
put which is identically zero.
1 More precisely, we say that the pair (A('),B('» is controllable on [to,td iff
:3 ul.."I.) that steers the phase (xo,to) to the phase (xI,t l ).
'V (xo,to) and 'V (xI,II)
Given the pair (A('),B('», both piecewise continuous, we know that U[to,l.J
3 Equation (2) shows that there will be an input ul.."l.] that will transfer an arbitrary
phase (xo,to) to an arbitrary phase (xI,tl) if and only if the linear map
II
4 L r : ul.."I,] f <l>(tl,t)B(t)u(t)d't : PC ([to,td) (fn
to
5 Reduction Theorem.
The pair (A('),B('» is controllable on [to,td
227
8 Given the representation R (.), or the pair (A('),B('», we say that the state Xo is
controllable to zero on [to,tIl iff :3 U[Io.tl) that steers (xo,to) to [9n ,t l ); we say that
the state XI is reachable on [to,ttl iff :3 u[to.ttl that steers (9 n ,to) to (xI,tl)'
The essence of the reduction theorem (5) is that, for linear system representa-
tions, controllability on [to,td, controllability to zero on [to,td of all states, and reacha-
bility on [to,tll of all states are equivalent. The reader should construct a one-
dimensional nonlinear example to show that this is not so for nonlinear systems.
We state now two closely related theorems that characterize the controllability of
the pair (A(·),B(·».
II
(ii) The set of reachable states on [to,ttl is the subspace R (L r), which is equal to
R [Wr(to.t l )].
fixed final time tl' all the singular values of Wr(to,t l ) decrease as to increases towards
tl' In particular if to < to', then
19 Wr(to,t\) Wito',t\).
13) The function tl Wr(to,tl) is the solution of the linear matrix d.e.
20 X(t) = A(t)X(t) + X(t)A(t)* + B(t)B(t)*
where (',')n denotes the inner product in ([n and C·) denotes the inner product in L;'.
Using matrix notations we have
21
Using this result and the definition (4) of Lr we conclude that the linear map L.Lr* has
the matrix representation Wr(to,tl) defined in (17).
(15) (16). Since L.Lr*: ([n ([n, L.Lr* is surjective iff it is a bijection,
so using its matrix representation (17), we see that (15) is equivalent to (17).
Statement (ii) follows immediately from (2) with Xo = Sn' I
1\
II
31 W c (to,t1) := I <I>(to,'t)B('t)B('t)*<I>(to,'t)*d't.
to
32 (ii) The set of all states controllable to zero on [to,td is the subspace R (L c)'
which is equal to R [Wc (to,t1)]'
33 Exercise. Show that, given tl >to and definitions (4) and (25),
34 Exercise. Obtain a matrix differential equation that will. have t W e (t,t1) as its
solution.
38 II u lIi=xtWr(lo.t1r1xl .
41 Xl -¢(tl.to)xO=L,u .
42
From (38) it follows that. for a unit cost (as specified by (35». we can reach any of
the points Vi 1-fA:. i=I.2 ..... n. In fact for a unit cost. starting from phase (8 n .tO) we can
231
reach any point on the ellipsoid whose semi xes are i=1,2, ... ,n. If we order the
eigenvalues as follows, Al ... then the direction vn is the most
expensive to reach and the direction VI is the cheapest. Roughly speaking, the eigen-
values of Wr(to,t l ) measure the effectiveness of the actuators in the task of reaching a
specified state.
Suppose that A(·) is not stable and that we wish to stabilize the system by state feed-
back.
For u 0, to and tl E IR+, let us define
I,
56 II x(t)e<Xt II ° exponentially as t 00
(i.e. all z-i state trajectories of the feedback system go to zero at an exponential rate,
which is faster than exp(-ut». •
Proof. To establish the claim we will show that the equilibrium point of
60 (t) = (A(t) +
61 =
By (54), we have
62 S S e4at.h;;;11,
63 .
64 V. * [A(t)* *
- + +
+ - B(t)B(t)*
+ A(t)Ha(t,t + Ha(t,t + + .
V. (60) -2a.!;(t)*Ha,(t,t +
65 S 112 .
From (61) and (65), we conclude that along any trajectory of (60)
. -2a.hMI
66 V (60)/V S 1 =-2a.(h m /h M )exp(-4M) .
h;;;
2 J
y(t) = C(t)cl>(t,to)xo+ C(t) cl>(t,t)B(t)u(t)dt + D(t)u(t).
10
Suppose that in addition to R (-) we know u(·), then we can calculate the last two
tenns; thus wJ.o.g. for the study of the relation between the state Xo and the output
Y[Io,I,l we need only consider the first tenn of (2):
5 We'll say that state Xo is unobservable on [to,t!] iff its z-i response is zero on
[to,t!].
In view of definition (5), we have
equivalently, the set of all states of R (.) or of the pair (C('),A('» that are unobservable
on [to,tll is the linear subspace N (Lo).
Recalling the definition (8.1.16) of a dynamical system D observable on [to,tl]'
and specializing to the special case of R (.), we say that
7 the pair (C('),A('» is observable on [to,tl] iff given R (.), V inputs U[Io,I,) and V
234
As a consequence, the pair (C(·),A(·» is not observable on [to,tIl iff there is some
nonzero state that is unobservable on [to,td, or equivalently, it has a nontrivial sub-
space (namely, N (Lo)) of unobservable states on [to,td.
14
15
t,
18
235
19 xo=(Lo*LoflLo*Y=Mo(to,tlflJ <l>(t,totC(ttY(t)dt .
10
20 Comment. Let A" > 0 be the smallest eigenvalue of the, positive definite Hermi-
tian matrix Mo(to,t) and en its corresponding normalized eigenvector: then for Xo = en'
II Xo liz = 1 and its z-i response is s,t. (y,y) = A". So if An < < 1, some states are barely
observable in case of noisy observations.
*
(Lo Lo)xo=Lo Y *
from which (19) follows.
•
23 Corollary. Given the pair (C('),A('», the set of all unobservable states on [to,ttl
is a subspace of ffn and, in particular, it is N(Lo)=N(Mo(to,tl»'
Proof. Exercise.
24 Exercise. Using the notations of Theorem (12), show that to -; MO(tO,tl) is the
solution of the linear matrix differential equation
8.4. Duality
If we compare theorems (8.2.26) and (8.3.12), it becomes clear that they are
closely related. In fact, they are related by duality.
Consider a linear time-invariant representation R (-)= [A('),B(·),q'),DO], i.e.
where, as usual, x(t)e «r n , u(t)e «r n \ y(t)e «rn., and the matrix functions
AC·),BC·),C(·),DC·) are piecewise continuous as specified in definition (8.3.1). As
before, (t,to) -+ ¢Ct,to) denotes the state transition matrix of A(·).
The dual representation R (-) of R (.) is closely related to R (.) in such a way that,
roughly speaking, controllability properties of R (-) are closely related to observability
properties of R C·), and vice-versa. The formal definition of R (-) is as follows.
4 R- 0= [-A(·)*,-CO*,B(-)*,0(-)*] .
Equivalently, R0 is described by
where xCt)e ([n, uCt)e ([n., and y(t)e ([n;. The state transition matrix of R (-) is
(see 2.1.118)
7 *
'P(t,t)= ¢(t,t) .
Note the minus sign in the d.e. (5): we could get rid of it by introducing a reverse
t
time =-t, (see [KaI.l]). However, this is not convenient for physical interpretation.
Moreover, denoting by R(.) the dual of the dual-system representation R (.), (4), we
obtain
8 R(·)=[A(·),-BC·),-C(·),DC·)] ,
i.e. R(.) equals R (.) modulo a change of sign for the state. Hence their reachability,
controllability to zero and observability maps, (8.2.4), (8.2.25) and (8.3.4) resp. are
related by
12
•
From (12), we immediately obtain the following corollary.
14 Proof of Theorem (10). a) Calculations show that th(: state Xo of R (.) is con-
trol/able to zero on [to,t 1] iff :3 u[to.l,l S.t.
I,
15 f
Xo =- <l>(to,'t)B('t)u('t)d't ,
10
Now by (8.3.6), the state x 0 of R (-) is unobservable on [1o,tIl iff i oX 0 = e, more pre-
cisely iff
But by (8.2.26), (17) is equivalent to: XOE N(Wc(lo,tl»=N(L c*)' Now by (A.7.57),
R(Lc)=N(L c*1. ) . Thus we have established (since Lo=Lc - *), R(Lc)=N(L- 0) .1 .
Hence the first equality in (11) and in (12) hold.
b) To obtain the others repeat the reasoning of a) to the dual 0, (4), and the dual of
the dual R (-), (8), for which (9) holds. Hence by the first equality (11)
i.e. the second equality (11) holds. The second equality (12) follows similarly. •
Thus the notion of "xo is controllable to zero on [Io,ttl" is related by duality with
"xo is unobservable on [to,ttl." Also "RO is controllable on [1o,t1]" is the dual of
"R (-) is observable on [Io,ttl."
238
Now what is the notion that is related by duality with "xl is reachable (from zero)
on [to,tIl"? It turns out that the required notion is "xl is unreconstructible on [to,td."
20 We say that the state Xl is unreconstructible on [to,tIl iff the z-i response on
[to,ttl which corresponds to the final phase (xI,tl) is zero, equiv.
So if we let
we have
23 Xl is unreconstructible on[to,td $;> Lrec(xI) = e
where
I,
or equivalently,
27
$;> f
XI = <1>(t l ,t)B(t)u(t)dt, for some u[to,t,l .
to
By the theory of the Adjoint (see section A.7.4). R(Lr)=N(Lr*)l. =N(i rec)l.. The
remainder of the theorem follows easily. •
1 the controllability matrix: C:= [B: AB: A2B: ... : An-IB] E (rnxnn;
CA
2 the observability matrix: 0:= E ernn"xn
CAn-I
Proof. In both cases the proof is based on the Cayley·Hamilton Theorem (3 .. 2.24).
The A-invariance of N(O) is straightforward. Let us prove it for R(C). By 0). we
have
We must show that Ax E R (C). Multiply the equation (4) by A on the left. Now. by
Cayley·Hamilton. An is a linear combination of (Ak );-1 hence. after some rearrange-
n-I
ments. we see that for some Wk'SE ern;. Ax= L AkBwk' i.e. AXE R(C).
o •
240
9 Theorem (Observability properties of the pair (C,A». Given the pair (C,A),
10 i) the set of all unobservable states is the A-invariant subspace N(O) c. (l:n,
iii) For C and A real, for any monic t real polynomial 1t of degree n, there exists
LE Rnxn.. S.t.
13 XA+LC=1t
is positive definite.
•
Before proving Theorem (9) let us consider some aspects which illustrate the
dynamic consequences of the assertions of the theorem.
20 Remark on the extraction of the unobservable part. Let us assume that (i)
holds and that dimN(O)=r<n, the unobservable subspace is r-dimensional. Choose
a basis for N (0) and precede these r basis-vectors with n-r vectors from (l:n so that
we have a new basis for (l:n. Suppose, for simplicity, that D=O. Since a) N(O) is
A-invariant and since, by (1), N (C) -=> N (0) i.e. the last r basis vectors are in the
nullspace of C, in the new basis, the system is represented by
21
where xl e (Cn-r and X2 e (Cr. Equations (21) and (22) imply the connections indi-
cated on the block diagram of Fig. 8.7.
As the figure shows, we have extracted an r-dimensional subsystem from R and the
state x2 of that subsystem is unobservable. Since by (i) the set of all unobservable
states is of dimension r the pair (CI,A ll ) is observable.
23 Comments. a) The rank test (11) is not numerically appealing: suppose that the
ratio of I Aroax/Aroin I =: 11 for A is, say, 10; then by the spectral mapping theorem
(4.7.1), for An- l the corresponding ratio becomes lln-l. Thus, when we calculate An - l
we are going to add and subtract numbers that differ by many orders of magnitude.
Consequently we should expect to lose a lot of significant information in the round-off
process.
b) The rank test (12) is more appealing because the QR algorithm will give reliable
eigenvalues and by singular value decomposition the rank can be realistically evaluated
[Gol.1].
c) The interpretation of (13) is the following: suppose that, in the original system
x = Ax + Bu, we add Ly = LCx to the input of the integrators, (this is called "constant
x
output injection"), we then obtain = (A+LC)x + Bu, thus the resulting spectrum is
now cr(A+LC), i.e. the set of roots of XA+LC' Note that if R is not observable it is
obvious from (11) and (12) and Fig. 8.7, that no amount of output injection of
y=C1xl will change the spectrum of the unobservable part! In fact, partitioning the
matrix L in two submatrices L I , we find that A+LC has the form
Al1+ L I C I o
24 [
1
Thus, for all L, the elements of cr(A 22 ) are elements of cr(A+LC), the set of closed-
loop eigenvalues. Thus we see that constant output injection does not affect the eigen-
values of the unobservable part.
unobservable on [to,ti1 iff its z-i response is zero on [to,t!] that is, in the present case,
iff
Expanding t --+ exp[A(t-to)] in Taylor series about to, we obtain a Taylor series that is
identically zero on [to,t!] with t[ > to. This is equivalent to having all the coefficients
of the Taylor series equal to zero, or equivalently,
26
ii) (C,A) is observable ¢ > N (0)= { en} by (8.3.6), (8.3.7) and i) above.
¢ > the n columns of 0 are linearly independent
37 i) the controllable subspace of the pair (A,B) is the A-invariant subspace R (C);
ii) the pair (A,B) is controllable
38 rk[C]=n
iii) For A,B real, for any monic real polynomial 7t of degree n, there exists Fe lRD;Xl1
such that
is positive definite.
•
44 Remark on the extraction of the controllable part. Assume that i) of Theorem
(36) holds. Let dimR (C) =y< n: the controllable subspace is y-dimensional. Choose
a basis for R (C) and complete it n-y vectors from ([D so as to obtain a basis for ([D.
Now R(C) is A-invariant by (3), and by (1), R(C)::>R(B), hence R(B) is in the sub-
space generated by the first y basis vectors. Consequently, in this new basis, the sys-
tem representation is of the form
244
45
46
u
o--Y-
+
49 Comments. a) The rank test (38) is not numerically appealing for the reasons
245
Thus (iii) asserts that the pair (A,B) is controllable we can always choose F so that
the closed-loop characteristic polynomial XA+BF has as roots a list of n preassigned
points in IT; of course, these n points must be located symmetrically with respect to
the real axis because the polynomial XA+BF has real coefficients. In particular given
any unstable A with (A,B) controllable we can always stabilize it by constant state
feedback.
d) (iii) says that for any polynomial 1t, i.e. for any configuration r of its n roots, we
may find an F such that cr(A+BF) = r. A word of caution is required. Suppose we
have a system with h(s)=(s+l)-', i.e. R =[1,1,1,0] and y == x. Suppose we wish to
broaden the bandwidth by a factor of 103, then we set up the system shown in Fig.
8.9.
999
The closed-loop transfer function is he(s) = - - 3 ; note that he(D)
A A
Proof of Theorem (36). (i) By Theorem (8.4.10) of duality theory, we know that for
any [Io,ttl, with t, > 10,
=N CO).1. by (10)
Thus RCLJ is a fixed Cindependent of to,tl) A-invariant subspace of (1n since N(O) is
a fixed A* -invariant subspace of ern.
By the theory of the adjoint
52 NCO).1. =R(O*)=RCC)
where the last equality follows by calculation (A=-'A*, B=C\ Now from (8.2.33)
and w.l.g. to=O,
53 RCLr)=eAIIRCLc)'
R(Lr)=R(LJ=RCC) .
54 x
Exercise (Stabilization by constant state feedback). Given = Ax + B u, show that
(A,B) is controllable ¢> there exist a constant state feedback that stabilizes R.
(Hint: Define for any a>O and 't>0,
Ha := f e-iX1e-A1BB*e-A*'dt= f m(t)dt
o 0
calculate fo Adt (m(t»dt in two ways. Choose the state feedback F=-B*H;;I.)
Theorem 8.1.6.
Analysis. The analysis rests on two pillars: 1) R(C) is the subspace of all controll-
able states and N (0) is the subspace of all states, 2) the second
representation theorem: R (C) and N (0) are A-invariant.
The idea is to create four subspaces, say, Lo' 14, and such that
We proceed in 4 steps
4 Le<»$Lrj$=N(O).
8 31 32]
A41 A42
= [0 01
0 0
and since R (i3) c R (C), the two bottom submatrices of S are zero.
9 [::]-
By (4), the A-invariant subspace N (ci) is spanned by the basis vectors of Lee!>
and L ¢II' the
second and the fourth subspaces in the direct sum (I). Hence A 22' A 24'
A 42 and A 44 are nonzero and
10 12 14] =
A 32 A 34
[0 0]
0 0 .
Since N(O) c N(C), all the basis vectors in N(O) are in N(C) hence
11 C= [c 1: 0 : C3 : 0 ] .
Thus in the new basis the given representation R becomes a new representation
R= [A ,S ,C ,0] where
All 0 AI3 0 BI
A21 A22 A 23 A24 B2
12 A :=
0 0 0 0
=: S
A33
0
0 0 A43 A44
13 C .- [C I :0:C 3 :0].
These equations lead to the block diagram shown in Fig. 8.11: note that the input
affects only blocks 1 and 2, the output is only affected by blocks 1 and 3; all arrows
going from any block to any other block are from right to left.
,..., N
u C 1 X1 Y
0
1 .. ...
x3 3
C3 "3
19 of dimension nl +n3,
[I ° ] ° '
is controllable;
21 iv) R 34 := A33
A 43 A44 ' [0] [C3 : 0] ,0 is completely uncontrollable
24
251
27 Exercise. Consider the linear time-invariant circuit shown in Fig. 8.12. Let
XI' x2, x3, X4 denote the voltages across the capacitors and currents in inductors as
shown in Fig. 8.12. Use the Kirchhoff laws and vL=Li L, ic=CVc to obtain the state
equations. The current source delivers u amperes to the circuit. The output voltage is
labeled y.
a) Determine the subspace of controllable states and that of unobservable states.
b) Obtain the Kalman decomposition and identify physically the unobservable modes.
1H
+
HI.
y
HI.
29
4 We say that there is an uncontrollable hidden mode at Ak iff there exists a general-
ized eigenvector at Ak which is not controllable, or equivalently,
6 We say that there is an unobservable hidden mode at Ak iff there exists a general-
ized eigenvector at Ak that is unobservable, or equivalently,
253
9 Comments. ex) The negation of (5) reads: every mode generated by a general-
ized eigenvector at the eigenvalue Ak can be reached by an appropriate control; more
precisely, using the state transition function of R,
V nonzero x E N k' there exists a control uO that is zero outside some bounded interval
[O,T] S.t. V t T x(t) =eA(t-T) x = s(t,O,e,u). Hence an uncontrollable hidden mode
at Ak is a mode that cannot be displayed by the zero-state transition function after
appropriate control action.
(3) Condition (7) reads: there is (a mode generated by) a generalized eigenvector at Ak
that is unobservable (at the output); more precisely, using the state transition and
response functions of R ,
yet
yet) = p(t,O,x,e) = CeA'x = e Vt 0 .
12 Exercise. Using the notations of (8.6.12) and (8.6.13) show that if there is a
hidden mode at Ak' then Ak E a(A 22) u a(A 33) u a(A 44)'
[Hint: use contradiction.]
16 Theorem [Hidden
R = [A,8,C,D] where alA] = {Ak
U.th.c.
r
modes]. Consider a time-invariant
is the spectrum of A E (Cnxn.
representation
18 rk [AkI - A : B ] < n.
c) There is a hidden mode at Ak E orA] iff either (17) or (18). or both (17) and (18)
hold.
all t J
ll* eA(l--'t)Bu('t)d't=O.
o
If A is simple. then. using a basis of eigenvectors (ek); and the notations
(3.3.48)-(3.3.50). R becomes R = [A .N*B.CE.D]. In this case. if there is a hidden
mode at Ak' i.e., according to (17) or (18). if there exists an eigenvector ek at Ak S.t.
Cek = e or a left eigenvector !1k at Ak S.t. 11* kB = a*. then in the partial fraction expan-
sion of the transfer function H(s), the kth term
drops out.
23 NkcR [C]
24 rk [AkI - A : 8 ] = n .
255
For this purpose use orthogonal complementation and duality, whereby (using
expanded notations)
25
(for (25) see Exercise (4.4.46); for (26) see (8.5.52». Moreover we shall need
(J
(Since (J:n= ED N/[A*], (4.3.41), and note that any A*-invariant subspace is spanned
/=1
by generalized eigenvectors, see (4.3.l9).) Therefore
(J
<=>
<=>
rk [AkI-A: B ]=n
[Hints: a) use (23) ¢:> (29); b) note that the dual-system representation of the dual is
the original system representation modulo a change of sign for the state.]
33 N _:= e
keK
N k with K:= (k E Q : Re A.k < 0) ,
35 We say that there are no unstable (uncontrollable, unobservable) hidden modes iff
there are no (uncontrollable, unobservable) hidden modes at every A.k E a[A] S.t.
ReA.k
We stress that by "no unstable hidden modes" we mean neither unstable uncon-
trollable nor unstable unobservable hidden modes. Moreover, we have
40
•
45 Exercise. Prove Corollary (36).
[Hints: use the definitions of hidden modes, (34), and Theorem (16).]
37 N+cR [C]
<=>
47 N_+R[C]= ern
<=>
48 N+[A*] (")N[O (B*,A*)]= {a}
•
[Hints: ern=N_ffiN+; R[C(A,B)]l. =N[O (B*,A*)], and, by (4.4.54),
N _[A]l. =N +[A*]. ]
are called, resp., the stabilizable subspace of the pair (A,B) and the undetectable sub-
space of the pair (C,A).
55 Moreover, we say that the pair (A,B) is stabilizable iff S(A,B)= ern and we say
that the pair (C,A) is detectable iff ND (C,A) = ( e ) .
258
57 Exercise. Consider the stabilizable subspace (53) of (A,B). Let s(· ...... ) denote
the state transition function of the time-invariant representation R [A,B,C,D]. Show
that
XOE S(A,B)
iff
there exists a control uO such that
59 Exercise. Consider the undetectable subspace (54) of (C,A). Let s(· ...... ) and
pC' .... ), resp., denote the state transition and response function of the time-invariant
representation [A,B,C,D]. Show that
'{) By Corollary (36). (C.A) is detectable iff there are no unstable unobservable hid-
den modes.
[Hints: follows from the definitions; <=: w.l.g. xoe Nk=N([A-AkI]m l ) with
Re Ak O. By Exercise (4.3.38) there exists a chain of generalized eigenvectors (ei ]
with e 1 an eigenvector and [A-Akl]ei=ei-1. where eO := e and em=xo. Hence
2 tm - 2
Cexp[At]xo=C[e l (m-l)! + e (m-2)! + ... + em]exp(Akt). Show that Cei=e
for all i=1.2..... m.]
We conclude this section by some theorems synthesizing the results of this sec-
tion. Certain results on linear constant state feedback and output injection are added.
They are proved in Chapter 10. Another added result on the relation between I/O sta-
bility and expo stability is proved in Chapter 9.
if and only if the pair (A,B) is stabilizable. [Stabilizability by linear state feedback.] •
if and only if the pair (C,A) is detectable. [Stabilizability by linear output injection.] •
function H(s) ,
i) there are no unstable hidden modes
¢;> N+cR[C] and N+nN[O]= Ie}
¢;> "rIf..Eo[A]n 41:\,
2 Wc(O,oo) = f eAtBB*eA"'tdt =: We
o
The positive definite matrix We is related to the minimal cost to reach x on (- 00,0]:
261
o
4 min
u
J (u,u)dt=
-00
The positive definite matrix Mo is related to the energy in the z-i response starting
from x at time 0:
5 u
o
6 AWc+WeA* +BB*=O
where the uniqueness follows from that X AX + XA* and Y A*Y + YA are injec-
o
tive because o(A) c tr_.
It turns out that there is a basis in state space such that, for an R satisfying (1),
the representation with respect to the new basis has the property that the controllability
grammian and the observability grammian are equal and diagonal: thus the new
representation is "balanced."
10 Theorem. Given R controllable, observable and expo stable, there exists a coor-
dinate transformation that gives a new representation if = [A,B,C,O] that is "balanced,"
that is, in the new coordinate system
12 Exercise. Use (4) and (5) to give an interpretation of (11) in terms of energies
in the input and the output.
Proof. *
Let Mo = R R (e.g. let Mo = .t Ak ek ek' Ak> 0,* V k; then take
R=.t Perform an SVD on RWcR*: noting that RWcR* is Hermitian, we
obtain (L is defined in (11».
16 RWeR*=U.t2u*, Uunitaryandok>O, Vk
and, as usual, 01 ...
Choose
262
18
•
19 Discussion. The SVD delivers the a?,
in (16), with OJ Suppose that
0< on < < 1: then to reach the nth unit vector in the new coordinates, costs at least
liOn' hence to reach is expensive. Furthermore, II pC' = on hence the
z-i response due to is "small." In other words. in terms of input-output properties
the nth coordinate does not contribute much. The following result is interesting for
system reduction [Glo.1].
Using the notations above, assume that for some k E fl, Ok> 0k+': hence
V' i k, V' j > k, OJ> OJ. Partition A,B,C, conformally so that All is kxk. Then
o
o(A l1 )c (L-
In other words, if one truncates in this manner an expo stable representation one
obtains an expo stable representation of reduced order: and if 0k+' is small compared to
0, ,02' ... , Ok' then the transfer function of the reduced system is close to that of
the given one.
where A('),B(') are piecewise continuous and bounded on R+. For stating the results,
we need two definitions.
2 We say that the pair (A(·),B(·» is uniformly controllable over T seconds iff
the reachability grammian W.(to,to+T) >0, equivalently, by (8.2.12),
V' V' xo.x, E R n there is an input u[Io.Io+T] that steers the phase (xo,to) to the
phase (x"to+T).
We also need a stronger condition.
263
3 We say that the pair (A('),B('» is strongly uniformly controllable over T seconds
iff :3 As> 0 such that 'v' to e lR+, Wr(to,to + T) A?I. (The point here is that the
same As> 0 works for all to e R+.)
Consider now a nonlinear system
4 x(t) = g(x(t),u(t),t)
where 'v'toE lR+, 'v'XOE R n, 'v'u[o.oo) piecewise continuous Eq. (4) has a unique
solution on [0,00). We say that (4) is uniformly controllable over T seconds iff
'v' to 0, 'v' XO,X} ERn, :3 u[Io.Io+Tl that steers the phase (xo,to) to (x},to+ T).
The nature of the robustness results is displayed by the following statement.
5 Theorem. Let the pair (A('),B('» of (1) be strongly uniformly controllable over T
seconds, then
i) the perturbed nonlinear system
(where the supremum is taken over all x e lRn, v E lRn" t e 1R+), is uniformly controll-
able over T seconds;
ii) the perturbed nonlinear system
where for some y(f) e R+, P(f) E lR+, 'v'v ERn" 'v' t 0
where for some Y(\jI) E R+ and P(\jI) E R+, 'v' x E lRn and 'v' t 0
In fact, specific bounds on y(f) and Y(\jI) are available in the literature (see, e.g.
[Sas.I]); the important point is that under reasonable conditions the controllability of
264
the linear model is not destroyed by nonlinear perturbations provided these perturba-
tions are "small."
As expected, it can also be shown that under reasonable conditions, the zero-input
observability of the linear model is not destroyed by nonlinear perturbations provided
these perturbations are "small" (see, e.g. [Sas.I] and the references therein).
CHAPTER 8d
Introduction
In this chapter we study the most important discrete-time analogs of the
continuous-time case. The fundamental difference is that controllability to zero does
not necessarily imply reach ability from zero.
1 More precisely, we say that the pair (A(·),B(·» is controllable on [ko,k 1] iff for all
(xo,ko) and for all (xl,k,) there exists a control sequence u[ko.k,-I) that transfers the
phase (xo,ko) to the phase (xl,k 1).
Given the matrix-sequence pair (AO,B('» we know that u[ko.k,-l) transfers the
phase (Xo,ko) to the phase (xl,k 1) if and only if
kl-I
2 XI = s(k1,ko,xo,u)= <l>(kl,ko)xo + L <l>(k 1,k'+1)B(k')u(k').
k'=ko
Equation (2) shows that there will be an input u[ko.k,-l) that will transfer an arbitrary
phase (xo,ko) to an arbitrary phase (xI,k l ) if an only if the linear map
kl-l
4 Lr(ko,k l ): Ud(ko,k,-l) -? ern: u[ko.k,-') -? L <l>(k,,k'+1)B(k')u(k')
k'=ko
Sometimes we do not want to specify k l : then we say that the pair «A'),B('» is
controllable at leo iff for some kl > leo the pair is controllable on [ko,kd.
Unlike in the continuous-time case,
has an inverse <1>(ko,k d = [<1>(k I,kOW I iff detA(k) i: ° V k E [ko,k I - 1]. Hence
A= [8000
b?], b= ,
°
is not controllable on [0,3J, yet every state Xo at is driven to zero at k = 3 by using
a zero control. [Hint: Let x 3(O) = E3, then the zero control will steer (E3,0) to (6,0).]
9 Exercise. With Ae R3x3 as in Exercise (8) and b=€3' show that the pair (A,b)
is controllable and satisfies the conditions (6) and (7) on [0,31.
10 Comment. This exercise shows that controllability on [ko,kd and Condition (7)
on [ko,ktl may hold in the case that A(k) is singular.
[Hint: show that L r (0,3), given by (4) is surjective.]
Motivated by the reduction above we must make a distinction between the fol-
lowing.
11 Given the representation R ,i') or the pair (A(-),B('» we say that the stale Xo is
controllable to zero on [Ieo,ktl iff :3U[k{).k,-I] that steers (xo,k o) to (en,k l ); we say that
the state xl is reachable on [ko,ktl iff 3u[4k,-I] that steers (en,k o) to (xl,k l ). More-
over we say that the pair (A('),B('» or R dO is controllable to zero (reachable) on
[ko,kd iff every state Xo (xl resp.) is controllable to zero (reachable resp.) on [ko,kd.
267
20
i.e. more states are possibly reached as the horizon k) increases. The same happens as
268
23 X(k+l)=A(k)X(k)A(k)* + B(k)B(k{ k
with X(ko) = O.
We have also by the definitions and Theorem (12) the following theorem.
ii) The set of states XOE (Cn that are controllable to zero on [ko,ktl is the subspace
In other words, XOE ern is controllable to zero on [ko,ktl if and only if <I>(kl,ko)xo is
reachable on [ko,ktl.
27 Exercise. Prove Theorem (24). [Hint: for (i) use Eq. (2) and Theorem (12,
(ii».]
Hence by (4b) the transition matrix <I>(k,/) is nonsingular for all k and / with
<I>(k,l)-!=<I>(l,k). Hence, using (2) and definition (lla), a state XOE (Cn is controll-
able on [ko,ktl iff for some ulko.k,-l]
269
k,-l
33 Xo=- L <I>(ko.k'+l)B(k')u(k').
k'=k o
34a Lc(ko,k 1):= [<I>(ko,k1)B(k1-l): <I>(ko,k 1-l)B(k J-2): ... : <I>(ko.ko+l)B(k o)] .
3S Exercise. Let A(k) be nonsingular for all k and let Lr(kO.k J ) be the reachability
matrix (4a). Show that
35b Comment. If A(k) is nonsingular for all k. then a) on any interval [ko.ktl the
controllability to zero and reachability matrices are equivalent and b) (A(·).B(·» is con-
trollable on [ko.ktl iff (A(·).B(·)) is reachable on [ko.kd iff (A(·),B(·» is controllable to
zero on [ko,kJl.
ii) the set of states that are controllable to zero on [ko,kd is the subspace
270
•
43 Exercise. Prove Theorem (40). Hint: for (41) note that
43a
46 Exercise. Show that under the conditions of Theorem (36), on any [ko,kd the
subspace of reachable states and that of states controllable to zero are related by the
nonsingular transformation
57 u(k) := B(k)*<I>(kl,k+l)*Wr(ko,klrl[x)-<I>(kl'ko)xo]
58 lIullz=xtWr(ko,k)-lxl
Short proof. The control u[ko,k,-]I transfers (xo,ko) to (x],k]) if and only if
59 x]-W(k1,ko)xo=Lr(ko,k1)u
where Lr(ko,k 1) is given by (4a). By (A.7.57) (Theory of the Adjoint) the least /2_
solution of (59) is of the form L.(ko,k1)*1; for some I; e (]:n. Substitution into (59)
together with (20) gives the minimum cost u[ko,k,-lJ as
1 Given R d('), for any initial state Xo at ko and any input Ulko.k,) e U d(ko,k 1), the
corresponding output Ylko,k.J is given by
k-J
2 y(k)=C(k)W(k,ko)xo + C(k) 1: W(k,k'+l)B(k')u(k') + D(k)u(k) .
k'=ko
Suppose that in addition to R dO we know ulko.k,l' then in (2) we can calculate the last
two terms; thus w.l.g. for the study of the relation between the state Xo and the output
Ylko,kd we need only consider the first term, i.e. the z-i response
C(ko)
C(ko+ 1)<I>Cko+ 1,ko)
5 We'll say that the slale Xo of R dO or CC(,),AC'» is unobservable on [ko,kd iff the
z-i response (3) is zero on (ko,kd.
In view of definition (5), we have
Xo is unobservable on (ko,kd
6
equivalently, the set of all states of R C·) or CC('),AC'» that are unobservable on [ko,kd
is the linear subspace N CLoCko.k I» =N CLoCko,k l ».
Recalling the definition (8.1.16) of a dynamical system observable on [to,td and
specializing to the special case of R dO, we say that
7 the pair (C('),A('» is observable on [ko,kd iff given R dO. V inputs u[k",k,-I]
and V corresponding outputs Y[ko,ktl, the state Xo at ko is uniquely determined.
Since Lo(ko,kl) is a linear map represented by the matrix Lo(ko,k l ), we have
As a consequence the pair (C('),A('» is observable iff the zero state is the only
state which is unobservable on [ko,kd, equivalently, the subspace of unobservable
states is the trivial subspace of IT
kl
17a Mo(ko,k 1):= L cI>(k',ko)*C(k')*C(k')cI>(k',ko).
k'=ko
•
18
Proof. Exercise.
23 Corollary. Given the pair (CC'),A('», the set of all unobservable states on
[ko,kd is a subspace of tr n and, in particular, it is
N (Lo(ko,k 1» =N =N [Mo(ko,k 1)] .
Proof. Exercise.
24 Exercise. Using the notations of Theorem (12), show that the sequence
ko Mo(ko,k 1) is the solution at of the backward matrix r.e.
with X(k\+l)=O.
described by
274
32 y(k)=C(k)A(k)-lx(k+l) + [D(k)-C(k)A(k)-lB(k)]u(k) .
Note that given RdO, for any state xl at k l +l and any input u[lco.k,] there
corresponds a unique output Yrko.kd'
35
37 Xl is unreconstructible on [ko,ktl
k,
38 Wrec(ko,k l ) = L <I>(k',k l+ l)*C(k'>*C(k')<I>(k',k l+ 1) .
k'=ko
•
275
8d.4. Duality
If we compare Theorems (8.2.36) and (8.3.12) it becomes clear that they are
closely related. In fact, they are related by duality. Throughout this section we
assume det A(k) '11= 0 'v' k for reasons of simplicity.
Consider a discrete-time linear time-invariant representation
Rd(·)=[A(·),BO,CO,DO], i.e.
2 y(k)=C(k)x(k) + D(k)u(k)
where as usual x(k)e ern, u(k)e er ni, y(k)e ern.. and A(·),B(·),C(·),D(·) are compati-
ble matrix-sequences. As before, (k,leo) -4 <I>(k,leo) is the transition matrix.
As in (2d.120), the reverse-time dual-system representation
4 -
Rd(')=[A* (-),c
AI<
('),B* ('),D* (.)]
is described by
whence [<I>(k,k')r' = <I>(k',k) exists \f k' and k. Moreover, we shall denote by U d the
appropriate space of input-sequences. Calculations based on (5) and (6) show that for
any state x(leo) e er n at time leo, for any k> leo, and for any ii (.) E U d, the state and
output of RdO is given by
k-l
8 x (k) = <I>(ko,ktx (leo) - I. <I>(k',ktC(k'tii(k'+l)
k'=ko
k
9 Y(k+l) = B(k)*<I>(ko,k+ ltx. (ko) - B(k)* L <l>(k',k+ 1>*C(k'>*U (k'+ 1) + D(k)*U (k+ 1)
k'=ko
Note that by the pairing Lemma (2d.1.126), given Rd(') and Rd(')' 'v'ko,k with ko<k,
276
k-I
10 = (it (k).x(k» + (ii (k'+l).y(k,) .
k'=ko
11 The pamng (10) shows that the input. state and output of RdO are the
sequences k ii (k+ 1). k x (k) and k Y(k+ 1) resp.. This the choice of
fonnulas (8) and (9) for describing the state and output trajectories of R d(·)' Moreover
using the convention (11) and Eqs. (8) and (9), on any [ko.k l ]. the controllability to
zero map (8d.2.34) and the observability map (8d.3.4) of Rd(') are given by. resp .•
kl-I
12 -Lc(ko,kl)ii= «ll(k'.ko)*C(k')*ii(k'+I)
k'=ko
and
14
and
15
where Lo and Lc are the observability and controllability to zero maps. (8d.3.4) and
(8d.2.34) of the system R i'). (1)-(2).
Hence Theorem (8.4.10) reads in the discrete-time case as follows.
ii) The subspace of all states of Ri') that are unobservable on [ko.kl-l] is the
orthogonal complement of the subspace of all states of its dual RdO that are controll-
able to zero on lko,kd. In terms of the maps (8d.3.4) and (II), that is,
277
21b
and
22b
where
k,-\
23a Wc(ko,k\) = 1\1 o(l<o,k\-I) = L <l>(ko,k'+I)B(k')B(k')*<l>(ko,k'+l)*
k'=ko
k,-\
23b M,,(ko,k\-I) = Wc(ko,k\) = L <l>(k',ko)*C(k'tC(k')<l>(k',ko)
k'=ko
(We' Mo and We' M 0 are the controllability to zero and observability grammians of
R dO and Ri') resp.).
26 Exercise. Prove Theorem (20). [Hints: use (14) and (15); note that We(ko,k\) is
a matrix representation of Lc*Lc , etc.]
26a Exercise. Let Ri') be the dual system representation (5)-(6) where
detA(k) '" 0 for all k with state x(k), input u(k+l) and output Y(k+I). a) Show that
the forward representation of RdO is given by
b) Show that under the same conventions as in a) the dual of the dual representation
II dO denoted by RdO is given by
29 x(k+l)=A(k)x(k)-B(k)u(k+2)
30 y(k+2)=-C(k)x(k)+D(k)u(k+2) 0
31 Comment. Under the conventions (II) the forward versions of RdCo) and R dO
read as followso
33 II dO=[ACo),-BO,-CO,-DCo)] 0
Hence a) the dual of the dual equals R dO modulo a change of sign of the state and b)
Corollary (24) can be made to read
•
279
40 Reconstructibility of Rd' Observe that Eqs. (5) and (6) are the backward equa-
tions of Rd(') reflecting the evolution map x(k+l),u(k+1) (x(k),y(k+l», where
as in (11) u (k+ 1) and y (k+ 1) are the control and output, resp. at time k. Hence, by
definition (8d.3.36),
where Lr(ko,k l ) is the reachability map (8d.2.4) and Lrec(kO,k1-l) is the reconstructibil-
ity map of RdO on [ko,k1-l] given by
-
Lrec(ko,k,-I): ern Yd(ko,kl-I): XI
* *1Ike [k",k - ' ]
B(k) <lJ(k l ,k+l) l
CA
2 the observability matrix 0 .-
3 Comments. a) The matrices (I) and (2) are obtained from the reachability
matrix (8d.2.4a) (for ko=O and kl =n) and from the observability matrix (8d.3.4a)
resp. (for ko=O and kl =n-l). For matrix (I) we use the notation C because con-
trollability is equivalent to reach ability on [O,nJ by Corollary (8d.2.11 a).
/3) The subspaces R [C] and N [0] are A-invariant by Lemma (8.5.3).
In this section we shall use the following tool.
Proof. i) Since by assumption detA "* 0, by Theorem (4.5.30) there exists a poly-
nomial p(.) S.t. A-I=p(A). Since V is A-invariant, '\IXE V, '\IkE N, AkxE V.
Since V is a linear subspace, 'It x E V , A-Ix = p(A)x E V.
ii) Since V is A-invariant, 'ltXE V and '\IkE N, AkxE V. Since V is A- I-
invariant, '\I x E V and 'It (-k) E N, Akx E V. •
and
281
CA-n
CA-n+1
7 0 1=
CA- 1
Then
8 (i) R [C] =R [C 0] ,
9 (ii) N[O]=N[Od·
10 Forward comments. i) will mean that for a pair (A,B) with A nonsingular, the
subspace of reachable states is identical to the subspace of states that are controllable
to zero; moreover rk [C) = n will mean that such pair (A,B) is simultaneously reach-
able and controllable to zero. See Comments (SOy) and (S5P) below.
ii) can be shown to mean that for a pair (C,A), with A nonsingular, the subspace of
states that are unobservable is identical to the subspace of states that are unreconstruc-
tible; moreover rk [0] =n will mean that such pair (C,A) is simultaneously observable
and reconstructible. Compare with the conclusions of Theorem (9) below.
Proof of Corollary (5). (i) By (1) and (6) C o=A-nC. Now since A is nonsingular,
by Lemma (4) the A-invariant subspace R [C) is also A-I-invariant. Hence
R [C 0] cR [C]. Moreover, with A nonsingular, R [C 0] is both A-1_ and A-invariant.
Hence R [C o]:::>R [C]. Thus R [C] =R [C 0]' (ii) follows similarly. Note that
01=OA-n and that with A nonsingular, both N[O] and N[Otl are A- and A- I-
invariant. •
12 that R d or the pair (C,A) is observable iff the zero state is the only state of
R d or of (C,A) that is unobservable.
Thus
282
xo e cr n is unobservable
13 <=>
CAkxo=6 'v'k=O.1.2 ..... n-l.
14 Theorem [Observability properties of the pair (C.A»). Given the matrix pair
(C.A),
15 i) the set of all unobservable states is the A-invariant subspace N (0) c <r n,
iii) For C and A real. for any monic real polynomial 1t of degree n. there exists
L e IRnxn. S.t.
18 XA+LC = 1t.
iv) Let. in addition, a(A) c D(O, 1); the pair (C.A) is observable if and only if Mo. the
unique solution of
19 W=A*WA+C'C
is positive-definite.
•
Proof of Theorem (14), part (iv). By Exercise (7d.2.52). the operator A : W - A*WA
is bijective because a(A) cD(O,l). Hence. (as in the proof of Lemma (7d.2.54». Eq.
(19) has a unique solution
20 Mo= 1: (A*)kC'C(A)k
k=O
that is Hermitian positive semi-definite. Note. for any Xo e (C n. with 0 the observa-
bility matrix (2).
Proof =>. If k < n, the proof is immediate; hence, consider only k n. Let xI be
reachable on [O,k] with k i.e. there exists a control U[O.k-1] S.t.
k-I
xI = L Ak-k'-IBu(k') .
k'=O
.
By the Cayley-Hamilton theorem (3.2.24) we know that, for all k' n, Ak ' is a linear
combination of Ai, for i=O,l, ... ,n-1. Hence there exists a control v[O.n-l) S.t.
n-I
xI = L An-k'-I Bv(k'). Hence xI is reachable on [O,n].
33 a state XI is reachable
284
n-l
xl = L An-k'-IBu(k') for some control u[O,n-lj
k'=O
By (4.3.1) the state space can be decomposed into A-invariant algebraic eigenspaces,
i.e.
43 Let N r be the direct sum of the Nk's corresponding to all the nonzero eigen-
values and let N d be the algebraic eigenspace corresponding to the (possible) eigen-
value at O. Thus by (42) ern is uniquely decomposed into
44 ern=Nr$N d
where the fixed A-invariant subspaces N rand N d are called the reversible subspace
and the deadbeat subspace resp, 0/ er n. States that belong to N rand N d are called
reversible and deadbeat states resp., and denoted by xr and xd, resp. Thus, by (44)
every state x e ern is uniquely decomposed into
46 Facts [Decomposition of the r.e. x(k+l) = Ax(k)]. Let Ae (CnXII and denote by
285
and
48 Ad : N d N d : xd Adxd:= AXd .
U.t.c.
a) The map AI' (47), is invertible.
(Since every eigenValue of Ar is nonzero by (43)).
b) The r.e. Xr(k+ 1) = on N r is time-reversible, i.e. V xr(O) E N r the r.e. has a
unique solution xr(k) = E N r for k 0 and for k O.
(This justifies the name reversible state for xr(O); note that because Ar: N r N r is
invertible, xr(k) E N r for all k E Z .)
c) The map Ad' (48), is nilpotent with index md n, (i.e. the smallest nonnegative
integer k S.t. At=O is md)'
(This follows because, by (43), a(A d)= (O }).
d) The r.e. xd(k+l)=Adxd(k) on Nd has a unique solution xd(k)=Ajxd(O) on k
such that for every xd(O) ENd' xd(n) = Ximd) = e.
(This justifies the name deadbeat state for xd(O».
e) For every X(O)E (tn, decomposition (45) applied to x(O) induces a unique decom-
position of the solution of the r.e. x(k+ I) = Ax(k) on k 0, viz.
where
a) (R[C]nN r) is both an Ar-invariant and an Ar-I-invariant subspace of N"
(R[C]nN r is the subspace of all reversible states that are reachable),
b) R[C]nN d is an Ad-invariant subspace of N d , (R[C]nN d is the subspace of all
deadbeat states that are reachable). •
[Hints: (51) holds because the A-invariant subspace R[C] is spanned by generalized
eigenvectors, (4.3.19); for a) and b) use (47) and (48); for a) use also Fact (46a) and
Lemma (4).]
286
S3 Exercise. Consider a constant pair (A,B) and decomposition (44), where the
dimension of N r is"r' Consider as in (47). Use decomposition (45) to decompose
each column of BE 0: nxn, to obtain finally
B=Br+Bd
S4 Comment. One could call (ApBr) the reversible part of the pair (A,B).
For controllability to zero we now have by definition (8d.2.11) and decomposition
(44).
Proof. Let k < n: bring Xo to zero in k steps; with zero inputs from k on, the state
will forever stay at zero. Let k n. Let C be the reachabiIity matrix (I) and decom-
pose Xo according to (45) into Xo = XOr + xOd' Observe that by Fact (46e) Akxo = ArkxOr
for all k Hence we have
Xo is controllable to zero on [O,k] for some k
k-l
=> -Akxo= L Ak-k'-l Bu(k') for some control u[O.k-1] (by definition)
k'=O
Proof. ::>. By Corollary (8d.2.11a), on a given integer time-set [O,k] with k>O,
controllability of (A,B) implies reach ability and controllability to zero. By Lemmas
(30) and (55) the latter imply reachability on [O,n] and controllability on [O,n]. Hence,
by definition, for any state xl at n and any state Xo at 0 there exists controls U(O.n-Ij
and v(O,n-Ij such that
n-l
XI = An-k'-I Bu(k')
k'=O
and
n-l
-Anxo= 1: An-k'-I Bv(k') .
k'=(j
Therefore, by addition
n-l
x)-Anxo= 1: An-lc'-l B(u(k')+v(k'» .
k'=O
In other words given any state xI at n and any state Xo there exists a control
(u+v)(O,n_Ij that steers the phase (xo,O) to the phase (xI,n). Thus by definition
(9d.2.11) the pair (A,B) is controllable on [O,n]. •
62 (A,B) is controllable
Now that we have nonnalized and are using the definitions (31) and (60), we have in
analogy with (8.5.36).
67 ii) The reachable subspace of the pair (A,B) is the A-invariant subspace R [C].
68 rk[C]=n
iv) For A and B real, for any monic real polynomial 1t of degree n, there exists
FE 1Rnixn such that
(v) Let cr(A)cD(O,I); the pair (A,B) is reachable if and only if Wr the unique
solution of the equation
71 W=AWA* +BB*
is positive-definite.
•
72 Comments. a) Deadbeat closed-loop systems. If in iv) 1t(A.)=A.n , then the
resulting closed-loop r.e. x(k+l)=(A+BF)x(k) is S.t. a(A+BF)= (0 I, i.e. every
closed-loop state x(O) E (en is a deadbeat state by decomposition (40): every closed-
loop state-trajectory k --t x(k) is zero for k n.
(3) Controllability to zero: It turns out that condition (69) holds for every
289
I.E o(A)\ (O) if and only if the pair (A,B) is controllable to zero, (see Theorem (75)
below).
Proof of Theorem (66), part (v). As in the proof of Theorem (9), part (v), the
unique solution Wr of (71) reads
Wr = (A)kBB*(A*)k
k=O
Hence, by Cayley-Hamilton
ii) R[C]nN r
( subspace of reversible states that are controllable to zero )
71
= (subspace of reversible states that are reachable) .
79 VA.Eo(A)\(O} rk[A.I-A:Bj=n.
•
80 Comments. a) (78) means that (A,B) is controllable to zero iff every reversi-
ble state is reachable.
(77) means that if x EN r (i.e. x is reversible), then x is controllable to zero iff x
is reachable.
y) If A is nonsingular then N r = fr nand (77) reduces to the equality of the sub-
space of reachable states and the subspace of states that are controllable to zero, more-
over (78) reduces to rk[C]=n. Compare with Eq. (8) and Comment (10).
290
Xo is controllable to zero
(by (61»
(ii) The equalities (77) follow by comparing the direct sum decompositions in (51)
and (76).
(iii) Equivalence (78) follows by (76) and trn=N rEf> N d' In view of (43).
equivalence (79) follows from condition (78) which is equivalent to
82 NdcR[C] rk[A:B]=n
90 Duality of reversible systems (i.e det A "# 0). Consider a discrete-time time-
invariant system representation R d = [A,B,C,D] where A is nonsingular.
Consider the reachability matrix C and the observability matrix 0 given by (1) and
(2).
By Corollary (5) and Exercise (81) the reachable subspace of (A,B) and the subspace
of (A,B) of states that are controllable to zero are both equal to R [C]. Hence we are
justified
92 N[OI]=N[C*]
is easily recognized as the unobservable subspace of the pair (B* ,A*). Hence we have
by a) and (92)
b) (N [0]).1. =R [C 0]
where Co is the to zero matrix (6) of the (forward) pair (A*-',- A*-'c!)
of the dual system Rd' Moreover,
292
is easily recognized as the controllable subspace of the pair (A*,C\ Hence we have
by b) and (94)
96 Comment. In (92) we have exchanged the forward pair (B* A*-',A*-') of Rd for
its backward pair (B*,A*). Similarly in (94) we exchanged the forward pair
LA*-',-A*-'c*) for its backward pair (A*,c*). Compare the forward representation of
R d (8d.4.27)-(8d.4.28) with its backward representation (8d.4.5)-(8d.4.6). Another
benefit of Corollary (5) is the ability to exchange forward pairs and backward pairs!
given by (8d.5.2). Consider the matrix A E (C nxn with spectrum (l(A) = {Ak rand
Nk =N (A-AkI)ffik) .
Observe that, for every Ak E (l(A) tl D(O,l)C R (C] tlN k represents the A-invariant
subspace of (unstable) modes at Ak that are both controllable to zero and reachable,
(prove this: observe that unstable modes are reversible states and use Theorem
(8d.5.75», therefore we call simply R [C) tl N k' the controllable subspace at such Ak'
293
8 We say that the pair (A,B) is stabilizable iff R d or (A,B) has no uncontrollable
unstable hidden modes or equivalently
9 We say that the pair (C,A) is detectable iff R d or (C,A) has no (nonzero)
unobselVable unstable hidden modes, or equivalently,
The following results parallel Theorems (8.7.62) and (8.7.65), and are proved similarly.
N+cR[C]
o(A+BF) c 0(0, 1)
'<IAE a(A)nD(O,I)C, rk [ c-
AI-A]
=n,
a(A+LC) cD(O,I)
REALIZATION THEORY
Introduction
In this chapter we study the main properties of the realizations of a given proper
transfer function matrix R(s) E CI:p(S)n..xnI. We show that the McMillan degree of R(s)
is the minimal dimension of the state space of any of its realizations. We prove that
two minimal realizations are algebraically equivalent. Vfe show that eigenvalues of A
of any minimal realization are dictated by the poles of H(s) .• We conclude this chapter
with a description of a controllable canonical realization of H(s) as in [Chao 1].
(5.4.22), (nqte that L [H(t)] = H(s». Concerning the equivalence a) ¢:> c) observe that
1) H(oo)=H(oo) ¢:> D=D, and 2) t -t Cexp[At]B as well as t -t C exp[At]B are
analytic on R and have a Taylor expansion, e.g.
8 for all te IR .
s.t. VI
13 H, = [CAi+jB J:.j=O=O/ . C/
where
297
CA
14 01 e ."
'" (1+llno"" t and C,:= [B :. AB :.
. .. :. Al B] E .".<(1+1).;
'U- ,
CAl
15 H
n-l
= rtR I+J.J iJ=O
n-l =0 . C E CI:1U\,XJUl1
where 0 = 0 n-l and C = C n-l are the observability and controllability matrices of R,
given by (8.5.2) and (8.5.1). Moreover, we have the following fact.
and
18 rk[H n-l] is independent of the given realization .
•
Proof. a) (17) is obtained as follows: let I with HI given by (13)-(14). Let
011-1' C11-1and if n-l be matrices (of dimensions (l + I )no x n, n x (l + I )nj and
(l+l)nox(l+l)nj, resp.) obtained by placing On-I' Gn-l and H II- 1, resp. in the upper
left-hand comer and bordering them with zero entries. Now, (exercise), by (14) and
the Cayley-Hamilton theorem (3.2.24), there exist nonsingular matrices Land R
(obtained by row and column operations, resp.) such that
LH/R=0n-lCn_l =H n- 1 ,
rk[Hd=rk[Hn-d=rkfHn_d· QED.
For this purpose let I {n-l,n-I} and observe that, by (13) and (7),
22 Theorem [Rank of Hankel matrices]. Let H(s) E ([: p(s)""xn i be a given proper
transfer function matrix; let its expansion at 00 be described by (10)-(11). Consider its
Hankel matrices HI of all orders described by (12)-(l3).
U.th.c.
a)
23 "II E N rk[HI ] rk[HI+t1,
b)
where nand A,B,c; are, resp. the dimension and parameter matrices of any realization
=
R [A,B,C,D] of H(s).
is called the McMillan degree of H(s). (The matrices HI are the Hankel matrices of
299
38 aM=rk[O . C],
where 0 and C are the observability and controllability matrices of R, resp. (see
(8.5.1)-(8.5.2». Hence by Sylvester's inequality (A.5.37), since 0 and C have, resp. n
columns and n rows
39 rk[O] + rk[C]-n $ BM $ min (rk[O],rk[C]) $ n
and
41 (A,B) is controllable ¢> rk[C]=n.
300
Therefore, (31) follows from the RHS of (39). Moreover, by (39)-(41), (32) is
equivalent to (33). •
43 invariant for all minimal realizations of a given transfer function; it has a con-
stant size 8M no x 8M nj and consists of realization independent Markov parameters Hi'
(11).
y) The McMillan degree 8M the minimum number of integrators needed to simu-
late the given transfer function H(s).
The next result establishes that minimal realizations of the same transfer function
are algebraically equivalent.
45 R is a minimal realization
if and only if
more precisely,
there exists a nonsingular matrix Te o:nxn s.t.
R -.::L R [i.e. x= Tx] .
Furthermore,
47a T=(O*O)-IO*O.
andR.
51 D=D,
52 OC=OC,
53 0 AC=OAC,
are well defined. By (52), T I T2=I, hence TI and T2 define a nonsingular matrix
T e o:n><n s.t.
From (55)-(56) it follows, on mUltiplying (52) separately, first on the left by 0* and
-*,that
second on the right by C
57 C=TC and 0=0 rl.
Hence by (8.5.1)-(8.5.2)
58 n=TB and c=crl.
302
Hence there exists a nonsingular matrix T defined by (55)-(56) s.t. (59), (58) and (51)
hold, i.e., by Exercise (5.4.61), R 2- R. QED.
Equivalence (45) (46) is now established. Finally (47a) and (47b) follow from
(55)-(56). •
The last main result of this section shows that the eigenvalues of A of any
minimal realization of a given transfer function B(s) are precisely the poles of B(s).
"# 0 for I =m
AE P[H(s)] and lim (S-A,)' B(s) {
s-->A. = 0 for I > m .
67 (d k is called the McMillan degree of the pole Ak of B(s); hence, by the direct
sum decomposition (4.3.2) of the state space into algebraic eigenspaces,
303
68 the McMillan degree of R(s) is the sum of the McMillan degrees of its poles,
b) Let R = [A,B,C,D] be a minimal realization of H(s) E tr p(S)"'X1\ where the basis
of the state space tr n is chosen to be a union of bases of the algebraic eigenspaces of
A, (see decomposition (4.3.2». Use notations induced by (4.3.6) and (4.4.1) et seq.
A
Show that (with appropriate matrices) the partial fraction expansion of H(s) reads
,... " a mit
69 H(s)-H(oo)= L L CdAk-AkI<I..1/-IBk(s-Ak)-1
where
and, V kE Q:,
[Hints: for a) use Theorem (62) and Theorem (44); for b) use (4.4.19c), (4.4.2),
Theorem (62) and a).]
72 Comment. Theorem (62) shows that by examining the entries of H(s) one can
determine alA] and the multiplicities of the eigenvalues (63) for any minimal realiza-
tion. The minim!!l polynomial of such A is the least common (i.c.d.) of
all elements of H(s). By examining the partial fraction expansion of H(s) (see (87)
below), one can obtain the dimension n of any minimal realization and the dimension
d k of the algebraic eigenspaces of its matrix A. It be shown that the characteristic
polynomial of such A is the l.c.d. of all minors of H(s), e.g. [Che. I].
76 Vj E ill with e O := e.
Note that e l is an eigenvector of A.
304
77
_r
-C le
1 (t-T)m-l
(m-I)! + e2
(t T)m-2 ]
+ ... + em exp[A(t-T)],
where by observability Ce l :F 6.
Observe now that the piecewise continuous functions Y[o.n and u[O.T) (set equal to
zero on (T,oo» have Laplace transforms that are analytic in the finite plane, (i.e. they
are entire). Moreover, by (2), since m is the size of largest Jordan block associated
with AE a[A], R(s), when it has a pole at A, has a pole of order at most m. Thus,
using (77),
78 y(s) = Y[O.T)(S) + e- sTC[e 1(s-A)-m + e 2(S-A)-m+1 + ... +em(s-A)-l]
and yes) has a pole at A. of order m (indeed Y[O.T) has no pole at A. and Ce l :F e).
Moreover, since
, , ,
79 y(s)=H(s)' Uro.T)(S)
(where has no pole at A.), R(s) must have a pole at A. of order m. Indeed the
negation of the last assertion leads to contradiction:
1) if R(s) has no pole at A. then, by (79), yes) has no pole at A: -H-.
2) if R(s) has a pole at A. not of order m, then, since its order is at most m, the order
must be strictly smaller than m; hence in view of (79), yes) cannot have a pole of
orderm
Hence we have established that H(s) has a pole at A. of order m.
¢:: We assume that R(s) has a pole at A. of order m and use contradiction:
1) if A is not an eigenvalue of A then, by (2), R(s) is bounded in a neighborhood of
A: hence A is not a pole of R(s) : f- .
2) if A is an eigenvalue of A with multiplicity Jl (as a zero of its minimal polynomial)
such that 11 :F m, then by the proof above, R(s) has a pole at A of order
Jl :F •
• 0
81 P[H(s)] c CL .
Then
o
x=Ax is expo stable (or equivalently, cr(A)c (1:_, (7.2.33»,
if and only if
R has no unstable hidden modes (or equivalently,
Short Proof.
<=. With N _ and N + resp. the stable and unstable A-invariant subspaces of ern,
(given by (8.7.33)-(8.7.34)), pick a basis according to ern=N _ $ N +. In that basis R
reads xI =AtxI +Blu, x2=A2X2+B2U, y=Clxl +C2x2+Du, where
o
cr(AI)=cr(AIN)=cr(A)n CI:_ and cr(A 2)=cr(AI N)=cr(A)n CI:+. Let
:= and let H I(s)=H(s)-H 2(s). By assumption (81),
P[H2(s)] = 0; furthermore [A2,B 2,C2,0] is minimal, (by Theorem (8.7.68), N + cR [C)
and N + nN [0] = { e }.) Hence, by Theorem (62),
• 0
cr(A)n (1:+ = cr(A 2) = P[H2(s)] =0. Thus cr(A)c CI:_.
:;'. Assume that R has an unstable hidden mode; then there exists an unstable mode
•
o
of x = Ax. Thus cr(A)¢ (I: _ •
The exercises below show how to compute the McMillan degree from the partial
fraction expansion of a given transfer function.
wh<;.re KI and K2 belong to (l:JloXIl; with K2 nonzero. Show that the McMillan degree
of H1(s) is given by
84
[Hints: use Exercise (65). a) Let [A,B,C,O] be any minimal realization of HI(s). Then
cr[A]= {O}, HI(s) = CBs-I +CABs-2, AI =0 for I ;::,2 and n ;::,2.
b)
306
where KI and K2 belong to ern.><n, with K2 nonzero. Show that (84) still holds.
[Hints: set o=s-A, then H 2(s)=H 1(0), see (83), ... , show that
rk[O (C,A)C (A,B)] =rk[O (C,A-AI)C (A-AI,B)]]
88 H(s)-H(oo)= Kk/(S-Akr1
r
k=1 1=1
{Ak = P[H(s)] and, for all k E Q, Kkm • # O. Show that the McMillan degree
of H(s) is given by
(see (5.4.61) and Sections A.5.1 and A.5.2 on matrix representation theory). Note that,
given A and B, the choice of II determines A and R.
2 (A,B) is controllable
and
4 Comment. Assumption (3) can be made without loss of generality: if (3) does
not hold, introduce a change of coordinate ti = Lu of the input space: e.g. choose a
basis L- I (representing elementary column operations upon B) such that R := BL- I is
in column echelon form (A.5.66), whence for all u E cr n, with ti = Lu
where RI has full-column rank. Hence assumption (3) holds by reducing the input to
ii I'
has rank n. Hence it has n l.i. columns of the form Aibj which, if nj> I, may be
selected in many ways. We select these columns by reading the columns of C from
left to right, rejecting each column that is t l.d. on previously selected columns. Note
308
. ] k.-\
8 the family [ AJbj .' is selected;
J=O
moreover
II;
9 "Lkj=n.
i=l
Let us express the nj linear dependence relations (7): for convenience, replace i by I,
and write:
k, n,
10 Ak,b/ = -"L L y/Aibj
j=o j=l
where
(these y's correspond to columns of Ak'B that do not precede Ak'b/), and, for all ie ni
such that kj k/
(because, for all m Ambj is l.d. on previously selected columns). Note that the nj
families (8) constitute a basis for cr n.
Note that the ni families (8) constitute a basis for ern. Replace in (8) i by I and j
for q. and add to each selected vector A%/ a linear combination of previously
selected vectors Aq-ib i• defining vectors by:
Moreover. by (lOa)-(lOb). only previously selected vectors are present in the RHS of
(12) and. by (9). (12) accounts for n vectors Therefore (12) defines a basis of
ern and the nxn matrix II defined by
is nonsingular. The basis (12). ordered as in (14). is the basis required for the con-
trollable canonical form (A.R) of (A.B). (by using (1). i.e. (20)-(21) below).
Since L is nonsingular, (20) shows that R[B]=Sp [el,el· ... Let now
L:= t: =1' then = 1 and = 0 for i > I; with these notations (20) becomes
I-I
22 b, = 1: el + ef V IE nj.
j=l
Note the position of the vectors el in the ordered basis (see (14». Equation (22)
expresses the Ith column of B. b in terms of the new basis (12). ordered as in (14).
"
From (22) the general form of B follows (for a special case, see (28) below).
Equation (21) exhibits for all I E OJ the image of under the map A. Let now
Lul (q+l):= then from (21), for q=O.l, ...• k/ -l
1 1 n, '1 .
23 A eq+l = eq+2 - 1: a(-<t el .
j=l
Equation (23) expresses the image of the basis vector under the map A. as
a linear combination of the new basis vectors. Hence the representation of A in the
new basis, namely A. follows immediately from the first representation Theorem
(A.5.3). For a special case, see (28) below.
0 0 0 0 0 0
-ap -aF -o.F -o.F 1 (l12
28 A= 0 0 0 1 0 B= 0 0
0 0 0 0 0 0
-all -o.}i -al2 -a.:p -a.:p 0
r
where A is a (k l +k2)x(k l +k2)=5x5 matrix and B is a (k l +k2)x2 matrix. Here
the (lil and al are computed from the Yjil, defined by (10), by
29 1 :1:
and
30
34 Exercise. Consider the single-input case, nj = 1, whence (28) reads e.g. for n=3
35 A= [ 0
0 1
0
0
1
1
-at -a3
b) Show that
312
37
c) With c := [11 12 13] and (A i» given by (35), give a block diagram realization of
h(s) =c(sl-A)-li) .
[Hints: a) consider (sl-A) and the e.c.o.'s 12 +-- 12 + s13 and 11 +-- 11 + s12; b) calcu-
late (sl-A) [I s s2 ... sn-I]T; c) consider a chain of integrators of length n .]
Therefore apply procedure (5) to the pair (A*,r!) and let f' be defined by the result-
ing RHS of (14). Of course Hermitian transposition converts this procedure into the
selection of rows of the observability matrix and the definition of the rows of T, (exer-
cise).
In the single output case, (i.e. no = 1), (co,A o) in observable canonical form, for
n=3, reads
41a co=[OOI],
b) Show that
45 Exercise [/.i. state trajectory]. Consider equations (12)-(13). Let, for all / E nj,
(u (j)
l denote the control sequences of Exercise (15). Set 0'0:= 0 and
(1, := k t + k2 + ... + k, for all / E n j. Consider nj shifted control sequences (Vi (j) ]j':o
S.t. for all / E nj
47 x(O'/-1 .
In other words. the basis (12) consists of n states reached by the recursion equation
above starting from x(O) = 9 and using the control sequence .
[Hints: by Exercise (15) each control sequence (u (j)
l generates the zero-state state
trajectory with ek/=9 (by (13»; then use time-invariance and superposition.]
Indeed (stacking the vectors u(j) and xU) as columns of the matrices denoted below by
square brackets) F is given by
where the last matrix is nonsingular because, by Exercise (45), the state trajectory
is a l.i. family.
Observe now that by (47), for 1=1 and q = 0, x(I)=el=b l (by (12». Hence,
by (46) and (51), for all j E n=1
52 x(j+I)=(A+BF)ib l .
Therefore (since the family (xU»j',!,l is l.i.) the controllability matrix of (A+BF, b I)
given by
is nonsingular. Recall now the standing assumptions (2)-(3), where the last assump-
tion can be made without loss of generality (see Comment (4». Hence we have
55 Comment. In other words, if the pair (A,B) is controllable and B has full-
column rank, then there exists a linear state feedback law u = Fx + v, (where vet) E (I: fl,
is the new external input), S.t. the state d.e.
Theorem. Let the constant pair (A,B) be controllable; let b be any nonzero vector in
R [B]. Then there is a constant state feedback matrix F such that the single-input pair
(A+BF,b) is controllable.
CHAPTER 10
2 y(t)= Cx(t)
whence FE RnjXII is called a state- feedback matrix and vet) E R nl is the new exogene-
ous control, as shown in Fig. 10.1.
r-----{A ...----,
+
v +
7 y(t) = Cx(t)
8 A r := A+BFe R nxn
generates the
In this section we show that, given any monic polynomial 7t(s) e R[s] of degree n,
viz.
(where, for all i e n, 7tj E R) there exists a state-feedback matrix Fe Rn;XJl S.t.
This means that, modulo this condition, any list of n closed-loop eigenvalues
(having the property of conjugate symmetry) can be assigned to R f by an appropriate
state-feedback law (4); this property is usually called "arbitrary spectral assignment by
state-feedback. "
We start our study by considering the pair (A,B) in the single-input case (i.e.
nj = 1) where A is in row-companion form.
Let £j denote the ith standard unit vector of R n and consider any monic real polyno-
mial 7t of degree n given by (10).
Show that
18 For all ie n,
19
20
[Hints: for (19), use (18) and observe that X:4.(A)=O, (3.2.24); for (20), use (18) and
(19).]
23 Exercise. Let A e R nxn and beRn be S.t. the pair (A,b) is controllable. Let
(A,b) be its controllable canonical form where A is given by (16)-(17) and b=En. See
Exercise (9.2.34), where the transformation matrix 11 e IRnxn S.t.
and
is given by (9.2.12)-(9.2.14) as
318
Let C E R nxn be the controllability matrix of (A,b), given by (8.5.1), and let q E R n
be defined by
26 q* =£n*C- 1 .
Show that
<l:2 UJ CXn 1
UJ o
o
27 [TCr l =
CXn
CXn 1
0 o
28 q* =£1*T.
32 Af := A+bfe IRnxn
and let XAr(S) e R[s] be its characteristic polynomial. Let x(s) e R[s] be any real
monic polynomial of degree n with coefficients x n ' ••• 'Xl as in (10).
V.l.c.
if and only if
34 f=-q*x(A)
where q e R n is defined by
26 q* := t;.*C- I .
t
In the proof below we show that if (A,b) is in controllable canonical fonn, then
f = [fi in (34) is given by fi = ai-xi for all i e n, where the ai and Xi are the
coefficients of lAo (17), and 1t, (10): we say that "f is a coefficient shift vector."
y) From an engineering point of view, we must keep in mind that if we try to change
too drastically the location of the eigenvalues we will run into trouble with saturation.
To make the point, consider the simple case where n=l, A = -I, b=c=1. The input-
output transfer function is h(s)= (s+lr 1. Suppose we apply a state feedback f=99, the
transfer function becomes h1(s)= . The response is 100 times faster, but the
I/O gain at zero frequency is 1/100! To remedy this problem, let us place the gain f
between the summing node and the plant: then h2(s) = 99/(s+ 1(0) and the I/O gain at
CJ) = 0 is 0.99 == I, i.e., it is essentially restored to its previous value and the new system
is 100 times faster. Now consider the original system starting from the zero-state and
with an input l(t) (where the input l(t) is about as large as the system can tolerate
without saturating): the maximum value of the plant input is 1. Now repeat the exper-
iment with the last closed-loop system considered above: an easy calculation shows
that the input to the plant takes the value 99 at time 0+. Clearly, this unit step input
will saturate the plant. The conclusion is that, in the real world, the amount by which
one may change the spectrum of a linear system is limited by physical considerations
such as saturation, noise sensitivity, etc.
37 Proof of Theorem (32). Since the pair (A,b) is controllable it can be brought in
controllable canonical form (A,b) by a state transformation x= Tx, (see Exercise
(9.2.34». This transfonnation relates A,b, and fe R n to A, b, and 1', resp. by
320
38 b=l l b f=fr
where Ae IRnxn has the row-companion form (16)-(17), b=En, and the matrix
rl e R nxn is described by (24)-(25). Hence, by defining
39 f:= [flf2 •.. fnl e IR lxn ,
and
40 Af := A + fir ,
Ar=A+bf=rl[A+bflT=rIAcT,
1
Hence, for any f E IRn, Af is similar to Ar in row-companion form, and therefore
Note that the coefficients eli of 'XA(S) have been shifted by the components fi of the
feedback vector f, (39). Hence, upon comparing (41) and (10) we get, successively,
33 is s.t.
if and only
42 for all ie n,
34 f=-q*1t(A) .
The last equivalence follows because by the third equation of (38), (42), (20), the first
equation of (38) and (28)
- * - * *
f=fT=-E, 1t(A)T=-e, T1t(A)=-q 1t(A) .
•
321
*44 Exercise. Let Ae R nxn and be R n S.t. the pair (A,b) is controllable and let
q e R n be given by (26) where C is the controllability matrix of (A,b). Let feRn
generate Ar := A+bf. Show that
i) For all
ii) For all
fe R n q*(sI-Ar)-lb= r\
feR lxn the pair (q* ,Ar) is observable;
iii) For any monic real polynomial1t given by (10) and for all feR lxn
* *
q 1t(Ar)=q 1t(A)+f.
[Hints: consider the controllable canonical form of (A,b) using (38)-(40) and Exercises
(15) and (23).]
We are now ready to tackle the general eigenvalue assignment problem (10)-(11).
with characteristic polynomial XA,(S). Let 1t(s) e R[s] denote any real monic polyno-
mial of degree n, as given by (10).
V.t.c.
if and only if
Define now a matrix F2 E Rn,)(n S.t. all rows of F2 are zero except for the first one,
which is f2; consequently, BF2 =b,f2. Hence setting F:= F,+F2 and
A f := A+BF=A+B 1F,+b,f2, we get XA r=1t. Hence (47) holds.
Only if. We use contraposition. So suppose (48) is not true, i.e. (A,B) is not con-
trollable. Then by Theorems (8.5.36) and (8.7.16) the pair (A,B) has an uncontrollable
hidden mode, i.e. there exists an eigenvalue A of A and a nonzero vector 11 E c: n S.t.
Le. A f := A+BF has the fixed eigenvalue A. Thus, V FE R nixn, A is a zero of XA r'
and for any polynomial1t of the form (10) with no zero at A, 1t(S)=XA,(S) is not possi-
ble. Hence (47) does not hold. I
We conclude this section by justifying the claim of Theorem (8.7.62) that stabil-
izability (no unstable uncontrollable hidden modes) is equivalent to stabilizability by
linear state-feedback (Le. such that the closed-loop system (6)-(9), is exponentially
stable).
if and only if
56 Comment. Claim (54) means that all eigenvalues of the closed-loop system
matrix A+BF can be placed in the open left-half plane by appropriate state-feedback.
Of course the eigenvalues of A associated with uncontrollable modes will not be
affected by state-feedback.
58 Proof of Theorem (53). With N_ and N+ resp. the stable and unstable A-
invariant subspaces of R n (given by (8.7.33)-(8.7.34», pick a basis of R n according to
Rn=N_(J)N+. In that basis the state d.e. x=Ax+Bu reads x,=A,x,+B,u and
results in
A+BF= [
Hence
o
o(A+BF) = o(A 1) u o(A 2+B 2F2) c tr _ .
Only if. We use contraposition. So suppose that (A,B) is not stabilizable. Then
(A,B) has an unstable uncontrollable hidden mode, i.e. by Theorem (8.7.62) there
exists an eigenvalue A. of A in tr+ and a nonzero vector" E tr n S.t.
T\*A=T\*A. and T\*B=e*.
Thus, for all FE Rn;xn, A+BF has the unstable eigenvalue I.E tr+, whence
•
o
'<:/ FE Rn;xn o(A+BF) c tr _ .
2 yet) = Cx(t)
324
where A e R nxn , Be R nxn1 and C e Rn.,xn. Consider Fig. 10.2 where the system R has
been transfonned by adding a linear function of the output y, (expressed by Ly = LCx
for some Le Rnxn.,), to the input of the integrators. Note that Fig. 10.2 is precisely
the dual of Fig. 10.1.
4 y(t) = Cx(t)
where Ae R nxn , Be R nxni , Ce Rn.,xn and Le Rnxn.. Note that the resulting spectrum
is a(A+LC). As already observed in Comment (8.5.23c) the eigenvalues correspond-
ing to unobservable hidden modes of the pair (C,A) are always in a(A+LC) for every
Le 1Rnxn.. Hence, in order to have arbitrary spectral assignability by linear output-
injection, it is necessary that the pair (C,A) be observable. This is also sufficient.
Indeed we have the following theorem.
+
_U_ _ __ )--_••
+
-_..
Fig. 10.2. Implementation of linear output-injection; L is the output-injection matrix.
325
if and only if
11 Comments. Ct.) The proof below shows that Theorem (7) and Theorem
(10.1.46) are related by the duality, (8.4.13), of observability and controllability.
\3) In the single-output case the output-injection matrix L in (9) is unique (see Exer-
cise (14) below).
13 Proof of Theorem (7). By duality, (8.4.13), the pair (C,A) is observable iff the
pair (A* ,f!) is controllable. Note that all matrices are real and hence for any
LE Rnxn., A*-tC'L*=(A+LC)T, whence XA*+C"L*=XA+LC' Thus, using Theorem
(10.1.46).
iff
iff
iff
the pair (C,A) is observable.
•
14 Exercise [Single-output case]. Let A E R nxn and c* E R n such that the pair
(c,A) is observable. Let 0 denote the observability matrix (8.5.2) of (c,A) and let Ej
denote the ith standard unit vector of Rn. Let 1t(s) E R[s] denote any monic real poly-
nomial of degree n given by (8).
Show that:
IE R n is S.t. XA+lc=1t
if and only if
I =-1t(A)q ,
where q E R n is given by
q=O-IE n •
•
326
if and only if
u x __.. Y
CJ--__- -
+ Y
21 y(t)= Cx(t)
The estimator, (shown in Fig. 10.3 surrounded by broken lines) is constructed by tak-
ing a copy of R and applying constant output-error injection using an output-error
injection matrix L e Rnxn.. Hence, with x(t) e R n denoting the estimated state, the
estimator is described by the d.e.
22 = (A+LC)x(t) + Bu(t)-Ly(t)
[U(.),y(.») -+ x(') .
We wish to find conditions such that the estimated state x(t) "follows" the state x(') of
the given system R, i.e. we would like the state estimation error eO given by
23 e(t) =x(t}-x(t)
24 e(t) = (A+LC)e(t),
e(t) = exp[(A+LC)t]e(O)
where
e(O) = x(O}-x(O)
The full order state estimator described by (22) assumes perfect knowledge of
A,B,C. Suppose now that we have only imperfect knowledge of the given system R =
[A,B,C,O]: more precisely suppose that the estimator is given by (22) but the system
under consideration is described by (20) and (21), in which A,(B,C, resp.) is replaced
by A+5A, (B+5B, C+5C, resp.) with 11M II «IIAII,
v+ u Y
R=[A,B,C,O]
x
+
u
ESTIMATOR
y
Fig. 10.4. State feedback of the estimated state.
329
5 e(t) := i(t)-x(t)
6
:, rX('l]= [A+BF BF ] rX(t)] + [B ] v(t)
le(t) 0 A+LC le(t) 0
7 y(t)=[C:O] rX(t)]
le(t)
Using (6) and (7) we find that I/O transfer function from v to y is given by
8 H(s)=C(sI-A-BFrIB.
10 Remark. In practice the choice of the gain matrices F and L is subject to many
constraints; typically optimization may be used to choose them. For example, for the
choice of F one may solve a time-invariant infinite horizon LQ optimal control prob-
lem, as is done in Sec. 10.4 below. For the choice of F and L in a more general sto-
chastic LQ-context, see [Kwa.l].
11 Exercise. For the system of Fig. 10.4, repeat exercise (10.2.25). In addition,
330
I,
where
and 11'11 denotes the Euclidean nonn. (As usual, it is understood that x(1) in Eq. (4) is
an abbreviation for s(t,to,xo,u(·».)
6 Problem. Minimize the cost J(tl'to,Xo'-) over all possible inputs u(·) of class PC
of piecewise continuous functions from [to,ttl to Rnl.
A quick review of the proofs of Theorems (2.1.67), p. 35, and (2.1.80), p. 37,
shows that we established in Chapter 2 the following result.
11 *
u(t)=-B P(t,tl,S)x(t) te (-oo,td
with
and generates the optimal closed-loop system dynamics described by the linear homo-
geneous d.e.
16 x(to)=xo
22 P(t,tl,S)=P(t+t,to+t,S).
Therefore we see (by (4) and (14» that the cost and the optimal cost are shift-
invariant, i.e. 'V t E R
Hence there is no loss of generality if we set to=O. This will be done below.
b) The unique solution of the RDE is necessarily symmetric: first P(t,)=S and S is
symmetric; second, visualizing Picard iterations to solve the RDE, every iterate is sym-
metric, hence so is the limit, namel}" the unique solution of the RDE.
c) The final state penalty tenn x(tl SX(tl) in the cost (4) is usually there for avoiding
that the final state x(tl) deviates too much from zero. As the horizon tl --+ 00, this
becomes clearly a pseudo-stability consideration. However, in that case it can be
shown that under generic conditions, [Kwa.l], an expo stable optimal state trajectory is
obtained for any final state penalty matrix S = S* O. So under these conditions
x(oo)=O and the final state penalty becomes zero, i.e. S=O is acceptable. This will be
done below. Note however that if one has duality in mind and one wants to design a
stochastic LQ-optimal estimator, [Kwa.ll, then S becomes the initial variance of the
optimal state estimation error and so for certain applications it might be imperative that
S be nonzero.
d) The infinite horizon problem below must be seen as the limiting case for t, --+ 00 of
the time-invariant standard LQ problem. This implies the following:
1) In view of (4), where to=O and S=O, the limit cost given by
333
00
25 = J[llu(t)ll2+IICx(t)lf]dt
o
exists only in R;. := u (oo ) , (the cost may blow up in the limit).
2) In view of Theorem (10) the solution of the infinite horizon problem depends criti-
cally upon the solution in the limit of the Riccati d.e. (12). It becomes therefory
important to study symmetric solutions of the equation obtained from (12) by setting P
equal to zero, i.e.
26 O=A*P+PA-PBB*P+CC
30 Standard LQoo-problem.
Data. We are given
a) a real time-invariant system representation R = [A,B,I,O] described by
32 x(O)=XOE Rn ,
where
34 C E Rn.,xn ,
11'11 denotes the Euclidean norm and xO denotes the solution of (31) to the control
u(·) and the initial condition Xo.
334
35 Problem. Minimize the cost J(oo,O,xo:) over all possible inputs u(') of class PC
•
The solution of problem LQoo is usually done in two steps. First one establishes
the existence of a unique positive semi-definite (p.s.d.) solution P+ of the ARE such
that A - BB*P+ is expo stable, and then problem LQoo is solved by an optimal constant
linear state-feedback law u=-B*P+x.
38 Theorem [The ARE has a p.s.d. solution that is unique and stabilizing]. Con-
sider the standard LQoo problem, (30)-(35). Assume that the pair (A,B) is stabilizable
and that the pair (C,A) is detectable.
U.t.c.
i) the ARE (26) has a positive semi-definite solution P+,
ii) this solution P+ is unique and stabilizing; more precisely it is unique and the state-
feedback control u(t)=-B*P+x(t) results in a system matrix A+E Rnxn given by
39
equivalently
41
•
42 Comments. a) The matrix P + will be shown to be a limit solution of the RDE,
more precisely, for any fixed t E R,
The two last equalities in (43) follow by the time-invariance relation (22) which
implies
Therefore the unique p.s.d. stabilizing solution P+ can be obtained by integrating the
RDE (12) backwards from P(O) = O. (This method is not recommended for computa-
tion).
/l) The matrix A+, given by (39), is obtained by substituting the constant linear state-
x
feedback law u = -B*P+x in = Ax + Bu. The latter feedback law will be shown to be
optimal for problem LQoo in Theorem (91) below. Therefore the expo stability, (40),
335
of x=A+x implies that the optimal closed-loop dynamics will always be expo stable.
Note that the spectrum of A+ contains every eigenvalue corresponding to the stable
hidden modes of the system [A,B,C"O]. This is natural for the (A,B) uncontrollable
modes and follows by Corollary (75) below for the (C,A) unobservable modes,
because the latter are in N (P+).
48 Proof of Theorem (38). The proof is done in three steps. In step 1 we establish
the existence of a p.s.d. symmetric solution P+ of the ARE (26) as a limit solution
(42) of the RDE (12). In step 2 we prove that such solution is stabilizing, Le. (39)-
(41) holds. Finally in step 3 we show that a symmetric stabilizing solution of the
ARE is unique.
Step 1. Existence: the ARE (26) has a p.s.d. solution p+=p/ such that (43)
holds.
Observe first that, because of (44) Eq. (43) reduces to
i.e., when the horizon recedes to infinity, the solution at time 0 of the RDE (12) with
P(tl)=O converges to a p.s.d. symmetric solution P+ of the ARE (26). We start to
prove this.
For this purpose we first claim that for any XOE R n
Now, for any fixed control u(·) of class PC on R+, the cost J(II'O,xO'u('» is increasing
as II increases. Hence claim (50) follows by (51) and (52).
Next we claim that
:3K<oo S.t. forany xoERn, forany tiER
336
53
i.e.
Indeed, since the pair (A,B) is stabilizable it follows by Theorem (10.1.53) Ihat there
exists a state-feedback matrix FE Rl1;xn s.1. the linear state-feedback control
55 u(t)=Fx(t)
56 X(I) = (A+BF)x(t)
where by (7.2.17) and (7.2.33) there exists an m<oo and an a>O s.t. for any
XOE R n and "It
Hence, for the specific control (55)-(56), the cost (52) satisfies for any Xo E IRn , for any
tl E R,
with
(For the last inequality of (58) we used Schwarz's inequality and (57); the matrix norm
in (59) is the norm induced by the Euclidean norm on Rn).
Claim (53) follows now by (51) and (58)-(59).
Now by claims (50) and (53) it follows that for every Xo there exists a p(xo) E R+
such that
337
and
Hence by (60) every element of the p.s.d. symmetric matrix P(O,t"O) converges as
tl i.e. there is a real p.s.d. symmetric matrix P+=P+* S.t.
We show now that P+ is a solution of the ARE (26). Indeed by (61), as tl 00,
the RHS of the RDE (12) converges to a constant matrix and hence the LHS, con-
verges to a constant matrix. The only possible limit is zero (indeed if lim P is a
nonzero constant, then, for tl large, P(O,t"O)=P(-tl,O.O) would not converge). Hence
P+ is a solution of the ARE, and step 1 is proved.
Step 2. The matrix P+, as a p.s.d. solution of the ARE, is a stabilizing solution, i.e.
(39)-(41) hold.
We prove (41). Consider any eigenvalue A. of A+ with corresponding nonzero
eigenvector e e c:: n. thus
We are done if Re A. < O. Now since P+ is a solution of the ARE (26) and
A+ := A-BBP+ we have
63 *
O=A+ P++P+A++P+BB P++c
.-II< C. *
Pre- and postmultiply (63) by e* and e, resp.; then use (62):
338
where e*P+e 0, (P+ is p.s.d. by step 1). Two cases occur: either e*P+e>O or
e*P+e=O.
Case 1: *
e P+e > O. By (64) ReI.. S O. We note that Rel..=O implies
B*P+e = Ce = 8. Recalling
39
we obtain
Ae=A+e=l..e with Rel..=O and Ce=8.
6S
Case 2. e*P+e=O, equivalently, ee N(P+). Then by (64) and (39), for the eigenvec-
tor e defined in (62) we have Ae=A+e=l..e with Ce=8. Hence e is an unobserv-
able hidden mode of (C,A) at A. By assumption, (C,A) is detectable (no unstable
unobservable hidden modes); hence Re I.. < O. Therefore
66
By (65) and (66), Re I.. < 0 in all cases, and every eigenvalue of A+ has a strictly
negative real part. So (41) is established.
satisfy
o 0
68 a(A+) c ([_ and a(A+) c ([_ •
339
Define now
71 oP+ := P+- p+
72
Now equation (72) is a linear homogeneous equation in oP+ of the fonn A x = 9 where
i) the vector x E Rn2 is obtained by stacking the columns of oP+ consecutively and ii)
A E Rn'xn" is a matrix representation of the operator
The following corollary and theorem i) reveal the structure of the null space of P+
and ii) indicate an algebraic way to compute P+.
75 Corollary [N (P+)]. Let (A,B) be stabilizable and (C,A) detectable. Consider the
symmetric p.s.d. solution P+ of the ARE (26), (hence P+ is unique and stabilizing).
Let N -(A) denote the stable subspace of A, (8.7.33), and let 0 (C,A) denote the obser-
vability matrix of the pair (C,A), (8.5.2).
V.t.c.
76 N(P+)=N_(A)nN(O(C,A».
Hence
77 P+ > 0 ¢> (C,A) is observable .
78 Comments. ex) Equation (76) means that the null space of P+ is spanned by the
stable unobservable modes of the pair (C,A).
Since (C,A) is assumed to be detectable (no unstable unobservable hidden modes),
(77) is a direct consequence of (76).
y) The RHS of (76) is the largest stable A-invariant subspace in N (C).
340
79 Short proof. Note that xe N_(A) iff lim e Alx=9 and xe N(O(C.A» iff
63
80 *
B P+x=9. Cx=9. P+A+x=P+Ax=9.
'V ie N Aix=A!x.
hence
o
where. by (41), o(A+)c (L. Hence
N(P+)cN(O(C.A» .
00
•
85 Theorem [Properties of the Hamiltonian] [Kuc.l,Lau.l]. Assume (A,B) stabiliz-
able and (C,A) detectable. Define
86
1 E
R2nx2n.
D.l.c.
a) H has no eigenvalues on the imaginary axis,
b) A E o(H) <=> - X" E o(H)
[i.e. the eigenvalues of H have quadrantal symmetry; they are symmetric W.f.t. the feal
axis (because H is real) and symmetric W.f.1. the imaginary axis, (by b»; hence N _(H),
the stable subspace of H, has dimension n, (by a) and b»).
c) Let ZE R2nxn have columns fonning a basis of N -(H). Set
87
[ : 1 E R2nxn
then
-
•
-\
88 P+=XX .
Proof. Let P+ = P +* 0 be the unique stabilizing p.s.d. solution of the (ARE) (26);
as before let A+=A-BB*P+. Consider the change of coordinates Z=y-1ZE )R2n
where
89
.-
[
-P+
I 0 1 E R2nx2n.
A+ -BB*
90 H .- rlHT :;;::
[ 0 -A +*
e R2nX2n.
o
-
Hence a(H)=a(H)=a(A+)ua(-A+). * By Theorem (38) a(A+)c (L; hence
N -(H) is n-dimensional and conclusions a) and b) follow.
Let z::;;:: (xT,,?l e R2n be a vector in N -(H), then, by (89),
z ::;;:: rlz=(xT,(i-P+x)T)T, and ze N_<H). Inspection of (90) gives i-p+x=e and
z=(xT,eT)T where xe N_(A+) = Rn. Now let Ze R 2nxn be any matrix as in (87)
whose columns constitute an n-dimensional basis of N _(H), then
z:= rlz=[XT:(X_p+X)T)T=[XT:OT)T has rank n: thus X is nonsingular and
We conclude that (88) holds for any such Z. •
We are now ready to solve the standard LQoo-problem by the sole knowledge of
where p+=p+* 0 is the (unique) p.s.d. solution of the ARE (26); furthermore,
* .
A+ = A-BB P+ IS s.t. a(A+) c (L.
0
b) The LQoo-problem has the optimal cost Jo(oo,O,xo) given by the quadratic form
and generates the optimal closed-loop system dynamics described by the time-invariant
expo stable linear homogeneous d.e.
94 x(t) = [A-BB*P+]x(t)
95 x(O)=xo,
100 Proof of Theorem (91). Note that, since (A,B) is stabilizable and (C,A) detect-
able, by Theorem (38) the p.s.d. symmetric solution of the ARE (26) is unique and
stabilizing. Let J(oo,O,xo,u('» be the cost of LQoo, (33), generated by any u(·) of class
PC Then by (25), for any t) E
2J(oo,O,xo,u('» 2J(t),O,xo,u('»
II
= J[lIu(OIl2 + IICx(OIl2] dt
°
where xO satisfies (31)-(32). Hence, by Theorem (10), using the optimal cost (14)
with S=O on [O,t)],
2J(oo,O,Xo,u('» *
2Jo(t),O,xo)=xo P(O,tl'O)XO .
Therefore, by the convergence (43) of P(O,tl'O) as tl 00, for any u(·) of class PC on
R+
101
where P+ is the unique symmetric p.s.d. stabilizing solution of the ARE (26). Thus the
RHS of (101) is a lower bound for any cost of problem LQoo. It remains to be shown
that, for u(·) given by the state-feedback law (92), we have
2J(oo,O,xo,u('»=xo P+xo. *
Now with the linear feedback control given by (92), x=Ax+Bu becomes x=A+x,
(where A+=A-BB*P+), which by (41) is expo stable; whence, for all Xo, .
= J(x(t),[P+BB*P++ C'C]x(t» dt
o
344
I,
J
=- (x(t),[A+*P++P-rA+]x(t»dt
o
I,
=- fo ddt (x(t),P+x(t)} dt
Le. the lower bound in (101) is attained for the constant state-feedback law (92). •
Denormalization
Let us start by emphasizing a well-known fact.
104 Fact. Let Re R nxn satisfy R=R*>O. Then from the expansion
n
R=I; "'jejej* (where the eigenvalues Aj>O, Vi and the eigenvectors ej are chosen
j=l
to form an orthonormal basis), we obtain
00
33N J= f [IIUN(t)112+IICx(t)ll2]dt
o
26N
92N
345
94N
101
Introducing these changes in the ARE (26N), the optimal control (92N) and the
closed-loop d.e. (94N), we obtain
26D O=A*P+PA-PBR-1BP+C"C
92D
94D •
x(t) = [ A - BR- 1B*P+] x(t) .
These three equations are the basic equations of the denonnalized fonnulation.
Equations (26D), (92D) and (94D) show that once Q and R are chosen and the
assumptions of Theorem (38) hold, the closed-loop d.e. is completely detennined.
An elementary way of choosing Q and R is the following: 1) choose Q and
R > 0 diagonal; 2) since the state variables represent physical variables (positions, velo-
cities, currents, voltages, ... ), large diagonal values are assigned to those state-variables
whose deviation from zero (Le. nominal) are undesirable and smaller ones are assigned
to the other state-variables; 3) the inputs also represent physical quantities (voltages.
flows, aileron angles, ... ) so large diagonal values are assigned to those inputs that are
costly and smaller ones to the other ones. The lesson of these considerations is that
the LQoo-solution is essentially a tool, once Q and R are chosen, the optimal closed-
loop dynamics is fixed but the LQoo-theory does not provide the designer with a
methodology for choosing Q and R. That choice is a matter of art and experience.
For an interesting contribution relating the LQ-design and closed-loop perfonnance see
[Fuj.l].
346
Final comment. Every aspect of the LQoo-problem has not been studied in this sec-
tion. For more infonnation see, e.g. [Wil.lJ, [Mol.lJ, [Cop. I], [Rod.1J, [Kuc.IJ.
[Lau.1J. [Cal.2J.
where
6 Problem. Minimize the cost J(k1.ko.xo") over all possible input sequences
uO : [ko.k1-1]-+lRn, . •
A quick review of the proofs of Theorems (2.1.167) and (2d.1.180) shows that in
Chapter 2d we established the following result
U.t.c. a) on any integer time-interval [ko.kl] the LQ-problem is solved by the linear
state1eedback law (independent of ko)
13
and generates the optimal closed-loop system dynamics described by the linear homo-
geneous r.e.
16 x(I<o)=xo
Thus w.l.g. we set 1<0=0. Moreover. we would like to know what happens as the hor-
izon kl for the case S=O (i.e. with zero final state penalty term in (4».
30 Standard LQoo-problem.
Data. We are given
a) a real time-invariant discrete-time system representation R d = [A,B,I,O], with
det A ':/: 0, described by
348
where
34 CE RfloXJl ,
11'11 denotes the Euclidean norm and x(') denotes the solution of (31) to the control
u(') and the initial state xo.
35 Problem. Minimize the cost J(oo,O,xo,) over all possible input sequences
u('):N -+Rnl. •
The solution of the LQoo-problem, (30)-(35), involves the study of the associated
discrete-time algebraic Riccati equation (abbreviated ARE), which is obtained from
the RRE (12) by setting P(k-l)=P(k)=P, i.e.
26 P=c*C+A*P[I+BB*pr'A.
This nonlinear equation is to be solved for p=p* Note that p=p* implies
that det(I+BB*P)=det(I+B*PB) '" O.
36 Exercise. [Equivalent forms of the ARE]. Prove that equations (26a), (26b). and
(26c) below are equivalent to (26).
26a P=CC+A*[I+PBB*r'PA,
and
26c P=CC+A* [I+PBB* r'(p+ PBB*P)[ 1+ BB*pr'A .
38 Theorem [The ARE has a p.s.d. solution that is unique and stabilizing). Con-
sider the standard LQoo-problem, (30)-(35), Assume that the pair (A,B) is stabilizable
and that the pair (C,A) is detectable.
349
V.t.c.
i) the ARE (26) has a positive semi-definite solution P+,
ii) this solution is unique and stabilizing; more precisely it is unique and the state-
feedback control
u(k)=-B*P+[I+BB*P+]-IAx(k) ke N
39 A+=[I+BB*p+r1A
equivalently
41
•
42 Comments. a) It will tum out that the matrix P+ is a limit solution of the RRE
(12), more precisely, for any fixed kE Z
(the equalities on the RHS follow by the shift-invariance of the solution of the RRE
(12), see (22». Hence the unique p.s.d. stabilizing solution P+ of the ARE can be
obtained in the limit by letting the RRE (12) run backwards from P(O)=O.
The matrix A+, given by (39), is obtained by substituting the constant linear state-
feedback law u(k)=-B*P+[I+BB*p+r1Ax(k) in x(k+l)= Ax(k)+ Bu(k). The latter
feedback will be shown to be optimal for problem LQoo in Theorem (91) below.
Therefore the expo stability, (40), of x(k+l)=A+x(k) implies that the optimal closed-
loop dynamics will always be expo stable.
48 Proof of Theorem (38). The proof is done in three steps as in the continuous-
time case. Details of the proof are presented only when the algebraic manipulations
are specific to the discrete-time case.
Step 1. The ARE (26) has a p.s.d. solution P+=P+* More precisely, as the
horizon kl recedes to 00, P(O,kl'O) i.e. the solution at 0 of the RRE (12) with P(k 1) = 0
converges to a p.s.d. symmetric solution P+ of the ARE (26). In other words, there
exists a solution P+= P+* 0 s.t.
350
A reasoning similar to the one used in the proof of Theorem (10.4.38) shows that for
every XOE R n the optimal finite horizon cost (4) on [O,k!] with S=O satisfies (by (14»
*
2Jo(k 1,0,xo)=xo P(0,k1,0)xo '
is increasing for kl increasing and is bounded for kl E N (because the pair (A,B) is
stabilizable). Hence for every Xo E R n there exists a nonnegative number p(xo) s.t.
(where the last equality follows by the shift-invariance of the solution of the RRE, see
(22». Finally P+ in (61) is a solution of the ARE: this follows from (61) by letting k
tend to - 00 in the RRE (12) subject to P(O) = O.
Step 2. The matrix P+, as a p.s.d. symmetric solution of the ARE (26) is a stabiliz-
ing solution, i.e. with A+ given by (39), conclusion (40), equivalent to (41), holds.
We prove (41). Consider any eigenvalue A of A+ and a corresponding (nonzero)
eigenvector e E (Cn; thus
We are done if I A I < 1 (which is equivalent to (41». Now since P+ is a p.s.d. solution
of the ARE (26), by (39) and (26c) we obtain
63 AI<
P+=c C+A+* [P++P+BB*P+]A+ .
*
Case 1. e P+e > O. By (64) I A I 1. We claim that I AI = 1 leads to a contradic-
tion. Indeed by (64) 11..1 = 1 implies B*P+e=Ce=O. Since A+=[I+BB*p+r!A we
351
have
64a
65
Case 2. e*P+e=O, equivalently ee N(P+). Then by (64) and (39) (:::> (64a», for the
eigenvector defined in (62) we have Ae=Ae with Ce=e. Hence e is an unobservable
hidden mode of the pair (C,A) at A. Since the pair (C,A) is detectable by assumption,
it follows that I AI < 1. Therefore
66
Thus, by (65) and (66), I AI < 1 in all cases, and every eigenvalue of A+ has magni-
tude strictly less than one. So (41) holds.
satisfy
69 P+=C"C+A*[I+p+BB*r1p+A
Define now
and note that we are done if we show that BP+=O. Now by subtracting (70) from
352
72
Now Eq. (72) is a linear homogeneous equation in SP+ of the fonn A x = e where i)
the vector x e IRnz is obtained by stacking the columns of SP+ consecutively and ii)
A e RnZxnl is a matrix representation of the operator
Xe R nxn -+ X-A+*XA+.
86
A-lBB* 1 E JR2nx2n •
A* +C'CKIBB*
V.t.e.
a) Hb has no eigenvalues on the unit circle,
b) AE a(Hb) ==- X"-l E a(Hb)
[i.e. the eigenvalues of Hb have double symmetry, viz. 1) w.r.t. to the real axis
(because H is real), and 2) w.r.t. to unit circle, (by b»; hence N +(H b), the unstable
353
then
88
•
Proof. Exercise. Hints: copy the proof of Theorem (10.4.85). With P+ the p.s.d. sta-
bilizing solution fo the ARE (26) we have with
89 ,1_ [ I
0
1 e
R2nx2n
-p+
A;I I
K ••' ] .
90 Hb-,IHbT- [
0 A*
+
We are now ready to solve the standard LQoo-problem by the sole knowledge of p+.
where p+=p+* is the (unique) p.s.d. solution of the ARE (26); furthennore
A+=[I+BB*p+rIA is S.t. cr(A+)cD(O,l).
b) The LQoo-problem has the optimal cost Jo(oo,O,xo) given by the quadratic fonn
354
and generates the optimal closed-loop system dynamics described by the time-invariant
expo stable linear homogeneous r.e.
94 x(k+ I) = [I + BB*P+rl Ax(k) keN,
95 x(O)=xo
reduces the control law (92) and the closed-loop r.e. (95) to
97 u(k)=-B*
and ke N
98 x(k+ 1) = A+x(k) .
Note that compared to the continuous-time case formula (97) includes the additional
factor A+.
100 Proof of Theorem (91). Note that, since (A,B) is stabilizable and (C,A) detect-
able, by Theorem (38) the ARE (26) has a unique stabilizing p.s.d. solution P+.
Let J(oo,O,xo,u(·» be the cost of LQoo generated by any input sequence u(·) (see (33».
Then, on replacing the infinite horizon by a finite horizon kl e N and on using the
optimal cost on [O,ktl dictated by Theorem (10) with S=(),
Note that P(O,kl,O) converges for kl to P+ by (43). Hence, for any input
sequence u(·) and for any Xo
355
101
Thus XO*P+xo is a lower bound for any cost of problem LQoo. Hence we are done if
we show that, for u(') given by the state-feedback law (92), (Le. (97) by (39))
2J(oo,0,xo,u('))=xo P+xo . *
Observe for this purpose that for such u(·), x(k+ 1) = Ax(k) + Bu(k) has the form
(98), which by (40) is expo stable. Thus for all Xo
Recall that, by (26c), the p.s.d. matrix P+ satisfies the equivalent ARE
63 *
...11< C+A+ (P++P+BB P+)A+.
P+=c *
Hence using (92) o¢> (97), (63) and (98) we obtain successively
2J(k 1,O,xo,u('»
k,-l
= 1: [lIu(k)11 2 + IICx(k) 112]
k=O
= 1: * *
[x(k) P+x(k)-x(k+l) P+x(k+l)]
k=0
2J(oo,O,xo,u('))=xo P+xo, *
i.e. the lower bound in (101) is attained for the constant state-feedback law (92). •
CHAPTER 11
Introduction
This chapter covers a number of the main techniques and results in MIMO linear
time-invariant feedback systems. There are three main reasons for this choice of sub-
ject: first, MIMO feedback systems are ubiquitous in modem industry (autopilots, con-
trol of auto and airplane engines, automated manufacturing systems, process control,
... ); second, the statement of and derivation of these main results constitute an excel-
lent demonstration of the power of the concepts and the techniques developed in the
previous chapters; third, a number of these results are basic to computer-aided design
procedures. In fact a good number of these concepts and techniques were invented to
understand and solve feedback problems.
For simplicity we restrict ourselves to the unity feedback configuration: that is the
given dynamical system (the plant) is driven by the output of the compensator; the
plant output Y2 is fedback and compared to the input ul to obtain the error el = ul-Y2;
hence the tenn unity feedback. For more complicated configurations see [Des.6, Net.1,
Vid.l, Des.7, Gun.l].
In section 1 we will develop the state-space representation of the MIMO unity-
feedback system (11.1.14, and 11.1.15), calculate its characteristic polynomial
(11.1.26), and establish that, in the absence of unstable hidden modes, the expo stability
of the system can be guaranteed by testing for stability four transfer functions
(11.1.40). Important special cases are treated in detail; note, in particular, the case
where R 2 is expo stable (11.1.43) and the Q-parametrization of all I/O maps HY1U1
(11.1.48).
The main result of Section 2 is the Nyquist theorem for MIMO feedback systems
(11.1.25). Its importance lies in that 1) it relates the characteristic polynomial to the
return difference, 2) it is the basis of many arguments used in robustness theory and of
many computer-aided design algorithms. Important aspects of the Nyquist theorem are
discussed in subsection 2.2.
Robustness is the subject of Section 3, the main properties of a well-designed
MIMO feedback system are derived. Provided the loop-gain PC is large (in all direc-
tions) we show that the feedback systems I/O map is relatively immune to changes in
the plant (11.3.9) and to exogeneous disturbances (11.3.21). We establish the key
features of set-point regulators and demonstrate their robustness (11.3.27). The final
theorem (11.3.44) establishes the trade-off between the achievable bandwidth of the
feedback system and the plant uncertainty.
Section 4 uses the Nyquist theorem to obtain a simple proof of the celebrated
theorem of Kharitonoy (11.4.9), it uses a well-known characterization of Hurwitz poly-
nomials (11.4.6).
From an engineering point of view, structured perturbations are the most realistic
ones; in Section 5 we obtain necessary and sufficient conditions for stability of a class
of structurally perturbed systems (11.5.23) and develop, for a speeial class, a procedure
for testing the stability of each member of the class; this procedure is well suited for
computer-aided design (11.5.31).
In Section 6, we derive necessary and sufficient conditions under which a unity
feedback system remains stable under arbitrary additive plant perturbations (11.6.6),
the additive perturbations are only required to be proper, i.e. unstable proper additive
perturbations of the plant are covered by the theorem.
It is possible to extend the results of Sees. 1,2,3 and 6 to more general classes of
linear time-invariant systems, e.g. systems involving delays. This requires the use of
advanced algebraic techniques (8-algebra, see, e.g. [Cal.3-6], [Vid.l], [Che.2]). It is
important to note that the theorems remain valid in this more general context.
In Section 7, we define the concept of zeros of transmission, we characterize
them and we illustrate their importance in feedback design.
3
u:= [ul,uI r and for its output, y:= r.
We choose for the state of L,the vector x:= [xl,xI r; for its input,
From an engineering point of view, if assumption (4) does not hold, there is a funda-
mental defect in the modeling of L.
10 Exercise. Show by example that the converse of fact (9) is not true.
359
14
15a
-BI(I+D2DlrID2]
ISh
B2(I+D 1D 2 rl
-Dl(I+D2Dl)-IC2]
15c (I+D 2D 1)-IC 2
15d
360
r; r;
Notation: it will be convenient to define
r; r.
and
for the state. the output, the error and the exogeneous input of Lc .
MN(I+MNrl = I-(I+MN)-l.
Proof of Theorem (13). Consider the closed-loop system 1:c shown in Fig. 11.1. the
subsystems 1:1 and 1:2 are described by: for i=1.2
17a
Using this expression. the notation defined in (15e) and (15f). and exercise (16) in Eq.
(18) we immediately obtain
361
20 y=Cx+Du,
where C and D are defined by (15c) and (15d) of Theorem (13); note that C and D are
simply the RHS of (19) multiplied on the right by and diag[D I,D2],
respectively.
Now use (20) and (17b) in the differential equations in (l7a) to obtain, after some
calculations,
218 xI = [AI-BI(I+D2DI)-ID2CI ] Xl -
238
23b = [::] .
25 Theorem (Characteristic polynomial of L) [Hsu.1]. Consider the feedback
configuration L of Fig. 11.1 where det [1+D 1D2] ":F- 0, and the R {s are specified by
(1) and (2). The characteristic polynomial of Rc= [A,B,C,D], namely,
XA(S) =det(sl-A), with A defined in (14) and (15a), is given by
det[I-tG I(s)G 2(s)]
26 det(sl-A) =det(sl-A I)' det(sl-A2) • A A
det[l+G 1(00)02(00)]
28 Comment. In the SISO case, we know that the "loop gain" gl(S)g2(S) and the
return difference 1 + g I(S)g2(s) play a crucial role in the stability analysis of Le'
362
equivalently, if det[I-tG 1(s)G 2(s)] has one or more zeros in <T+, then R c is not expo
stable.
30 Remark. To gain insight, rewrite Eq. (26) as follows: call Aci the closed-loop
eigenvalues, (i.e. the eigenvalues of A), and Aki the eigenvalues of R k; then
31
<\
Equation (31) shows that any zero of det[I + (S)02(S)] is a zero of
XA(S) =det(sI-A) = n(s-Aci) (as we have already seen in Corollary (29». It also
shows that some closed-loop eigenvalues, say AC I' may not be a zero of
det[I+G 1(s)G 2(s)]; indeed the factor (S-Ac1) may be canceled by one of the denomina-
tor polynomials. This exhibits the fallacy that lies in trying to detennine the stability
of Lcby examining only the rational function det[I+G 1(S)G 2(s)]. From Eq. (31). we
obtain the following corollary.
•
33 Z [det[I+G 1(s)02(S)]] c
Proof of Eq. (26). To simplify notations we use the x. y. e and u (defined in (l5e)
and (15£). Let
363
with Bo, Co and Do defined similarly. Finally, introduce the (square) symplectic
matrix
17b e=u-Jy.
Note that A o, Bo, Co and Do are block diagonal matrices. From the last two equations
we obtain
x= [A o-Bo1(I+Do1)-ICo] x.
Hence
sI-A = sI-Ao+ Bo1(I+Do1r 1Co
Now use det(PQ) = det P det Q and det(I+MN) = det (I+NM) to write
Referring to the definitions of Ao, Bo, Co and Do, the last determinant is easily seen to
be
Hence, finally
364
=
A A
37 I\u(s) 1- J f\.u(s)
that we say that the transfer function H(s) E lR(sr xn is expo stable if and
A
Explicit expressions for li,u have been obtained in Exercise (6). It is important to keep
in mind that Fact (39) holds for the unity-feedback configuration L; for more compli-
cated configurations, equivalence (39) need not be true! ([Des.6]; see also Fig. 11.5 in
Section 6 below.)
365
P) Assumption (36) allows us to infer the stability of R c from that of the closed-loop
transfer functions either I\u or ltu. Obviously the presence of unstable hidden modes
in either R 1 and/or in R 2 would destroy that inference. Theorem 40 has a very useful
special case:
44 R c is expo stable
Proof. (=» By remark (41), R c expo stable => I-\-u is expo stable => f{,u is expo
stable => f{,lU, is expo stable.
A A
hence
'"
(I +
" " '"
=
1
is Ii"u
expo stable. From this we obtain:
(I + = -I-\"u2 is expo stable. Thus we have established that all four sub-
A
of I-\,u are expo stable once He2u , is expo stable. Hence by Theorem (40)
R c is expo stable. •
I-\"Ul by Q.
A A
49
A A
He2u, = Q= 0 1(I +
A
°°
A
2
A
1)
-1
.
A A A
51 0 1=QJ-02Q .
52 Theorem (Design for G 2 expo stable and strictly proper). Consider the
configuration L defined by (1) and (2). Assume that both R I and R 2 have no
unstable hidden modes. Suppose that O2 is expo s!able and strictly proper. Under
these conditions, a) for all proper transfer functions 01> which result in an expo stable
R C' the I/O map is given by
(51) is proper. achieves the I/O map (48) and the resulting feedback configuration Rc
is expo stable.
53 Comments. a) Theorem (52) can be summarized as "for any given expo stable
and strictly proper plant O2 • the set of all I/O maps achievable by an expo stable R c is
parametrized by the expo stable Q's (see Eq. (48». The required controllers 0 1 are
given by (51) and are proper." The use of (48) and (51) in design is usually referred
to as the Q-parametrization method [Zam.1.Des.5.Mor.1].
Note that the 01 given by (51) is proper; there is no guarantee. however. that it
will be expo stable; of course :E, is expo stable.
'I) Consider Eq. (48): 62 is given. Q is free subject to being expo stable. Hence any
O H
tr+-zero of 2 will be a tr+-zero of y2u (because, roughly speaking. Q may not
Gv.
AI
have tr+-poles to cancel the tr+-zeros of Thus. some dynamical characteristics
of the "plant" 62 necessarily impose limitations on the achievable JlO maps of :E,.
A A
8) Since Q is proper, the JlO map Hy2uI has a behavior as Is I -+ 00, which is con-
strained by that of 62, (see (48».
x=Alx+Blu sx(s)=Atx(s)+Blu(s)+x(D-)
55 { {
y=Ctx+Dtu yes) = Ctx(s)+D 1ii (s)
368
The Eqs. (55) and (56) are identical in fonn except for the z factor multiplying
x(O) in (56).
Consequently, the results of Theorems (13), (25), (40), (43) and (52) apply
o
without change except that the open-left half-s-plane CL is replaced by the open-unit
disc D(O,I).
For a very detailed and elementary treatment of the discrete-time case see Ogata's
book. and the references therein. [Oga.l]
Gk(s)= Ck(sI-AkrIBk+Dk
Note that this power series converges absolutely for I s I > p(A k), the spectral radius of
Ak. We assume that L is well posed. Le. that
In the complex s-plane. (s = o+jro). consider the curve D defined as follows: let R be
369
6 R c is expo stable
•
and arg f(s) is well defined for all SED
'"
.-pl.n. "
• 11
D'
..
1. If R 1 and R 2 are expo stable, (equivalently, P<H- = 0), then, by Theorem (5), we
need only to check that f(D) does not encircle the origin. This special case is very
important; because (i) the values of <\ (jro) and for ro e 1R+ may be measured
easily and accurately (by a sinusoidal steady-state experiment) and hence modeling
errors avoided; (ii) more generally, the theorem sti11 applies to systems where <\(s)
and/or represent expo stable distributed systems (e.g. delays, distributed dynamic
elements, etc ... ): in that case <\ (s) and/or are transcendental functions, but they
are analytic in tr +.
Hence for R sufficiently large, the image of the right half-circle of D is an arc of a
small circle centered on (doo'O) and of radius O(1IR), (see (9) above). Since d oo i= 0,
(see (3) and (10», the number of encirclements of the origin is determined by the
image under f of the jro-axis segment [-jR,+jRJ, (duly indented if required), where R is
taken sufficiently large.
3. Left indentations. In the SISO case, if jrol is a pole of l+gl(S)g2(S), then jrol is a
zero of (l+gl(S)g2(S»-I; in particular, (l+gt(S)g2(S»-1 cannot have a pole at jrol'
Thus in the SISO case, (assuming no tr+ pole-zero cancellation in the product
g 1(s)g 2(S», if So e tr+ is a pole of 1+g 1(s)g2(S), then the closed-loop transfer func-
tions (l+gl(S)g2(S»-I, gl(S)(l+gt(S)g2(S)r l , g2(S)(l+gl(S)g2(S»-1 and
gl(S)g2(S)(1+g 1(s)g2(S»-1 do not have a pole at so' i.e. these four transfer functions
are analytic at so. Consequently, So need not be encircled by the contour D, (in order
to guarantee the analyticity of the closed-loop transfer functions at so), and D may be
indented either to the left or to the right.
In the MlMO case, if So is a pole of f(s)=det[I-tG 1(s)G 2(s)], it may be a pole of
[I-t0 1(s)G 2(SW I . Consider the following example
371
s+1
A A [ ..
s+12 S2 s+2
-s(s+2)
--
1+0 1(s)02(s) = 0 s+1 , [It<'i.(,lG,('Jr' [ s
0
s s+1
Here f(s), [I+<3 I (s)02(S)] and [I+<3 I (s)02(S)r1 have a pole at s=O, a point on the jro-
axis. It is for this reason that in the MIMO case, jco-aKis poles of f(s) must be
included inside the contour D , i.e. the indentation must be taken on the left.
A A
the Rk's are in (Cnixn", where R3 is not the zero matrix. That pole PI' as a pole
of 01(S), has a McMillan degree (9.1.90) given by
RI Rz R3
rk [ R2 R3 0
R3 0 0
Note that in the Nyquist criterion (11.1.5), we are concerned with Po+, defined in (4),
the number of (C+-zeros of det(s!-A I ) det(sI-Az}. In the context of present dis-
cussion, Po+ is the sum of the McMillan degrees of all the 4r+-poies of O\(s) and that
of all the (C+-poles of 02(S).
7. Suppose we have the list of all jco-axis zeros of det(s!-A 1) and of det(s!-A 2); it is
easy to see that no harm is done if left-indentations are used for each such zero, i.e.
there is no need to check whether at each such point the matrix 01(S)02(S) has a pole.
372
where the z;'s and the Pk's are the zeros and the poles of f with Zj Pk, * V i,k.
Assume that f(s) has neither poles nor zeros on D l' Consider s traversing clockwise
the closed curve D \. Since, by (13),
referring to Fig. 11.3, we see that if zl is outside the curve D l' the net change in
arg (s-z\) as s traverses D \ once is zero; on the other hand, if Zz is inside D l' then the
net change in arg (s-Zz) is 21t radians in the clockwise sense, that is -21t radians in the
counterclockwise sense. Similarly, poles outside D \ contribute nothing to the net
change of phase, but if P2 is inside D \ the net change in arg (s-P2) is 21t radians in the
counterclockwise sense. Hence by (14) the net change in phase in the counterclock-
wise sense is (p+-n+) 21t = 21t x (the number of poles inside D 1 - number of zeros
inside D 1)' counting multiplicities. Thus we have:
The Argument principle. Let f(s) be a proper rational function with f(s) finite and
non-zero for all SED \; as s traverses D \ clockwise, f(s) traverses feD \) and encircles
the origin p+-n+ times counterclockwise.
Now XA' XA, and XA, are polynomials, whereas f(s) := det[I+O I (S)02(S)] is a proper
rational function, hence every pole of f(s) is completely canceled by some zeros of XA,
and/or of XA" i.e. if f(s) has a pole of order 5 at PI then the product XA, XA, has a zero
at PI of order at least 5.
With this in mind together with Eq. (11.1.26), we can write
o
16 R c is expo stable ¢:> Z[XA(s)] = Z[det(sI-A)] c (L
21 Remark. From the proof and the use of the argument principle, it is clear that
the shape of D can be changed, the encirclements simply che:ck that the characteristic
polynominal of R C' namely, XA(s), has no zeros inside or on D. For design purposes,
we might want all the zeros of XA(S) to lie to the left of the vertical line Re(s)=-a,
for some given (;(>0, then we simply shift the left boundary of D to Re(s)=-a.
Similarly if we want the zeros of XA(s) to lie to the left of some simple curve C in
o
([ _, (the simple curve C is assumed to partition the ([ -plane in two parts; one to the
left of C and one to the right of C), we choose this curve C as the left boundary of
the closed curve D; as before D is closed by an arc of a circle of sufficiently large
radius R centered at the origin. Because of these possibilities, the Nyquist criterion is
a very flexible tool in the hands of imaginative engineers.
374
22 Exercise. Give a heuristic explanation for the following recipe for counting the
counterclockwise encirclements of the origin by the closed oriented curve f(D). By
assumption f(D) does not go through the origin. i) Draw a half line L from the origin
to infinity (in any direction); ii) choose L so that it intersects the oriented curve f(D)
at a number of points say PI> ...• Pn • furthermore L is chosen so that P I.P2 ..... Pn are
simple points of the curve f(D). i.e. points where f(D) does not cross itself; iii) for k ==
1..... n. if. at Pk • f(D) intersects L in a counterclockwise. (clockwise). fashion. the count
is 1. (-1. resp); iv) add all the counts for PI to Pn : this sum is the number of counter-
clockwise encirclements of the origin by f(D). (For a rigorous proof see Vid.2].)
11.3. Robustness
The main purpose of feedback is to obtain a closed-loop system that is insensitive
to plant perturbations and that attenuates effectively external disturbances. These
insensitivity properties are rather obvious in the SISO case by simple Nyquist diagram
considerations, [Hor.l]. The same holds for the trade-off between uncertainty and
achievable performance. In the MIMO case, the correct formulation of the correspond-
ing properties took some time to be discovered but. fortunately for us. they are now
easily derived if we use some of the basic tools developed previously.
For simplicity. we consider the simple configuration S(P.C) shown in Fig. 11.4
below: S(P.C) is the same configuration as L of Fig. 11.1 except that we use now the
control terminology: P designates the "plant," i.e. the dynamical system to be con-
trolled and C designates the "controller." (For a study of more involved
configurations, see [Des.6,7], [NeLl], [Vid.I,Gun.I].)
91
(the superscript "0" is used to remind us that it refers to the equivalent open-loop sys-
tem). Now consider the I/O map variation due to the change from P to
is given by
m yzu, = (l+pC)-lpC-PC(I+PC)-1
= (I+pC)-1 [PC(I+PCHI+PC)PC](I+PC)-I
5 = (I+pcrlM> C(I+pC)-I.
Hence
i.e. the I/O map variation of the closed-loop system is the I/O map variation of the
equivalent open-loop system premultiplied by (1+ P C)-I.
376
Now for fixed w, using the nonn induced by the Euclidean vector nonn,
7 I1(I+PC(jW»-11 I= cr[(l+PC(jW»-I]
= 1 / Q[I+PC(jW)]
where a and Q denote the maximum and the minimum singular values of the matrix in
the brackets (A,7.81). Now since
8 :::>
10 Comment. Where do the M come from? First from modeling, models are
simplified representations of the physical reality: a) in electrical circuits one neglects
some propagation effects (lumped circuits), some small inductances, small capacitances
and small resistances; b) in mechanical systems, one assumes that a bar is infinitely
rigid or that a beam does not bend, etc .... ; c) the mass and the moments of inertia of
a crane change when it picks up a load, same for a communication satellite when it
consumes fuel, or when it reorients its solar panels, etc.... Second, in mass produc-
tion, M comes from manufacturing deviations. Third, in operations, the equipment is
exposed to temperature changes that may significantly change their physical properties,
etc .....
18 Hy:At=(I+PC)-I.
This relation is very important for it shows the constraint imposed by the configuration
377
S(P,C):
In other words, if, over the frequency band n, the system S(P,C) achieves excellent
disturbance rejection then the I/O map H y1U, is close to I and there is a considerable
decoupling between, say the first component of U1 and the second, third, ... com-
ponents of Y2; and conversely.
From a design point of view, we see that if V coe n
23 Remark. Suppose that for some (bad) design, for some COo e n,
.Q(PC(jcoo» =: 0'110 < < 1; then, using the singular value decomposition of PC(j roo) ,
namely,
PC(jcoo) - O'nollnovno* .
then, for this perturbed system, any disturbance do in the direction of Vno is not
attenuated by the feedback! In other words, at that frequency, for that perturbation,
the feedback is useless!
26 y(t) -+ Uo as t -+ 00 •
Physically, this means that after some transients, (whose duration depends on the
spectrum of the closed-loop system), the output Y2(t) becomes a constant precisely
equal to the input. The theorem below shows that such regulators have remarkable
robustness properties.
378
27 Theorem. Consider the system S(P,C) where assumptions (1)-(2) hold. Assume
28 C(s)=Co(s)/s
29 det[P(O)Co(O)] 'F 0
then, for all initial conditions, for all inputs of the form uol(t), the output yet) Uo as
t 00.
Furthermore, if Co and P are penurbed to Co and P subject to the only conditions that
S(P,Co(s)/s) is expo stable and (29) holds for the penurbed system, then the asymptotic
regulator propeny still holds.
32 Proof. By linearity, the output Y2(t) consists of two terms, one due to the initial
conditions and one due to the input uol(t). The first term goes to zero in view of (30).
Use Laplace transforms to compute the second term
The partial fraction expansion of this expression involves only poles in the open left-
half-plane - since Hy2u1 is expo stable - and one pole at s=O. The residue at s=O is
given by
where we used assumption (29). Hence, under the conditions stated above, we have
Y2(t) Uo as t 00.
The second statement is proved similarly.
•
33 Exercise. For the system described by Theorem (27), calculate Hcju1 ' Evaluate
it at s=O and derive from it the tracking propeny (26).
379
where
40 Preliminary Calculations.
Consider any P and C satisfying (1):
= [1+(P-Po)C(I+PoC)-l]. (I+PoC).
As in (11.1.49), we set
42 Q := C(I+poCr 1
so
and
44 Theorem (Stability Robustness of A(Po;w». Let Po satisfy (35) and let the
compensator C be S.t. S(Po,C) is expo stable; then
This theorem was first stated by Doyle and Stein, [Doy.l], then proved in a more
general context in [Che.2]. The proof below uses a technique in [Vid.lj.
Proof.
'*'. by assumption, condition (46) holds.
For simplicity assume that C(s) is analytic on the jco-axis.
48 det(sI-A) = det(sI-Ap>det(sI-Ac)det[I+Po(s)C(s)] .
Let 'Yp and 'Yc denote the number of (i:+-eigenvalues of Ap and Ac. resp. (counting
multiplicities). Since. by assumption. S(Po.C) is expo stable. (48) shows that
det[I+Po(s)C(s)] has 'Yp+'Yc (i:+-poles and no (i:+-zeros; hence the Nyquist plot of
co -+ det[l+PoCUco)] encircles the origin 'Yp+'Yc times counterclockwise (Theorem
11.2.5).
Step 2. Since '1/ P e A(Po;w). P has 'Yp (t +-poles. it remains to show that for all
such P·s. the Nyquist plot co -+ det[I+PCUco)] encircles the origin 'Yp+'Yc times counter-
clockwise.
Now by (36). (38), and assumption (46) we have
50 det[IH(P-Po)Q(jro)] '#: O.
Call f(tjro,P) the LHS of (50): given the assumptions on P, (35-37), and those on C,
'1/ P e A(Po;w) the function
(t,ro) -+ f(tjco,P)
Now, the Nyquist plot of f(I,jro,P) does not encircle the origin; also, by step I, the
Nyquist plot of det[I+PoC] encircles the origin 'Yp+'Yc times; hence, by (51), that of
det[I+PC] encircles the origin 'Yp+'Yc times. Finally by Theorem (11.2.5),
'1/ PE A(Po;w), S(P,C) is expo stable.
=>. We use contraposition. So we assume (45) and we negate (46); i.e. we assume
(45) and that for some roo
If roo=O, (coo = 00, resp.), by continuity of Q(.) and w(') we replace roo in (52) by
some roo>O arbitrarily close to 0 (arbitrarily close to 00, resp.). Hence w.l.o.g. we
382
53 QUroo)=ULV*=L OjUjvt,
hence U and V are constant unitary matrices, the only nonzero elements of L are on
its diagonal, and, as usual, 01 02 ,.... For later use, note that (52) and (53) imply
that
We are going to use (54) to construct a perturbed P specified by P=Po+L'1 such that
P E A(Po;w) and such that
In view of (43), (55) implies that, for that P, S(P,C) is not expo stable: indeed the
closed loop characteristic polynomial has a zero at s= jroo.
58
59 (U(S»)-=UI' -
1 1
3 We say that the polynomial p(s.a) is Hurwitz iff p(s.a) ¢ 0 V s e CI:+. (i.e.
V s in Res
4 We say that the set A is Hurwitz iff p(s.A) ¢ O. V s e CI: +. equivalently. iff
every polynomial in A is Hurwitz.
384
6 Lemma. a) If the real monic polynomial p is Hurwitz, then all its coefficients
are positive and arg(p(jro)) is a strictly increasing function of roo
b) The real polynomial p of degree n is Hurwitz
where the -(X/s are the real zeros of p. (hence (Xi> 0). and the (-13k ±jYk)'s are the
complex conjugate pairs of zeros of p. hence 13k> O. Equation (8) shows that all the
coefficients of p are positive. The fact that arg pUro) is strictly increasing is immediate
by calculation using (8). or geometrically obvious by drawing a diagram showing the
o
constellation of zeros in <r_.
b) =:> i) is immediate since there are no zeros on the jro-axis. ii) As ro increases from
o to 00, argUro'Hx j ) increases by and that of [U<o+l3k)2+Y(1 increases by It. Since
o
there are n zeros in <r _. the net increase is nlt/2.
¢ (Sketch of proof). Suppose that p had one real zero zl with zl > 0; then
argUo>-zl) would decrease by as ro increases from 0 to 00. Hence, ii) would be
violated since p has degree n, and the net increase in argument of the LHS of ii)
would be strictly less than nlt/2. •
In assumption (2) above. we required that "II i, O<l!j: in view of lemma (6)
and since each polynomial p(s,a), aE A, is monic, the requirement O<.!!,j entails no loss
of generality.
9 Theorem. 'The set A of monic polynomials (defined in (1) and (2» is Hurwitz if
and only if the following four polynomials are Hurwitz:
10 kll(S)=l!O+.l!ls+1lzs2+a3s3+ ...
385
14 Comment. Kharitonov discovered this amazing fact that the Hurwitzness of four
polynomials are necessary and sufficient to guarantee the Hurwitzness of the whole
class A. [Kha.l,2J. Step I of the proof and Fig. U.5 will give an intuitively clear rea-
son why these four polynomials are crucial to the stability of the set A of polynomials.
We will refer to the four polynomials as the K-polynomials. The proof below follows
[Min. I].
Proof.
Step 1. Claim: \;;/ ro E JR, p(jro,A) is a rectangle in € whose sides are parallel to
the coordinate axes.
Note that \;;/ aE A, \;;/ roe R,
With obvious definitions of gj(jOO) and hj(joo), (i=I,2), we rewrite (16) and (17) as:
\;;/ aeA and V ooe
Note that, \;;/ 00 e the sides of the rectangle p(joo,A) are parallel to the axes of
«r. By inspection, we note that the four polynomials defined in (10-13) are given by
386
21 i,j= 1,2.
For convenience, call the rectangles p(jO),A), defined by (16) and (17) "r/ 0) E Rt, the
K-rectangles. Note that they are bounded away from the origin.
=> By assumption the whole class A is Hurwitz, hence, in particular, the four polyno-
mials kij are also Hurwitz. •
Fig. 11.5. p(jco,A) is a rectangle with edges parallel to the axes; its
vertices are defined by the K-polynomials.
387
.;: For ro=O. the K-rectangle reduces to [.Ilo.liQ] with .Ilo>O. By assumption k21 is
Hurwitz. hence arg[k21(jro)] is stricdy increasing and increases by 3x/2. Also k21 (jco)
is the lower right-hand corner of the K-rectangle which. V roe has its sides
parallel to the coordinate axes. Finally. for ro large. not only k21 (jro) but the whole
K-rectangle enters and remains in the third quadrant; indeed. for n=3. since all polyno-
mials in A are monic. for ro large. any peA is such that
since k21 is Hurwitz and .Il2> O. for ro large. both the real part and the imaginary pan
of any such p(jco.a) are negative with Im[p(jro.a)] :: ro Re[p(jco.a)]; hence. V peA.
the net change in argument is 3x/2. Hence all polynomials in A are Hurwitz. •
The proof for n=4 is similar except that to guarantee that V peA the net
change in argument be 2x we need both k21 and k22 to be Hurwitz.
26 ...
where the (Xi's and are real and for given 1l.i.ai.fti.Pi
(Hint: now a E c:
n and we lose the property that p(jro.a)* = p(-jro.a); hence. we must
consider the change of argument from co=-oo to ro=oo. Consider ro :::?. 0 and
ro 0 separately; each case gives you a set of four K-polynomials. The necessary
and sufficient conditions will involve eight K-polynomials. (See. e.g. [Min. I].) Use
these results and the transformation z = soia to find conditions so that the set A has its
zeros in the sector +e arg(s) S; - - e. where e is a positive angle <1t/2.)
28 Exercise. Let us now consider the set Aof real polynomials of degree n of the
form
and .Il i > O. V i. Show that the set A is Hurwitz if and only if four special
388
29 The discrete-time case. The formulation of the Kharitonov type theorem for the
discrete-time case is quite involved (see [Man.1]).
2 vo=Ri+Li+KO)
3 't=-Ki+Jro+bO).
Equations (2) and (3) define our plant with Vo as the control input and the velocity 0)
is the output. The torque 't is set to zero because we view it as a disturbance. Using
Laplace transforms and abusing notations viz. we write i(s) for the Laplace transform
of i(t) .... we obtain from (2) and (3)
n (s)
4 O)(s) = P(s)vo(s) = -p- vo(s)
where
6 = LJ s2 + (JR+Lb)s + Rb+K 2 .
2
7 C(s)=k s +as+a
S(S+P)
where a and P are positive numbers chosen so that S(P,C) is expo stable. The charac-
teristic polynomial of S(P,C) is given by
where
ao=Kka
a1 =
10 a2=
Note that except for the terms and K2 in a1 and a2, resp., every coefficient of the
characteristic polynomial pes) is affine in each of the physical parameters (R, L, J, b
and K) and in the design parameters (k, a and 13).
I J J
where 1. J, n, b are known constants: thus we have two variable parameters ql =J,
q2 = b and the parameter vector q := (ql,q2) is restricted to belong to the rectangle
QcR2 with vertices (I,b), (l,b), (1.12), (J,b). Let us label the polynomial p of
(8)-( 10) as p(s,q). Show that p(s,Q) is a parallelogram in lr. (Hint: for a fixed s,
consider successively only J varying, then only b varying: you obtain two straight-line
segments that are, in general, not parallel to the axes nor orthogonal to each other.... )
or
U= {S E (t I either Re s 2:.. -a for some fixed a> O,or
15
I arg s I "2n; + 9 for some fixed 9 E (0,1t/2) )
where
21 Fact. a) '<;j q E Q, the zeros of p(s,q) are in some fixed compact disc contained
in, say, D(O,r);
b) '<;j s E (t, p(s,Q) is compact and arcwise connected, (as the continuous image of Q
compact and arcwise connected, [Die. 1, ChapJ]).
22 Exercise. Let Il:= and Ilj:= maxa;.(q), i=O,l, ... ,n-l, where the
minimum and the maxima are taken over all q E Q. Show that > and that r can be Il 0
taken to be any number larger than
n-l
(Hint: if s is a zero of p(s,q), I s I L I I . I s I k+l-n).
o
Let Qp denote the set of polynomials defined in (17-20); we say that the set Qp is
U-stable iff '<;j q E Q and '<;j s E u, p(s,q) '# O.
¢. By assumption. p(s.qo) is U-stable. i.e. all its zeros are in the open set
Uc = tr\U. For a proof by contraposition. assume that. for some qe Q. p(s.q) is not
V-stable; hence p(s.q) has at least one zero in V. say. z.
By (20). there is an arc in Q
that joins qo to q. Consider a point q moving continuously from qo to q along this arc;
during this motion. by (18) and (19). the coefficients of p(s.q) vary continuously;
hence the zeros of p(s.q) move continuously in fr. (Indeed the zeros of a polynomial
are continuous functions of its coefficients; note (19) and [Die.!. Thm 9.17.4].) In
particular one zero of p(s.q) starts from a point in VC (when q = qo). moves along a
z
continuous curve C to end at in V (when q=q). If ze av.
then p(z.q)=O. which
z
contradicts ii). If is in the interior of V. then the curve C starts (for q =qo) at a
point in the open set UC and ends (for q=<}) at z in the interior of U; consequently.
the curve C intersects the boundary of V. au. at some Zj. which corresponds to some
q) e Q along the arc joining qo to q. Hence, p(z),ql)=O with zi e au and q) e Q,
which is a contradiction. I
25 Q=co { e\.e2.....er }
r r
= { q e Rm I q= L '-;ej. Aj 0 It i, L Aj = I }.
) )
26 f: q -t b+Aq
r
Proof. Consider an arbitrary q e Q. with q = L Aiei' (where L Ai = 1, Ai 0 It i).
) \
392
then
r r
= L Ie; (b+Ae;)= L A;f(e;).
i=1 i=1
Consequently,
r r
29 f(Q)= ( ZE lRn I z=L le;f(ej); L Aj= 1; Aj 0" Vi},
t i=1
31 Test. For a sufficient number of points s E av, calculate n(p(s,Q», the nearest
point from the origin to p(s,Q). If, for all such s, n(p(s,Q» > 0, then the set of polyno·
mials is V-stable, else it is not V-stable.
33 V SE av,
where
= min arg(p(s,ej»
j
a(s)=max arg(p(s,ei» .
,
11.5.3. The Discrete Time Case
The methods used in Sections 5.1 and 5.2 above are based on the continuity of
the zeros of a polynomial, on a connectedness argument and on properties of affine
maps. Clearly these methods apply to the discrete-time case; note, however, that V is
now chosen in the z-plane, is symmetric with respect to the real axis and must include
393
1. Suppose that we have designed aU-stable (11.5.16) control system S(P,C) (see
Fig. 11.4) which has the following properties: a) P and C are proper rational matrices;
b) det[I+PC](oo) :# 0; c) the representations of P and of C have no U-unstable hidden
modes; and d) the closed-loop transfer functions
are U-stable, equivalently, they are proper and have no poles in the closed set U. By
(11.1.39), Heu is U-stable ¢:> Hyu is U-stable.
Now in the process of modeling, certain dynamical aspects of the physical plant
are usually neglected for simplicity; furthermore in the course of the operations some
parameters of the plant may vary; we model these changes by replacing P(s) by
P(s)+.-lP(s). In certain applications some exogenous inputs are coupled to the unmo-
deled dynamics and not to the original model. To capture this aspect we introduce a
third input u3 in the new configuration S(P,M,C) shown in Fig. 11.6. The system
S(P,M,C) has three inputs UI,U2,U3 and three outputs YioY2,Y3; hence we consider
Hyu: (uI,u2,u3) (YI,Y2,Y3) and Heu: (uI,u2,u3) (e"e2,e3)'
2 We assume that
3 M(s) E Mat(Rp(s»
+.. 6P
b +
P +
Y2
u
2
+
4 Given that there are no U-unstable hidden modes in P, C and tJ>, we say that
S(P,tJ>,C) is V-stable iff the transfer functions Hyu and Heu are V-stable.
5 It is easy to see that Hyu is V-stable Heu is V-stable, but the converse is not
true: Hence to establish V-stability we need only show that Hyu is V-stable.
We are ready to state the necessary and sufficient conditions for V-stability of
S(P,M,C) [Bha.l].
11 M(I+Qt1pr l = R[(s-p)I+QRr l .
Since the matrix expression in the brackets is analytic in V, (see (10) above), by (7)
we have
12 S(P,RJ(s-p),C) is V-stable det[(s-p)I+Q(s)R] '# 0, VSE U .
In case R is a dyad, say cbT , the second expression in (12) reduces to the scalar con-
dition, (using det(I+MN) = det(I+NM»
395
13 (s-p)+bTQ(s)c ¢ 0, \tSE U.
This equation is very useful: it shows how the closed-loop control (through Q(s»
affects the open-loop pole p.
8) Consider equivalence (8): the condition on AI> in (8) says in particular, that AI>
will not destabilize S(P,C) if, \t s e U, Q(s) is small whenever P(s) is large: more
precisely, if
15
In (15)-(17), P,C and LlP are matrices of proper rational functions. Let us perform the
following elementary operations, first, (16) (16)-P'(15) and, second,
(17) (17) + Q . (16). If we write the result in matrix form we have
I -C
18
[ o
o
I+PC
o
19 Y3 =LlPC) =.1P(I+QLlP)-I[U3+QUI-QPU21.
In view of (to), namely, Q and QP are U-stable, and of the assumption that
LlP(I+Q.1P)-1 is U-stable, Eq. (19) shows that the transfer function (ul,u2,u3) Y3 is
U stable. Next
Y1 = Ce1 = C(I+pC)-1[_Y3+U1-PU21,
hence
Thus (20) shows that the transfer function (u1,u2,u3) -+ Y1 is U-stable. Finally,
396
shows that because of (10) the transfer function (Ul,U2,U3) Y2 is V-stable. Hence we
have shown that Hyu is V-stable, hence S(P,IlP,C) is V-stable.
22 Y3=Hy,u,U3=L1P(I+QllPr1u3'
¢:. Follows immediately from Cramer's theorem and the fact that both Q and L1P are
V-stable.
==-. By the assumption in (23) and the V-stability of Q, we have
24 I-QL1P(I+QL1P)-1 = (I+QL1P)-1
is also V-stable. Since both Q and IlP are V-stable, the last expression (24) is V-
stable iff det(l+QL1P)(s) "# 0, 'rI s E V. So necessity is established. •
4
SI-A -b
n(s)=det [ -c -d
1,d(s)=det(sI-A) .
o
b) Let a(A) c (L and Re z 0; show that
p(t;O,xo,exp(zt» 0 as t 00.
8 S(A):= [ ;\.I-A
-c
-B
-D
1 is ful/-normal rank
i.e. the polynoinal matrix S(A) is full rank except at a finite number of points in C: . •
Note that S(A) is an (n+no) x(n+ni) matrix. From (7) and (8), we conclude
11 Definitions [Transmission zeros]. Let R be given and satisfy (7) and (8):
a) if n + ni ::; n + no. we say that z is a transmission zero of Riff :3uo e ni
and a corresponding unique Xo en such that
13 *
11 pet; O,xo,uo exp(zt» =0 Vt 0.
14 Comments. a) For n+ni ::; n+n o' (12) says that the input space c: ni
has a
special direction Uo such that the input uo exp(zt) and a uniquely chosen initial state xo
produce an output that is identical to zero. Thus, R blocks completely the transmission
of the input uoexp(zt).
For n + nj n + no, (13) says that the ouput space € n" has a special direction 11
such that for all inputs Uo exp(zt), (with Uo arbitrary), and a corresponding initial state
xo, the output yet) remains in the subspace orthogonal to 11. Thus, R blocks the
transmission of any input uoexp(zt) in the sense that the corresponding output is con-
strained to remain orthogonal to 11.
Thus, in the MIMO case, the notion of transmission zero consists of not only a
point z in the complex plane but also of either a direction in the input space or a direc-
tion in the output space. If, in addition, ni = no' then, for the special uo, the output yet)
is completely blocked out, and for any input uo exp(zt), the output y( t) is orthogonal to
11·
A We could have formulated the problem starting with the transfer function
H(s), considering any of its minimal realizations and require (8): then we would say
399
This theorem is the extension to the MIMO case of the definition (2) of the SISO
case. (See, in particular, Eq. (4) of Exercise (3).)
17 [ZI-A -B]
-C -D
[Xo]
Uo
[88 ].
Note that neither Uo nor Xo can be zero; this follows from (7) and (9). Furthermore, by
(7), to any such uo, there is a unique Xo that satisfies (17). For the Uo and Xo in (17),
consider the input Uo exp(zt) and the state trajectory x(t) = Xo exp(zt). Referring to the
differential equation x= Ax + Bu, we see that
20 (sI-A)x(s)-Buo/(s-z)=xo
From (20), (21), and assumption (7), xes) is uniquely defined and has only one
x x
pole at s=z. Call k the residue of (s) at z, so (s) = k/(s-z). But x(O) = Xo by (12),
hence k = xo. Substituting this result in (20) and (21) and letting s 4 z we obtain
(zI-A)xo+ B uo=8
{
22 C Xo + Duo=8'
i.e. the nonzero vector (xo,uo) is in the nullspace of S(z); hence, rk [S(z») < n + ni; i.e.
SeA) drops rank at s=z. •
23 I T]* ) [ = [8* I 8* ) .
Again, from (7) and (8), neither nor T] can be zero vectors.
Now since, by assumption, z If. a(A), zI-A is nonsingular; hence V UOE (Cn; there
is a unique Xo such that
Consequently, V t <::.. 0
25 * *
T] pet; o,xo,uo exp(zt» =T] (Cxo exp(zt) + Duo exp(zt» = °
where the last equality follows from (23) and (24). By definition (13), Eq. (25) estab-
lishes that z is a transmission zero of R. •
*
T] (Cxo+Duo)=O
and
26
Hence
401
where" ¢ e.
Now since (zI-A) is nonsingular and since the two matrices
28
ZI-A
S(z)= [ -c
-B
-D
1 and
are equivalent (i.e. either one can be derived from the other by elementary (block) row
and column operations), the two matrices in (28) have the same rank. Now (27)
implies that the rank of the second is smaller than n + no; hence
H(s)=diag
A [S-I
-, -S+IJ
s+1 s-1
;
the corresponding output Y2(t) -t el10 as t -t 00. In particular, if ZEjIR, (i.e. z=O or
z=jroo for some nonzero roDE R), S(P'C) will not respond asymptotically to some
constant inputs (in case z = 0) or some sinusoidal inputs (in case, z = jwo), and this
will happen irrespective of the choice of the stabilizing C.
Proof. S(P,C} is expo stable and ZE (1:+ imply that C(I+PC}-I=Hy,u, is analytic at Z.
Hence, p. C(l+PC}-1 is the product of two matrices analytic at Z.
Since nj and P has a transmission zero at z, the no rows of P(z) are linearly
dependent; hence rkP(z) < no and det[PC(z)] =0. Now, by (31),
Now He,u, = (I+pcr l is analytic at z because S(P,C) is expo stable and z E <r +; con-
sequently,
34 det(I+PC(z» oF- O.
From (33) and (34) we conclude that detH y2u ,(z)=0, equivalently, H y2u , E IRp(S)I1o X I1o
has a transmission zero at Z. I
We describe here the main concepts and results useful in linear system theory.
We start with basic algebraic concepts, functions, rings, fields, linear spaces and linear
maps. Next we tackle the question of representation of linear maps, a crucial concept
in many applications. Normed linear spaces are next, with the key concept of conver-
gence. The final section covers adjoints. It leads naturally to the singular value
decomposition.
We assume that the reader is familiar with the notion of sets, intersections of sets
and unions of sets; and with the symbols e",c .¢,U ,andrl; and with the logic sym-
bols 'fI, 3, 3!,=>,<=, and <=>. (See, e.g. [Loo.1], [Rud.1].)
I Some sets are given standard labels: for example Z, N, R,R+, €, (J: + denote the
sets of integers, nonnegative integers, real numbers, nonnegative real numbers, com-
plex numbers and complex numbers with nonnegative real part, resp.
2 Given two sets X and Y, by their cartesian product we mean the set of all ordered
pairs (x,y) where x e X and ye Y. The cartesian product of X and Y is denoted by
X x Y. Consequently the set of all ordered n-tuples of real (complex) numbers is
denoted by R n ( n).
3 Next the notion of function. Given two sets X and Y, by f: X -4 Y, we mean that
to every xe X, the function f assigns one and only one element f(x)e Y called value
of fat x. X and Y are resp. called the domain and codomain of f and we say that f
maps X into Y. f(X):= {f(x) I xe X} is called the range of f. The words "map,"
"operator," and "transformation" have the same meaning as function; a more complete
specification for a function is f: X -+ Y : x -+ f(x) where x -+ f(x) means that f sends
xe X into f(x)e Y. The latter can also specify f when the domain and codomain are
known; for example t -4 cost defines the function cosine: R -4 R.
the composition of g and f (in that order); h is also said to be a composite function.
X fog z
\1 y
Fig. AI. A commutative diagram.
405
t Bounded at infinity.
tt (Zero at infinity).
406
(a) Addition is
Associative: (a+p}+y=a+(l3+y) Va,p,y
Commutative: a+p = l3+a V a,p
:3 identity 0:
°
cH.()=a Va
:3 inverse: Va, :3 element (-a) S.t. a+(-a.) =
(b) Multiplication is
Associative : (a 'P) .y= a . (P .y) V a,p,y;
Note that we require our rings to have an identity element 1: some algebraists do not
•
require this and call our rings, "unitary rings," e.g. [Sig.l]. There is a standard pro-
cedure to add a 1 to any ring (Jac.l].
From the axioms above four important facts follow.
10 Fact. In any ring and in any field, the identities 0 and 1 are unique.
(This is easily shown by contradiction.)
11 Fact. In a ring, the cancellation law does not necessarily hold; more precisely, in
a ring
13 Remark. We know some rings for which the cancellation law holds: e.g.
Z, R[s], cr [s], lRv(s), Rv.o(s), R(O), R 0(0), RU' Such rings are called integral
domains, or better yet, entire rings.
•
16 Exercise. Show, from the axioms, that in any ring R,
(± (I:
i=1
ai ].
k=!
/3k ] = i: I: a i/3k .
i=! k=1
•
17 The ring K is called a commutative ring iff, in addition to the standard axioms
(9) we have
18 pq=qp 'ip,qe K.
defines the matrix P+Q that is, in fact, in Krnxn. If Pe K rnxn and Qe KTlXp then
n
(PQ)ik := L Pij qjk V ie 01, V ke I!
j=l
21 Exercise. Show that for n > I, K nxn is a noncom mutative ring. [Hint: check that
the axioms are satisfied].
22 Fact. For matrices with elements in K, the definition and properties of deter-
minants hold as in the case of elements in a field as long as one does not take
inverses! For example, if P,Q e Knxn , then det(PQ) e K and det(PQ) =det(P)·det(Q).
23 Fact [Cramer's rule]. Let PeKnxn , hence detPe K. Let Adj(P) denote as usual
the "classical adjoint", [Sigl. p.282], of P. By direct calculation we have:
In that case,
27 Comments. From (24) and (26) it follows that P has a right inverse iff it has a
left inverse; the common right and left inverse of P is called the inverse of P, (cf.
(AU 1).
(2) Cij is the cofactor of element Pij of P, i.e., cij=(-I)i+jmij with mij
denoting the determinant of the matrix obtained by crossing out row i
and column j of P.
(b) If P has an inverse p- i e Knxn, then by the axioms of K, det(p- i ) e K; now
pp-I = implies [det(P)]. [det(p-I)] = 1; hence (25) holds. Conversely if (25) holds,
then the RHS of (26) e Knxn and is the inverse of P according to (24). •
28 Note. A matrix Pe Knxn is said to be nonsingular iff detP *" 0, where 0 is the
additive identity of K. Hence if K is a field then condition (25) is equivalent to
det P *" O. Therefore, we have the following coroJlary.
P is nonsingular.
•
30 Comment: If the ring K is not a field then there may exist nonsingular matrices
P e Knxn having no inverses in K n xn : however corollary (29) still holds for inverses
in F nxn where the field F has K as a subring."
31 Example: Let K = R[s] be the ring of polynomials; R[s] is a subring of the field
F= R(s) of rational functions.
Hence, according to Corollary (29), P e R[s]nxn c R(s)nxn has an inverse in
R(s)nxn iff Pis nonsingular.
However, according to Fact (23), Pe R[s]nxn has an inverse in R[s]nxn iff det Pis S.t.
(detp)-I e R[s], i.e. such that det P is a nonzero constant (i.e. a polynomial of order
zero): such polynomial matrices are called unimodular, (equiv. invertible in
R[s]n xn ).
It follows that unimodular polynomial matrices are nonsingular but the converse is not
true.
To wit: PI(s) e R[s]2x2 and P2(s) e R[s]2x2 are both nonsingular but only P2(s) is uni-
modular: det P2(s) = I, det PI (s) = I-s3 (not a nonzero constant).
1 Definition. We call linear space (V,F) the object consisting of a set (of vectors)
Y, a field (of scalars) F and two binary operations viz. addition of vectors + and
multiplication of vectors' by scalars ., which obey the following axioms:
Addition is
Assqciative: (x+y)+z = x+(y+z) V x,y,ze V
Commutative: x+y = y+x V x,ye V
3! identity e, (called the zero vector), S.t.
x+e=e+x=x VxeV
3! inverse: V xe Y, 3!(-x)eY S.t. x+(-x)=9;
(c) Addition and multiplication by scalars are related by distributive laws viz.
V xe V, V F =
2 Canonical Example I. The linear space (P.F): the linear space of n-tuples in F
over the field F, with elements x = (Si );. Y= (lli); where each Si:'lie F for ie n·
Addition and scalar multiplication are defined by
x+Y:=(Si+lli)P and aX:=(uSi)P V ae F .
The most common examples are ( ern, er), (Rn,R), (R(s)n,JR(s» or ern,Rn,R(s)n for
short.
3 Exercise. Show that (P,F) is a linear space. [Hint: use the axioms of the field
411
5 Exercise. Using the definitions of a function and of a linear space, show that
F(D,V) is a linear space. Describe precisely what the zero vector is in this case.
(Denote it by 9F , and that in V by 9 v ).
6 Example III. The function space PC ([ro,td,Rn): it is the set of all functions map-
ping the bounded interval [to,ttl into R n which are piecewise continuous, i.e. they are
continuous everywhere, except that they may have a finite number of discontinuity
points 'tk where the one-sided limits f('tk+) and f('tk-) are well defined and finite. An
Rn-valued function defined on an infinite interval is said to be piecewise continuous
iff it is piecewise continuous on every bounded subinterval.
The prototype of a function in PC ([O,oo),R) is the square wave.
00
15 Examples. The set of all vectors in R n whose first component is zero. The set
of all functions fe F (D,V) that are equal to 8 y (the zero vector in V) at some fixed
point doe D or on some fixed subset Dl of D. The set of all functions f: R+ --+ R,
integrable over whose Laplace transform is zero at some fixed point Zo with
Re(zo) > O.
18 Fact [Sig.1. p.196]. Let (V,F) be a linear space. Then the smallest subspace
generated by a subset S cV is the span of S denoted by Sp(S) viz. the set of finite
linear combinations
19 Definition. Let (V,F) and (W,F) be linear spaces over the same field F. The
linear space (VxW,F) is called product space of V and W: it consists of vectors
413
and
a' (v.w) := (av.aw) 'v' ae F. 'v' ve V. 'v'weW.
21 Definitions. Let (V.F) be a linear space. The family of vectors (Vi)r. where
each ViE V. is said to be linearly independent iff any relation of the form
implies
The family of vectors (vi)f. where each ViE V is said to be linearly dependent iff there
exist scalars ...• Un. not all zero S.t.
1 x x x
o 1 x
o0
M .-
·-
000 1 x x
where the x's denote arbitrary real numbers. Note that the leading nonzero element in
each row is 1. Show that the family of row vectors. (ordered from top to bottom). is
linearly independent.
414
2S Note. In view of Fact (18) condition (b) reads also Sp ( (b i );] = V. Hence, in
view of the definitions, if XE V and (b i ); is an (ordered) basis, then there exists a
;=1
26 Exercise. Show that the Si'S are uniquely defined in terms of x and (b i ];.
27 Fact [Sig.1,p.215]. Let (bit be a basis of (V,F). Then any other basis of
(V,F) has also n elements. Thus the number of elements of a (finite) basis of a linear
space is independent of the basis considered. Hence the following definitions.
28 Definitions. If a linear space (V,F) has a basis of n elements then (V,F) is said
to be finite dimensional or of dimension n. (We write dim V=n). Otherwise is
said to be infinite-dimensional.
29 Examples. a) (Rn,R) and (Rnxn,R), i.e. the space of n x n real matrices, are
finite dimensional of dimension, n and n2, resp., with corresponding bases [ei and t
(Eij)i,jeD where a) e i is the n-vector with every element 0 except for the ith which is 1
and b) Eij denotes the n x n matrix with every element 0 except for the (i J)th which is
1. PC ([O,I],R) is an infinite-dimensional linear space: indeed subspace
30 Modules. Careful examination of Definitions (1), (14), (17), (19), (21) and (24)
shows that one may replace the field F by a commutative ring K: if this substitution is
carried out according to Table I below then we have generalizations of the notions in
the column on the left to those in the column on the right.
In system theory, we shall have many uses of the module (R[s]n, R[s)), i.e. the
module of polynomial n-vectors over the ring of polynomials. Another one is
415
Table 1
subspace, submodule,
basis, basis
32 Note. Not all modules have a basis; the modules that we shall use have a basis.
Modules (V,K) that have a basis are free modules [Sig.l.p.200].
In the sections below we shall be involved with linear maps over linear spaces
and their matrix representations over a field. Note that, for example, for polynomial
matrices one must introduce linear maps over modules and their matrix representa-
tions over a ring see, e.g., [Sig.1], [Cal.1].
A4 Linear Maps
1 Let (V,F) and (W,F) be linear spaces over the same field F. Let A be a map
from V -+ W. We say that A is a linear map (or linear operator) iff
2
•
Note the abuse of notation: the value of the function A at v is denoted by Av rather
than A(v).
That condition (2) holds here is immediate from the rules of matrix addition and multi-
plication by a scalar.
f
wet) := (A v)(t) = exp[-(t-1:)]V(1:)d1:
o
8 N (A) := { ve V : A v = 6w } c V
9 R (A): ( A v: ve V ) cW
16 Comments. a) In case dim N (A) > 1, (ii) shows that Ax = b has an infinite
number of solutions. In particular, if is a basis for N (A), then
k
x=Xu + L cibi
i=l
is a solution.
b) In general, the set of solutions is an affine space denoted by Xo + N (A) (note that if
every vector of that space is shifted by -Xu, then the affine space becomes a linear
space),
z := x - XOE N (A ) .
M := Sp(S)
t
when S is a subset of V.
Note that if M has a basis [bi then
M .- (v E V I B v = 8 u )
419
equivalently M:= N (8 ).
•
18 Note. IT the linear space V above is of finite dimension n then the subspace M
of V is also finite dimensional and the integer n-dim( M) is called the codimension of
M w.r.t. V.
where is called the component vector of x w.r.t. the basis Now by linearity
m
1 A Uj = l: 3jjVj '\:t je n ,
j=l
note that (3jj). is the component vector of A Uj' Hence, by stacking column vectors
A= m rows
3ml
n columns
where AeFrnxn.
420
n n m
y = Ax =L Uj = L a;jVi
pI pI i=1
m
= L11iv i'
i=1
3 Theorem. Let (U.F) have. a basis (Uj); and let (V.F) have a basis Let
A : U , V be a linear map. Tnen. w.r.t. these bases. the linear map A is represented
by the m x n matrix
A= [a;). . E pnxn
j lEll.jEn
where the jth column of A is the component vector of A Uj w.r.t the basis
4 Note. In most applications we replace vectors and linear maps by their represen-
tations. viz. component vectors and matrices. e.g. we write N(A) instead of N(A).
5 Exercise. Let A be a linear map of (U.F) into itself. where dim U = n. i) Sup-
pose that ... -Un-IA -UnI. where I is the identity map:
U, U and the ai's are elements of F. Let b be a vector in U. Suppose that
}
o 0 0 -ex,.
1 0 0 -<Xn-I
0
0 o} 0 -<Xn-2
el= A= 0
o o0 o
The matrix A above is said to be in column-companion form.
ii) Suppose that A.E F and (bi )." is a basis of U such that
1=1
and
A bk = Ab k + b k - I for k = 2,...,n.
7 Exercise. Let A be a linear map from R3 into R3. Consider the usual mutually
orthogonal axes Ox I' and Ox 3, (Fig. A2). The map A first rotates any vector by
an angle Cli about the aXIs Ox} and then rotates the resulting vector by Cl3 about Ox3.
Obtain the matrix representation of the linear map A w.r.t. the given axes.
Fig. A2. The map A is the composition of the two rotations shown.
422
where dim U = m, dim V = n and dim W = p. Let U,V, and W have bases fUk
ht (Wit and with corresponding component vectors s= 11= (11j and
A= (a;.]J iellJell
and B= (b·J
k]
jen.kern
and n=Bs.
Consider the composed linear map
C= (Cik] iell,kern
where \;:fi,k
n
cik = L aijbjk
pI
i.e.
C=AB
where I is the identity map on U. Show that if A and A -1 have matrix representations
A and B, then
423
BA=AB=In
h)
11 Exercise [Linear spaces and bases]. Let (V,F) be a linear space S.t. dim V = m.
Let'V have bases and (Vi For ye V let the corresponding component vec-
tors be 1'\ = (1'\i ]mand -1'\= l'\ both in pm: thus y = Lm1'\ivi= Lm_l'\Vi'
I ! i=! i=!
Show that
a. Each choice of basis (Vi ) defines a bijective linear map
b. Each change of basis defines a bijective linear map: vi Vi' ie ill given by the
matrix
t
12 Let A : (V,F) (V,F) be a linear map with dim U = m and dim V = n.
Let U, with elements x, have bases [Uj and [iij ) generating component vectors
and S, resp. in F". According to Exercise (11) we know that and Sare related by
13 S=PS
where the matrix PeF"xn is nonsingular, (A2.29). Similarly, let V, with elements y,
have bases (Vi ) and generating component vectors 1'\ and 11 in Fm related by
424
V,y
•
Fig. A3. Change of basis in a linear space. (Note that a double arrow
represents a bijection.)
14 ll=QrJ
where the matrix QEPrnxm is nonsingular. Finally let according to Theorem (AS.3)
and (AS.2) A have matrix representations A and A in Fmxn corresponding to com-
ponent equations
15
16
Note that the process of generating Eq. (15) is depicted in the commutative diagram of
Pig. A4 where the double arrows indicate bijections generated by the choice of bases
in domain and codomain. Of course, a similar diagram can be drawn when generating
Eq. (16).
It follows now immediately from Eqs. (13)-(15) that using matrix multiplication
17 A=QAP
Equations (17)-(18) specify the relations between the matrix: representations A and A
of the map A under the change of bases described above.
19 Two matrices A and A S.t. (17)-(18) holds are said to be equivalent [Sig. 1,
p.256].
Thus we see that under change of basis, two matrices A and A, both in pnxn,
represent the same linear map A : U Y with dim U = n and dim Y = n if and only
if A and A are equivalent.
The whole process is visualized by the commutative diagram of Fig. A5. Recall
the diagrams of Fig. A3 and Fig. A4. In Fig. A5 the right part of the diagram depicts
the change of basis in the codomain Y, the left part does so for the change of basis in
the domain U. The lower part and the top part show the matrix representations
A: '11 and A: ii. Reading off the outer part of the commutative diagram a)
clockwise, we obtain A=Q-1AP-"1 and b) counterclockwise, we obtain A=QAP.
Thus the commutative diagram of Fig. AS summarizes Eqs. (13) to (18).
Special Case. Suppose A : U U. Then we may use the same basis for domain
and codomain. The diagram simplifies then and becomes that of Fig. A6. Using the
commutative properties of its left-most part two matrix representations A and A of A
are related by:
-
A
p-1
A
F".,
Fig. AS. The diagram of change of bases for A : U Y.
426
21 A=Pi\p-' .
Note A and A are square matrices in pxn and that PeF nxn is nonsingular.
21 Two square matrices A and A S.t. Eqs. (20)-(21) hold are said to be similar
[Sig.l,p.257]. Hence, when the same basis is used in domain and codomain, two
matrices A and Aepxn represent the same linear map A :U U, with dim U = n if
and only if A and A are similar.
25 Remark. One might expect that N (A) and R (A) be disjoint. This is not so.
Various possibilities may occur. For example let U=V=R2
If A=
If A = then R (A ) (') N (A ) = { e} .
0 0 0
Let U=V=lR3 andA= [ 001 . Then
1
000
427
Proof of (24). Consider Eq. (24). Let (Uj) be a basis of N (A). Complete the basis
26 Ax=A t
j=l j=1 i=k+l
Now A uj=9 u for i = 1.2•...•k because those ujeN(A). Clearly. since x in (26) is arbi-
trary. fA Uj) n spans R (A) and we claim that this family is a linearly independent
t k+l
family. Suppose it is not: then it would be linearly dependent. i.e. for some scalars
<Xk+ l' • • •• <Xn (not all zero)
9y = t
k+l
<Xr4 Uj =A [ t
j=k+l
<XjUij.
i.e. the vector t <XjUj is in N (A). However. this vector is nonzero since (Ui] n is a
w w
n
linearly independent family. Hence l; <XjUj. a nontrivial linear combination of the
k+1
basis [Ui]; of U. is zero. This is clearly impossible and hence (a) (A Ui] n is a basis
t
•
k+l
for R(A); (b) dimR(A)=n-k; and (c) Eq. (24) follows.
27 By rank of the matrix Ae pnxn (denoted by rkA) we mean dimR (A) and by
nUllity of the matrix Ae pnxn (denoted by nlA) we mean dim N (A). Hence
28 n=rkA+nIA.
30 0 S; rkA S; min(m,n) ;
32 Exercise. Let Aepxn be a square matrix. Show that A has an inverse if and
428
Sylvester's inequality: Let Aepxn and Bep><P be two matrices, then ABePrnxp
and
37 rkA+rkB-n :!> rkAB :!> min {rkA, rkB} .
Hints: Let W be the codomain of A and let A I R (B) : R (B) -+ W be the restriction of
A to R (B). Note that
Sylvester's inequality and (24) are the main tools for the following result which is left
as an exercise.
38 Theorem [Rank and nullity invariance under equivalence]. Let AeF rnxn be a
matrix and Pe pnxn and Qe pmxm be nonsingular matrices. V.tc.
40 nIA=nIAP=nIQA=nIQAP.
44 Exercises. a) For each e.r.o. compute its corresponding left elementary matrix L
and check that the transformed matrix A is related to A by A== LA.
b) Show that each e.r.o. is invertible and, for each one, obtain Cl.
(Hint: Pj Pi ; Pi t-- c-1Pi etc ... ).
t Pi stands for row i; with re F, rpi stands for the product of row i by r.
tt Pj denotes row j; 'Yj denotes column j.
430
47 Exercise. For each e.c.o. compute its corresponding right elementary matrix R
and check that the transformed matrix A is related to A by A = AR.
50 Theorem [Row echelon form]. Let Aeprnxn. Then there exists a nonsingular
matrix Qe prnxm, (obtained by e.r.o.'s) S.t.
kr
1 0
2 0
0
51 QA=A= 0
0
r 0
--- ---
0 0
0 0
where A is said to be in row echelon form and has the following properties:
To reduce a given matrix Aeprnxn to row echelon form we use the following
algorithm. Here the field F is or R or tr, the algorithm is a special case of a more
general algorithm that is valid for the case that F is any field (see below).
Step 1. Search for 'Yk" the first column from the left that is nonzero. If no such
column is found. stop.
Step 2. Choose the entry which is the largest in absolute value among the entries
of 'Yk,. and. by row permutation. bring it in position (l.k l ).
Step 3. For t = 2.3 .....m. from Pt subtract PI multiplied by entry (t.kl ) and divided
by entry (l,kl)'
(Hence there is only one nonzero entry left in 'Yk, viz. entry (l.k l ).)
For i = 2.3 ....
Step 4. Search for 'Yk" the first column from the left that is nonzero below Pi-I' If
no such column is found. or if i=m+l. or stop.
Step S. Choose the element that is the largest in absolute value among the
nonzero entries of 'Yk. below Pi-I; by row permutation bring that entry into position
(i.ki )·
Step 6. For t = i+ l .....m. from Pt subtract Pi mUltiplied by entry (t.ki ) and divided
by entry
(Hence there is only one nonzero element left in 'Yk, below Pi-It viz. entry (i.ki ).) •
S6 Remarks. a) In the case of a general field F. the notion of absolute value may
not be available (e.g. R(s». then in Steps 2 and 5 we pick any appropriate entry of 'Yk,
that is not the zero element of that field.
13) However if F is R or 0: then careful analysis of the propagation of round-off
errors has shown that it is preferable in Steps 2 and 5 to choose as nonzero entry of 'Yk,
one that is largest in absolute value. This is referred to as partial pivoting
[Gol.l.p.IIO]. This is done in Algorithm (55).
57 Exercise. Use algorithm (55) on the matrix A (given below) to obtain A (in row
echelon form):
6 12 6
o -1 -1
58
000
000
In the proof of Theorem (50) below we will apply Algorithm (55) under the conditions
of Remark (56. a».We consider any field F.
432
PA=LU.
LUx=Pb.
Ly=Pb
Ux=y
by backward substitution.
If all the elements of A and b are nonzero, the solution requires O(n 3/3) opera-
tions. This estimate is not valid if A is sparse (i.e. made of mostly zeros) and if
sparse-matrix software is used the solution requires typically O(n1. 5) operations.
433
63 Ax=9
n-r times by setting successively all components of x with indices in KC equal to zero
except for one component which is set equal to one. The resulting family [Xi) of
solutions of (63) is a basis for N(A). •
Por example if A and A are given by (58) then K= { 1,3 } , KC = (2,4) and a
basis for N(A) is given by where t Xl = (-2,1,0,ol, X2=(2,O,-I,ll
."
are linearly independent because
n-r
66 Transposition, Theorem (50), Algorithm (55) and Remark (56a) show that for
any matrix AE pnxm there exists a nonsingular matrix PE p,nxm (obtained by e.c.o.'s)
S.t. A= AP is in column echelon fonn.
67 Exercise. State precisely the properties of matrix A in column echelon fonn (cf.
(52)-(54».
68 Exercise. State precisely an algorithm that will reduce any matrix Ae Fnxm in
column echelon fonn using partial pivoting (cf. Algorithm (55».
Let Ae pnxn be a matrix, then there exist nonsingular matrices Qe FIlOOn and
Pepxn (obtained by e.r.o.'s and by e.c.o.'s, resp.) S.t.:
70 QAP=D
with
r
71
m-r
and
72 r=rkA.
74 Exercise. Let A e pn xn. Let L(R) denote a product of left elementary matrices,
(right elementary matrices, resp.)
Show that:
(i) N(LA)=N(A) ,
(ii) dimR(LA)=dimR(A),
(iii) R (A) =R (AR) ,
(iv) dimN(A)=dimN(AR).
A6.1. Norms
Intuitively the nonn of a vector is a measure of its "length." Of course, since ele-
ments of a linear space can be vectors in R n, matrices, functions, etc ... , the concept of
nonn must be defined in general tenns.
2 Example. n.
(F ,F) with elements x = [ )n
Xj IE P.
n
3 IIxllt:= L IXil
I
5 II x II := max I xi I
00 i
are called sum norm, Euclidean norm, and sup-norm, resp. See [Nob.l.p.13ll
8 (Frobenius norm),
Note the correspondence between (3)-(5) and (7)-(9). The following are so-called
induced norms and will be discussed later, (see (73) below):
12 (row sums).
where II f(t) II. stands for any nonn in (Fn,F), (see equivalent nonns, especially (47)).
15 Exercise. Show that the nonns defined in (5), (9), (12), and (14) with II f(t) II
given by (5) are nonns, i.e. satisfy the axioms (a), (b), and (c) in (1).
16 Exercise. Let KeF', Aepntxn, and BePnxp. Using the nonns defined by (5) and
(12) show that
18 II AB II II A II II B II.
19 Exercise. Let f:[to,td F2:t f(t)=(f1(t),f2(t», (fl and f2 are the com-
ponents of f). Let f be piecewise continuous. Consider
v := i f(t)dt:= [i j
fl (t)dt, f2(t)dt ].
II 11
22 Comment. The definitions (21) are special cases of similar definitions in metric
spaces [Die. 1].
437
23 Exercise. In (1R2,R) draw the unit balls for the nonns 11'11 b 11'112, 11'11 00
defined by (3)-(5).
24 Exercise. Let B be the unit ball in the nonned linear space (V,F, 11'11). Show
that
(a) B is convex (Le. v"v2E B =:> AVI+(I-A)v2E B 'v' AE [0,1]),
(b) B is balanced (Le. VE B =:> -VE B),
(c) 'v' VE V there is a finite r>O S.t. VE B(9;r).
(d) Give an example of a set C that satisfies (a) and (b) but not (c).
A6.2. Convergence
A typical engineering use of nonns is in deciding whether or not an iterative
technique converges or not. Again, since we use nonns, the field F is either 1R or (C.
26. Given a nonned linear space (V,F, 11'11) and a sequence of vectors [Vi); C V,
we say that the sequence converges to the vector VE V, (denoted by vi. --+ v
1-->00
or lim Vi = v), iff the sequence of nonnegative real numbers II Vi-v II tends to zero as
i --+ 00.
Thus, thanks to the notion of nonn, the concept of convergence of vectors is
reduced to that of convergence of nonnegative real numbers.
In iterative techniques the "limit vector" v is not known (otherwise, why iterate?).
So we need to be able to decide the question of convergence without knowing v. As
with real numbers, we need the concept of Cauchy sequence,
27 A sequence (Vi] in (V,F, 11'11) is said to be a Cauchy sequence in V iff for any
DO, there exists an integer N, depending on E, such that
Reference to the definitions shows that every sequence (Vi] C V that converges
to some VE V is a Cauchy sequence. For the converse we ne.ed a new concept.
28 A nonned linear space (V,F, II'ID is said to be complete or a Banach space, (or
the linear space (V,F) is said to be complete in the norm II''') iff every Cauchy
sequence in V converges to an element VE V.
Finally, for purposes of approximation we need the notion of dense set.
29 A subset X of a nonned linear space (V,F, 11'11) is dense in V iff for every
438
30 Fact (F = R or (C). Let (V,F) be any finite-dimensional linear space. Let 11'11
denote any norm on V. V.l.c. (V,F,II'II) is a complete normed linear space, or equiv.,
a Banach space.
32 Fact. Let (Rn,R,II'II) be the Banach space of n-tuples of real numbers. Let Q"
be the subset of n-tuples of rational numbers. V.l.c. Q" is a dense subset of Rn.
33 Fact. Let F=Ror (C. (a) The normed linear space C«[to,td,F"),F,II'1I 00 ) of
Example (13) is a Banach space.
(b) Its subset P ([to,til,F") of n-tuples of polynomials in tE [to,til c R with coefficients
in F is dense in C ([to,til,F") (in the norm 11'11 00),
41 Let (V,F) be a linear space. Let 11'11 a and 11'11 b be two norms defined on V.
We say that the norms 11'11 a and 11'11 b are equivalent iff there exist two positive
numbers mt and mu S.t.
42 VVEV.
It is crucial to note that the same ml and mu must work for all VE V. Note also that
equivalence is an equivalence relation (reflexive, symmetric, and transitive). (Prove it.)
43 Exercise. Let (V,F) be a linear space. Let 11'11 a and 11'11 b two equivalent
norms. Let (vi);CV be a sequence,let X a subset of V and let veV. V.l.e.
4S Fact [Tay.1,pp. 55,62]. Let (V,F) be a finite-dimensional linear space, then any
two norms on V are equivalent. Hence for F=R or «I:, any two norms on the linear
spaces (p,F) and (F'"xn,F) are equivalent. •
46 Exercise. Consider the linear space (P,F) of Example (2) with norms 11'111'
II' 112' resp. II' 1100 given by (3)-(5). Show that
II x II 00 S II x lit S n II x II 00 'v' XE Fn
II x II 00
S II x 112 S ..filII x II 00
'v' XE F"
and
Show that the nonns II' II oo.a and II' II oo.b are equivalent.
(Hint: observe first that because of Fact (45) the nonns II '11 a and II' II b are
equivalent.) (This exercise justifies why in the definition of II f II 00' in (14), Ilf(I)II
may be chosen 10 be any P-nonn.)
b) for p=oo,
c) for p E [1,00],
is called the space of pth power summable sequences (1 p < 00), resp., space of
bounded sequences (p=oo). For n = 1 these spaces are denoted by IP. •
53 Fact. For pE [1,00], (If: ,F,II'II P ] is a Banach space, i.e. a complete nonned
linear space.
b) for p=oo,
56 II f II := ess sup ( II f(t) II ) := inf { K> 0: ( tel: II f(t) II > K )) = 0 )
00 tel
where I![A] and ess sup denote the measure of the set AcR and the essential
supremum, resp., (the essential supremum measures the least upper bound K > 0 s.t.
II f(t) II K except for a set of measure zero). In the text we shall write "sup" for
"ess sup."
57 LP(I,p n):= {f: f : 1-+ F" is measurable and II f II p < oo} is called the space of
pth power integrable functions on I, (1 P < 00), resp., the space of essentially
boundedt functions on I, (p=oo). LP(I,F") is also denoted LJ(I) or simply LJ. Por n=l
the latter are denoted LP(I) or LP.
58 Fact. Por pe [1,00], (LP(I,F"),F, 11'11 p) is a Banach spac(: i.e. a complete normed
linear space if we identify two functions f and g that are equal almost everywhere, i.e.
equal except on a set of measure zero.
66 Let F = R or IT and consider two normed linear spaces (U,F, 11'11 u) and
(V,F, 11'11 v). Let F be a map (or operator) S.t. F : U -+ V.
a) [Local continuity]. We say the F is continuous at the point u E U iff, for every
£>0, there exist a possibly depending on £ and u, S.t. considering points u'eU
ueU.
c) [Induced (operator) nonn]. The induced (operator) norm of F is defined to be
•
67 IIF II := ( IIF(u)lIv/ lIullu) .
68 Fact [Local sequential continuity]. Let U and V be as defined above and let
F :U V be a map. U.t.c. F is continuous at ue U iff, for every sequence
F(u)=.lim F(uj).
i-JOO
69 Comment. Definition (66a) of local continuity is a special case of the usual 10-0
definition when U and V are metric spaces. Fact (68) remains valid for metric spaces.
If: Let (uj]; cUbe any sequence s.1. Uj u. By the 10-0 definition of continuity,
for every DO, there exists an integer I S.t. Vi IIF(uj)-F(u)II <10, i.e.
Only if: If F is not continuous at ueU, then :310>0 and a sequence
S.t. lIui-ull < yet IIF(uj)-F(u)l1 >10... •
71 II A II := { II A u II v I II u II u } = sup ( II A u II v: II u II u = I ) .
72 Comment. Note that the induced nonn II A II is the "maximum gain of the map
A over all directions"; moreover II A II depends on the choice of the nonns 11·11 u and
II . II v in the domain and the codomain, resp ..
r
10 (column sum)
2
11 IIA A./A*A) ,
JEn
This justifies the use of the name "induced nonns" in Example (6).
82 Theorem [Rud.1,p.102]. Let (U,F, 11·11 u) and (V,F, lI·n v) be two nonned linear
spaces and let A : U -+ V be a linear map. Then, by the linearity of A, the following
three statements are equivalent:
a) A is continuous on U;
b) A is continuous at one point ueU, (for example, u=8 u );
c) A has a finite induced nonn i.e. II A II < 00. I
444
83 Comments. u) Note that (77) and (81) generalize the results of Exercise (16).
With V and V as given above, consider the set L(U ,V) of continuous linear maps
A : V --+ V, upon which we define addition and scalar multiplication by
(A+B)v:=Av+Bv VveV and (M)v:=MV Vue F VveV resp. From
Theorems (76) and (82) it follows at once that under the induced norm, (71)
(L(V,V),F,II'ID is a normed linear space. Moreover, it is known that [Tay.l,p.189] if
V is complete then L(U,V) is complete; i.e. Cauchy sequences of continuous linear
maps will converge to such a map in the induced norm.
y) From Example (73) and Fact (45) we see that matrices have finite induced norms.
Hence by Theorem (82), matrices represent continuous linear maps.
84 Theorem [Extensions of equalities]. Let (U,F) and (V,F) be normed spaces; let f
and g map U into V.
If f and g are continuous (equivalently, for all convergent sequences in V, [Xi
limf(x)=f(limXi» ,
and if f(x) = g(x) V X in a dense subset DeV then f(x) = g(x), V XE V.
Proof. Let xe U\D. Since D is dense, there is a sequence [Xi] e D such that xi --+ X.
By assumption, since each xiE D,
and by continuity
•
= .lim g(Xi) = g [.lim Xi] = g(x).
1-+00 1-+00
unknowns. More precisely, let P=R or C:; let AeF'xn, be F' and xe F'. Consider
the equation
86 Ax=b
We assume that det A 0 and b en; hence, (86) has a unique nonzero solution
xo=A-lb. Suppose that as a result of noisy measurements, round-off errors, etc., we
only have an approximate value A+3A for A and b+3b for b: thus we have a per-
turbed system of equations and calling its solution Xo + 3x, we write
OAxo + Aox = Ob .
Calculating 8x, taking norms of both sides, and using the properties of the norm and of
the induced norm, we obtain
89
IIxoll
:S II A II . II KIll [Mill
IIbll
+ 11M II
II All
1.
90 1(A):= II A II . II KIll
446
is called the condition number of A. Note that it depends on the nonn chosen. How-
ever, regardless of the nonn K(A) 1. In fact, K(A) = 1 if A is unitary and if II· Ii2
is used.
Matrices for which K(A) is large are called ill-conditioned; for such matrices, the
solution is very sensitive to some small changes in A and in b. Since all nonns on F n
are equivalent, it can be shown that if a matrix is very ill-conditioned in some nonn, it
will also be ill-conditioned in any other nonn.
Note that (89) gives an upper bound on II ax 11111 Xo II. In some cases, the right-
hand side gives a gross overestimate of the error.
The following exercise shows that small errors in A and b can lead to large errors
in x when K(A) is large.
First a sign bit, then t binary digits fonning the mantissa, and the integer e is the
exponent. Typical!?' e may be any integer in [L,U] with L of the order of - 103 and U
of the order of 10 . Thus, there is a largest possible positive number and a smallest
positive number that the computer can handle: they are
If, in the course of a computation, the result is larger than the first we have overflow
and if it is smaller than the second we have underflow; in both cases calculations stop.
In the following we assume that neither overflow nor underflow occurs. In the
process of representing a real number x, (say 7t, e, 1/3, log 2,... ), by a sequence of t
binary digits as in (94) above, an error usually occurs; clearly, the best choice is for
the computer to choose the number of the fonn (94) that is closest to x. This number
is denoted by flex). Clearly
The number Em is called the machine E. For scientific computers, t is typically 48,
then rt= 3.55 x lO-15. Note that replacing x by flex) causes a very small relative
error.
447
Multiplication. Given two real numbers x and y, we obtain first their binary represen-
tations, then multiply these representations and round off the result:
fI(x x y) := flrf/(x) x f I(y)), i.e.
So the relative error is at most 3Em. Hence computer multiplication is subject to very
small relative errors. It is easy to check that the same conclusion holds for division.
XE1+yE2 ]
fl (x+y)= (x+y) [ 1+ + £3 .
x+y
Here the relative error may be considerable: consider x>O,y=-x+d with 0 < d « x
and £1 =-£2=£3 = Em, then the relative error is, neglecting higher-order terms,
Note that in this case, the round-off errors in x and y, namely, £1 and are amplified
by the ratio x/d which is very large. When this happens we say that there is a catas-
trophic cancellation. The conclusion is: computer addition may involve very large
relative errors.
Finally, note that computer addition is not necessarily associative and that in the
course of the computatign, valuable data may be rounded off. For example, let a = I,
b = -1, c = 0.l2345 10- . Carrying five decimal digits in the computation, we see that
fI«a+b)+c) = 0.12345 10-5 butfl(a+(b+c» = O. It is for this reason that it is important
to scale problems and to use normalized data so that, say, in evaluating polynomials or
scalar products, one does not add and subtract numbers whose magnitudes are very
different.
3 A linear space (H,F) equipped with an inner product viz. the triple (H,F,(','» is
said to be an inner product space. The norm 11'11 defined by (1) is said to be the
norm defined by the inner product. It is a norm because of Schwarz's inequality.
5 I (x,y) I S; II x II . II y II VX,ye H.
Comments. a) This inequality implies the triangle inequality for the norm defined by
(2); (prove it).
For F = R the angle between two vectors x,y e H is defined by
y) For F=R
(x+y,x+y)-(x-y,x-y) =4(x,y).
Proof of (5). Choose a E F, with 1a 1= 1, S.t. a(x,y) = 1(x,y) I. Then, for all
'A. E R, obtain, after some calculations,
•
7 An inner product space (H,F,(','» that is complete in the norm defined by the
inner product is called a Hilbert space.
449
10 Example. Let L2([to,ttJ, F") be the Banach space of square integrable F'-valued
functions on [to,t 1], (A6.57). Let f and g be two such functions and define the L2_
inner product by
I,
11 (f,gh := f f(t)"'g(t)dt
10
12 Example. Let C ([to.tll. 1"') and PC ([to,ttl, 1"') be the linear space over F of pO-
valued continuous. resp., piecewise-continuous functions on [to,ttl c R On both
spaces define the L2-inner product given by (11). The linear spaces
are inner product spaces that are dense in the Hilbert space (L2([to,td. F,(''')2) in the
L2_nonn, (see Fact (A6.59». •
13 Comment. For all practical purposes inner products over (piecewise) continuous
functions on [ta,t 1] may be replaced by inner products over square integrable func-
tions; indeed the latter can be approximated by the fonner because of the following
Fact.
Hint: by Schwarz's inequality and adding and subtracting tenns one has
450
15 Let (H,F,(-.. » be an inner product space. Two vectors x,ye H are said to be
orthogonal iff (x,y)=O. In that case
17 Exercise [Annihilation property]. Let (H,F,(','» be an inner product space and let
Y be a dense subset of H under the inner product nonn. Then "i/ x e H
"i/ye H
19 Let (V,F,II'I!) be a nonned linear space. A subset FeV is closed iff every
ve V, for which there is a sequence in F that converges to v, belongs to F. A
subset G e V is open iff it is the complement of a closed set, equiv. G C := V\G is
closed.
"i/i,ke m.
451
Such an orthononnal basis can be obtained from (Ilj t as follows: start with
The idea is to proceed step by step. Suppose we have obtained b l .b2..... bk_1 using
al.a2 .....ak_l; in other words (b i ] are orthononnal and span Sp [(aj] ;-1 ]. To
obtain we first compute from ak a vector Uk orthogonal to the bi's for ie k-l
By calculation. Uk is orthogonal to bi for all i e k-l, The last step is to nonnalize Uk:
k-I
(Note that II Uk II > O. for otherwise. ak would be a linear combination of bi ] • or
equivalently of r. ] 1k-I • which is impossible since LIlj
LIlj r. ] 1k is a linearly independent1 fam-
ily.) Thus the procedure leads to an orthononnal basis (bi ) of the subspace M.
24 Let M and N be two subspaces of the linear space (V.F). The sum of two sub-
spaces M and N. viz. M+N = ( u+v;u e M.ve N) is called a direct sum iff
MnN= (9). The direct sum is denoted by MEaN. •
27 H=MEaM.i .
x-ye Ml.
(equiv. in the inner product nonn II x-y II = inf ( II x-u II.u eM) ).
tE
30 N
30b Exercise. Let F= R or C::. Consider the Hilbert space (P. F. (.,.»
defined by
(8). Let Sand T be arbitrary linear subspaces of P. (Since Sand T are finite dimen-
sional. they are necessarily closed.) Show that:
a) (S1.)1. =S,
b) SeT <==> T1. eS1. •
c) (S+T)1. =S1. nT1. •
d) (SnT)1. =S1. +T1. .
31 Let F be R or er. Let (U,F.(·,,)u) and (V,F.(·,,)v) be Hilbert spaces i.e. complete
inner product spa,fes. Let A : U V be a continuous linear map. Then the adjoint of
A, denoted by A • is the linear map A *: V U s.t.
453
where
11
35 J
Au := O(t)u(t)dt
10
with
It follows easily by using Schwarz's inequality that A has a finite induced nonn, so A
is continuous, (A6.82). The adjoint A * : F" L2([to,tt1,prn) is given by
where O*(t) e poon is the Hennitian transpose of O(t). Indeed, for any ve Fm ,
J J
(v,A u)pn= v*O(t)u(t)dt= (O(t)*v)*u(t)dt
to 10
=(A*v,uh·
I,
39 A : PC ([to,tll,pm) -t P: u -t Au = f G(t)u(t)dt
I()
where G(') satisfies (36): indeed since PC([to,td,pm) is dense in L2([to,td,Fm ) in the
L2-norm, (see Example (12), for every u E there exists a sequence
[u i J cPC n S.t. Ui -t u and Au = lim AUi' We shall therefore also call A>I< the adjoint
of A.
39a Exercise. Let (U,F) and (V,F) be two normed linear spaces. Show that if
(V,F) is finite dimensional, they any linear map A : V -t V is continuous.
(Hint: if cbi]n cV is a basis of V then UE V iff u= ±<ljbi and A u= ±C:X0 ui
1 i=! i=!
Compare with (37).)
40 Comment. If V is not finite dimensional then not all linear maps A : V -t V are
continuous.
To wit let (V,IR) be the linear space of polynomials with coefficients in R for which
we choose the norm II u II 00 = max { I u(t) I : t E [0,1] }, and let V = IR. Consider the
linear evaluation map A : U -t R defined by
Au=u(4).
Consider then the sequence (Uk) where Uk(t):= (t/2)k. Clearly, in terms of the norm
defined above as k -t 00 uk -t 9u the zero polynomial. However,
A Uk=Uk(4)=2k -t 00. Hence A is not continuous. •
Self-Adjoint Maps
V x,ye H.
(x,AY)p=(Ax,y)p.
455
and
*
Proof. (a) From A Vi = Aivi and A = A, we obtain successively
(where in the last RHS, the overbar indicates the complex conjugate). It follows that
Ai is real.
(b) From A Vj=AjVj and A vk=Akvk we obtain
Aj(Vk,Vj}= (Vk,A Vj}
•
A7.3 Properties of the Adjoint
4S Fact. Let A be a linear continuous map from U into V, where U and V are Hil-
bert spaces. Then
Au=A**u forallueU.
•
A7.4. The Finite Rank Operator Fundamental Lemma
The following lemma is crucial for the study of controllability and leads to the
singular value decomposition of a matrix.
55 Lemma [Finite rank operators]. Let F=IR or (1:. Let (H,F'(-')H) be a Hilbert
space and consider pm as the Hilbert space (ptn,F,(-,>F"')' Let A : H F m be a continu-
ous linear map with adjoint A * : pm H. U .I.c.
56 A*A:H-->H
are continuous linear maps with AA * and A *A self-adjoint. Furthermore (see Fig.
A7).
57 a) H=R(A*)iN(A), pn=R(A)iN(A*);
61 N(A*A)=N(A), R(A*A)=R(A*).
•
458
A'"
--
Fig. A7. The orthogonal decomposition of the domain and codomain of a finite rank
operator A : H Fm and its associated bijections.
Comments. (X) Conclusions (a) and (b) of the theorem above display the following:
Let P: H -+ R (A *) be the orthogonal projection of H onto R (A *) and J: R (A ) -+ Fm
be the natural injection of R (A) into pm. V.l.c. the map A is depicted by the com-
mutative diagram of Fig. A8 where A IR(A") is a bijection of R(A"') onto R(A).
Note that, modulo restriction in the domain and codomain, A becomes bijective.
*
A similar diagram can be drawn for the adjoint A (do it).
*:
It is crucial to notice that AA pm -+ pm and hence, by (59), the study of the range
*
of A and the null space of A is equivalent to study of the range and null space, resp.,
of any (Hermitian) matrix representation M of AA *; (cf. controllability ... )
"I) [Tay.l,p.244]. If A V where V is a Hilbert space, then the theorem statement
applies with all range spaces replaced by their closures.
* * *
Proof of Theorem (55). A is continuous by Fact (45). AA and A A are obviously
self-adjoint and they are continuous as the composition of continuous maps. Hence
(56) holds. Moreover:
a) R (A *) cHand R (A ) c Fm are finite-dimensional subspaces, (for R (A *) note that
dim R (A *) s: dim Domain(A *) = m). Hence, by Fact (20), both R (A *) and R (A) are
closed subspaces of Hand Fm, resp. Therefore, by the orthogonal projection Theorem
(26):
:OF:
RIA·) AtRIA.) RIA)
and
xe R(A).1
xeN(A*).
66 A matrix U E FOX" is said to be unitary iff U·U = UU* = In' (equiv., the n
columns and the n rows of U form orthonormal bases of P). If F=R such matrix is
called orthogonal.
67 Unitary matrices preserve the inner product and hence length; more precisely, if
Ue Fnxn is unitary, then 'r/x,ye P, (Ux,Uy)p=(x,y) and hence IIUxI12= IIxll2,
(prove it).
460
70 Recall that a matrix AE F"xn is called Hennitian iff A=A*; moreover such
matrix A is said to be positive semidefinite iff (x,Ax)p V x E F".
Hennitian positive semidefinite matrices are associated with nonnegative quadratic
fonns [Nob.l,p425].
where A is a real diagonal matrix whose diagonal elements are the corresponding
eigenvalues Ai of A; in dyadic fonn the last equation reads
n
73 A= 1: AjUjut;
j=I
Proof of (a) and (b). From (44), we know that A being Hennitian has only real
eigenvalues and that if)., t:: Ak' then any eigenvector of )., is orthogonal to any eigen·
vector of Ak'
We establish the existence of a basis of orthononnal eigenvectors by using induction.
Step II. Pick an orthonormal basis [bJ: for M t ; then (u t ,b 2, ... ,b n) is a basis for F"
and ul.l M I with M I A-invariant and M 1= Sp[b 2 ,b 3, .... bn). By the second matrix
representation theorem of Chapter 4, with respect to this new basis A has the form
1...1 0 . 0
o
M
Step III. Repeat Steps I and II successively (n-l) times; the result is an orthonormal
basis of eigenvectors (u i ] and the representation of A with respect to that basis is
diag(A I,"-2, ...
,A,J Clearly P is the matrix whose ith column is Uj E P, i En. Equa-
tion (73) follows immediately from the orthonormality of the Uj's.
The proofs of (c) and (d) are left as an exercise.
•
We consider now F m and F n as Hilbert spaces and the matrix A E F mxn as a
linear continuous map A: F" P with as adjoint A * the Hermitian transpose of A.
Hence the conclusions of Lemma (55) apply with H and A replaced by F n and A.
resp.. In particular AA* E pnxm and A* A E F"xn are Hennitian positive semidefinite
matrices. For these we have the following.
75 Exercise [Common rank and nonzero eigenvalues of AA* and A*AJ. Let
A E pxn have rank r.
Show that the Hermitian positive semidefinite matrices AA* and A* A obey:
with
s.t.
where the ai' for i e r. are the square roots of the common positive eigenvalues of
A*A and AA*; the aj's for ie r. are called positive singular values of A.
d) A has a dyadic expansion
r
88 A=UILIVI*' or equivalently, A=Lajujvj*
j=1
where the ui' vi' for i e r. are the columns of U I and VI, resp.,
e) Ae pnxn has a singular value decomposition (s.v.d.)
89 A=ULV*
where
r
90
n-r
where the ai' i E L are those given in (92). From properties (82), especially the
columns of V I form a set of orthonormal eigenvectors asmciated with the nonzero
eigenvalues of A*A, A*AVI=VILf, whence (AVILj'I)*(AVILj'I)=Ir • This defines
an mxr matrix
Let ui' i E L be the columns of U I, then ui = O'i- l AVi' i E r. From the properties above
we obtain by calculation:
'V i,k E r,
95
which follow from it uses also the fact that the nonzero eigenvalues of A *A in
(92)-(93) are those of A A.
465
d) The dyadic expansion (88) follows from the first part of (95).
e) The singular value decomposition (89)-(90) follows because
A[V I :OJ=[U I
•
96 Definition. Let A e F mxn have rank r. Then the n nonnegative square roots cri
...
of the eigenvalues cr'f of A A are called the singular values of A. When ordered
according to
the first r singular values are called the positive singular values of A.
\lien.
Thus A[Sn]' the image of the unit sphere Sn under A, is an ellipsoid whose principal
axes are along the Uj'S and are of length OJ, viz. the positive singular values of A.
This is shown in Fig. A9 for the case that F=IR and n=2. The action of A consists
of, first, a "rotation" (Vi Ui), then a scaling (Ui Uicr), i.e. the size of the action of A
is dictated by its singular values.
466
Fig. A9. A e R2x2 is nonsingular, the unit sphere is on the left; the image under
A is the ellipsoid is on the right.
where H 1,H2 e ptxn are Hermitian positive semidefinite and W I'W 2 are unitary. «99)
is called polar decomposition of A, in analogy with a+jb = pel<il for a complex
number.)
(Hint: Consider A=ULU*UV* .... )
100 Sensitivity analysis of Ax =h. As above let F= R or tr. Let now A E ptxn,
detA '¢ 0, A=ULV*,
1. From (97) we have
and
whence
because, in the present case, A-1=VL-1U*. Thus, using 2-norms, the condition
number of A is
467
Thus, from (A6.89), we see that if 01;» On> 0, for some and the resulting
II Ih I II Xo 112 may be very large.
II. The smallest singular value on of A is a measure of how far the nonsingular
matrix A is from being singular. First, note that ujvtis an n x n matrix of rank 1
whose range is Sp[ Uj] and
II UjVj* Ib = 1 .
Furthermore, since
101
with M defined by
we see that
det(A+aA) = o.
Furthermore that particular aA is the aA with least norm, II aA 112, such that A+aA is
singular, (prove it). In conclusion, an measures how far A is from being singular.
III. The solution Xo of Ax = b is given by
It is clear from (103) that a of length £ along the vector un will cause the largest
change in x: Un gives the direction of maximum sensitivity of x for changes in b; and a
penurbation II ab 112 = £ in that direction causes a change in ax in x of length
468
118x 112 = Elan and in the direction vn. In conclusion. the SVD a) shows how close A
is from being singular. b) detennines the least nonn perturbation in A that makes A
singular (see (102»; c) gives a geometric interpretation of the effect on x by a pertur-
bation 8b (see (104»; and d) identifies the "worst" 8b.
APPENDIX 8
DIFFERENTIAL EQUATIONS
Since the opening and closing of switches and since square waves are common
occurrences in engineering we shall allow discontinuous functions of time in our
differential equations. Section B I discusses the existence and uniqueness of solutions;
Section B2, the dependence on initial conditions and parameter perturbations. Finally
in Section B3, we discuss briefly the concept of flow and numerical calculations.
Essential references are [Cod.l], [HaUl, [MiUl, [Die.I,Ch.xl.
81.1. Assumptions
Our basic differential equation (d.e.) under consideration is written as follows:
1 x=p(x,t),
where X(t)E 1R n, for t 0 and p(.,'): 1R n x 1R+ ---> 1R n. We are given "initial conditions"
(to,xo) and require that
2 x(to)=xo.
3 for each fixed x ERn, the function: t E IR+\D p(x,t) E IRn is continuous and for
any tED the left-hand and right-hand limits p(x,t-) and p(x, t+), resp., are finite vec-
tors in IRn.
b) There is a piecewise continuous function kO: R+ IR+ S.t.
4 II II k(t) II
This is called a global Lipschitz condition because it must hold for all and in IRn.
II p('I'(t),t) - p('I'(t),t) II
where k(-) being piecewise continuous is bounded in any compact interval [to,td c:::.
Rr. Therefore by (3) and the continuity of '1'0, it follows that for all tE IR+\D,
470
lim p(\jI(t),t) = p(\jI('t);t). Hence the function t --; p(\jI(t),t) is continuous at such 'to
t-4t
Now if 't E D then the inequality above still applies with p(\jI('t),'t} replaced by
p(\jI(t),t-), respectively, p(\jI('t),t+), which are well defined by (3). A similar reason-
ing shows that both one-sided limits exist, i.e. 'v' 't E D, with t increasing to 't
lim p(\jI(t),t) =p(\jI('t),t-) and with t decreasing to 't lim p(\jI(t),t) = p(\jI('t),'t+). It fol-
1-4t l-4t
lows therefore that for any continuous function \jI: R+ --; IRn the function t --; p(\jI(t),t)
is piecewise continuous on with discontinuity points in D.
Therefore, for any such \jI, we can integrate p(\jI(t),t) versus time. Also the func-
tion
t
f
t --; p(\jI('t),'t)d't
o
c) When (4) does not hold, it may happen that the solution cannot be continued
beyond a certain time. For example, the scalar equation lIc, c 0
has the solution = lI(c-t) defined on (-oo,c). As t --; c, I S(t) I blows up; we say
that we have a finite escape time at time c.
By imposing the global Lipschitz condition (4), (i) we can construct the solution
on and (ii) we greatly simplify the description of the iterations without losing the
key features of the reasoning.
5 Exercise. Show that, given R > 0, if there is a piecewise continuous k(') such
that
then the inequality (4) holds 'v' S,S' E B(en;R), 'v' t E R+.
6 Theorem [Existence and uniqueness of the solution of a d.e.]. Consider the d.e.
(1) under the initial condition (2). Let p satisfy conditions (3) and (4).
Then i) For each (to,xo) E R+ x R n there exists a continuous function 4> : R+ --; IRn S.t.
471
and
ii) This function is unique. The function cp is called the solution through (to,xo) of
the d.e. (1). •
x
In other words, given any (to,xo) e R+ x Rn, the d.e. = p(x,t) defines a unique solu-
tion cp(t) that is defined for all t in (This soluti?n is often written as cp(t,to,xo»'
The solution is continuous on R+ and, in addition, cp is continuous at al1 te R+\D,
because t -+ p(cp(t),t) is continuous at such.t. Now consider some. 1: e D, then t -+ «P(t)
i.s continuous but at 1:, the function t-+cp(t) jumps at 1: from c!>(1:-)=p(c!>(1:),1:-) to
cp(t+)=p(cp(t),t+), if p(cp(t),t-) "" p(cp(t),t+).
The proof of the theorem is in two steps; first, a solution is constructed by itera-
tion, and second, uniqueness is established.
9 f
"m+l(t) := xo + p(xm(1:),1:)d1: for te R+
10
with
xo(t) := Xo
Let [tl,t2] be any closed interval of containing to. We shall show that, on any
such the sequence of continuous functions Gm(')]; is a Cauchy sequence of
the Banach space [c ([tl,tv,lRn),R, 11'11 00 ] . where
and "." is any norm in IRn, see Fact (A6.33); (the norm in R n is arbitrary by
equivalence of norms, see (A6.47». By completeness of the Banach space, there is a
continuous function cp: JR+ -+ JRn to which the sequence [xm(·»); converges in 11'11 00
;
thus xm(t) -+ cp(t) uniformly on [tl,t2]' This function cp will be shown to be a solution
of the d.e. (1).
We start by studying estimates where 11'11 is any norm in R n and we use
(A6.20), which is valid for any such norm. To wit: for m=I,2,... and te JR+
472
I
II Xm+1 (t)-Xm(t) II = II p(xm('t),'t)-P(Xm_1 ('t),'t)d't II
10
It I
:j II p(xm('t),'t)-P(Xm_l ('t);t) II d't: (by (A6.20»
110 I
(by(4» .
Let k be the supremum of k(t) over [tl.t21. then for m=1,2 •... and for all tE [t l ,t2]
I[ I
11 IIxm+I(t)-x m (t)II :!> k:J IIxm('t)-Xm_I('t)lld't:.
110 I
where M is known since Xo is specified. Hence by (11), 'It t E [tl,t21, we obtain suc-
cessively
M[klt-tolf
II x3(t)-x2(t) II :!>
12
(triangle inequality)
p-) [ktr+k
(by (12»
S M I, (m+k)!
k=O
[ since ekT = I, - -
- 00 [k"T]k) .
k=O k!
the Banach space [CC[t),t2],Rn),]R, 11'11 00]. whence [XmO ); converges to a continu-
ous function cjl on [t l,t 2 ] in 11'1100'
Recall that [t 1,t2 ] is an arbitrary closed interval of containing to and note that
convergence in 11'11 00 implies pointwise convergence at every te [t),t21; indeed, for all
such t's, II II S II II 00.1[ ••1", -+ O. Therefore if <II and", are con-
.1
Therefore there is one continuous function cjI defined on R+ to which the sequence
(xmC') ]; will converge in 11'11 00 on any [t),t2] containing to·
474
We now show that 4> is a solution of the d.e. (1) on any such interval [t 1,t2) and
hence on R+. Consider therefore the iteration fonnula (9), namely
9 "m+1 (t) = Xo f
+ p(xm(t),t)dt
to
I t I
:f II p(xm(t);t)-p(4>('t),t) II dt :
I 10 I
I t I
: Ik(t) II xm(t)-4>(t) II dt :
I to I
f
4>(t)=xo + p(4)(t),'t)dt foraB tE [tl,t2)'
10
Hence (by the fundamental theorem of calculus) at every point tE [tl,tz) \ D, (where
p(4)(t),t) is continuous),
$(t) = p($(t),t)
475
Since the interval [t 1,t2] is an arbitrary closed interval of containing to' we con-
clude that the proposed iterative scheme converges to a solution cj> defined on 1R+.
It I
16 I f
u(t) S cl + I k(t)u(t)dt II 'ite
Ito I
then
17
{"f '}
u(t) S cl exp : k(t)dt:
I to I
'ite .
•
Proof. For reasons of symmetry we may restrict ourselves to the case that t) to.
Call U(t) the right-hand side of (16). Hence (16) reads u(t) S U(t) 'i t e R. Multiply
both sides by the nonnegative function
Integrating between to and t, noting that U(to)=Cl and that at t=to the exponential is
one, we obtain for all t to
476
•
18 Exercise. Let u('), $0, k(') be real-valued piecewise-continuous functions on
R+. Let u(·), $0 and kO be ) 0 on 1R;.. If uO satisfies
It I
u(t) f
$(t) + : k('t)u(t)dt: V tE ,
I to I
then
It [It 1]1
u(t) S;
I
f
$(t) + I $(t)k(t) exp
I I I
I fk(a)da I dt I
I 10 I 't I I
B1.5. Uniqueness
Let us return to the d.e. x=p(x,t). where p(.,) satisfies assumptions a) and b) of
B 1.1. Suppose there are two solutions $ and 'I' satisfying x = p(x.t) and
$(to) = 'I'(to) = xo· By integrating the d.e., we obtain
f
$(t)-'I'(t) = [p($(t).t)-pC'I'(t).t)]dt
10
As before. restricting our attention to any closed interval [t.,t 2 ] of IR+ containing
to and using the Lipschitz condition, we obtain
I I I
II$Ct)-'I'(t) II S; f
k : 11$(t)-'I'Ct) Ildt:
I 10 I
Hence by the Bellman-Gronwall inequality, taking u(t) = II «pCt) - 'l'Ct) II. we obtain
Since this holds for any ci by taking cl=O. we see that 1I<I>(t)-'I'(t)II =0.
477
'v'te [tl,t2]' Hence for each te [t1,t2]' cjl(t)=",(t). Hence the solution is unique on
[t1,t2]' Since [t1,t2] is an arbitrary closed interval of R+ containing to, the domain of
uniqueness can be extended to cover all Hence cjl(t) = ",(t), 'v' t e •
Heuristic Introduction
Suppose we have a system described by a d.e. that depends on m real parameters:
Call its solution "'00+5",(·). Using (2) and a Taylor expansion we obtain successively
3 Wo(t)+aw(t) = f("'o+a""t,zo+az)
where a) for i = 1,2,3, Dl denotes the derivative of f with respect to its ith argument
and "h,o.t." denotes the higher-order terms in az and a",. Dropping the h.o.t., using
(1) we have approximately
where the matrices A(t,Zo) and B(t,Zo) are known once the nominal solution "'o(t) is
known. So with denoting the state transition matrix associated with A(t,Zo),
478
we have
5 8",(t) = f <I>(t;t)B(t,Zo)d't . 8z
10
where we used the "approximately equal" symbol to remind ourselves that Eq. (5) is
an approximation caused by our dropping the h.o.1. in (3).
Perturbation Theorem
The heuristic derivation above gives us a feel for the nature of the result. Now in
order not to clutter the statement of the theorem we will assume that the Lipschitz
conditions hold globally, etc. The theorem below specifies the properties of the solu-
tion. ((10) and (11) below). and gives expressions for the derivatives of the solution
with respect to t,!o,Xo and zoo (see (12), (15) and (17) below).
7 x(t) = f(x(t),t;Zo)
where without loss of generality 'to=O and 'tn=T. For all T > 0, f(· .... ) is of class
C k on IRnx['ti-!,'tj]xB for ie n where at the boundary points 'tj the function value and
the values of the derivatives are defined to be the appropriate one-sided limits.
c) f(· .... ) is globally Lipschitz in x ; more precisely there is a piecewise-continuous
function k(·) : R+ such that
9 II - II :5 k(t) II II.
V.t.c.
10 i) "I (xo.!o.Zo) e IR n x 1R+ x B. Eq. (7) has a unique continuous solution defined
on say ",(t,!o.xo.Zo);
479
In the following. we abbreviate the solution specified in (i) by 'l'o(t). Consider the
variational equation; let U: x B -+ R nxn : (t,Zo) -+ U(t,Zo) be the (matrix) solution of
13
where the derivatives of f are both evaluated at ('I'o(t),t,Zo)' Note that the RHS of (13)
is of the form A(t,Zo)U(t.Zo)+B(t,Zo) where A(',Zo) and B(',zo) are known functions
that are piecewise C k- 1 with discontinuity points in D; let t -+ Cll(t,t;Zo) be the
corresponding state transition matrix, namely, the solution of
a
at [Cll(t.t;Zo)] = A(t.Zo)Cll(t.t;Zo), Cll(t.t;z,,) = I.
With these notations in mind we have (all partial derivatives below are evaluated at
(t,to.xo,Zo»:
15 iv)
16 v)
•
17 vi)
18 Comment. Conclusion (ii) follows from the C k version of the implicit function
theorem, [Die.l,Thm 10.2.3], applied to
t
Numerical Solutions
Under assumptions (1) and (2), the most obvious way of calculating an approxi-
mate solution of the differential equation is to use the forward Euler method. The idea
is that
where h is "small" and positive; essentially, we assume that the velocity of the state is
constant in [to,to+h).
481
More precisely the algorithm for computing the solution on the compact interval
[to,to4oT]
is as follows.
Step 2. = Xo
for i=O,I,2, ... ,m-1
Si+l = Si+hmP(Si,ti)
1j+l = 1j+hm·
•
The output is a sequence of m+1 Rn-vectors the interpretation is that
the calculated solution is the polygonal line joining to for
i = 0, 1,2, ... ,m-1.
It can be shown that (I) and (2) guarantee that, as m 00,
LAPLACE TRANSFORMS
4 I
00
Call O'f the infinum of alIa's satisfying inequality (4); O'f is called the abscissa of
absolute convergence of f(·).
·
For some f unctions, nne, e.g. tior eat, Of= a; tior 0 thers, Of= 00, e.g. et 2 :
. fi'
Of IS
such functions are not Laplace transfonnable.
The Laplace transfonn of f, [denoted by L[f] or f(s)], is defined for all s in the
open half-plane Re[s] > O'f by
J f(t)e-stdt .
00
S f (s) :=
(}-
6 Analyticity. The integral in (5) is a well-defined finite complex number for. all
Re[s] > Of. Furthennore since, for each fixed t, the integrand is analytic in s, f is
analytic in s for Re[s] > O'f, [Die.2,Thm13.8.6].
7 In the RHS of (5), we choose 0- to be the lower limit of the integral; this is
because in engineering we often consider generalized functions such as
f All integrals are Lebesgue integrals. In most practical cases it does not mauer; the standard
Riemann integral suffices.
483
n
8 f(t) = fe(t) + L CjS(Hj)
j=O
where fe is locally integrable with Of. < 00. S(·) denotes the Dirac delta function.
O=to < tl < ... tn. and the c{s are in R; then. since we integrate from 0-. we
obtain from (5)
n
+ L Cie-st. .
A
9 f(s)=fc(s)
i=O
b) For k=l •...• n. let 7tk(t) be a polynominal in t of degree mk-1 and let Ake Cl:;
finally let
L [1:
k=1
7tk(t)eAt.1 1=: £(s) .
Then £(s) is a strictly proper rational function in s with a pole of order mk at Ak}or
k= I •...• n. It is easy to see that. in this case. the defining integral (5) defines f(s)
only for Re[s] > max ( Re[Ak] ). It is well known that f(s) may be continued
k •
analytically so that the rational function f (s) is defined for all s in Cl: except at its
poles AI.···. Am .
12 Exercise. Let f(t) = 1(t)e1sin(el ). i.e. f(t) oscillates with an exponentially increas-
ing amplitude and with an exponentially increasing frequency.
a) Use the defining integral to show that Of= 1.
b) Set t=el and use successive integration by parts to establish that £(s) is analytic in
the whole plane! (It has. in fact, an essential singularity at infinity.)
Thus this example shows that f may be analytic in the whole plane and have f(t)
unbounded as t 00.
15 [(s) =
k=O
As is expected from O"f= 0, f (s) is analytic in Re[s] > 0, but f has ,!Jranch points
dense on the jo>-axis so that it is impossible to extend the definition of f into the left
half-plane.
f I f(t) I dt < 00
o
f If(t)-g(t)ldt=O,
o
i.e. whenever f and g differ on a set t of measure zero (equivalently, whenever f=g
almost everywhere). The resulting set of equivalence classes also forms a linear space;
furthermore it has a norm
2 II fill := f I f( t) I dt.
o
The normed space thus defined is denoted by LI (R+), or L I. It can be shown that L 1
is complete, i.e. Ll is a Banach space; indeed, any Cauchy sequence of LI functions
has, as limit, an L 1 function.
(ii)
or equivalently,
6 I £(s) I -+ 0 as I s I -+ 00 in (t+;
and
= = alf
. .
.
9 Differentiation. Let f: 1R+ -+ ([ and let f denote its derivative taken in the dis-
tribution sense. (In particular if f(·) has a finite "jump" at to from f(Io-) to f(Io+), then
f(.) includes the tenn [f(to+) -f(to-)]· 8(t-lo) .) If f is Laplace transfonnable with
abscissa of absolute convergence O"f' then, for Re[s] > 0"(
L[O = s[(s)-f(o-) .
L[f*g] = f(s)·g(s).
11 Inversion integral. Let f(s) be analytic in Re[s] > 0"(. If, in the neighborhood
of t, f is of bounded variation, then t, for 0" > 0"(,
a+joo
12 2- 1 ·[f(t+)+f(t-)] = (21tj)-I. f f(s)eS1ds.
a-joo
In applications, it is important to note that the inversion integral gives the average
of f(t+) and f(t-): for example, since L [I (t)] = s-I, for t=O and a > 0
a+joo
13 Z-l = (21tW' f s-l'ds, so f(0)=2-'.
a-joo
14 Jordan Lemma [Doe.I,VoU, p.224J. Let rl denote the left half-plane semicir-
cle of radius R centered on the origin:
lim r f (s)estds = °.
R--+oo f,
°
15 Application. For f(s)=s-', for t > and cr > 0, the evaluation of (12) using
the Jordan lemma is done as follows: a) close the vertical integration path from a-jR
to a+jR by the left half-plane semicircle of radius R and centered at the origin; b) now
use (14): as R 00, this closed contour integration tends to the integral required by
(12); c) note that the integrand is analytic in (C except for the pole at s=O; d) use
Cauchy's theorem, note that the residue of the integrand at s = 0 is I, hence obtain
L-'[S-l]=1 forallt > O.
16 Initial value theorem. If, as t decreases to 0, f(t) has afinite limit f(O+), then
Examples show that the RHS limit may exist but that f(t) tends to no limit as t
decreases to 0, [Doe.l,p. 476].
18 Final value theorem. If, as t 00, f(t) tends to finite limit f(oo), then
23
where the polynominals nand d are coprime. For 1s 1 sufficiently large, we may
f(s) in a power series in s-l: let, for k= 1, ... ,n, Pk denote the kth pole of
f('), then, for any p > max 1Pk I, this power series converges absolutely and uni-
k
formly for all I s I > p. (Note that the power series is easily obtained by long divi-
sion.) Let
24 res) = bo+CXos-l+ttls-2+tt3s-3+ ... +ttks-(k+l)+ . . . .
Using the inversion integral with (J p, and integrating term by term we obtain
Since for t > 0, f is a sum of polynominals in t times exponentials the power series
(25) converges absolutely for each t > O.
THE z-TRANSFORM
In many engineering applications signals are sampled periodically, say, with sam-
pling period T. Thus the continuous-time signal is replaced by a sequence of real
numbers: fo , fl .... , fn , . .. . We denote this sequence by (fn );.
f
i.e. fez) is specified as a power series in z-I. In order for the expression (1) to mak
sense, the power series must have a finite radius of absolute convergence; i.e.
Pf < 00, where Pc is the least nonnegative number such that the power series in (I)
converges absolutely for all I z I > Pc.
The z-transfonn maps sequences f= (fn];, for which the series (1) converges abso-
lutely for some finite z, into functions of the complex variable z.
We write
d) For = (nan); = (0. a. 2a2 • 3a3 • . . • • nan •... ). f(z) = az(z-ar2 and
Pr= 1a I.
e) For [fn); = (1. e. e4 • e9 • . • . • en\ ... ]. Pr=oo and the sequence has no z-
transfonn.
has a z-transfonn fez) which is a proper rational function with poles Ak of order mk in
the closed disk D(O,Pr) := {z e (C: I z I S Pc ) where Pr := max I Ak 1.
k
1 In the following, we consider exclusively sequences for which the power series
(Dl.l) converges absolutely for a finite z:f(z) = fnz-n.
o
The classical theory of power series [Hil.l Ch. 5] [Con.1. p. 31] yields the following
facts:
3 If 1 z 1 > Pr, then the power series converges absolutely and the sum is an ana-
lytic function of z defined for all 1z 1 > Pr.
4 For 1 z 1 > Pr. derivatives of fez) of any order may be obtained by differentiating
the series tenn by tenn and summing the results.
5 If Iz 1 < Pr, then the tenns of the series become unbounded as n -+ 00; hence
490
the series diverges.
6 For any P > Pc, the series (1) converges uniformly in I z I 2:. P .
The following properties are very useful in applications; they follow easily from the
definition (1).
7 Linearity. For all e € and for all sequences (fn) , (gn) with finite radius of
convergence Pc and Pg , resp.,
=
= .
9 Delay by k steps. Let us extend the sequence (fn]; to the left by setting C n= 0
for all integers n > O. Let keN. let rfn_k] - be the sequence (fn) delayed by k
r.
steps. i.e. the sequence lO.O •...• O.fo.fl •...
n;()] ; then
and
491
where r is any closed rectifiable curve in I z I > Pc that encircles the origin once in
counterclockwise sense, (equivalently, n(r,O)= I, the index of r with respect to 0 is
equal to one, [Con.I,p. 81]).
f lim (z-l)f(z) ,
00
z-+ 1
Che. 1 C.T. Chen, "Linear System Theory and Design," Holt, Rinehart and Winston,
New York, 1984.
Che. 2' M.J. Chen and C.A. Desoer, "Necessary and Sufficient Conditions for Robust
Stability of Linear Distributed Feedback Systems," Int. J. Contr., Vol. 35, pp.
255-267, 1982.
Cod. 1 E.A. Coddington and N. Levinson, "Theory of Ordinary Differential Equa-
tions," McGraw-Hill, New York, 1955.
Con. 1 J.B. Conway, "Functions of One Complex Variable," 2nd ed., Springer-
Verlag, New York, 1978.
Cop. 1 W.A. Coppel, "Matrix Quadratic Equations," Bull. Austral. Math. Soc., Vol.
10, pp. 377-401, 1974.
Cur. 1 R.F. Curtain and A.J. Pritchard, "Infinite Dimensional Linear Systems
Theory," Springer-Verlag, Berlin, 1978.
Del. 1 D.F. Delchamps, "State Space and Input-Output Linear Systems," Springer-
Verlag, New York, 1988.
Des. 1 C.A. Desoer, "Notes for a Second Course on Linear Systems," Van Nostrand
Reinhold, New York, 1970, (paperback).
Des. 2 C.A. Desoer and K.K. Wong, "SmaIl-Signal Behavior of Nonlinear Lumped
Networks," Proc. of the IEEE, Vol. 56, pp. 14-22, 1968.
Des. 3 C.A. Desoer and M. Vidyasagar, "Feedback Systems: Input-Output Proper-
ties," Academic Press, New York, 1975.
Des. 4 C.A. Desoer and W.S. Chan, "The Feedback Interconnection of Linear
Time-Invariant Systems," J. Franklin Inst., Vol. 300, pp. 335-351, April
1975.
Des. 5 C.A. Desoer and MJ. Chen, "Design of Multivariable Feedback Systems
with Stable Plants," IEEE Trans. Auto. Control, Vol. AC-26, pp. 408-415,
1981.
Des. 6 C.A. Desoer and C.A. Lin, "A Comparative Study of Linear and Nonlinear
MIMO Feedback Configurations," Int. Jour. Systems Sci., Vol. 16, No.7, pp.
789-813, 1985.
Des. 7 C.A. Desoer and A.N. Gundes, "Algebraic Theory fo Linear Time-Invariant
Feedback Systems with Two-Input Two-Output Plant and Compensator," Int.
Jour. of Control, Vol. 47, No. I, pp. 33-53, January 1988.
Die. 1 J. Dieudonne', "Foundations of Modern Analysis," Academic Press, New
York, 1969.
Die. 2 J. Dieudonne', "Treatise on Analysis," Vol. II, Academic Press, New York,
1970.
Doe. G. Doetsch, "Handbuch der Laplace-Transformation," 3 Vols., Birkhaiiser.
Basel, (German), 1971.
Doe. 2 G. Doetsch, "Introduction to the Theory and Application of the Laplace
Transformation," Springer-Verlag, Berlin, 1974.
494
Don. J. Dongarra, J.R. Bunch, C.B. Moler, and G.W. Stewart, "LINPACK Users
Guide," SIAM Publications, Philadelphia, 1978.
Doy. J.C. Doyle and J.C. Stein, "Multivariable Feedback Design: Concepts for a
Classical/Modern Synthesis," IEEE Trans. Auto. Control, Vol. AC-26, pp.
4-16, February 1981.
Eng. 1 J.C. Engwerda, "Stabilizability and Detectability of Discrete-Time Time-
Varying Systems," IEEE Trans. Auto Control, Vol. AC-35, pp. 425-429,
1990.
Fie. 1 W.H. Fleming and R.W. Rishel, "Deterministic and Stochastic Optimal Con-
trol," Springer-Verlag, New York, 1975.
Fuj. 1 T. Fujii, "A New Approach to the LQ Design from the Viewpoint of the
Inverse Regulator Problem," IEEE Trans. Auto. Contr., Vol. AC-32, No. 11,
pp. 995-1004, November 1987.
Gar. 1 B.S. Garbow, J.M. Boyle, J.1. Dongarra and C.B. Moler, "Matrix Eigensys-
tem Routines: EISPACK Guide Extension," Springer-Verlag, New York,
1972.
Glo. 1 K. Glover, "All Optimal Hankel-Norm Approximations of Linear Multivari-
able Systems and Their Bounds," Int. Jour. Control, Vol. 39, pp.
1115-1193, 1984.
Gol. 1 G. H. Golub and C.F. Van Loan, "Matrix Computations," Johns Hopkins
University Press, Baltimore, 1983.
Gol2 G.H. Golub and J.H. Wilkinson, "Ill-Conditioned Eigensystems and the Com-
putation of the Jordan Canonical Form," SIAM Review, Vol. 18, pp. 578-
619, 1976.
Gold 1 R.R. Goldberg, "Fourier Transforms," Cambridge University Press, Cam-
bridge, U.K., 1961.
Gru. 1 W.A. Gruver and E. Sachs, "Algorithmic Methods in Optimal Control," Pit-
man, Boston, 1981.
Gun. 1 A.N. Giindes and C.A. Desoer, "Algebraic Theory of Linear Feedback Sys-
tems with Full and Decentralized Controllers," Springer-Verlag, New York,
1990.
Hal. 1 J.K. Hale, "Ordinary Differential Equations," Wiley, New York, 1969.
Hal. 2 J.K. Hale, "Functional Differential Equations," Springer-Verlag, New York,
1975.
Hil. 1 E. Hille, "Analytic Function Theory," Vol. I, Blaisdell, Waltham, Mass.,
1959.
HoT. I.M. Horowitz, "Synthesis of Feedback Systems," Academic Press, New
York,1963.
Hor. 2 R.A. Horn and C.A. Johnson, "Matrix Analysis," Cambridge University
Press, Cambridge, U.K., 1985.
495
Hsu. 1 C.H. Hsu and C.T. Chen, "A Proof of the Stability of Multi-Variable Feed-
back Systems," Proc. IEEE, Vol. 56, No. 11, pp. 2061-2062, 1968.
In. 1 M.K. Inan, "On the Perturbational Sensitivity of Solutions of Nonlinear
Differential Equations," Memo. No. ERL-M270, Electronics Research
Laboratory, University of California, Berkeley, 1970.
Jac. 1 N. Jacobson, "Basic Algebra I," Freeman, San Francisco, 1974.
Jur. 1 E.I. Jury, "Theory and Application of the z-Transform Method," Wiley, New
York,1964.
Kal. 1 R.E. Kalman, "Mathematical Description of Linear Dynamical Systems,"
SIAM Jour. Control, Vol. I, pp. 152-192, 1963.
Kag. 1 B. Kagstrom and A. Ruhe, "An Algorithm for Numerical Computation of the
Jordan Normal Form of a Complex Matrix," ACM Trans. Math. Software,
Vol. 6, pp. 398-419, 1980.
Kai. 1 T. Kailath, "Linear Systems," Prentice Hall, Englewood Cliffs, NJ, 1980.
Kat. 1 T. Kato, "A Short Introduction to Perturbation Theory for Linear Operators,"
Springer-Verlag, New York, 1982.
Kha. 1 V.L. Kharitonov, "On a Generalization of a Stability Criterion," Izvestia Aka-
demii Nauk Kazakhskoi SSR, Seria Fiziko-Matematicheskaia, Vol. I, pp.
53-57, 1978 (in Russian).
Kha. 2 V.L. Kharitonov, "Asymptotic Stability of an Equilibrium Position of a Fam-
ily of Systems of Linear Differential Equations," Differential Equations, Vol.
14, pp. 1483-1485, 1979.
Kuc. 1 V. Kucera, "A Contribution to Matrix Quadratic Equations," IEEE Trans. on
Automatic Control, Vol. AC-17, pp. 344-347, 1972.
Kwa. 1 H. Kwakernaak and R. Sivan, "Linear Optimal Control Systems," Wiley,
New York, 1972.
Kwa. 2 H. Kwakernaak and R. Sivan, "Modem Signals and Systems," Prentice Hall,
Englewood Cliffs, 1990.
Lau. 1 A. Laub, "A Schur Method for Solving Algebraic Riccati Equations," IEEE
Trans. on Automatic Control, Vol. AC-24, pp. 913-921, December 1979.
Loo. 2 L.H. Loomis and S Sternberg, "Advanced Calculus," Addision-Wesley, Read-
ing, Mass., 1968.
MacL. 1 S. MacLane and O. Birkhoff, "Algebra," 2nd Ed., McMillan, New York,
1959.
Man. 1 M. Mansour and F. Kraus, "Strong Kharitonov Theorem for Discrete Sys-
terns," Proc. 27th IEEE Conference on Decision and Control, pp. 106-111,
December 1988.
Mil. 1 R.K. Miller and A.N. Michel, "Ordinary Differential Equations," Academic
Press, New York, 1982.
Min. 1 R.J. Minnichelli, J.J. Anagnost, and C.A. Desoer, "An Elementrary Proof of
Kharatinov's Stability Theorem with Extensions," IEEE Trans. on Automatic
496
Frequently used mathematical symbols are defined briefly in the five listings below.
1. Set theory
2. Sets
3. Algebra and linear spaces
4. Analysis
5. System theory
1. Set Theory
B(O,p) the open ball of the complex plane centered at 0 of radius p, [B(9,p)
denotes an open ball of a linear space].
field of complex numbers.
:= { s E (£: Re s 0 ) ; equiv. the closed right half of the complex plane.
:= { ZE 0:: I Z I < 1 } , equiv. the open unit disk of the complex plane.
the set of integers (1 ,2, ... ,k).
set of nonnegative integers, namely, (O,1,2, ... ).
field of rational numbers.
field of real numbers.
:= { x E R : x 0 } ; equiv. the set of nonnegative numbers.
an undesirable subset of 0: that is symmetric w.r.t. the real axis and
contains (£ +.
Z ring of integers; equiv. ( ... ,-2,-1,0,1,2, ... ).
[a,b], (a, b) a x b, resp. a < x < b.
[a,b), (a,b] a x < b, resp. a < x b.
4. Analysis
5. System Theory