Professional Documents
Culture Documents
Abshlct-in this paper, algorithms are presented for obtaining rela matrix and the inverse of another polynomial matrix, e.g.,
tively prime matrix fraction descriptions WD's) of Linear time-invariant any rational function matrix W(s) can be written as
systems. orthogonalcoordinate translormations are used to reduce a given
state-space model to one whose state matrsX is in "block Hessenberg"
form. The observabiity (controllability) index= of the system are also W(s)=N(s)[d(s)z]-'=[d(s)l]-'N(s) (1.2)
obtained in the process. A recorsive algorithm is then desaIbed for
obtaining a relatively prime MFD from the block Hessenberg representa- where d(s) is theleastcommondenominator of allthe
tion. It is then shown that some simplifcation can be made in the d v e
terms of W(s).The difficulty arises when one is interested
algorithm by f d redacing the block Hessenberg form, by means of a
nolrsingalar ( triangular)coordinate bansformation, to a more compact in relatively
a prime MFD. The approachusedby
form, such as the %I& Fmbenius" form. A permutation of the state Rosenbrock [ 11 requires elementary row operationson the
variables of the block Frobenias form yields a canonical representatioa matrix [N(s) d(s)l] involvingmultiplication and divi-
similar to the Luenberger canonical form. Some nnmerlcal and other sion by polynomials. The method is therefore difficult to
p~~ofthealgorWmsarediscussed,andtheuseofthealgorithmsis implement on a computer and can be numerically unsta-
Wmtmted by nnmerlcal examples. The numerical performance of the
algorithm¶ is also compared with that of the struchn;e algorithm of ble. In [6]-[lo] the problem of solving a proper rational
Wolovich and Falb. functionmatrixequation is castinterms of findinga
minimal polynomial basis (see [7] for definition) for the
right or left nullspace of a polynomialmatrix-the
INTRODUCTION
I. minimal design problem (MDP). This is particularly rele-
T iswell known [l], [2] that a strictly proper rational vant to our problem in that the determination of relatively
I function matrix W ( s )can be represented by the follow-
ing matrix fraction description (MFD):
prime MFDs canalsobeformulatedas
example, starting with W ( s )in the form
an MDP. For
(1.2), we can find
a set of relatively right prime factors NR(s)and DR(s)by
W(s)=N,(s)D,-'(s)=D,-'(s)N,(s) (1.1) solving
where NR(s),DR(s) are relatively right prime and NL(s), [d(s)l]-'N(s)=NR(s)o,-'(S),
D,(s) are relatively left prime polynomial matrices with
degree (NR(s)) <degree (DR(s)) and degree (NL(s)) < 1.e.,
degree (D,(s)). Such MFDs can be regarded as exten-
sions to the multivariable caseof the classical single-input,
single-outputtransferfunctionswithrelativelyprime
numerator and denominator polynomials. In recent years,
In otherwords, we can obtain arelativelyrightprime
] =o.
Authorized licensed use limited to: University of Leicester. Downloaded on September 3, 2009 at 10:06 from IEEE Xplore. Restrictions apply.
PATEL: MATRIX FRACTION DESCRIPTIONS OF TIME-INVAIUANT SYS"S I49
solving the MDP operate on the transfer function descrip- remarks concerning the
three algorithms and their
tion of a system. It is well known that conversion of a numerical properties. Several examples are given in Sec-
state-space description to a transfer function description tion V to illustrate the use of the algorithms and, more
can introduce numerical instability and is by no means a importantly, to compare the proposed methods for com-
computationally easy task. The converse, i.e., obtaining a puting relatively left prime MFD's with the structure
(minimal order) state-space realization fTom a transfer algorithm of Wolovich and Falb [2], [ 111. We conclude in
function description, is muchmore straightforward and Section VI with some comments about the results of this
can be achieved in a numerically stable manner [12]-[14]. paper.
Therefore, in general, for computing MFD's, and for that
matter systempoles and zeros,it is preferable to have Notation
algorithms with good numerical properties which operate
on state-space descriptions rather than transfer function Throughoutthepaper,unless stated otherwise,all
descriptions. matrices are real (with elements in the field of real num-
Wolovich and Falb [2], [ l l ] obtain a relativelyprime bers W). The notation Z' is used to denote the transpose of
MFD from a state-space representation of W(s).This is a matrix Z , and wherever matrices are partitioned in
the method which is most commonly used for computing blocks, the individual blocks w l
i be identified by sub-
MFD's since it is the easiest to implement on a computer scripts i , j, i = 1,2; -,k , j = 1,2,-- - ,k. -
Therefore, by the
and can be used for high-order systems. However, it has a matrix Zi,, we mean the matrix in the ith "block row" and
disadvantage in that the state-space system must first be j t h "block column" of Z . The algorithms to be described
transformed to the Luenberger canonical form [ 151 which in the following sections involve sequential operations on
can be quite sensitive numerically. matrices. The matrices at each sequential step are identi-
In this paper we present algorithms for obtaining rela- fied by subscripts i= 0, 1, * ,k, except for partitioned - -
tively left prime MFD's for systems in the state-space matriceswhere the notation is superscript ( i ) . The no-
form tation I, denotes the n X n identity and the Z, is used to
denote an IX I diagonal matrix with elements u1> u2 >
f(t)=Ax(t)+Bu(t) (1.3a) u, 2 * > a, >0 along the diagonal.
Authorized licensed use limited to: University of Leicester. Downloaded on September 3, 2009 at 10:06 from IEEE Xplore. Restrictions apply.
150 EEE TRANSACTIONSON AUTOMATICCONTROL, VOL. AC-26,NO. 1, FEBRUARY 1981
Thedecomposition(1.4) is called the singularvalue the ith row and j t h column of 2. Similarly, 2 is an upper
decomposition (SVD). Algorithms for computing the SVD, Hessenberg matrix if zij =0, i > j 2. +
i.e., the matrices M , T, and X,, havebeen described By analogy with Definition 1.5, wegivethefollowing
extensively in the numerical analysis literature (see, e.g., definition for “block Hessenberg” matrices.
[ 161-[181). It suffices to mention here that numerically Definition 1.6: An n X n matrix Z is a lower block Hes-
stable algorithms exist for computing the SVD and that +
senberg matrix if Zj, = 0, j > i 2. Similarly, Z is an upper
use of the SVD greatlyalleviates the problem of rank block Hessenberg matrix if Zii =0, i >j 2. +
determination [ 1617 [ 191- [2 11 since only orthogonal
matrices are used to transform a given matrix whose rank
is required to a diagonal matrix. On a digital computer, 11. A BLOCKHESENBERG FORM
FOR S( A , B, C )
the computation of the SVD is, of course, subject to In this section we present an important result obtained
rounding errors so that the resulting Adecompositionis recentlyuponwhich the numerical algorithms will be
exact for some perturbed matrix Z = 2 + A, where based. The result is concerned with applying a nonsingu-
11 A, \I/ 11 Z (1 is a modest multiple of the machine precision lar coordinate transformation on the state vector of the
[ 191, i.e., state-space model S ( A , B, C ) to reduce it to a model
n S ( F , G , H )where the matrix F is a lower (upper) block
II A, II zII
€2 = +A II Hessenberg matrix and H ( G ) has the form
where 9 is the machine precision and +, is a function of p
and n.
From the definition of SVD, it is clear that the number
of nonzero singular values of Z determines its rank. How- The sigmficance of the result from the point of view of
ever, since in practice the computed SVD is exact for a computation lies in the fact that this coordinate transfor-
perturbed matrix 2, the problem of deciding the rank of Z mation can be achieved by means of orthogonal matrices.
is not so straightforward.Let us denote the singular values In this paper we shall state the results for the lower BHF
-
of Z by 4, i = 1,- ,p with 6l > -
2 4 2 0. In general, only. The corresponding results for the upper BHF can be
all the computed values for e;, i = 1;. , p wdl be per- - obtained by duality.
turbed so that rank z = p . The problem is to choosea Theorem 2.1: Givena state-space model S ( A , B , C ) ,
“zero threshold” such that all less than or equal to this there exists an orthogonal matrix T such that
0 ... 0 0
0 ... 0 0
F;4 ... 0 0
(2.la)
given in [20]. For more details concerning the use of SVD p X I , matrix. The integers I i , i-0; ‘ ,k are defined by
in the problem of rank determination and other applica-
I, = rank ( C )
tions, the reader is referred to [21].
In the rest of this paper, unless stated otherwise, it will li=ra~(E;,i+l),
be assumed that all rank determinations are for numerical
rank defined above, and are carried out using the SVD. and k is the smallestinteger
Definition 1.5: An n X n matrix Z is a lower Hessenberg I, . . + I k =n. Furthermore, +
matrix if zj, =0, j > i + 2 where zji denotes the element in 1,. * , k and H I have the form -
Authorized licensed use limited to: University of Leicester. Downloaded on September 3, 2009 at 10:06 from IEEE Xplore. Restrictions apply.
PATEL: MATRIX FRACTION DESCRIPTIONS OF TIME-INVARIANT SYSTEMS 151
and
v = v , + v ~ + * *+ .V , = Z ~ + Z ~ + - . - +lk. (2.5)
where &, is an orthogonal matrix.
The followingtheoremestablishes an important result
Remarks concerning these integers. A detailed proof of the theorem
can be found in [27] and will not be repeated here.
The above result (with some slight variations) has been
obtained in recent years, apparently independently in
Theorem 2.2: The integers vi, j= 1, ,p are the ob- ---
servability indices of the system S ( A , B yC ) .
connection with different applications byseveralre-
searchers[13],[22]-[25]. The algorithms in [13] and [23]
Remarks
(Algorithm1) use the SVD to generate the orthogonal
transformations to obtain the BHF,whereas in [22] 1) The integers li, i = 0,. ,k can be calculated from vi, --
Householder transformations are used. In [24] the result is j= 1,. ,p as follows:
obtained by carrying out elementary row and/or column
operations on a “tableau” formed from the given state- li=O, i=O;-.,k
space description. The algorithm in [25] uses nonorthogo-
Zi=Zi+l, i = O , . - . , v j - l , j = l , - - . P. (2.6) Y
nal transformations which are generated by reductions of
the form This clearly shows that the integer k 2 v 1 - 1 wherev 1 is
the largest observability index of S ( A , ByC ) . If k > v , - 1,
then ZV1= ZV1+ = , --
= = 0. In other words, k = v , - 1 is
the largest value of k such that zk Po.
2) FromTheorem 2.1we note that k is the smallest
where P and Q are square nonsingular matrices. In [25] integer such that v =n or l k = 0. If v =n, then Theorem 2.2
the BHF is called the “generalized Hessenberg representa- implies that S ( A , B, C ) is observable. If, on the other
tion.” hand, l k = O , then F k , k + , =o and from @.la) and (2.lb)
To date the BHF has beenusedinseveral different we see that Fk+1, k + denotes the unobservable modes of ,
applications, e.g., incomputing Kronecker‘s canonical the system. Furthermore, it follows directly from Theorem
form for singularpencils[22], in obtaining irreducible 2.2 that these are the only unobservable modes of
state-space realizations in anumerically stable manner S ( A , B YC).
[ 131, [ 141, in computing canonical forms [23], [24], and in 3) The relationship between the reduction to the BHF
model reduction [25]. It should be noted that reduction to and some results of Silverman and Payne [26] has been
the BHF is not unique, although as pointed out in [22], pointed out by Tse et al. [25]. Theresults of Silverman
[23], and [25], there are certain invariant, and therefore and Payne identify the unobservable subspace by succes-
unique, features relating to this representation. A some- sively decreasing the null space of Cywhereas the reduc-
whatdifferentblockHessenberg structure hasbeen ob- tion to BHF obtains the observable subspace by succes-
tained by Van Dooren [ 131 in connection with the compu- sively extending the range space of C.
tation of suprema1 ( A , B)-invariant and controllability In Theorem2.1wehavegivena lower BHF for
subspaces contained in ker C. The algorithms employed in S ( A , B , C ) . By duality we can obtain an upper BHF
e,
[13] for this purpose are very similar in principle to the S(k, I?) where P is an upper block Hessenberg matrix
“structurey’ algorithm of Silverman and Payne[26]al- and [ 6; i 01’. e=
though, of course, the latter is not numerically oriented.
Thealgorithmsin[13],[22], and [23]giveways of Comments on Reduction to the BHF (Algorithm I )
constructing the orthogonal transformation matrix T and
matrices F, G , and H , thereby providingproofby con- As mentioned earlier, reduction to the BHF is not
struction of Theorem 2.1. We shall therefore not give a unique and canbeachieved in several different ways
formal proof of this theorem here. However, for reference using orthogonal as well as nonorthogonal (elementary)
later on in this paper, we shall denote by AZgorithm I the transformations. Orthogonal transformations areto be
reduction to the BHF and assume henceforth that the preferred because of their attractive numerical stability
reduction has been carried out using Algorithm 1 in [ 131 property [16],[28]. In [13] and [23] the orthogonal trans-
or [23]. formations are generated by means of the SVD and in [22]
Authorized licensed use limited to: University of Leicester. Downloaded on September 3, 2009 at 10:06 from IEEE Xplore. Restrictions apply.
152 IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. AC-26, NO. 1, FEBRUARY 1981
byHouseholder transformations. Sincerank determina- and observable state-space model S(F, G , H ) in lower
tion plays an important part in the reduction to the BHF, BHF, i.e.,
the use of the S V D isrecommended as it is the most - -
stable and perhaps the only reliable method for determin- 4 1 4 2 0 ... 0 0
ing the numerical rank of a matrix. However, it requires F21 4 2 F, ... 0 0
more computingtime than other methods such as F= :
Householder or stabilized elementary transformations [28],
[29] whichmay be used as alternatives to the SVD Fk3 in Fk2 Fkl
.. .
Fk,k+l Fkk
Authorized licensed use limited to: University of Leicester. Downloaded on September 3, 2009 at 10:06 from IEEE Xplore. Restrictions apply.
PATEL: MATRIX FRACTION DESCRIPTIONS OF TIME-INVARIANT SYSTEMS 153
detailed proof of the theorem can be found in [27]. The But from (3.9)
proof is rather involved and we shall not repeat it here.
Theorem 3.1: The polynomial matrices D(s) and N(s) deg(det(D(s))}=deg{det(l(s))}
in Algorithm 2 are the factors of a (left) MFD of
S(FYG , H ) .
+deg{ det( E ( $ ) ) ) .(3.1 1)
The other result which we need to establish concerns Therefore, if deg{det(L ( s ) ) }> 0, then
therelativeprimeness of thepolynomialmatrices D ( s )
and N(s). deg{det(fi(s))}<deg{det(D(s))} =v (3.12)
Theorem 3.2: The MFD of S(F, G , H ) obtained using
Algorithm 2 is relatively left prime. and we have a contradiction. Hence, deg{det(l(s))} =O,
Pro08 We shall prove the theorem by showing that implying that L ( s ) is a unimodular matrix. This completes
thegreatest common leftdivisor of D ( s ) and N ( s ) is the proof of the theorem.
unimodular,
Let $(Z(s)) denote the highest degree of s in the j t h - Discussion of AZgorithm 2
row of a polynomialmatrix Z(s). Then in Step 2 of
Algorithm 2 we note that The main step in Algorithm 2 is Step 2. This step
involves the recursive computation of the coefficients of
$(D,(s))>$(N,(s)), j = l , - . - , Z , ; q = k , k - l , - - . , O . the polynomial matrices P,(s), Dq(s),and N,(s),q=k, k-
(3.5) 1,. - . ,O. At first sight the computation may appear rather
complicated and unattractive, but on closer examination
Note that for Z, < jQ Z4- ( I - = p ) , the j t h row of Nq(s) is we see that the only operations involved are addition and
zero and the only nonzero entry in the j t h row of D,(s) subtraction of polynomial matrices and postmultiplication
occurs in the j t h column which contains a 1. It is also of polynomial matrices by constant matrices. This can be
clear from this step that accomplished on a digital computer by addition and sub-
Comparing (3.6) with (2.4), we get traction of the appropriate coefficientmatrices of the
polynomialmatrices and postmultiplyingthecoefficient
$(D,(s))=v,, j = l , . . . P.
Y (3.7) matrices by constant matrices. Notethat thematrices
Denote by [Dq],, q=O, l;.. , k the Z 4 - l x Z ~ - high-order
~
involved are of dimensions Ii x Zj and Zi x m yi=O, * ,k, -
j = O , - - . , k where Z, +Il +... + / k = n , the order of the
constant coefficient matrix of the polynomial matrixDq(s),
state-space model S(F, G , H ) . Also, the degree of the
i.e., the j t h row of [D,], consists of the coefficients of the
polynomial matrices involved is between 1 and k = v1 - 1
$(D,(s))power of s in thej t h row of Dq(s).We note from
where v 1 is the largest observability index of S(F, G , H ) .
Step 2 of Algorithm 2 and expression (3.6)that thematrices
This gives us the number of operations in Step 2 of the
[ D,], are nonsingular. In particular, [Dolhis nonsingular,
order of n3.
and since f i 0 is orthogonal (and therefore nonsingular), it
It is worth noting that for single-output systems, expres-
follows that
sions (3.2) and (3.3) give the characteristic polynomial of a
deg{det(D(s))} =deg{det(Do(s))} lower Hessenberg matrix. Similar expressions have been
=vl +v2+ +Y,=Y. (3.8) given by Wilkinson [28] for an upper Hessenberg matrix.
In Algorithm 2 it is clear that in order to derive the
Now let L ( s ) be the greatest common left divisor of D ( s ) coefficient matricesof D J s ) , we must store thoseof D,.(s),
and N ( s ) , i.e., i=q+ 1, -- ,k. This follows from expression (3.2). How-
ever, we can reduce the number of such matrices to be
D ( s )= L ( s ) D ( s ) (3.9a) stored by reducing some of the matrices 4,k - j + i =k -j
and + l , - . . , k + l ,j = O ; . - , k in expression (3.2) to null
N(s)=L(s)N(s). (3.9b) matrices of appropriate dimensions. This can be accom-
plished by means of similarity transformations on F. We
Then O-'(s)N(s) is a relatively left prime MFD of shalldiscussnextonesuch transformation which will
S( F, G , H ) . Therefore, since S( F, G , H ) is controllable enable us to apply Algorithm 2 without having to store
and observable (by assumption), any of the matrices D,(s), i = q+ 1, --.
,k for the computa-
tion of Dq(s).The resulting formfor the sbte-space model
deg{det(B(s))} =deg{det(sI, -I;)} = Y . (3.10) is also of some interest.
Authorized licensed use limited to: University of Leicester. Downloaded on September 3, 2009 at 10:06 from IEEE Xplore. Restrictions apply.
154 IEEE TRANSACTIONS ON AUTOMATIC CONTROL,NO.
VOL. AC-26, 1, FEBRUARY 1981
Xq =
and
Remarks Remark
1) Note that the matrix X,-' is also a unit lower uian- Note that ?+ is a diagonal matrix whose diagonal
gular matrix. In fact, xq-'is the same as X , except for the elements are the first 5 diagonal elements of %j multiplied
( r , k + 1 - 4 )block matrices r = k + 2 - q ; k+ 1. These - e ,
by the diagonal elements of El,. The matrix V 1s therefore
also a diagonal matrix with positive diagonal entries. The
matrices are given by
effect of carrying out Step 5 is to reduce the matrices E,
-
(in N a n d F ) to I[',i=O,. . ,k. We shall call the resulting
system S ( P , G*, H*)a blockFrobeniw form (BFF) of
S ( 4 B, C).
2) Theeffect of carrying out Steps 2 and 3 for 4= Step 6: Carry out a permutation of the rows of P
l , . . . , k is to reduce the matrices e,,
i = 2 , - - . ,k+1, and G* and the c o l u m n s of FL and H* so as to rearrange
j= 2, --- ,i to null matrices of appropriate dimensions. theminthefollowingorder: l , ~ O + l , ~ l + l , ~ ~ ~ , a k
Note that the matrix H is not affected by the coordinate 2 , 0 1 , + 2 , . . . , ( ~ ~ - ~ + 2 ;I. .,. ,; ~ ~ ~ + l ~ ; ~ ~ , ( ~ ~ ~ ~ +
transformation in Step - -3. a, = I o + I , + * * +I,,j=O; * * ,k - 1.
3) The matrices F, G , and Kresulting from Step 4 have We denote the resulting system by S( F,, G,, H,) where
the following form: the matrices & and H, have the form
0 ... 0
Authorized licensed use limited to: University of Leicester. Downloaded on September 3, 2009 at 10:06 from IEEE Xplore. Restrictions apply.
PATEL: MATRIX FRACTION DESCRIPTIONS OF TIME-INVARIANT SYSTEMS 155
x 1 0 - + e 0 x x ... X
x 0 ..* 0 x
1 X X
... ... ... . ... ...
x 0 0 .*. 1 x X
x 0 0 o x X
x 1 .*. 0 x .--
0
x 0 1
o x
... ... ... . . .. ..
e..
- *
F,=
x 0 0 -.. 1 x
x 0 0 -.- 0 x e - .
. .. -
x 1
x 0
... ...
x 0
x 0
Hc=
[:;x
. .
0 ...
...
...
x
x
x
.. ..
*
0
0
0
.
.-.
...
...
x
x
x
...
...
...
x
x
x
0
0
... ...
0 ... 81.
0
The x’s in the above matrices represent possible nonzero only and eachdiagonal entry is of the form aisr
elements. where ai, i=ly---,l,-j-lare constants and O<<r<j+l.
Therefore, we do not need to store D k - j ( s ) , j = 0, ,k- 1 ---
Discussion of Algorithm 3 in order to compute Do(s). Note, however, that Po(s)#O.
In fact,
Algorithm 3 can be regarded as an extension to the case
of block Hessenberg matrices of the reduction from the
lower Hessenberg form to the Frobenius form. In fact, we
note that the matrix Fc resulting from Step 5 is in “block
Frobenius form.” For the sake of clarity, we have
expressed the algorithmin terms of the similarity transfor- and can be easilycomputedsincethematrices Di(s),
mation matrices explicitly. We do not, of course, need to ---
i= 1, ,k + 1 are diagonal with diagonal entries of the
form thematrices X q and -Xq-’. In practice, wewould form ap’ where aj,j = 1; , --
are constants and 0 < r
carry out the transformations represented by Xq and Xq-’ <k-i+l.
by performing the appropriate Operations on the “block The effect of carrying out the transformation in Step 5
rows” is to scale the ith columns of Fand N b y l/uii and the ith
- and “block columns” o f I q - and the “block rows” rowsof F and G by uii where V = diag[ q i ] , i = 1 , . ,n.
of G4-’. Note that the matrix H q - ] is not altered by these
transformations and is equal to the matrix H . Instead of From Step 5 we note that
operating on the block rows and block columns, it is also
possible to implement the desired transformations by per-
forming elementary rowand column operations on F and
row operations on G. There are less than n3 operations where the subscripts 44 denote the 4th diagonal elements
involved in this reduction. of the corresponding matrices. Therefore, it is clear that
We noted in the discussion of Algorithm 2 that we need even when the elements of EIJyj = 0,-* ,k are not very -
to store the coefficient matrices of Di(s), i= 1,- k in e , large or very small, the product of some of the elements
order to compute the coefficient matrices of Do(s). We may become very large or very small, resulting in a very
now observe
_ _ - that by applying Algorithm 2 to the system large condition number for V. Therefore, it is possible to
S(F, G, H ) , we can avoid this problem. This is because, introduce numerical instability by carrying out Step 5.
from (3.2),we note - that the matrices Pk-,(s)=O, j = However, by delaying the operations specified by this step
O , - . . , k - l since &-j+l=O, j=O,,--,k-l, i=k-j+ until almost the last step, or until absolutely necessary,we
+
1,- . ,k 1. Then from (3.3),we see that the matrices can probably reduce the buildup of errors due to reduced
- -
D, -,( s), j =0, , k - 1 are diagonal matrices involving numerical stability.
Authorized licensed use limited to: University of Leicester. Downloaded on September 3, 2009 at 10:06 from IEEE Xplore. Restrictions apply.
156 IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. AC-26.NO. 1, FEBRUARY 1981
ALG. 1
Hessenbergmatrix to the Frobenius form. A detailed
discussion of the latter algorithm can be found in [28].
T lorthogonoll We note that in Algorithm 1 the use of singular value
decompositions, although attractive from the point of
view of numerical stability, does require more computing
time than the use of Householder or stabilized elementary
transformations which may achieve the desired reduction
in many cases without any appreciable loss of numerical
stability. This isbeing investigated further and w li be
discussed in another paper. The main difficulty in Algo-
rithm 2 is that the matrices Di(s), i=q+ l; -
,k have to
be stored in order to compute D,(s).Otherwise, the num-
ber of operations is not significantly large-of the order
of n3. Algorithm 3 providesa means of avoiding the
problem of storing D,(s),but possibly at the cost of
introducing numerical instability. The problem of ill con-
ditioning is one which is inherent in the Frobenius form
[28], and we wouldthereforeexpecttheproblem to be
present in the BFF as well.Weshall not discuss this
problem further except to point out that in going to the
BFF via the BHF, we have separated the reduction into
Fig. 1. Main functions performed by Algorithms 1-3. twostages:anumerically stable reduction to the BHF
and apotentially unstable reduction of the BHF to the
BFF. The interested reader is referred to [28] for a detailed
By performing the operations specified by Steps 5 and discussion of the problem of ill conditioning in the
6, we get a form S(&, G,, H,) which is similar to one of Frobenius canonical form.
the Luenberger canonical forms. The only significant dif- It is worth noting, particularly in viewof the system
ference is the structure of the _nonzerocolumns of H,. S(F,, G,, H,) resulting from Steps 5 and 6 of Algorithm 3,
Note, however, that the matrix M,H, is in the appropriate that the problem of obtaining the Luenberger canonical
form.There are approximately 3n2 multiplications in- forms essentially involvesthe reduction of the state matrix
volved in Step 5. A to a BFF. From the foregoing discussion, we see that
this can be achieved in two stages:by first obtaining a
BHF and then reducing it to a BFF. In the direct method
Iv. SOME GENERAL
REMARKS CONCERNING THE
of computing the Luenberger canonical forms [15], these
ALGORITHMS
two stages are combined, and therefore no allowance is
made for the difference in the numerical stability of the
The functions of the algorithms whichwe have de- two stages.
scribed in the preceding sections are summarized in Fig. l.
In developing these algorithms, we have relied heavily on
many developments in numerical linear algebra. In fact, V. EXAMPLES
theresultswhich we have described in the preceding
sectionscanbeconsidered as extensions of some quite In this section a few examples are given to illustrate the
standard concepts in the numerical analysis literature; use of the algorithms described in the preceding sections
more specifically, we have the following. andto providesome numerical results for comparison
1) The transformation of S ( A , B , C ) to S(F, G , H ) in with other methods. The computations were carried out at
thelowerBHFis analogous to the well-knownidea of the Delft University of Technology on an IBM 370/158
reducing a general matrix to the Hessenberg form. Such a using the Fortran HExtendedcompiler (OPT=2) and
reduction can be achieved in a numerically stable manner double precision arithmetic.
byusing either Householder or stabilizedelementary For comparison, the well-known structure algorithm of
transformations [28], [29] and is very extensively used, e.g., Wolovich and Falb was used to compute relatively prime
as apreliminary step in solving the eigenvalueproblem MFDs. The principal computational burden in this algo-
using the QR algorithm. rithm and the major source of numerical difficulties is the
2) Forsingle-outputsystems,expression (3.3) used in computation of the Luenberger canonical form. This
Algorithm 2 for calculating D ( s ) reduces to the expression canonical form denoted byS( F,, G,, H L ) is obtained by a
for the characteristic polynomial of a lower Hessenberg nonsingular coordinate transformation on the state vector
matrix in terms of the elements of the matrix [28]. of the given state-space model S ( A , B , C ) i.e.,
,
Authorized licensed use limited to: University of Leicester. Downloaded on September 3, 2009 at 10:06 from IEEE Xplore. Restrictions apply.
PATEL: MATRIX FRACTION DESCRIPTIONS OF TIME-INVARIANTSYSTEMS 157
Element -
Exact Algorithms 1 and 2 Structure Algorithm
0
8 7.999999999999958 7.999999999852012
Dl1
I 14 13.99999999999986 13.99999999990649
2
7 7.000000000000008 6.999999999987714
3
1 1.000000000000008 0.9999999999999270
0 43.42857142857143 43.42857142834353
N11 43.42857142857128
I
23.5 23.50000000000030 23.49999999985268
2 2.999999999979656
3 3.000000000000053
0
0 0.0000000000004394 0.0000000000829807
D2I
I
S 0 0.0000000000006712 0.0000000000496407
2
0 0.0000000000001583 0.0000000000086549
3
0 0.0000000000000045 0.0000000000000121
N21
0
-10 9.999999999999207 - 9.999999999972651
1
- I 0.9999999999999485 - 0.9999999999746491
2
0 0.0000000000000083 0.0000000000061312
0
012 - 0.8571428571428571 0.8571428571421240 - 0.8571428571218676
1
- 1.285714285714286 1.285714285713602 - 1.285714285703008
2
- 0.428571428571428t 0.4285714285711726 - 0.4285714285695891
3
0 0.0000000000000345 - 0.0000000000000553
0
Nl2 5.714285714285714 5.714285714285729 5.714285714102967
1
12.07142857142857 12.07142857142829 12.07142857132304
2
6 6.000000000000048 5.999999999989812
Authorized licensed use limited to: University of Leicester. Downloaded on September 3, 2009 at 10:06 from IEEE Xplore. Restrictions apply.
158 IEEE TRANSACTIONS ON AUTOMATICCONTROL, VOL. AC-26, NO. 1, FEBRUARY 1981
TABLE I1
OF 'IHE STATE~~ATRIcFS
EIGENVALUES FOR I%A?.WLE 2
-
+j 4.781296533270743 -
+j 4.781296533270153 -
+j 4.781506145907923
-
+j 1.311993408686664 -
+j I .3l I993408689803 -
+j 1.311977138814522
- 2.618496272311535 - 2.618496272313465 - 2.618602830529899
to 16 sigmficant figures. For lack of space, we have only rithms are numericallymorestable.For this example
shown someof the elementsof N ( s ) and D ( s ) . rc(Q)=5.4x 1 0 1 3 .
The complete numerical results for this and the other It must be pointed out that, in general, it is difficult to
examples are availablefrom the author. We noted that compare the relatively left prime MFDs resulting from
our algorithms gave an overall improvement in accuracy Algorithm 2 with those obtained using the structure algo-
of about two-threesignificantfiguresfor this example. rithm of Wolovich and Falb since the factors N ( s ) and
For this example K ( Q ) - ~ . ~ Xlo5. D ( s ) are only unique upto premultiplication by unimodu-
lar matrices, and for high-order systems itis often difficult
Example 2 to determine the common unimodular left divisors. How-
ever, we checkedtheaccuracy of ourdoubleprecision
For this example we consider the state space model of MFD computations by carrying out the same computa-
the FlOO Turbofan engine [32]. This model has 16 state tions using extended precision arithmetic. We found that,
variables, 5 inputs, and 5 outputs. The data given in [32] in general, the double precision results of our algorithm
have obviously been rounded to four significant figures. wereaccurateto about 12 significantfigures,whereas
However, for the sake of illustration, we assumed that the those of the structure algorithm were accurate to about
data as given are exact. Also, as no exact results for an four significant figures.
MFD or a canonical form are available for this example,
we checked the accuracy of our results indirectly: we first Example 3
computedtheeigenvalues of the state matrix A of the
model as given; we then applied Algorithms 1 and 3 and As the final example, we consider the 9th-order model
computed the eigenvalues of the matrix F, resulting from of aboiler as given in [34]withthecorrectionforthe
Step 6 of Algorithm3. For comparison, the Luenberger typographicalerrorfor A g 5 . This modelhas two inputs
canonical form S( F,, G,, HL)was computed and the ei- and two outputs. As in Example 2, we computedthe
genvalues of FL were determined. The eigenvalue compu- eigenvalues of thestatematrix of this model and then
tations were camed out using the QR algorithm [33]. The applied Algorithms 1 and 3 and computed the eigenvalues
results for the three setsof eigenvalues are shownin Table of F,. Wealsocomputed the eigenvalues of the state
11. We note that theeigenvalues of F, areaccurate to matrix F, of the Luenberger canonical form. The results
between 11 and 14 significant figures, whereas thoseof F, are shown in Table 111. We note that eight of the nine
are accurate to between 4 and 6 significant figures. These eigenvalues of F, are accurate tobetween12 and 14
results imply that the transformations used in our algo- sigmficant figures, whereasthe corresponding eigenvalues
Authorized licensed use limited to: University of Leicester. Downloaded on September 3, 2009 at 10:06 from IEEE Xplore. Restrictions apply.
PATEL: MATRIX FRACTION DESCRIPTIONS OF TIME-INVARIANT S y s l E M S 159
TABLE III
OF THE STATEM A ~ C E SFOR
EIGENVALUES EXAMPLE 3
Model (A) Algorithms I and 3 (Fc) Luenberger Canonical Form(FL:
0.05674267363488311 0.05674267544936184
0.1511045611241676 x 0.1511045201678978 x
- 6320.450920832599 - 6320.450922716307
- 0.07403142245301157 - 0.07403142134638936
2905.953467247269 2905.953467741293
0.05439483470987738 0.05439483456375491
- 16.97882270508176 - 16.97882270594451
- 0.4199369784374313 x - 0.4199369784293333 x
5.994355684307683 5.994355680013523
- 29133.90702005274 - 29133.90701051355
0.3511703139228007 0.3511703149209043
1597.144529915056 1597.144529763461
0.01884761466507784 0.01884761457919057
16.56657092766203 16.56657092566126
0.4879206626556721 x 0.4879206624445921 x
Authorized licensed use limited to: University of Leicester. Downloaded on September 3, 2009 at 10:06 from IEEE Xplore. Restrictions apply.
160 IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. AC-26, NO. 1, FEBRUARY 1981
TABLE V
SOMECOWPICENTS OF D ( s ) AND N ( s ) FOR EXAMPLE
3 USING
THE STRUCTURE
ALGORITHM
0.06076122102811582 0.06076122102817948
0.07403141767135694 0.07403141767135729
1323.971330536166 1323.971330536125
- 0.05439483399914120 - 0.05439483399914135
- 15.67688565322756 - 15.67688565322693
0.4199369772748431 x 0.4199369772748371 x
3.790501949439258 3.790501949438864
56442.08924297256 56442.08924293174
- 0.3511703185091978 - 0.3511703185091926
- 131.4908052910283 - 131.4908052910373
- 0.01884761425357449 - 0.01884761625357452
21.37577674848618 21.37577674848450
Authorized licensed use limited to: University of Leicester. Downloaded on September 3, 2009 at 10:06 from IEEE Xplore. Restrictions apply.
PATEL: M
A= FRACTION DESCRIPTIONS OF TIME-INVARIANT SYSTEMS 161
least squares solutions,” Numer. Math., vol. 14, pp. 403-420, 1970.
any In contrast, &Onthrn is prone to [I81 B. S. Garbw et a[., “Matrix eigeasystem routines-EEpACK
numerical instability. The decision as to whether Algo- guideextension,” in Lecture Notes in Conputer Science, vol. 51.
rithm 2 shouldbe applied to the lowerblockHessenberg Berlin: Springer-Verlag, 1977.
[19] G. H. Golub and J. H. Wilkinson, “IUIL1-Conditionedeigensystems
form or to the block Frobenius form w li depend on the and the computations of the Jordan canonical form,” SZAM Reo.,
systemmodel. As we have noted in aprecedingsection, VO~. 18,
pp.578-619,1976.
[20] G. H.Golub, V. C. Klema, and G . W. Stewart, “Rank degeneracy
Algorithm 2 has the disadvantage that the coefficients of and least squares problems,” Dep. Comput. Sci., Stanford Univ.,
thematrices Di(s), i= 1, ,k have to be stored in order Stanford, CA,Tech.Rep.
[21]V.C.Klema
STAN--76-559, 1976.
and A. J. h u b , ‘ T h esingular value decomposition:
to compute D,(s).This problem - is- avoided
- if Algorithm 2 Its computation andapplications,”
some ZEEE Trans. Automat.
is applied to the system S(f,Y G , H ) or S(PY G*Y H*) p. vanvol.
[221 Contr., AC-25,A.pp.Emami-Naeini,
Dooren, 164-176, 1980. and “stable
resultingfromAlgorithm 3. However, the reduction to extraction of the Kronecker structure of pencils,” in Proc. 17th
either of theseforms can lead tonumerical instability. ZEEE Con5Decision Contr., 1979, pp. 521-524.
[U] R V. Patel, “On computing matrix fraction descriptions and
Consequently, for systems which have an ill-conditioned canonical forms of linear time-invariant systems,” Univ. Manches-
block Frobenius form, Algorithm 2 should be applied to 489, Sci
ter Inst.
Rep. . Technol., Manchester, England, Contr. Syst. Cen.
1980.
the block
Hessenberg
form
directly. [24] J.Aplevich,
D. “Tableau methods for analysis and design of linear
A comparison of the numerical performance of our systems,” Automatica, vol. 15, pp. 419-429, 1979.
[25]E.C.Y. Tse, J. V. Medanic, and W. R Perkins, “Generalized
Algorithms 1 and 2 with that of the structure algorithm of Hessenberg transformations for reduced-order modelling of large
scale systems,” Znt. J. Contr., vol. 27, pp. 493-512, 1978.
wO1Ovich and showed that Our gave [26] L. M. Silverman and H. J. Payne, “Input-output s m c m e of hear
better numerical accuracy. The differencein the numeri- systemswithapplication to the decouplingproblem,” SZAM J.
cal accuracy was more sigdicant for ill-conditioned sys- Contr,, voL 9,
pp.199-233,
1971.
[27] R V. Patel, ‘‘Matrix fraction descriptions of linear time-invariant
terns.Thesamewas true of a comparison between our systems,” in Proc. 18th ZEEE Conf: Decision Contr., FL, Dec. 1979.
Algorithm 1 and 3 and the algorithm of Luenberger for 1281 J. H. Wilkinson, Algebraic Eigenoalw Problem. London:
OxfordUniv Press 1965.
computing canonical forms, These results and others from ~291 R S. M& and’J. H. w w n , of a
several test examples indicate that our Algorithms 1 and 2 general matrix to Hessenbergform,” Numer. Math.,. voL12,pp.
349-368,1968.
are considerably more stable IlUmeriCally than the StruC- 1301 R E. ~h~ “Mathematical description of linear dynamical
ture algorithm of Wolovich and Falb. systems,” SZAM J. Contr., vol. 1, pp. 152-192, 1963.
..
1311 N. Munro and R S. McLeod, “Minimal realization of transfer
function matrices using *he system matrix,“ Proc. Znst. Elec. Eng.,
VO~. 118, pp. 1298-1301, 1971.
REFERENCES [32]M. K. Sain, “Thethemeproblem,” in Alternatksfor Linear
Muitioaraable Control, M. K. Sain, J. L. Penkowski, and J. L.
H. H. Rosenbrock, State-Space and MuitiWrable Theoty. New Melsa, Eds. Chicago: Nat. Eng. Consortium, 1978.
York: Wiley,1970. [33] B. T. Smith et al., “Matrix eigensystemroutines-EISPACK
W.A.Wolovich, Linear Multiwufable @stems. Berlin: Springer- guide,”2nd ed., in Lecture Notes in Computer Science, vol. 6.
Verlag,1974. Berlin:Springer-Verlag,1976.
C. A. Desoer and J. D. Schulman, “Zeros and poles of matrix [34]E. J. Davison and S . H.Wang,“Properties and calculation of
transfer functions and their dynamicalinterpretation,” ZEEE Trans. transmission zeros of linearmultivariablesystems,” Automatica,
Circuits Syst., vol. CAS-21, pp. 3-8, Jan. 1974. VOL 10, p ~ 643-658,
. 1974.
W. A. Wolovich, ‘The determinations of state-space representa-
tions for linear multivariable systems,” Automatica, vol. 9, no. 1,
pp. 97-106, 1973.
B. W. Dickinson, M. Morf, and T. Kailath, “A minimal realization
algorithm for matrix sequences,” ZEEETrans. Automat. Contr.,
vol. AC-19, pp. 31-38, Feb. 1974. Rajnikant V. Patel (”76-SM80) was born in
S . H. Wang and E. J. Davison, “A -tion algorithm for the Nairobi, Kenya, on March 26, 1948. He received
design of multivariablesystems,“ ZEEETrans, Automat. Contr.,
vol. AG18, pp. 220-225, June 1973. the B.E. degree in electronicswithfirstclass
G. D. Forney, Jr., “Minimal bases of rational vector spaces, with honors from the University Liverpool,
of
applications to multivariable linear systems,” SZAM J. Contr., voL Liverpool, England, in 1969 and the Ph.D. d e
13, pp. 493-520, 1975. gree in electrical engineering from the Univer-
M. K. Sain, “A free-modul,? algorithm forminimaldesign of sity of Cambridge, Cambridge, England,in 1973.
linearmultivariablesystems, 1975 ZFAC Congr. Preprints, paper From 1972 to 1975 hewasaPostdoctoral
9.1, part 1B. Research Fellowin the Department of Engineer-
L.V. Foster, “A practical solution to the minimal design problem,” ing, University of Cambridge. In 1973- 1974 he
ZEEE Trans, Automat. Conir., vol. AG24, pp. 449-454, 1979.
S. .Kung, T. Kailath, and M. Morf, “Fast and stable algorithms for was a visiting Scientist at the Division of Autc-
.
nunuaal desien Droblems” inProc. 4th ZFAC Znt. Sym. Multimk
~~~ ~ ~ ~~
matic Control, Lund Institute of Technology, Lund, Sweden. From 1975
able Techno[ &st., D.’P. Atherton, Ed. LondbnfPergamon, to 1977 hehelda National ResearchCouncilPostdoctoralResearch
1977. c - 97-
DD.
. ,c - - 104. Associateship at NASA’s Ames Research Center, Moffett Field, CA.He
W:A. Wolovich and P. L. Falb, “On the structure of multivariable was a Visiting Professor in the Department of Electrical Engineering,
systems,” SIAM J. Contr., vol. 7,pp. 437-451, 1969. University of Waterloo, Canada, from 1977 to 1979 and aSenior
L. S . de Jon& “Numerical aspects of realizationalgorithms in Research Scientist in the Department of Applied Physics, Delft Univer-
linear systems theory,” Ph.D. dissertation, Eindhoven Univ. Tech-
nol., Eindhoven, The Netherlands, 1975. sity of Technology, Delft, TheNetherlands, in 1979- 1980. At present he
P. Van Dooren, “The generalized eigenstructure problemin linear is aLecturerin Control Engineering in the Control SystemsCentre,
system theory,” ZEEE Trans. Automat. Contr., submitted for publi- University of Manchester Institute of Science and Technology,
cation. Manchester, England. He has conducted research on problems of struc-
R V. Patel, “Computation of minimal order state-space realiza- ture and design of multivariable systems, robustness in estimation and
tions and observability indices using orthogonal transformations,” control theory, adaptive stochastic control, and computational methods
Znt. J. Contr., to be published. for largescale systems. He has published numerous technical papers in
D. G;,Luenberger, “Canonical forms for linear multivariable sys- these areas, and is the author, with N. Munro, of a forthcoming book:
tems, ZEEE Trans. Automat. Contr., vol. AG12, pp. .. 290-293,
Muitimriable System T h e o ~and ~ Design (Oxford: Pergamon, 1981). His
June 1967.
G. W. Stewart. Introduction to Matrix Comtations. NewYork: current research interests are in numerical analysis and its applications
Academic, 1973. in developing computational methodsfor the analysis and designof
G. H. Golub and C.Reinsch, “Singular value decomposition and multivariable systems.
Authorized licensed use limited to: University of Leicester. Downloaded on September 3, 2009 at 10:06 from IEEE Xplore. Restrictions apply.