Professional Documents
Culture Documents
Abstract-A Kalman filter requires an exactknowledge of the the number of unknown elements in Q must be less than
processnoise covariance matrix Q andthemeasurement noise n X r where n is the dimension of t,he state vector and T
covariance matrix R. Here we consider the case in which the true
values of Q and R are unknown. The system is assumed to be
is t.he dimension of the measurement vector. It, is shown
constant, and the random inputs are stationary. First, a correlation that in spite of this limitation, the optimal steady-state
test is given which checks whether a particular Kalman filter is filter ga.in can be obtained by an it,erative procedure.As a
working optimally or not. If the filter is suboptimal, a technique is corolhry, it is shown that the Ka1ma.n filter gain depends
given to obtain asymptotically normal, unbiased,and consistent only on n X r linear relationships between the elements
estimates of Q and R. This technique works only for the case in
which the form of Q is known and the numberof unknown elements
of Q.
in Q is less than n X r where n is the dimension of the state vector A numerical example is included to il1ustrat.e the appli-
and r is the dimension of the measurement vector. For other cases, cation of the results derived in the paper. The extension
the optimal steady-state gain KO,is obtained directly by an itera- of the results to t,he continuous case is st.raightfonvard
tive procedurewithout identifying Q. As a corollary, it is shown and is given in the last sect.ion.
thatthesteady-state optimal Kalman filter gain KO, depends
only on R X r linear functionals of Q. The results are &st derived
for discrete systems. They are thenextended to continuous systems. 11. STATEMEKT
OF THE PROBLEM
A numerical example is given to show the usefulnessof the approach.
XYStt???t
the measurement noise covariance ma,trix R. I n a number where x, is 7% X 1 state vector, 9 is n X n nonsingular
of practical situations, Q and R are either unknown or are transition matrix, r is n X q const.ant input matrix, zi
knownonlyapproximately.Heffes [3] andKishimura is T X 1 measurement vector, and H is r X 77. constant
[4] have considered the effect of errors in Q and R on the output matrix.
performance of the optimal filter. Several other investi- The sequences ui(q X 1) and vi(r X 1) are uncorre-
gators [5+[9] have proposed on-line schemes to identify lated Gaussian white noise sequences n-ith means and co-
Q a.nd R. Most of these schemes do well in identifying R variances as follows :
but run into diaculties in ident.ifying Q. Moreover, their E ( u ; ) = 0; E(uiujT]= Q S ,
extension to continuous cases is not clear. A different
approach ha.s been taken in this paper. It is assumed that ] R&j
E { v ~=) 0; E ( v , v , ~ =
the system under consideration is time invariant, com- E (u,$'} = 0, for all i,j
pletely controllable, and observable [a]. Both the system
andthe filter (optimal or suboptinlal)areassumed to where E { denotes the expectation, and 6 i j denotes the
a }
The true values of Q and R a.re unlmown. It is required Let K denote the steady-state filter gain. m e will show
that under steady state, the innovation sequence vi is a
check whether the Ka1ma.n filter construct,ed using stationary Gaussian sequence;
some estimates of Q a.nd R is close to optimal or = - H2ili-1
not (hypothesis testing),
obta.in unbiaseda.nd consistent estimates of Q and R = Hei + vi
(statistical estimation), and
adapt the Kalman filter a t regular intervals using
E [ v s i _ k T }= HE(eiei-t}HT + HE(eivi-;,T), f o r k > 0.
all the previous information (adaptive filtering). A recursive relationship can be obtained for e; by using
(11, (21, (5), and (6) :
To solve these problems, we ma.ke use of the innovat.ion
property of an optimal filter [ l o ] . ei = @ ( I- KH)ei-l - +Kuidl + (9)
Therefore,
E(vpi--kT) = H[@((IT- K H ) Y - ’
- @[,.MHT - K ( H X H T + R )1, k > 0.
When k = 0, E ( v i v i T j = H M H T R. +
It. is seen that the autocorrelation funct,ion of v i does
not dependon i. Therefore, vi is ast.ati0na.n.Gaussian
random sequence (Gaussian because of linearity) and we
can define
Ck = E { V i V i A * } .
Then
C/: = H N H T + R, h: = 0 (12)
= H[@CI - KH j]k-l@[MHT- KC0]? IC > 0. ( 1 3 ) Fig. 1. Kormalized autocorrelation function of innovation process.
(a) Srtboptimal filter. (b) Optimal filter. (Arro.crs indicate points
Furthermore, for which 95 percent confidence limits do not, enclose zero,)
c-, = CkT.
Kotice thatthe opt,imal choice of K , viz. K = It is seen from ( 1 3 ) that CI;+ 0 for large k. It can be
+
3 f H T ( H 3 f H T I?)-’ makes Cc vanish for all k # 0 (the shown4 that the infinite series in ( 1 6 ) has a finite sum so
innovation property). that the covariance of e k is proportional to l/N. Thus,
the estimates are asynzptotically unbiased and con-
IT. A TESTO F OPTIhL4LITY FOR -4I<-kL&1L41V’ FILTER sist.ent. Moreover, since all the eigenvalues of @((IT- K H )
From the discmsion of the preceding two sections, it is lie imide the unit. circle, v i belongs to t,he class of linear
clear tha.t a necessary and sufficient condition for the processes [I41 for which Parzen [lrj] has shou-n that, t k
opt,irna.lity of aKalman filter is thatthe innovation a.re asympt.otically norma,l.
sequence vi be white. This condit,ion can be t,ested statis- For t.he white noise case, ( 1 6 ) is greatly simplified by
tically by a number of different methods [13], [l6]-[19]. putt,ing C k = 0, for all k # 0:
Here lve consider a particular method given in Jenkins
and Wat.ts [13]. cov ( [ t k ] i j , [ t l l P P )= 0, k # 2
In this method, we obtain an estimate of Cr, denoted = ( 1 / N )[Co]ip[Co]jqq k = > 0
a.sel:, by using the ergodic propertg of a st,ationary ran-
dom sequence = (1/N) [C~lipCC~ljq[Coliq[Coljp, +
N k = I = 0. (17)
t k = ( l / N ) 2 YjVi-kT (14)
i=k
Estimates of t,he normalized autocorrelation coefficients
where N is the number of sample points. pe a.re obta,inedbydividing the elements of 6~. by the
eo,
A
The estinlat,es C k are biased for finite sample sizes: appropriate elements of e.g.,
n
E{C,I = ( 1 - J;/NJc~. (15) r.r. 7
M H T = B" Cp
rcl + H@KCo
+ H@KC1+ HWKCo
1
I (20)
the optimal gain K , ca.n still be obtained.
Restricting ourselves to the case in which the number
of unlmonm in Q is n X T or less, we can solve for the
unknown elements of Q by rewriting (1 1) as follows :
LC, + H@KC,-1+ + H@,"KCo] M = @X@T + n + rQrT (24)
I-@.
where B* is the pseudo-inverse of matrix B [l] defined as myhere
1
B =@ ;
LH@~-~J M
f2
= @w(@)T
= @[
Substituting back
for
-K H M - &IHTKT + KC&']@'.
+ ~
Ill on the right-hand
P T
side
+ Q + @rQrTQT
+ rQF.
of (X),
+ K1(H$lzHT + R ) K I ~ ]+@rQrT.
~ (31)
6 The proof is similar t.o the one by Kalman [l]for showing the
positive definiteness of M in (11).
180 AUTOMATIC COSTROL,
IEEE TRANGACTIOXS OXAPRIL 1970
tion for ( 3 1 2 - M I ) does not involve Q. We need Q only For N >> n, the bias in is negligible. The covariance
to calculate M I H T . This leads us to the following corol- ofA?aTfor large N is
lary.
of
corollmy: It, is sufficient to Icnon- n_xr linear functions
order
Q in to obtain the optimal
gain Kalman
of a
cov (:YfiP) w K var ( ~ O ) K T + -4::cov ([q ) P
C,
iilter.
Proof: Consider (30) which can be written a.s + K cov eo, (* [?j)
M1 = @ ( I- KOH)Ml(I - K o H ) T @ T cn
M1HT =
m
]=o
[ @ ( I- KoH)]j(@K&KoT@T + rQrTj
Expressions for
,)I:[(
cov
Cn
etc.
M I H T depends on n X r linear functions of Q ; x&., can be obta.ined from ( l G ) . It can be seen that cov (I@@’)
decreases as l / N for h g e sa,mple sizes. Similarly,
5 [@(I
i=O
- K a ) ] j r Q r T [ ( I- KJI)TPI~HT.
I 1,
Equat,ion(44) gives usa simple expression for the The startingvalues of Q a.nd R are taken as
minimumvalianceinestimating K. It can be used in
lo
0.25 0 0
deciding upon the minimum sa.mple size N .
0.4 0 ;
We now consider the asymptotic convergence ( N 1a.rge)
of theiterative scheme of Section VII.Equation(34)
QO = 0.5 0 RO= o.6] .
shows that E[8JI1] dependson the second- and higher
0 0 0.75
ordermoments of k1 which for a. normal process are
finite andtend
to zero asymptotically.
Therefore, Using these \dues, the innomtion sequence vi = (2, -
E [ ~ $ T J= 6 . ~ 1 . H&,iLl) is generated from (3) to ( 6 ) . Theestimates
Similarly, the comriance of 8 A f l asymptothlly tends eo,el,- -
.,6k of the autocorrelation are calculated using
to zero. Thus, 8L@l tends to 6 M 1 withprobability one. (14). For a typical sample of 950 points, Fig. 1(a) ahom
Extending the same argument, K 2 -+ K?, K 3 -+ Ka, * -, a plot of the first diagonal element, of fik for k = 0>40.
k,, -+ KO,with probabilit>yone. The 95 percent confidence limits are +0.0636 and four
points lie outside this band (i.e., 10 percent of the tota.1).
E. ,4NUMERICAL EXAMPLE FROM Therefore, we reject the hypot.hesis t,hat v i is n-hite. The
INERTIAL KAVIGATION same conclusion is reached by looking at the second diag-
onal element of PI;.
The results of Sections T’ and VI are
applied toa da.mped We now proceed to the ident,ification of Q and R. Since
Schuler loop forced by an exponentiallycorrelated sta- t,he number of unknowns in Q is less than n X r = 10,
tionary random input. Two measurements a.re made on we can identify Q completely. The set of equat,ions (28)
thesystem, bot,h of which arecorruptedby exponen- gives us alargenumber of linear equa.tions for G1, &,
tially correlated as well a.s white noise t,ype errors. The and G3. However, the most important of these occur along
state of the syst,em is augment,ed to include all the cor- the diagonal for k = 1 and k = 5.
relat,ed random input,sso that the augmented &ate vector For k = 1 t.he left-hand side of (28) is
z is 5 x 1, the ra.ndom input vector u is 3 X 1, a.nd the
measurement noise vector v is 3 X 1. The sgstem is
discretized using a timestep of 0.1 andthe result.ant
system mat.rices are
+= 0 0 0.95 0 0
J
The diagonal elements of the first. equation are used to
0 0 0.55 0 calculate i j 3 and i2.The first diagonal elementof the second
equation is then used to calcula.te 61.
0 0 0 0.905 It is possible to use a. few other equations a.nd to make
a lea,st-squares fit for 41, &, and 4 3 . This, however, does
0 0 0 not. alter the resultssignificantly in the present example.
The results obtained by using the identification scheme
repeatedly onthe same bat,chof data areshown in TableI.
It is seen that most of t.he identification is done during
the first iteration. Further it.erations do not increase the
likelihood function7 much, even though the changes in Q
and R are significant. ,4check case using true values of
Q and R is also sholm in TableI. It is seen that the value
of t.he likelihood function in the check case is very close to
that in thefirst iteration. This indicatesthat theestimates
R = r
0
‘1.
T2
obtained are quite close to the maximum likelihood esti-
mates. It was further noticed that even if different starting
valuesare used for Q and R, the identificationscheme
converges to the same va.lues.
The actual values of ql, q2, qa, rl, and r2 are unity, but The ljkelihood function L(Q,R) has beengiven
[12]:
by SchwePPe
t.hey are assumed unknown. It is required to identify these s
values using n1easurement.s { z i , i = l,Ar]. L(Q,R) = Z:
- ( l / S ) i=l viT(HXHT + R) - 111 I H M H T + R I.
182 IEXE TMSACTIONS OR AUTOMATIC CONTROL, APBIL 1970
TABLE I
ESTIUTESO F Q -am R BASEDOK A SETOF 950 POINTS
Percentage of
Points Lying
Outside the 95 Percent
Con6dence Limits
Estimat,e
First Second of -4ctual Calculat.ed
Likelihood Measure- h.1easul.e- h*.lean- h,lean-
Number of Func$on
Square
Square ment ment
A A A *
Iterations ?1 fi !73 71 72 L (Q,R) (percent)
(percent) Error* Errori
t Calculated mean-square error is t r (AT,) There X, is obtained from the variance equation using Q and R [cf. (4)].
h
1.5 q;
Similarly,
-. ......;i-""-
t 0
0.5
J.
........L.....:..................-.
h : z T7
=-.
Rb+l = EP + C l l ( k + 1 ) ] ( 2 M l , k -
0.d
X. CONTINUOUS
SYSTEM
1.5
..... ........ .
......9 c. 01 The results of the Drevious sections can
tended t o continuous systems. We simplystate theresults
below.8
0.d
System
1.0 ...... -..- *=Fx+Gu (47)
0.5
z = Hx+v. (48)
: 2 3
%tch h'.xzber
4 5 b 7 E 9 l c Filter
Fig. 3. On-line ident,ification of Q and R.
2 = FP + Ko(z - HP) (49)
where
We now check the optimality of the filter after i d e n a - KO = PdY'R0-l (50)
cation. Fig. 1(b) shows a plot of the &st diagonal element and
of p k , for k = 0140. It is seen that only one point lies
outside the band of 95 percent confidence limits (2.5 per-
FPo PoF' GQoG' - P&'Ro-'HPo = 0. (51) + +
cent of the total). This supports the hypothesis that vi The error covariance P1 is given as
is white.
The a,symptotic convergence of Q and R towards their (8' - K J I ) Pi+ PI( F - KdY) GQG' K8KoT = 0. '+ +
actual values is s h o r n in Fig. 3. The estimates of Q and (52)
R are updated after every batch of N points ( N = 950).
In the absence of any knowledge about the variances of Innovation Process
the estimates, a simple averagingof all the previous values v=z-HP
is performed. This is equiva.lent to the following stochastic =He+v (53)
approximation scheme[5] :
where e = ( x - 2 ) . For an optimal filter, v is white with
the same covariance as 4) [lo]. For a suboptimal flter,
where k denotes the batch number, d= (F - Ka)e + Gu - Kov (54)
Qk theestimate of Q after k batches
QP+~,L the estimate of Q based on the ( k 1)th batch +
* These result.s havenot been applied to apractical problem
Qk+l the estimate of Q after ( k 1) batches. so fa.r. +
MEHRA: IDENTIFICATIOH OF TARIANCES A X 3 ADAPTIVE EALMAN FILTERlNG 183
j=O
(- l)jHFjGQGTFk-jHT, for k = 0,1, - to show the usefulness of the approach. The results a.re
first derived for a discrete system. They are then extended
to continuous systems.
[the set of equations analogous to (28)].
If the number of unknowns in Q is more than n X r, XOMENCLATURE
KO,is obtained directly without identifying Q. The pro-
cedure is as follon-s. D e h e Matrices
@,I',H,F,G,B,A System matrices
K1 = PIHTR-l. (57) covariance matrices
Q,R,N,P,M1,Pl
Let PP be the error cova.riance corresponding to K1. Qo,Ro,Jfo,Po initial values of Q,R,M,P
Then it canbe shown that K,K@l$op, * * Kalman filter gains
Q,&w,K,. .. estimated values of Q,R,M,K (caret
( F - K1H) (Pz - P I ) + (P2 - P I )( F - K I H ) over any quantity denotes an esti-
mate)
- ( K I - Ko)R(Kl- = 0. (58) autocorrelation function
Ck,&,C(T)
Therefore Pkrh normalized
autocorrelation
function
P, < P1. S(W) power spectral density
8~1,8~,8.ii?1,8P increment matrices.5
Similarly, define K2 = P2HTR--Iand let P3 be the error
covariance for K2. Then Vectors
xi,ii+~ Actual and estimated states
P3 < P2 < PI. ui,vi white noise sequences
I n this way, P is decreased at each step and the sequence zi measurements
--
Ko,Kl,Kz- converges to Kop. Vi innovations
Equation (58) is now used to obtain kOp, an estimate ei error in state estimation.
of K,. After obtaining = FlBT&l, me substitute it
Scalars
in (58) to get an estimate of 8P1 = P2 - P1: n,q,r Dimension variables
( F - &H) 8@1+ (F - k1H) N sample size
[Ck]ij element in the ith row and the jth column of
- (KI - KO)&(&- K o )=~ 0. (59) the matrix Ck
Then i d (. r.) Kroneckerdeltaand the delta function
6 ..
F,BT = fill?' + 8$1HT (60) L(&,R) likelihood function.
184 IEEE TRAWSACTIOXS O X .4UTQJUTIC COXTROL, APRIL 1970