You are on page 1of 6

BAYESIAN SMOOTHING ALGORITHMS IN PAIRWISE AND TRIPLET MARKOV CHAINS

B. Ait-El-Fquih & F. Desbouvries

GET/INT/CITI and CNRS UMR 5157


9 rue Charles Fourier, 91011 Evry, France
Boujemaa.Ait Elfquih@int-evry.fr, Francois.Desbouvries@int-evry.fr

ABSTRACT propose smoothing algorithms for general PMC and TMC.


An important problem in signal processing consists in esti- Finally in section 4 we consider the particular case of linear
mating an unobservable process x = {xn }n∈IN from an ob- and Gaussian TMC; the algorithms of section 3 then reduce
served process y = {yn }n∈IN . In Linear Gaussian Hidden to an extension of the RTS algorithm [6] by Rauch, Tung
Markov Chains (LGHMC), recursive solutions are given by and Striebel, on the one hand, and of the Two-Filter algo-
Kalman-like Bayesian restoration algorithms. In this paper, rithm by Mayne [7] and Fraser and Potter [8], on the other
we consider the more general framework of Linear Gaussian hand.
Triplet Markov Chains (LGTMC), i.e. of models in which
the triplet (x, r, y) (where r = {rn }n∈IN is some additional
2. PARTIALLY OBSERVED MARKOV CHAINS
process) is Markovian and Gaussian. We address fixed-
interval smoothing algorithms, and we extend to LGTMC
2.1. The Kalman filter and its extensions
the RTS algorithm by Rauch, Tung and Striebel, as well as
the Two-Filter algorithm by Mayne and Fraser and Potter. Let us consider the classical state-space system :

1. INTRODUCTION xn+1 = F n xn + G n u n
, (1)
yn = H n xn + J n vn
An important problem in signal processing consists in recur-
sively estimating an unobservable process x = {xn }n∈IN in which xn ∈ IRnx is the state, yn ∈ IRny is the obser-
from an observed process y = {yn }n∈IN . This is done clas- vation, un ∈ IRnu is the process noise and vn ∈ IRnv is
sically in the framework of Hidden Markov Chains (HMC), the measurement noise. The processes u = {un }n∈IN and
which have been extensively studied for many years (see v = {vn }n∈IN are assumed to be independent, jointly inde-
e.g. the recent tutorial [1]). pendent and independent of x0 .
In this paper we deal with the recently introduced Pair- Let x0:n = {xi }ni=0 and y0:n = {yi }ni=0 . Let also
wise [2] (PMC) and Triplet [3] Markov Chains (TMC). In p(xn ), p(x0:n ) and p(xn |y0:n ), say, denote the probabil-
the PMC model we assume that the pair (x, y) is a Markov ity density function (pdf) (w.r.t. Lebesgue measure) of xn ,
Chain (MC), and in the TMC model that the triplet (x, r, y) the pdf of x0:n , and the pdf of xn , conditional on y0:n , re-
(in which r = {rn }n∈IN is some additional process) is an spectively; the other pdf are defined similarly. The filtering
MC. These models are more general than the HMC model problem consists in computing the posterior pdf p(xn |y0:n ).
and yet enable the development of efficient restoration algo- If furthermore x0 and wn = (un , vn ) are Gaussian, then
rithms of the hidden process x. In particular, a Kalman-like p(xn |y0:n ) is also Gaussian and is thus described by its
filtering algorithm for PMC (resp. for TMC) has been pro- mean and covariance matrix. Propagating p(xn |y0:n ) over
posed in [4] (resp. in [5]). The aim of this paper is to pro- time amounts to propagating these parameters, and the fil-
pose Bayesian smoothing algorithms for PMC and TMC. tering algorithm reduces to the celebrated Kalman filter [9],
The paper is organized as follows. In section 2 we briefly see also [10].
recall the Kalman filter as well as some of its many exten- Since this pionnering work, the Kalman filter has been
sions. We next recall the three embedded HMC, PMC and generalized in many directions. To name just a few ex-
TMC models, and for illustrative purposes we show that two amples, robust (i.e., square-root type) or fast (i.e., Chan-
classical extensions of the standard state-space system used drasekhar type) algorithms have been proposed; restoration
in Kalman filtering (namely the jump-Markov processes, algorithms for smoothing or prediction problems have been
and the state-space systems with colored process and mea- developed; the independence assumptions on u and v have
surement noises) are some particular TMC. In section 3 we been dropped, leading to linear models with colored process
and/or measurement noises; and the extension of (1) to non- can have different physical meanings (see e.g. the examples
linear and/or non-Gaussian systems has been addressed, lead- given below). By definition, the triplet t = (x, r, y) is a
ing to approximate solutions such as the extended Kalman TMC if (x, r, y) is a (vector) Markov chain. The interest of
filter and more recently particle filters. The literature on all TMC is twofold :
these extensions is vast (see e.g. [11] [12] [13] as well as
the references therein). • As far as modeling is concerned, if (r, (x, y)) is an
MC then the marginal process (x, y) is not necessar-
ily an MC, so TMC are not necessarily PMC;
2.2. Embedded Markovian models : HMC ⊂ PMC ⊂
TMC • As far as restoration is concerned, the TMC (r, x, y)
can be viewed as the PMC ((r, x), y); so x∗ = (r, x)
On the other hand, yet another direction in which it is pos- can be restored from y by a PMC algorithm, and
sible to extend the Kalman filter consists in releasing some finally x is obtained by marginalization (such algo-
conditional independence assumptions among x and y. Let rithms have been proposed in the discrete [3] or linear
us first come back to model (1). From (1), we get Gaussian [5] cases).
p(xn+1 |x0:n ) = p(xn+1 |xn ) ; (2) Finally, let us notice that in practice computer experi-
Yn ments have demonstrated the superiority of PMC [15] (resp.
p(y0:n |x0:n ) = p(yi |x0:n ) ; (3) TMC [16]) over HMC in the context of image segmentation.
i=0
p(yi |x0:n ) = p(yi |xi ) for all i, 0 ≤ i ≤ n . (4) 2.3. Some classical TMC
In other words, x is an MC, and since x is known only It happens that some classical generalizations of the state-
through the observed process y, (1) is an HMC. space system (1) are particular TMC. For an illustrative pur-
In an HMC x is first assumed to be an MC (by the very pose, let us leave in this section the general discussion and
meaning of the words ”HMC”), and next the stochastic in- consider the following two examples (the first one with a
teractions of x and y are designed in such a way that x can discrete latent process r, the second one with a continuous
be efficiently restored from y. On the other hand, a PMC is latent process r). Other examples of particular TMC models
a model in which the pair (x, y) is assumed to be an MC. can be found in [5] [17].
So in a PMC x and y are modeled altogether (and in a sym-
metric way), and a PMC can indeed be seen as a partially 2.3.1. Jump-Markov model
observed vector MC, in which we observe one component
y and we want to restore the other one x. Let us consider the jump-Markov model (see e.g. [18] [19]):
Now if (2) to (4) hold, then (x, y) is an MC, so any 
HMC is also a PMC. The converse is not true, because if xn+1 = F(rn+1 )xn + G(rn+1 )un
, (5)
(x, y) is a (vector) MC then the marginal process x is not yn = H(rn )xn + J(rn )vn
necessarily an MC; moreover, conditionnally on x0:n , the
in which u = {un }n∈IN and v = {vn }n∈IN are zero-
variables {yi }ni=0 form an MC and thus are not necessarily
mean, Gaussian, independent, jointly independent and in-
independent [4]. On the other hand, due to the symmetry of
dependent of x0 , and r = {rn }n∈IN is a scalar discrete
the PMC model, the conditional law of x0:n given y0:n is
MC. Conditionnally on r, (5) is a classical linear state-space
also Markovian. This key computational property (which in
system (1), and as such an HMC. Now if we consider the
the context of HMC is well known, see e.g. [1, eq. (5.21) p.
whole model x, y and r, then the process t = {tn } with
1539]), in turn, enables the derivation of efficient HMC-like
tn = (xn , rn , yn ) is a particular TMC.
restoration algorithms. In particular, in the linear Gaussian
case, the extension to PMC of the Kalman filter has been
considered in [14, Corollary 1 p. 72] and [4]1 . 2.3.2. State-space model with colored noises
The PMC model can be further generalized to the TMC Let us once again consider model (1), but in which we now
model [3] which we now recall. A TMC is a stochastic dy- assume that
namical model which describes the interactions between 3    u,u    u
processes : the hidden process x, the observed process y, un+1 An 0 un ξn
= + ,
and a third process r which, depending on the application, vn+1 0 Av,v
n vn ξnv
| {z } | {z } | {z }
1 more precisely, if (2) to (4) hold then both {(x , y )} rn+1 An ξn
n n n∈IN and
{(xn+1 , yn )}n∈IN are MC; the difference between [14] and [4] stems
from these two candidate definitions of a PMC (see also the remark at the where ξ u = {ξnu }n∈IN (resp. ξ v = {ξnv }n∈IN ) is zero-mean,
end of [4, section 3]). independent and independent of u0 (resp. of v0 ), and ξ u
and ξ v are independent. Each one of the two processes following, we focus on the computation of p(x∗n |y0:N ) for
u = {un }n∈IN and v = {vn }n∈IN is thus an MC, and u some n, 0 ≤ n ≤ N ; finally the pdf p(xn |y0:N ) of interest
is independent of v. Such a model has been introduced by is obtained by marginalization.
Sorenson [20] (see also [21, ch. 5]). It is no longer an HMC
(x is not an MC), but the whole model tn = (xn , rn , yn−1 ) 3.1. The Triplet RTS algorithm
can be rewritten as
       Proposition 1 Let t be a TMC. Then p(x∗n |y0:N ) can be
xn+1 F n Gn 0 xn 0 computed recursively as
 rn+1  =  0 An 0   rn  +  ξn  Z
yn H n Jn 0 yn−1 0 p(xn |y0:N ) = p(x∗n+1 |y0:N )p(x∗n |x∗n+1 , y0:n )dx∗n+1 ,

(with Gn = [Gn , 0] and Jn = [0, Jn ]), and so t = {tn } is (6)


a TMC. in which p(x∗n |x∗n+1 , y0:n ) can be computed as
p(x∗n |y0:n )p(x∗n+1 |tn , yn )
3. SMOOTHING ALGORITHMS FOR PMC AND p(x∗n |x∗n+1 , y0:n ) = R .
TMC p(x∗n |y0:n )p(x∗n+1 |tn , yn )dx∗n
(7)
Let again x = {xn }n∈IN be the hidden state process, y = Proof 1 We start from
{yn }n∈IN the observed process and r = {rn }n∈IN the ad- Z
ditional process. From now on we say that the process p(x∗n |y0:N ) = p(x∗n |x∗n+1 , y0:N )p(x∗n+1 |y0:N )dx∗n+1 .
t = {tn }n∈IN , with tn = (xn , rn , yn−1 ) = (x∗n , yn−1 )
is a TMC if t is an MC. The aim of this section is to derive p(x∗ ∗
n ,xn+1 ,y0:N )
Bayesian fixed-interval smoothing algorithms for TMC. Since t is an MC, p(x∗n |x∗n+1 , y0:N ) = R
p(x∗ ,x n+1 ,y0:N )dxn
∗ ∗
n

Let us first briefly recall the existing fast Bayesian smooth-


p(yn+1:N |tn+1 )p(tn+1 , tn , y0:n−2 )
ing algorithms in the HMC framework : =R
p(yn+1:N |tn+1 )p(tn+1 , tn , y0:n−2 )dx∗n
• In the case of a discrete hidden process x, the forward-
= p(x∗n |x∗n+1 , y0:n ), whence (6). On the other hand,
backward (Baum-Welch) algorithm [22] [23] enables
to compute p(xn |y0:N ) (for an arbitrary n, 0 ≤ n ≤ p(x∗n |y0:n )p(x∗n+1 |x∗n , y0:n )
N ), while the Viterbi algorithm [24] [25] computes p(x∗n |x∗n+1 , y0:n ) = R ,
p(x∗n |y0:n )p(x∗n+1 |x∗n , y0:n )dx∗n
arg max(p(x0:N |y0:N ));
x0:N
and
• In the case of a continuous hidden process x, a num- p(x∗n+1 , yn |x∗n , y0:n−1 ) p(tn+1 |tn )
ber of fixed-interval smoothing algorithms have been p(x∗n+1 |x∗n , y0:n ) = ∗
= ,
p(yn |xn , y0:n−1 ) p(yn |tn )
proposed (see e.g. [12, ch. 10] [26] for recent sur-
veys), among which the RTS algorithm [6] by Rauch, whence (7).
Tung and Striebel, and the Two-Filter smoother by
Mayne [7] and Fraser and Potter [8]; 3.2. The Triplet Forward-Backward algorithm
• Though originally derived independently, these dis- Proposition 2 Let t be a TMC. Let
crete and continuous range approaches have been rec-
onciled recently; in particular, the forward-backward α(x∗n ) = p(x∗n , y0:n−1 ), (8)
algorithm is implemented by the Two-Filter smoother, β(x∗n ) = p(yn:N |x∗n , yn−1 ). (9)
and the Viterbi recursions are implemented by the
Then p(x∗n |y0:N ) can be computed as
RTS smoother [27] [28].
α(x∗n )β(x∗n )
From now on we consider smoothing algorithms for PMC p(x∗n |y0:N ) = R , (10)
α(x∗n )β(x∗n )dx∗n
or TMC. Generalizations of the forward-backward algorithm
for PMC [2] and TMC [3] with discrete hidden process have in which the triplet forward (resp. backward) pdf α(x∗n )
already been developped. On the other hand, Kalman-like (resp. β(x∗n )) can be computed recursively via
filtering algorithms have been proposed for PMC [4] and Z
TMC [5] with continuous hidden process, but no smoothing α(x∗n+1 ) = p(tn+1 |tn )α(x∗n )dx∗n , (11)
algorithm has been derived yet. From now on we shall thus Z
propose the TMC RTS smoother (see §3.1 and §4.1) and β(x∗n ) = p(tn+1 |tn )β(x∗n+1 )dx∗n+1 . (12)
the TMC Two-Filter smoother (see §3.2 and §4.2). In the
∗ ∗
Proof 2 Equations (8)- (12) hold because of Bayes’s rule x∗n|0:N , Pxn|0:N
So (b ,x
) can be computed recursively (in the
and because t is an MC. Firstly, x ,x
x∗n|0:n , Pn|0:n
backward direction) provided (b

x∗n+1|0:n ,
) and (b

p(x∗n , y0:n−1 , yn:N )


∗ ∗
Pxn+1|0:n
,x
) are known; these, in turn, can be computed recur-
p(x∗n |y0:N ) =
p(y0:N ) sively (in the forward direction) by the TMC Kalman filter
∝ p(x∗n , y0:n−1 ) p(yn:N |x∗n , yn−1 ) . algorithm, see [4] [5].
| {z }| {z }
n)
α(x∗ n)
β(x∗
Proof 3 We first need to compute (7). From (13), we get
Next (11) and (12) are obtained from
Z p(tn+1 |tn ) ∼ N (Fn tn , Qn ). (19)
α(xn+1 ) = p(x∗n+1 , yn |x∗n , y0:n−1 ) p(x∗n , y0:n−1 )dx∗n ,

Using Proposition 6 (see section 5 below), we get
| {z }
p(tn+1 |tn )
x∗ ,x∗
Z p(x∗n+1 |x∗n , y0:n ) ∼ N (An x∗n +bn , Qn ), (20)
β(x∗n ) = p(tn+1 |tn ) p(yn+1:N |tn+1 , tn ) dx∗n+1 .
| {z } with
n+1 )
β(x∗

,x∗ ∗
−1 y,x ∗
An = Fnx −Qnx ,y
(Qy,y
n ) Fn ,

−1
bn = Qnx ,y (Qy,y
n ) yn
∗ ∗
−1 y,y
4. THE LINEAR GAUSSIAN CASE + [Fn − Qn (Qy,y
x ,y x ,y
n ) Fn ] yn−1 ,
x∗ ,x∗ ∗
,x∗ ∗
−1 y,x ∗

From now on we further assume that the TMC process t is Qn = Qnx − Qxn ,y
(Qy,y
n ) Qn .
also Gaussian. More precisely, we assume that
On the other hand, from (15)
 ∗   x∗ ,x∗ ∗   ∗   x∗ 
xn+1 Fn Fnx ,y xn wn
= +
∗ ∗
∗ , (13) p(x∗n |y0:n ) ∼ N (b x ,x
x∗n|0:n , Pn|0:n ).
yn Fny,x Fny,y yn−1 wny
| {z } | {z } | {z }
tn+1 Fn wn Using (20) and Proposition 7, we get
in which w = {wn }n∈IN is independent and independent " #" x∗ ,x∗ ∗
,x∗ T
#
b∗n|0:n
x Pn|0:n Pxn|0:n An
of t0 , y−1 = 0, and ∗ ∗
p(xn , xn+1|y0:n ) ∼ N ( ∗ , ∗ ∗ ∗ ∗ )
bn+1|0:n An Px ,x Px ,x
x
 x∗ ,x∗ ∗  n|0:n n+1|0:n
Qn Qxn ,y
wn ∼ N (0, ∗ ). (14)
Qny,x Qy,y
n b∗n|0:n + bn , Pxn+1|0:n
b∗n+1|0:n = An x
with x ,x
= Qn
∗ ∗ x∗ ,x∗
+
| {z }

,x∗ T
Qn An Pxn|0:n An . Using Proposition 6, we get
Then p(x∗n |y0:N )
is also Gaussian, and computing this pdf
amounts to computing its parameters. Let us thus introduce p(x∗n |x∗n+1 , y0:n ) ∼ N (b
x∗n|0:n +K∗n|0:N (x∗n+1 −b
x∗n+1|0:n ),
the following notation. For 0 ≤ i ≤ j ≤ N , let yi:j = ∗ ∗ ∗ ∗
{ym }i≤m≤j , and let Pxn|0:n
,x
− K∗n|0:N Pxn+1|0:n
,x
K∗T
n|0:N ) (21)

x∗n|i:j , Pxn|i:j
p(x∗n |yi:j ) ∼ N (b ,x
).
∗ ∗
(15) in which K∗n|0:N is given by (16).
It remains to compute equation (6). Using (15), (21) and
Proposition 7, we finally get (17) and (18).
4.1. The Triplet RTS algorithm
Proposition 3 (The Triplet RTS algorithm) Let (13) hold. 4.2. The Triplet Two-Filter algorithm
Let
Proposition 4 (The Triplet Two-Filter algorithm) Let (13)
∗ ∗ ∗ ∗
−1 y,x T x ,x −1
K∗n|0:N = Pxn|0:n
,x ∗
,x∗ ∗ ∗
[Fnx −Qnx ,y
(Qy,y
n ) Fn ] Pn+1|0:n . hold. Then (8) - (10) reduce to
(16) ∗ ∗
,x −1 ∗ ,x −1 ∗ ∗ ∗
,x −1 ∗ ∗ ∗

Then (6) reduces to Pxn|0:N bn|0:N = Pxn|0:n−1


x bn|0:n−1 +Pxn|n−1:N
x bn|n−1:N
x
∗ ∗
,x −1 ∗
b∗n|0:N
x = b∗n|0:n+K∗n|0:N[b
x x∗n+1|0:n], (17)
x∗n+1|0:N−b − Pxn|n−1 bn|n−1 ,
x (22)

,x∗ −1 ∗
,x∗ −1 ∗
,x∗ −1 ∗
,x∗ −1

Pxn|0:N
,x ∗
= Pxn|0:n
,x ∗ ∗
− K∗n|0:N Pxn+1|0:n
,x ∗
Kn|0:N
∗ ∗ T
Pxn|0:N = Pxn|0:n−1 +Pxn|n−1:N −Pxn|n−1 . (23)
∗ ∗ T
+ K∗n|0:N Pxn+1|0:N
,x
K∗n|0:N . (18) Proof 4 The proof is omitted for want of space.
As we can see, the TMC Two-Filter algorithm can be The algorithm is initialized at n = N + 1 by
used for computing p(x∗n |y0:N ) once we know how to com- x∗ ,y y,y −1
x∗ ,x∗
x∗n|0:n−1 , Pn|0:n−1
pute (b ), (b

,x∗
x∗n|n−1:N , Pxn|n−1:N x∗n|n−1 ,
) and (b b∗N +1|N = x
x b∗N +1 + PN +1 (PN ) [yN − y bN ], (37)
∗ ∗ ∗ ∗ ∗ ∗
∗ ∗ x ,x x ,x x ,y y,y −1 y,x
Px ,x ). Let us now consider these points. PN +1|N = PN +1 − PN +1 (PN ) PN +1 . (38)
n|n−1
∗ ∗
x ,x
Firstly, as in the TMC RTS algorithm, (bx∗n|0:n−1 , Pn|0:n−1 ) Proof 5 Using (24), (19) and Propositions 7 and 6, we get
can be computed recursively by the TMC Kalman filter [4]
[5]. p(tn |tn+1 ) ∼ N (Fen+1 tn+1 + (I − Fen+1 Fn )b en+1 ),
tn , Q

,x∗
Let us now address the computation of (b x∗n|n−1 , Pxn|n−1 ). (39)
Let in which Fen+1 and Q en+1 are given respectively by (29) and
   x∗ ,x∗ ∗  (30).
b∗n
x Pn Pnx ,y Next, from (15) we have
p(tn ) ∼ N ( , ). (24)
Py,y

bn−1
y Pny,x n−1
| {z } | {z } p(x∗n+1 |yn:N ) ∼ N (b

x ,x
x∗n+1|n:N , Pn+1|n:N

). (40)
b
tn Ptn
Using (39), (40) and Proposition 7, and marginalizing w.r.t.
From (13), we can compute btn and Ptn recursively by x∗n+1 , we get
btn+1  
= Fnbtn , (25) b∗n|n:N
x
p(tn |yn:N ) ∼ N ( , Ptn|n:N ),
Ptn+1 = Fn Ptn Fn + Qn . (26) bn−1|n:N
y

Using Proposition 6, we get from which we deduce (31), (32) and (33). Using Proposi-
tion 6, we get (34), (35) and (36).
(Py,y

b∗n|n−1
x = b∗n + Pxn
x ,y −1
n−1 ) (yn−1−b
yn−1),(27)

x ,x ∗ 5. SOME PROPERTIES OF GAUSSIAN R.V.
(Py,y

,x∗ ∗
−1 y,x ∗
Pn|n−1 = Pxn − Pxn ,y
n−1 ) Pn . (28)
The derivations in this paper rely on the following results,
x∗n|n−1:N ,
Let us finally address the computation of (b which are recalled for convenience of the reader :
∗ ∗
x ,x
Pn|n−1:N ). These parameters can be computed recursively
Proposition 6 Let
by the TMC backward Kalman filter :
   
µ Σ1,1 Σ1,2
Proposition 5 (The TMC Backward Kalman filter) Let (u1 , u2 ) ∼ N ( 1 , ).
µ2 Σ2,1 Σ2,2
" ∗ #" ∗ ∗ ∗ #
Fex ,t Fen+1
x ,x e x ,y
Fn+1 Then conditionally on u2 , u1 ∼ N (µ1 + Σ1,2 Σ−1
2,2 (u2 −
Fen+1 = Pn Fn (Pn+1) = en+1
t T t −1
y,t = ey,x∗ e y,y (29)
Fn+1 Fn+1 Fn+1 µ2 ), Σ1,1 − Σ1,2 Σ−1 Σ
2,2 2,1 ).
" ∗ ∗ ∗
#
ex ,x Q
Q ex ,y Proposition 7 Let u1 ∼ N (µ1 , Σ1 ) and conditionally on
e t e t
Qn+1 = Pn− Fn+1Pn+1Fn+1=e T n+1∗
n+1 .(30)
ey,x Q
Q ey,y u1 , let u2 ∼ N (Au1 + b, Σ2|1 ). Then
n+1 n+1
   
Then we have µ1 Σ1 Σ1 A T
(u1 , u2 ) ∼ N ( , ).
Aµ1 + b AΣ1 Σ2|1 + AΣ1 AT
∗ ∗ ∗
b∗n|n:N = Fen+1
x x ,x
x∗n+1|n:N −b
[b x∗n+1 ]+ Fen+1
x ,y
[yn −b
yn ]
+x ∗
bn , (31) 6. REFERENCES
e y,x∗ ∗ ∗ e y,y
bn−1|n:N = F
y [b
x n+1 −bxn+1 ]+ F
n+1|n:N [yn −byn ] n+1 [1] Y. Ephraim and N. Merhav, “Hidden Markov pro-
+ybn−1 , (32) cesses,” IEEE Transactions on Information Theory,
Pxn|n:N
,x ∗ ∗
e x∗ ,x∗ e x∗ ,x∗ x∗ ,x∗ e x∗ ,x∗ T
= Qn+1 + Fn+1 Pn+1|n:N (Fn+1 ) , (33) vol. 48, no. 6, pp. 1518–69, June 2002.

ex ,y + Fex ,x Px ,x
K∗n|n−1:N = [Q ey,x T
∗ ∗ ∗ ∗ ∗
[2] W. Pieczynski, “Pairwise Markov chains,” IEEE
n+1 n+1 n+1|n:N (Fn+1 ) ]
∗ ∗ ∗ ∗
Transactions on Pattern Analysis and Machine Intel-
ey,y + Fey,x Px ,x
× [Q ey,x T −1 , (34)
n+1 n+1 n+1|n:N (Fn+1 ) ] ligence, vol. 25, no. 5, pp. 634–39, May 2003.
b∗n|n−1:N = x
x b∗n|n:N +K∗n|n−1:N (yn−1 −b
yn−1|n:N ), (35) [3] W. Pieczynski, C. Hulard, and T. Veit, “Triplet Markov
x∗ ,x∗
Pn|n−1:N =

,x∗
Pxn|n:N ey,y
−K∗n|n−1:N [Q chains in hidden signal restoration,” in SPIE Interna-
n+1
∗ ∗ ∗ T ∗
tional Symposium on Remote Sensing, Crete, Grece,
+ Fen+1
y,x x ,x
Pn+1|n:N (Fen+1
y,x T
) ]K∗n|n−1:N . (36) September 22-27 2002.
[4] W. Pieczynski and F. Desbouvries, “Kalman filter- [16] P. Lanchantin and W. Pieczynski, “Unsupervised non
ing using pairwise Gaussian models,” in Proceed- stationary image segmentation using triplet Markov
ings of the International Conference on Acoustics, chains,” in Advanced Concepts for Intelligent Vision
Speech and Signal Processing (ICASSP 03), Hong- Systems (ACVIS 04), Brussels, Belgium, August 31 -
Kong, 2003. September 3 2004.

[5] B. Ait-El-Fquih and F. Desbouvries, “Kalman filter- [17] W. Pieczynski and F. Desbouvries, “On Triplet
ing for triplet Markov chains : Applications and exten- Markov chains,” in Proceedings of the International
sions,” in Proceedings of the International Conference Symposium on Applied Stochastic Models and Data
on Acoustics, Speech and Signal Processing (ICASSP Analysis (ASMDA 2005), Brest, France, May 17-20
05), Philadelphia, USA, March 19-23 2005. 2005.
[18] M. Mariton, Jump linear systems in automatic control,
[6] H. E. Rauch, F. Tung, and C. T. Striebel, “Maximum
Marcel Dekker, New York, 1990.
likelihood estimates of linear dynamic systems,” AIAA
J., vol. 3, no. 8, pp. 1445–1450, August 1965. [19] A. Doucet, N. J. Gordon, and V. Krishnamurthy, “Par-
ticle filters for state estimation of jump Markov linear
[7] D. Q. Mayne, “A solution to the smoothing problem systems,” IEEE Transactions on Signal Processing,
for linear dynamical systems,” Automatica, vol. 4, pp. vol. 49, no. 3, pp. 613–24, March 2001.
73–92, 1966.
[20] H. W. Sorenson, “Kalman filtering techniques,” in
[8] D. C. Fraser and J. E. Potter, “The optimum linear Advances in Control Systems Theory and Appl., C. T.
smoother as a combination of two optimum linear fil- Leondes, Ed., vol. 3, pp. 219–92. Acad. Press, 1966.
ters,” IEEE Transactions on Automatic Control, vol.
7, pp. 387–90, August 1969. [21] C.K. Chui and G. Chen, Kalman Filtering with Real-
Time Applications, Berlin, DE: Springer, 1999.
[9] R. E. Kalman, “A new approach to linear filtering and
[22] L. E. Baum and T. Petrie, “Statistical inference for
prediction problems,” J. Basic Eng., Trans. ASME,
probabilistic functions of finite state Markov chains,”
Series D, vol. 82, no. 1, pp. 35–45, 1960.
Ann. Math. Stat., vol. 37, pp. 1554–63, 1966.
[10] Y. C. Ho and R. C. K. Lee, “A Bayesian approach to [23] L. E. Baum and J. A. Eagon, “An inequality with
problems in stochastic estimation and control,” IEEE applications to statistical estimation for probabilistic
Transactions on Automatic Control, vol. 9, pp. 333– functions of a Markov process and to a model for ecol-
339, October 1964. ogy,” Bull. Amer. Meteorol. Soc., vol. 73, pp. 360–63,
1967.
[11] B. D. O. Anderson and J. B. Moore, Optimal Filtering,
Prentice Hall, Englewood Cliffs, New Jersey, 1979. [24] G. D. Forney, “The Viterbi algorithm,” Proceedings
of the IEEE, vol. 61, pp. 268–78, March 1973.
[12] T. Kailath, A. H. Sayed, and B. Hassibi, Linear es-
timation, Prentice Hall Information and System Sci- [25] A. J. Viterbi, “Error bounds for convolutional codes
ences Series. Prentice Hall, Upper Saddle River, New and an asymptotically optimal decoding algorithm,”
Jersey, 2000. IEEE Transactions on Information Theory, vol. 13, pp.
260–69, April 1967.
[13] A. Doucet, N. de Freitas, and N. Gordon, Eds., Se-
quential Monte Carlo Methods in Practice, Statistics [26] H. L. Weinert, Fixed interval smoothing for state-
for Engineering and Information Science. Springer space models, The Kluwer international series in engi-
Verlag, New York, 2001. neering and computer science. Kluwer Academic Pub-
lishers, Boston, 2001.
[14] R. S. Lipster and A. N. Shiryaev, Statistics of Random
[27] B. Delyon, “Remarks on linear and nonlinear filter-
Processes, Vol. 2 : Applications, chapter 13 : ”Con- ing,” IEEE Transactions on Information Theory, vol.
ditionally Gaussian Sequences : Filtering and Related
41, no. 1, pp. 317–22, January 1995.
Problems”, Springer Verlag, Berlin, 2001.
[28] P. L. Ainsleigh, N. Kehtarnavaz, and R. L. Streit,
[15] S. Derrode and W. Pieczynski, “Signal and image “Hidden Gauss-Markov models for signal classifica-
segmentation using pairwise Markov chains,” IEEE tion,” IEEE Transactions on Signal Processing, vol.
Transactions on Signal Processing, vol. 52, no. 9, pp. 50, no. 6, pp. 1355–1367, June 2002.
2477–89, 2004.

You might also like