Professional Documents
Culture Documents
Abstract—We introduce a distributed cooperative framework sensors, wireless communication interfaces, a processing unit,
and method for Bayesian estimation and control in decentralized and actuators, all together forming a cyber-physical system
agent networks. Our framework combines joint estimation of [10], [11] with a tight coupling between sensing, comput-
time-varying global and local states with information-seeking con-
trol optimizing the behavior of the agents. It is suited to nonlinear ing, communication, and control. A common task in mobile
and non-Gaussian problems and, in particular, to location-aware agent networks is seeking information, either about external
networks. For cooperative estimation, a combination of belief phenomena or about the network itself. This task relies on
propagation message passing and consensus is used. For cooper- estimation (quantifying, fusing, and disseminating information)
ative control, the negative posterior joint entropy of all states is and control (configuring the network to increase information).
maximized via a gradient ascent. The estimation layer provides
the control layer with probabilistic information in the form of A common theme in previous works is the reliance on position
sample representations of probability distributions. Simulation information for estimation and/or control.
results demonstrate intelligent behavior of the agents and excellent Estimation methods for mobile agent networks (our focus
estimation performance for a simultaneous self-localization and will be on distributed Bayesian estimation) address estimation
target tracking problem. In a cooperative localization scenario of common global states [7], [12]–[16], estimation of local
with only one anchor, mobile agents can localize themselves after
a short time with an accuracy that is higher than the accuracy of states [17]–[22], or combined estimation of local and global
the performed distance measurements. states [23]–[25]. In the first case, the agents obtain local mea-
surements with respect to external objects or the surrounding
Index Terms—Agent networks, distributed estimation, dis-
tributed control, information-seeking control, distributed target environment, which are fused across the network. Global fusion
tracking, cooperative localization, belief propagation, message methods that require only local communication include consen-
passing, consensus, sensor networks, sequential estimation. sus [26] and gossip [27]. Example applications are distributed
target tracking [7], cooperative exploration of the environment
I. I NTRODUCTION [15], and chemical plume tracking [3]. In the second case
(estimation of local states), the agents cooperate such that
A. Motivation and State of the Art each agent is better able to estimate its own local state. Here,
the dimensionality of the total state grows with the network
R ECENT research on distributed estimation and control
in mobile agent networks [1]–[4] has frequently been
motivated by location-aware scenarios and problems includ-
size, which leads to more complex factorizations of the joint
posterior probability density function (pdf). When the factor
ing environmental and agricultural monitoring [5], healthcare graph [28] of this factorization matches the network topology,
monitoring [6], target tracking [7], pollution source localization efficient message passing methods for distributed inference can
[8], chemical plume tracking [3], and surveillance [9]. The be used, such as the belief propagation (BP) [28] and mean field
agents in a mobile agent network are generally equipped with [29] methods. Example applications are cooperative localiza-
tion [17], synchronization [19], [22], and simultaneous localiza-
tion and synchronization [20], [21]. In the third case (estimation
Manuscript received August 15, 2014; revised November 27, 2014; accepted of both global states and local states), a message passing
March 18, 2015. Date of publication May 6, 2015; date of current version algorithm can be combined with a networkwide information
October 19, 2015. This paper was presented in part at the IEEE International dissemination technique. An example application is cooperative
Conference on Acoustics, Speech and Signal Processing (ICASSP), Brisbane,
Australia, April 2015. This work was supported in part by the Austrian Science simultaneous self-localization and target tracking [23], [25].
Fund (FWF) under Grants S10603 and P27370, and by the European Commis- In many cooperative estimation scenarios, it is advantageous
sion under ERC Grant 258418 (COOPNET), the EU FP7 Marie Curie Initial to control certain properties of the agent network, such as
Training Network MULTI-POS (Grant 316528), and the Newcom# Network of
Excellence in Wireless Communications. the agent positions or the measurement characteristics (“con-
F. Meyer and F. Hlawatsch are with the Institute of Telecommunications, trolled sensing”) [1]–[9]. In particular, here we will address
Vienna University of Technology, Vienna 1040, Austria (e-mail: florian.meyer@ the problem of combining distributed estimation and distributed
nt.tuwien.ac.at; franz.hlawatsch@nt.tuwien.ac.at).
H. Wymeersch and M. Fröhle are with the Department of Signals and Sys- control in mobile agent networks. We will limit our discussion
tems, Chalmers University of Technology, Gothenburg 412 96, Sweden (e-mail: to information-seeking control, which seeks to maximize the
henkw@chalmers.se; frohle@chalmers.se). information carried by the measurements of all agents about
Color versions of one or more of the figures in this paper are available online
at http://ieeexplore.ieee.org. the global and/or local states to be estimated. The use of
Digital Object Identifier 10.1109/JSAC.2015.2430519 information measures for the control of a single agent or a
0733-8716 © 2015 IEEE. Translations and content mining are permitted for academic research only. Personal use is also permitted, but republication/redistribution
requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.
2440 IEEE JOURNAL ON SELECTED AREAS IN COMMUNICATIONS, VOL. 33, NO. 11, NOVEMBER 2015
time n. Furthermore, let x(1:n) [x(1)T, . . . , x(n)T]T , u(1:n) In the correction step, CA l determines the marginal posteriors
[u(1)T, . . . , u(n)T ]T , and y(1:n) [y(1)T, . . . , y(n)T]T . Each CA f (xl |y) and f (xm |y), m ∈ T , which are given by
l ∈ C estimates its local state x(n) l and all the target states
(n) f (xk |y) ∝ f (xk ) f (yl,k1 |xl , xk1 ) dx∼k ,
xm , m ∈ T from the measurements of all CAs up to
k ∈A l ∈C k1 ∈Al
time n, y(1:n). This estimation is based on the posteriors
k ∈ {l} ∪ T . (10)
f (x(n)
k |y
(1:n) ; u(1:n) ), k ∈ {l} ∪ T , which are marginals of the
(p−1)
A. Sequential Calculation × f (yl,m |xl , xm ) ψm→l (xm ) dxm , (11)
For a review of the sequential calculation of (6) proposed in m∈Tl
[24], [25], we now switch to the following simplified notation
for the sake of readability. In the conditions of the various con- which is initialized as b(0)(xl ) = f (xl ). Similarly, the belief of
ditional pdfs, we omit the measurements up to time n −1, i.e., target state xm at message passing iteration p is given by
y(1:n−1), and the control vectors up to time n, i.e., u(1:n), because
y(1:n−1) has already been observed and u(1:n) has already been b(p)(xm ) ∝ f (xm )
(p−1)
f (yl,m |xl , xm ) ψl→m (xl ) dxl , (12)
determined; hence both are considered fixed. Furthermore, we l∈Cm
suppress the time index n, and we write the current and previous
(n) (n−1)
states of CA l as xl = xl and xl = x−
l , respectively. Sim- (p−1)
(n+1) + with initialization b(0)(xm ) = f (xm ). Here, ψm→l (xm ) and
ilarly, we write ul = ul . For sequential calculation of (6), (p−1)
CA l ∈ C employs a basic Bayesian recursive filtering method ψl→m (xl ) are the “extrinsic informations” from target m to
[17] consisting of a prediction step and a correction step. In the CA l and from CA l to target m, respectively, at the previous
prediction step, CA l ∈ C computes a predictive posterior of its message passing iteration p − 1. These extrinsic informations
current state, are calculated recursively as
− − (p) b(p)(xm )
f (xl ) = f xl |x− l f xl dxl . (8) ψm→l (xm ) = (p−1)
(13)
f (yl,m |xl , xm ) ψl→m (xl ) dxl
(n)
Here, f (xl ) and f (x−
l ) are short for f (xl |y
(1:n−1) ; u(1:n) )
(p) b(p) (xl )
(n−1) (1:n−1) (1:n−1) ψl→m (xl ) = (p−1)
, (14)
and f (xl |y ;u ), respectively. Furthermore, CA f (yl,m |xl , xm ) ψm→l (xm ) dxm
l computes predictive posteriors of the target states
(0)
with initialization ψm→l (xm ) = f (xm ) and ψl→m (0)
(xl ) = f (xl ), re-
−
f (xm ) = f xm |x− −
m f (xm ) dxm , m∈T. (9) spectively. In (11), the beliefs b (p−1) (xl ) of neighbor CAs l ∈ Cl
MEYER et al.: DISTRIBUTED ESTIMATION WITH INFORMATION-SEEKING CONTROL IN AGENT NETWORKS 2443
are used instead of the extrinsic informations. This is part of additional condition that has been observed previously and is
the sum–product algorithm over a wireless network (SPAWN) thus fixed:
scheme for cooperative localization [17], which results from a
specific choice of the message schedule in loopy BP and has −h x(n+1) y(n+1) ; y(1:n), u(1:n+1)
been observed to provide highly accurate estimates [17], [24],
[25], [37].
= f x(n+1), y(n+1)y(1:n) ; u(1:n+1)
A sample-based distributed implementation of (11)–(14)
has been proposed in [24], [25]. A problem
for a distributed
implementation is that the products l∈Cm f (yl,m |xl , xm ) × log f x(n+1)y(n+1), y(1:n); u(1:n+1)
(p−1)
ψl→m (xl ) dxl , m ∈ T involved in (12) are not available at the
CAs. However, an approximation of these products can be pro- × dx(n+1)dy(n+1), (16)
vided to each CA in a distributed manner (i.e., using only local
communications) either by a consensus algorithm performed in where
log denotes the natural logarithm. Note that
parallel for each sample weight [25], [38], [39] or by the like- h x (n+1) y (n+1) ; y(1:n), u(1:n+1) depends on the random vectors
lihood consensus scheme [7], [24], [40]. For the calculations in x(n+1) and y(n+1), i.e., on their joint distribution but not on their
the control layer—to be described in Sections IV–VII—all CAs values. Our notation indicates this fact by using a (1:n+1)
sans serif
require a common set of samples representing the target beliefs font for x (n+1) and y (n+1) in h x (n+1)y (n+1); y(1:n), u .
(n+1) (n+1) (1:n)
b(p)(xm ). This can be ensured by additionally using, e.g., a max- According to expression (16), −h x y ;y ,
consensus and providing all CAs with the same seed for random
u(1:n+1) is a function of the control vector u(n+1) . This function
number generation [39], [41].
will be denoted by Dh u(n+1) , i.e.,
The output of the estimation layer is the set of beliefs
b(P)(xk ), k ∈ {l} ∪ T at the final message passing iteration
Dh u(n+1) − h x(n+1)y(n+1); y(1:n), u(1:n+1) , (17)
p = P. These beliefs are handed over to the control layer,
which calculates the control variables u+l (see Fig. 2(a)). Each and it will be used as the objective function for control at
(P) (j)
belief b (xk ) is represented by J samples {xk }Jj=1 , which is each CA. This objective function is holistic in that it involves
(j)
briefly denoted by {xk }Jj=1 ∼ b(P)(xk ). From these samples, an all the next states (of both the CAs and the targets), x(n+1),
approximation of the MMSE estimate (7) is calculated as and all the next measurements,
(n+1) y(n+1). Instead of a full-blown
maximization of Dh u , we perform one step of a gradient
ascent [42] at each time n. Thus, u(n+1) is determined as
1 (j)
J
x̂k = xk .
J
j=1
û(n+1) = u(n+1)
r + c(n+1) ∇Dh u(n+1) u(n+1) =u(n+1) , (18)
r
(n+1)
where ur is a reference vector and c(n+1) > 0 is a step
(n+1)
IV. C ONTROL L AYER size. The choice of ur depends on the manner in which
(n)
the local control vectors ul (which are subvectors of u(n) )
Next, we present the information-seeking controller. We (n)
temporarily revert to the full notation. appear in the state evolution functions gl x(n−1) l , u(n)
l , ql in
(n+1) (n+1)
(1); two common choices are ur = u(n) and ur =0
(cf. Section IX-A).
A. Objective Function and Controller Since u(n+1) = u(n+1)
l l∈C
, we have
According to our definition at the beginning of Section III, (n+1) ∂Dh (u(n+1) )
the vector comprising all the measurements in the network at ∇Dh (u )= ,
∂u(n+1)
the next time n + 1 is y(n+1) = [y(n+1)
l∈C
l,k ]l∈C,k∈A(n+1) . However, l
l
(n+1)
in this definition of y(n+1), we now formally replace Al and thus the gradient ascent (18) with respect to u(n+1) is
(n) (n+1) equivalent to local gradient ascents at the individual CAs l with
by Al since at the current time n, the sets Al are not yet (n+1)
known. Thus, with an abuse of notation, y (n+1) is redefined as respect to the local control vectors ul . At CA l, the local
gradient ascent is performed as
(n+1)
(n+1)
y(n+1) yl,k (n)
. (15) (n+1) (n+1) (n+1) ∂Dh u
l∈C,k∈Al ûl = ur,l + cl , (19)
∂u(n+1)
l
(n+1)
u(n+1) =ur
Fig. 3. Flowchart of the flooding-based implementation of the control layer at CA l (see Section VI-A).
Fig. 4. Flowchart of the consensus-based implementation of the control layer at CA l (see Section VI-B).
Invoking [15, Th. 1], we obtain The likelihood function f (y+ |xC , x+ +
T ; u ) involved in (28) can
be written as
∂DI (u+ ) ∂f (y+ |xC , x+ +
T ;u )
= f (xC , x+
T) f (y+ |xC , x+ +
f (y+ + +
T ;u ) = l,l |xl , xl ; ul , ul )
∂u+l ∂u+
l l∈C l ∈Cl
+
f (y+ |x +
C , xT ; u ) × f (y+ + +
× log dxC dx+ + l,m |xl , xm ; ul ) , (29)
f (y+ ; u+ ) T dy . (28)
m∈Tl
2446 IEEE JOURNAL ON SELECTED AREAS IN COMMUNICATIONS, VOL. 33, NO. 11, NOVEMBER 2015
with q(y+, xC , x+ + +
T ) f (xC , xT )f (y |xC , x
+ +
T ; ur ) viathe following
(j) +(j) J
f (y+ + + two-stage procedure. First, samples xC , xT are drawn
l,l |xl , xl ; ul , ul )
j=1
+ +
from
= f (y+
l,l |xl ,xl ) x+ = g̃ (x ,u+ ),x+ = g̃ (x ,u+ ) , l ∈ C, l ∈ Cl (30)
l l
f (xC , x+ f (x+
l l l l l l
T)= f (xl ) m) . (35)
f (y+ + +
l,m |xl , xm ; ul ) l∈C m∈T
= f (y+ |x+, x+ ) +
l,m l m xl = g̃ (x ,u+ )
, l ∈ C, m ∈ Tl . (This factorization expresses the conditional statistical indepen-
l l l
dence of the xl , l ∈ C and the x+ m , m ∈ T given y
(1:n) . This
(The latter expressions are obtained using (3) and (20).) is a common approximation, which was introduced in [17]
Let ỹ+
l denote the subvector of y + = y+
l,k l∈C,k∈Al (cf. (15)) and is also used in the estimation layer [24], [25].) Then,
whose likelihood function includes all those factors of (j) +(j) J
for each sample xC , xT , samples y+(j,j ) j =1 are drawn
f (y+ |xC , x+ +
T ; u ) in (29) that depend on the local control vector (j) +(j)
+
ul . This subvector is given by from the conditional pdf f y+ xC , xT ; u+ r . The distributed
calculation of these samples will be discussed in Section VI
+ T T
ỹ+
l y + T
l,k k∈A yl ,l l ∈C , (31) and in Appendix C. Finally, we note that using (32), one
l l
easily obtains a (rather unwieldy) expression of the derivative
∂f (ỹ+ + +
l |xl ,xCl ,xT ;uC )
and its likelihood function is obtained as l l
occurring in (33). This expression involves
∂u+
l
f (ỹ+ + +
l |xl , xCl , xTl ; uCl )
∂ g̃l (xl ,u+
l )
∂f (y+ |x+,x+ )
l,l l l
the factors in (32) and the derivatives + , for
∂ul ∂x+ l
= f (y+ + + + + +
l,l |xl , xl ; ul , ul ) f (yl,l |xl , xl ; ul , ul )
∂f (y+ + +
l,m |xl ,xm )
l ∈ Cl , and ∂xl + for m ∈ Tl .
l ∈Cl
× f (y+ + +
l,m |xl , xm ; ul ) , (32)
VI. D ISTRIBUTED P ROCESSING
m∈Tl
+ In this section, we present two alternative schemes
for a dis-
with u+
Cl ul l ∈{l}∪C . By comparing (32) with (29), it is ∂DI (u+ )
l tributed, sample-based computation of ∂u+ + + according
seen that ỹ+ +
l is also the subvector of y whose likelihood
l u =ur
to (33) and (34). Both schemes are distributed in that they
function includes all those factors of f (y |xC , x+
+ +
T ; u ) in (29) require only local communication, i.e., each CA l ∈ C transmits
+
that involve the state xl . data only to its neighbors l ∈ Cl .
Using (29) and (32), it is shown in Appendix B that a Monte
Carlo (i.e., sample-based) approximation of (28) evaluated at
u+ = u+ r is given by
2 A. Flooding-Based Processing
∂DI (u+ ) We first discuss a distributed scheme where each CA l ∈ C
∂u+ + + performs an exact (“quasi-centralized”) calculation of (33) and
l u =ur (j)J
(34). As a result of the estimation layer, samples xk j=1 ∼
1
J J
1 f (xk ), k ∈ {l} ∪ T are available at CA l (see (10), noting that
≈ +(j,j ) (j) (j) +(j) + f (xk ) was denoted f (xk |y) there). A flooding protocol [44]
JJ f ỹ x ,x ,x ;u
j=1 j =1 l l Cl Tl r,Cl is now used to make available (j)to
J each CA l the reference
+(j,j ) (j) (j) +(j) + vectors u+
x , x , x ; u and the samples x l j=1 ∼ f (xl ) of all the other
∂f ỹl r,l
l Cl Tl Cl
× CAs l ∈ C\{l}. The flooding protocol requires each CA l to
∂u+ l u+ +
C =ur,C communicate with its neighbor CAs l ∈ Cl . In addition, CA l
l l
(j) +(j) +
+(j,j )
locally calculates predictive marginal posteriors for all target
f y x , x ; u
× log C T + r , (33) states via the following prediction step (which is (9) with n
f y+(j,j ) ; ur replaced by n + 1):
with f (x+ ) = f (x+
m m |xm )f (xm ) dxm , m∈T. (36)
1 J
(j ) +(j )
f y+(j,j ) ; u+ ≈ f y+(j,j ) xC , xT ; u+ (j)J
r r . (34)
An implementation of (36) using the samples xm j=1 , m ∈
J
j =1
T produced by the estimation layer (which are available at
+(j) J
(j) +(j)
Here, y+(j,j ) , xC , and xT are samples of y+, xC , and x+ CA l) and yielding samples xm j=1 ∼ f (x+ m ), m ∈ T is de-
T, (j)J
respectively that are drawn from the importance density [43] scribed in [24], [25]. At this point, samples xl j=1 ∼ f (xl )
+(j) J
for l ∈ C and xm j=1 ∼ f (x+ m ) for m ∈ T are available at
2 With an abuse of notation, the superscript (j) now indicates the jth sample,
whereas previously, in our full-blown notation, the superscript (n) indicated the CA l. Because all states xl , l ∈ C and x+ m , m ∈ T are con-
nth time step. ditionally independent given y(1:n) (see (35)), CA l can now
MEYER et al.: DISTRIBUTED ESTIMATION WITH INFORMATION-SEEKING CONTROL IN AGENT NETWORKS 2447
where
B. Consensus-Based Processing
1 (l)
Next, we present an alternative distributed computation of Fj,j,j Fj,j,j (40)
|C|
(33) and (34) that avoids the use of a flooding protocol and l∈C
does not require each CA to know the state evolution and
with
measurement models of all the other CAs. As a first step, CA l
(j)J (l) +(j,j ) (j ) (j ) +(j ) +
broadcasts its own samples xl j=1 ∼ f (xl ) calculated in the Fj,j,j log f yl x , x , x ; ur,Cl , (41)
l Cl Tl
estimation layer to all neighbor CAs l ∈ Cl , and it receives
(j)J
from them their own samples xl j=1 ∼ f (xl ), l ∈ Cl . In ad- for j = 1, . . . , J, j = 1, . . . , J , and j = 1, . . . , J. To com-
+(j,j ) J
+(j) J pute Fj,j(l)
,j in (41), CA l needs samples yl ∼
dition, CA l locally calculates samples xm j=1 ∼ f (x+ m ) for + (j) (j) +(j) +
j =1
all m ∈ Tl via the prediction step (36) (with T replaced by Tl ), f yl xl , xCl , xTl ; ur,Cl and the reference vectors ur,l for +
+(j,j ) J
using the sample-based implementation described in [24], [25].
(j) (j) +(j)
l ∈ {l} ∪ Cl . The samples yl j =1
are already available at
Thus, after the stacking operations xCl = xl l ∈C and xTl =
+(j) (j) (j) +(j) J l CA l since y+ +
l is a subvector of ỹl (see (31) and (37)) and
xm m∈T , samples xl , xCl , xTl ∼ f (xl , xCl , x+ +(j,j ) J + (j) (j) +(j) +
j=1 T ) are samples ỹl ∼ f ỹl xl , xCl , xTl ; ur,Cl have already
l
(j) (j) l +(j) j =1
available at CA l. Then, for each sample xl , xCl , xTl , been computed as described above. The u+
+(j,j ) J (j) (j) +(j) r,l can be obtained at
J samples ỹ l
j =1
∼ f ỹ+ x , x , x ; u+ are com-
l l Cl Tl r,Cl
CA l through communication with the neighbor CAs l ∈ Cl .
(l)
puted as described in Appendix D. This only involves commu- Once the Fj,j ,j have been calculated at CA l, their av-
nication between neighboring CAs. erages Fj,j,j in (40) can be computed in a distributed way
2448 IEEE JOURNAL ON SELECTED AREAS IN COMMUNICATIONS, VOL. 33, NO. 11, NOVEMBER 2015
of the marginal posterior of CA l, the correction step (10) be ensured by additionally using, e.g., a max-consensus and
simplifies to providing all CAs with the same seed for random number
generation [39], [41].
2) Control Layer: Since there are no unknown CA
f (xl |y) ∝ f (xl ) f (yl1 ,l |xl1 , xl ) dx∼l , (44) states, the objective function in (26) simplifies in that
l ∈C l1 ∈Cl DI (u+ ) = I(xx + + + +
T ; y ; u ) and Gl (ul ) = 0 for all l ∈ C. The
expression of the gradient of DI (u+ ) in (33) and (34)
while the prediction step (8) remains unchanged. A feasible simplifies because f (ỹ+ + + + + +
l |xl , xCl , xTl ; uCl ) = f (ỹl |xl , xTl ; uCl )
+ + + + + +
and, typically, accurate approximation of f (xl |y) in (44) can be and f (y |xC , xT ; u ) = f (y |xT ; u ); furthermore, sampling
obtained by evaluating from f (xC , x+
T ) reduces to sampling from f (xT ).
+
Next, we discuss the case where the local states of the CAs The following aspects of the simulation setup are common
are known, and thus our task is only the distributed, cooperative to all three scenarios. The states of the CAs consist of their 2D
estimation of the target states. position, i.e., x(n) (n) (n) T
l [xl,1 , xl,2 ] in a global reference frame. In
1) Estimation Layer: The marginal posteriors corresponding addition to the mobile CAs, there is one anchor CA (indexed by
to the CAs are no longer calculated, and the correction step (10) l = 1), i.e., a static CA that broadcasts its own (true) position to
in the calculation of the marginal posterior of the mth target the mobile CAs but does not perform any measurements. The
simplifies to CA network is fully connected. The states of the mobile CAs
evolve independently according to [35]
f (xm |y) ∝ f (xm ) f (yl,m |xl , xm ) , (46) (n) (n−1) (n) (n)
xl = xl + ul + ql , n = 1, 2, . . . . (47)
l∈Cm
(n)
where f (xm ) is calculated according to (9). A computationally Here, ql ∈ R2 is zero-mean Gaussian with independent and
feasible sample-based approximation of sequential state esti- identically distributed entries, i.e., q(n)
l ∼ N (0, σq I) with σq =
2 2
(n) (n )
mation as given by (46) and (9) is provided by the particle filter 10−3 , and ql and ql are independent unless (l, n) = (l, n ).
[7], [40], [49]. The admissible set Ul of the control vector u(n) is defined
The product of local likelihood functions l∈Cm f (yl,m |xl , xm ) (n) l
is not available at the CAs. However, as in Section III-B, an by the norm constraint ul ≤ umax l . For the interpretation
(n)
approximation of these products can be provided to each CA of ul within (47), it is assumed that the CAs know the
in a distributed manner by a consensus (or gossip) algorithm orientation of the global reference frame. In the initialization
performed in parallel for each sample weight [25], [38], [39] or of the algorithms, at time n = 0, we use a state prior that is
by the likelihood consensus scheme [7], [24], [40]. uniform on [−200, 200]×[−200, 200].
For the calculations in the control layer (described presently), The mobile CAs acquire distance
measurements according
a common set of samples is required at each CA. This can to (4), i.e., y(n)
l,k = x(n) − x(n) + v (n) , where the measurement
l k l,k
2450 IEEE JOURNAL ON SELECTED AREAS IN COMMUNICATIONS, VOL. 33, NO. 11, NOVEMBER 2015
(n)
noise vl,k is independent across l, k, and n and Gaussian with
variance
⎧ (n)
⎨σ 2 , x − x(n) ≤ d0
(n)2 0
(n) (n) κ l k
σl,k = (n) (48)
⎩σ 2 xl −xk −1 +1 , x − x(n) > d0 .
0 d0 l k
Fig. 10. Performance of three different methods for simultaneous self-localization and target localization. (a) Self-localization RMSE. (b) Target-localization
RMSE.
Inserting (52) into (51) and the resulting expression of Setting u+ = u+ r , and multiplying and dividing the integrand in
h(xx+, y + ; u+ ) into (50) gives (55) by f (ỹ+
l |x l , xCl , x+ +
Tl ; ur,Cl ), we obtain further
∂DI (u+ )
h(xx+ |yy+ ; u+ ) = h x l , x + +
A\{l} , y ; u
+
+ Gl (u+ y+ ; u+ ).
l ) − h(y + +
(53) ∂u+ l u =ur
1
= q(y+, xC , x+ T)
Next, we repeat this transformation procedure but apply it to f (ỹ+ |x + +
l Cl , xTl ; ur,Cl )
, x
l
the term h x l , x + , + ; u+ in (53) instead of h(x x+, y + ; u+ ). +
A\{l} y ∂f (ỹ+ +
l |xl , xCl , xTl ; uCl )
+
Consider an arbitrary l ∈ C\{l}, and note that xA\{l} consists × + +
+ ∂u+ l uC =ur,C
of x+ + + +
l and xA\{l,l } xk k∈A\{l,l } , where xl = g̃l (xl , ul ) ac-
l l
+
f (y+ |x +
C , xT ; ur )
cording to (20).Proceeding as above and inserting the resulting × log + dxC dx+ +
T dy , (56)
expression of h x l , x + + + into (53) yields
A\{l} , y ; u f (y+ ; ur )
where
h(xx+ |yy+ ; u+ ) = h x l , x l , x + +
A\{l,l } , y ; u
+
+ Gl (u+
l ) q(y+, xC , x+ + + + +
T ) f (xC , xT ) f (ỹl |xl , xCl , xTl ; ur,Cl )
+ Gl (u+ y+ ; u+ ) . × f (y̆+ + +
l |xC , xT ; ur ) .
l ) − h(y
1 + (j ) +(j ) +
J
f (y+ |xC , x+ +
T ;u ) f (y+ ; u+
r ) ≈ f y xC , xT ; ur .
f (y̆+ +
l |xC , xT ; u
+
= , (54) J
f (ỹ+ + +
l |xl , xCl , xTl ; uCl )
j =1
which, according to (29) and (32), involves all factors of Evaluating this for y+ = y+(j,j ) (again see Section V) yields (34).
f (y+ |xC , x+ +
T ; u ) that do not depend on the local control vector
+
ul . Using (54) in (28) yields
A PPENDIX C
(j) +(j)
D RAWING S AMPLES F ROM f y+ xC , xT ; u+
r
∂DI (u+ )
= f (xC , x+ + + +
T ) f (y̆l |xC , xT ; u ) We consider the setting of Section VI-A. As discussed there,
∂u+ (j)J +(j)J
l
samples xl j=1 ∼ f (xl ), l ∈ C and xm j=1 ∼ f (x+ m ), m ∈ T
∂f (ỹ+ + +
l |xl , xCl , xTl ; uCl )
are available at CA l, and it is assumed that the state evolution
× and measurement models of all CAs l ∈ C are known to CA l.
∂u+
l We start by noting that by combining (15) and (3), the compos-
ite measurement vector y+ can be written as
f (y+ |xC , x+ +
T ;u )
× log dxC dx+ +
T dy . (55) + +
f (y+ ; u+ ) y+ = dl (x+
l , xk , vl,k ) l∈C,k∈A . (58)
l
MEYER et al.: DISTRIBUTED ESTIMATION WITH INFORMATION-SEEKING CONTROL IN AGENT NETWORKS 2455
+(j) J
First, CA l obtains samples xl ∼ f̃ (x+
l )
R EFERENCES
j=1
f (x+ ) +
[1] T. Shima and S. Rasmussen, UAV Cooperative Decision and Control:
l xl = g̃l (xl ,u+ )
(see (20)) for all l ∈ C by evaluating Challenges and Practical Approaches. Philadelphia, PA, USA: SIAM,
r,l
(j) 2009.
g̃l (xl , u+ + +
l ) at xl = xl and ul = ur,l , i.e.,
[2] F. Bullo, J. Cortés, and S. Martínez, Distributed Control of Robotic
Networks: A Mathematical Approach to Motion Coordination Algorithms.
+(j) (j) Princeton, NJ, USA: Princeton Univ. Press, 2009.
xl = g̃l xl , u+
r,l , j = 1, . . . , J. (59) [3] A. Nayak and I. Stojmenović, Wireless Sensor and Actuator Networks:
Algorithms and Protocols for Scalable Coordination and Data Communi-
+(j) J cation. Hoboken, NJ, USA: Wiley, 2010.
Thus, at this point, samples xk j=1
for all k ∈ A are avail- [4] F. Zhao and L. J. Guibas, Wireless Sensor Networks: An Information
Processing Approach. Amsterdam, The Netherlands: Morgan Kaufmann,
able at CA l. Next, for each j ∈ {1, . . . , J}, CA l draws samples 2004.
+(j,j ) J
vl,k j =1
∼ f (v+
l,k ) for l ∈ C and k ∈ Al . Finally, CA l ob-
[5] P. Corke et al., “Environmental wireless sensor networks,” Proc. IEEE,
vol. 98, no. 11, pp. 1903–1917, Nov. 2010.
+(j,j ) J + (j) +(j) +
tains samples y j =1
∼ f y xC , xT ; ur by evaluating [6] J. Ko et al., “Wireless sensor networks for healthcare,” Proc. IEEE,
vol. 98, no. 11, pp. 1947–1960, Nov. 2010.
(58) using the appropriate samples, i.e., [7] O. Hlinka, F. Hlawatsch, and P. M. Djuric, “Distributed particle filtering
in agent networks: A survey, classification, and comparison,” IEEE Signal
+(j) +(j) +(j,j ) Process. Mag., vol. 30, no. 1, pp. 61–81, Jan. 2013.
y+(j,j ) = dl xl , xk , vl,k l ∈C,k∈A
, j = 1, . . . , J . [8] T. Zhao and A. Nehorai, “Distributed sequential Bayesian estimation
l
of a diffusive source in wireless sensor networks,” IEEE Trans. Signal
Process., vol. 55, no. 4, pp. 1511–1524, Apr. 2007.
[9] H. Aghajan and A. Cavallaro, Multi-Camera Networks: Principles and
A PPENDIX D (j) (j) +(j) +
Applications. Burlington, MA, USA: Academic, 2009.
D RAWING S AMPLES F ROM f ỹ+ [10] K.-D. Kim and P. R. Kumar, “Cyber physical systems: A perspective at
l xl , xCl , xTl ; ur,Cl the centennial,” Proc. IEEE, vol. 100, pp. 1287–1308, May 2012.
(j)J [11] S. Haykin, Cognitive Dynamic Systems: Perception-Action Cycle, Radar
In the setting of Section VI-B, samples xl j=1 ∼ f (xl ), l ∈ and Radio. New York, NY, USA: Cambridge Univ. Press, 2012.
+(j)J [12] A. D. Ryan, H. Durrant-Whyte, and J. K. Hedrick, “Information-theoretic
{l} ∪ Cl and xm j=1 ∼ f (x+ m ), m ∈ Tl are available at CA l. sensor motion control for distributed estimation,” in Proc. IMECE,
We start by noting that combining (37) and (3) yields Seattle, WA, USA, Nov. 2007, pp. 725–734.
[13] G. M. Hoffmann and C. J. Tomlin, “Mobile sensor network control using
mutual information methods and particle filters,” IEEE Trans. Autom.
+ + +
y+
l = dl (xl , xk , vl,k ) k∈A . (60) Control, vol. 55, no. 1, pp. 32–47, Jan. 2010.
l [14] M. Schwager, P. Dames, D. Rus, and V. Kumar, “A multi-robot control
policy for information gathering in the presence of unknown hazards,” in
Proc. ISRR, Flagstaff, AZ, USA, Aug. 2011, pp. 1–16.
Based on the analogy of this expression to (58), CA l
+(j,j ) J (j) (j) +(j) + [15] B. J. Julian, M. Angermann, M. Schwager, and D. Rus, “Distributed
∼ f y+
first obtains samples yl l xl , xCl , xTl ; ur,Cl
robotic sensor networks: An information-theoretic approach,” Int. J.
j =1 Robot. Res., vol. 31, no. 10, pp. 1134–1154, Sep. 2012.
by carrying out the steps of Appendix C with obvious [16] N. A. Atanasov, J. L. Ny, and G. J. Pappas, “Distributed algorithms for
modifications—in particular, y+ is replaced by y+ l , C by {l} ∪ stochastic source seeking with mobile robot networks,” Trans. ASME, J.
Cl , and T by Tl . More specifically, CA l obtains samples Dyn. Syst. Meas. Control, vol. 137, no. 3, Mar. 2015, Art. ID. 031004.
+(j) J [17] H. Wymeersch, J. Lien, and M. Z. Win, “Cooperative localization in
xl j=1
for l ∈ {l} ∪ Cl according to (59). Then, for each wireless networks,” Proc. IEEE, vol. 97, no. 2, pp. 427–450, Feb. 2009.
+(j,j ) J [18] T. Sathyan and M. Hedley, “Fast and accurate cooperative tracking
j ∈ {1, . . . , J}, CA l draws samples vl,k j =1
∼ f (v+
l,k ) for in wireless networks,” IEEE Trans. Mobile Comput., vol. 12, no. 9,
+(j,j ) J pp. 1801–1813, Sep. 2013.
k ∈ Al and, in turn, obtains samples yl j =1
by evaluating [19] Y.-C. Wu, Q. M. Chaudhari, and E. Serpedin, “Clock synchronization of
wireless sensor networks,” IEEE Signal Process. Mag., vol. 28, no. 1,
(60) using the appropriate samples, i.e., pp. 124–138, Jan. 2011.
[20] F. Meyer, B. Etzlinger, F. Hlawatsch, and A. Springer, “A distributed
+(j,j ) +(j) +(j) +(j,j )
yl = dl xl , xk , vl,k k∈A
, j = 1, . . . , J . particle-based belief propagation algorithm for cooperative simultaneous
l localization and synchronization,” in Proc. Asilomar Conf. Signals, Syst.,
Comput., Pacific Grove, CA, USA, Nov. 2013, pp. 527–531.
[21] B. Etzlinger, F. Meyer, A. Springer, F. Hlawatsch, and H. Wymeersch,
It remains to obtain samples of those entries of ỹ+ l that “Cooperative simultaneous localization and synchronization: A distrib-
+ uted hybrid message passing algorithm,” in Proc. Asilomar Conf. Signals,
are not contained in yl (cf. (31) and (37)). More specifically,
+(j,j ) +(j,j) Syst., Comput., Pacific Grove, CA, USA, Nov. 2013, pp. 1978–1982.
for each sample yl , CA l needs to obtain samples yl,l , [22] B. Etzlinger, H. Wymeersch, and A. Springer, “Cooperative synchroniza-
l ∈ Cl . This is done through communication with neighbor tion in wireless networks,” IEEE Trans. Signal Process., vol. 62, no. 11,
pp. 2837–2849, Jun. 2014.
CAs: CA l transmits to each neighbor CA l ∈ Cl the samples [23] J. Teng, H. Snoussi, C. Richard, and R. Zhou, “Distributed varia-
+(j,j ) J
yl,l j =1
, j = 1, . . . , J, and it receives from CA l ∈ Cl tional filtering for simultaneous sensor localization and target tracking
in wireless sensor networks,” IEEE Trans. Veh. Technol., vol. 61, no. 5,
+(j,j ) J pp. 2305–2318, Jun. 2012.
the samples yl,l j =1
, j = 1, . . . , J. Thus, finally, samples [24] F. Meyer, E. Riegler, O. Hlinka, and F. Hlawatsch, “Simultaneous dis-
+(j,j ) J + (j) (j) +(j) +
∼ f ỹl xl , xCl , xTl ; ur,Cl are locally available
tributed sensor self-localization and target tracking using belief propa-
ỹl j =1 gation and likelihood consensus,” in Proc. 46th Asilomar Conf. Signals,
at CA l. Syst., Comput., Pacific Grove, CA, USA, Nov. 2012, pp. 1212–1216.
[25] F. Meyer, O. Hlinka, H. Wymeersch, E. Riegler, and F. Hlawatsch,
“Distributed localization and tracking of mobile networks including non-
cooperative objects,” submitted. [Online]. Available: http://arxiv.org/abs/
ACKNOWLEDGMENT 1403.1824
[26] R. Olfati-Saber, J. A. Fax, and R. M. Murray, “Consensus and coop-
The authors would like to thank Dr. Günther Koliander for eration in networked multi-agent systems,” Proc. IEEE, vol. 95, no. 1,
illuminating discussions. pp. 215–233, Jan. 2007.
2456 IEEE JOURNAL ON SELECTED AREAS IN COMMUNICATIONS, VOL. 33, NO. 11, NOVEMBER 2015
[27] A. G. Dimakis, S. Kar, J. M. F. Moura, M. G. Rabbat, and A. Scaglione, [55] I. Arasaratnam and S. Haykin, “Cubature Kalman filters,” IEEE Trans.
“Gossip algorithms for distributed signal processing,” Proc. IEEE, Autom. Control, vol. 54, no. 6, pp. 1254–1269, Jun. 2009.
vol. 98, no. 11, pp. 1847–1864, Nov. 2010. [56] Y. Deville and A. Deville, “Exact and approximate quantum independent
[28] F. R. Kschischang, B. J. Frey, and H.-A. Loeliger, “Factor graphs and component analysis for qubit uncoupling,” in Proc. LVA/ICA, Tel Aviv,
the sum–product algorithm,” IEEE Trans. Inf. Theory, vol. 47, no. 2, Israel, Mar. 2012, pp. 58–65.
pp. 498–519, Feb. 2001.
[29] C. M. Bishop, Pattern Recognition and Machine Learning. New York,
Florian Meyer (S’12) received the Dipl.-Ing. (M.Sc.)
NY, USA: Springer-Verlag, 2006.
and Ph.D. degrees in electrical engineering from
[30] W. Burgard, D. Fox, and S. Thrun, “Active mobile robot localization by
Vienna University of Technology, Vienna, Austria, in
entropy minimization,” in Proc. EUROBOT, Brescia, Italy, Oct. 1997,
2011 and 2015, respectively. Since 2011, he has been
pp. 155–162.
a Research and Teaching Assistant with the Institute
[31] B. Grocholsky, “Information-theoretic control of multiple sensor plat-
of Telecommunications, Vienna University of Tech-
forms,” Ph.D. dissertation, Dept. Aerosp., Mechatron. Mech. Eng., Univ.
nology. He was a Visiting Scholar at the Depart-
Sydney, Sydney, Australia, 2002.
ment of Signals and Systems, Chalmers University of
[32] T. M. Cover and J. A. Thomas, Elements of Information Theory.
Technology, Gothenburg, Sweden, in 2013 and at the
New York, NY, USA: Wiley, 2006.
Centre of Maritime Research and Experimentation
[33] S. M. Kay, Fundamentals of Statistical Signal Processing: Estimation
(CMRE), La Spezia, Italy, in 2014. His research
Theory. Upper Saddle River, NJ, USA: Prentice-Hall, 1993.
interests include signal processing for wireless sensor networks, localization
[34] F. Morbidi and G. L. Mariottini, “Active target tracking and coopera-
and tracking, information-seeking control, message passing algorithms, and
tive localization for teams of aerial vehicles,” IEEE Trans. Control Syst.
finite set statistics.
Technol., vol. 21, no. 5, pp. 1694–1707, Sep. 2013.
[35] X. R. Li and V. P. Jilkov, “Survey of maneuvering target tracking—Part I:
Dynamic models,” IEEE Trans. Aerosp. Electron. Syst., vol. 39, no. 4, Henk Wymeersch (S’99–M’05) received the Ph.D.
pp. 1333–1364, Oct. 2003. degree in electrical engineering/applied sciences
[36] Y. Bar-Shalom, X.-R. Li, and T. Kirubarajan, Estimation With Appli- from Ghent University, Ghent, Belgium, in 2005.
cations to Tracking and Navigation. New York, NY, USA: Wiley, He is currently an Associate Professor with the De-
2001. partment of Signals and Systems, Chalmers Univer-
[37] J. Lien, J. Ferner, W. Srichavengsup, H. Wymeersch, and M. Z. Win, sity of Technology, Gothenburg, Sweden. Prior to
“A comparison of parametric and sample-based message representation joining Chalmers, he was a Postdoctoral Associate
in cooperative localization,” Int. J. Navig. Observ., vol. 2012, 2012, with the Laboratory for Information and Decision
Art. ID. 281 592. Systems, Massachusetts Institute of Technology.
[38] S. Farahmand, S. I. Roumeliotis, and G. B. Giannakis, “Set- Dr. Wymeersch served as an Associate Editor of the
membership constrained particle filter: Distributed adaptation for sensor IEEE C OMMUNICATION L ETTERS (2009–2013),
networks,” IEEE Trans. Signal Process., vol. 59, no. 9, pp. 4122–4138, IEEE T RANSACTIONS ON W IRELESS C OMMUNICATIONS (2013–present), and
Sep. 2011. Transactions on Emerging Telecommunications Technologies (2011–present).
[39] V. Savic, H. Wymeersch, and S. Zazo, “Belief consensus algorithms
for fast distributed target tracking in wireless sensor networks,” Signal
Process., vol. 95, pp. 149–160, Feb. 2014. Markus Fröhle received the B.Sc. and M.Sc. degrees
[40] O. Hlinka, F. Hlawatsch, and P. M. Djuric, “Consensus-based distributed in telematics from Graz University of Technology,
particle filtering with distributed proposal adaptation,” IEEE Trans. Signal Graz, Austria, in 2009 and 2012, respectively. He is
Process., vol. 62, no. 12, pp. 3029–3041, Jun. 2014. currently working toward the Ph.D. degree in elec-
[41] C. Lindberg, L. S. Muppirisetty, K.-M. Dahlen, V. Savic, and trical engineering with the Department of Signals
H. Wymeersch, “MAC delay in belief consensus for distributed tracking,” and Systems, Chalmers University of Technology,
in Proc. IEEE WPNC, Dresden, Germany, Mar. 2013, pp. 1–6. Gothenburg, Sweden. From 2012 to 2013, he was a
[42] R. Fletcher, Practical Methods of Optimization. New York, NY, USA: Research Assistant with the Signal Processing and
Wiley, 1987. Speech Communication Laboratory, Graz University
[43] A. Doucet, N. de Freitas, and N. Gordon, Sequential Monte Carlo of Technology. His research interests include indoor
Methods in Practice. New York, NY, USA: Springer-Verlag, 2001. positioning and navigation, UWB, and the interplay
[44] H. Lim and C. Kim, “Multicast tree construction and flooding in wireless between estimation theory, communications, and control in multiagent systems.
ad hoc networks,” in Proc. MSWIM, New York, NY, USA, Aug. 2000,
pp. 61–68.
[45] L. Xiao, S. Boyd, and S. Lall, “A scheme for robust distributed sensor Franz Hlawatsch (S’85–M’88–SM’00–F’12)
fusion based on average consensus,” in Proc. IPSN, Los Angeles, CA, received the Diplom-Ingenieur, Dr. techn., and
USA, Apr. 2005, pp. 63–70. Univ.-Dozent (Habilitation) degrees in electrical
[46] R. G. Bartle, The Elements of Integration and Lebesgue Measure. engineering/signal processing from Vienna Univer-
New York, NY, USA: Wiley, 1995. sity of Technology, Vienna, Austria, in 1983, 1988,
[47] S. Thrun, W. Burgard, and D. Fox, Probabilistic Robotics. Cambridge, and 1996, respectively. Since 1983, he has been with
MA, USA: MIT Press, 2005. the Institute of Telecommunications, Vienna Univer-
[48] A. T. Ihler, J. W. Fisher, R. L. Moses, and A. S. Willsky, “Nonparametric sity of Technology, where he is currently an Associate
belief propagation for self-localization of sensor networks,” IEEE J. Sel. Professor. During 1991–1992, as a recipient of an
Areas Commun., vol. 23, no. 4, pp. 809–819, Apr. 2005. Erwin Schrödinger Fellowship, he spent a sabbatical
[49] B. Ristic, S. Arulampalam, and N. Gordon, Beyond the Kalman Filter: year with the Department of Electrical Engineering,
Particle Filters for Tracking Applications. Norwood, MA, USA: Artech University of Rhode Island, Kingston, RI, USA. In 1999, 2000, and 2001, he held
House, 2004. one-month Visiting Professor positions with INP/ENSEEIHT, Toulouse, France
[50] G. E. Garcia, L. S. Muppirisetty, E. M. Schiller, and H. Wymeersch, and IRCCyN, Nantes, France. He (co)authored a book, three review papers
“On the trade-off between accuracy and delay in cooperative UWB lo- that appeared in the IEEE Signal Processing Magazine, about 200 refereed
calization: Performance bounds and scaling laws,” IEEE Trans. Wireless scientific papers and book chapters, and three patents. He coedited three books.
Commun., vol. 13, no. 8, pp. 4574–4585, Aug. 2014. His research interests include statistical and compressive signal processing
[51] R. Douc and O. Cappe, “Comparison of resampling schemes for particle methods and their application to sensor networks and wireless communications.
filtering,” in Proc. ISPA, Zagreb, Croatia, Sep. 2005, pp. 64–69. Prof. Hlawatsch was a Technical Program Co-Chair of EUSIPCO 2004 and
[52] B. W. Silverman and P. J. Green, Density Estimation for Statistics and served on the technical committees of numerous IEEE conferences. He was an
Data Analysis. London, U.K.: Chapman & Hall, 1986. Associate Editor of the IEEE T RANSACTIONS ON S IGNAL P ROCESSING from
[53] D. Q. Mayne, J. B. Rawlings, C. V. Rao, and P. O. M. Scokaert, “Con- 2003 to 2007 and the IEEE T RANSACTIONS ON I NFORMATION T HEORY from
strained model predictive control: Stability and optimality,” Automatica, 2008 to 2011. From 2004 to 2009, he was a member of the IEEE SPCOM Tech-
vol. 36, no. 6, pp. 789–814, Jun. 2000. nical Committee. He currently serves as an Associate Editor of the IEEE T RANS -
[54] D. L. Alspach and H. W. Sorenson, “Nonlinear Bayesian estimation using ACTIONS ON S IGNAL AND I NFORMATION P ROCESSING OVER N ETWORKS .
Gaussian sum approximations,” IEEE Trans. Autom. Control, vol. AC-17, He coauthored papers that won an IEEE Signal Processing Society Young Au-
no. 4, pp. 439–448, Aug. 1972. thor Best Paper Award and a Best Student Paper Award at IEEE ICASSP 2011.