Professional Documents
Culture Documents
p. 1
find different types of activities that involve physical motion deviation σ. Accordingly, the application of EM leads to an
throughout the natural environment. The input to our routine extension of the k-Means Algorithm (see e.g. [12]) to trajec-
is a collection of trajectories d = {d1 , . . . , dN } (called: the tories. Given the individual Gaussians for a model θ we can
data). The output is a number of different types of motion pat- compute the joint likelihood of a single trajectory di and its
terns θ = {θ1 , . . . , θM } a person might exhibit in their natural correspondence vector ci as follows:
environment. Each trajectory di consists of a sequence
T Y
M
Y 1 1 t t 2
di = {x1i , x2i , . . . , xTi } (1) p(di , ci | θ) = √ e− 2σ2 cim kxi −µm k . (4)
t=1 m=1
2πσ
of positions xti covered by the person. In our current system, Thereby, we exploit the fact that only one of the correspon-
these positions are computed based on a grid-based discretiza- dence variables cim in the inner product is 1, and all others are
tion of the environment: Each xti represents the position of the 0. Accordingly, the total likelihood over all values of i is given
cell covered by the person after t steps. Accordingly, x1i is by the product of the individual joint probabilities:
the first cell covered by the person and xTi is the final destina-
tion. Throughout this paper we assume that all trajectories di p(d, c | θ)
have the same length. In our current system we choose T as N T Y
M
!
the maximum length of all trajectories. Trajectories of length
Y Y 1 − 12 cim kxti −µtm k2
= √ e 2σ . (5)
T 0 < T are extended to length T by adding the final location i=1 t=1 m=1
2πσ
of that trajectory for exactly T − T 0 times.
Since the logarithm is a monotonic function we can maximize
the log likelihood instead of the likelihood. The logarithm of
2.1 Motion Patterns
(5) is given by:
We begin with the description of our model of motion pat-
terns, which is subsequently estimated from data using EM. N
X 1
Within this paper we assume that a person engages in M dif- ln p(d, c | θ) = T · M · ln √
ferent types of motion patterns. A motion pattern, denoted θm i=1
2πσ
with 1 ≤ m ≤ M , is represented by probability distributions T M
!
t 1 XX t t 2
p(x | θm ) specifying the probability that the person is at lo- − 2· cim kxi − µm k . (6)
cation x after t steps given that he or she is engaged in this 2σ t=1 m=1
motion pattern. Accordingly, we calculate the likelihood of a
Finally, we notice that we are not really interested in the log
trajectory di under the m-th motion model θm as
likelihood of the correspondence variables c, since those are
T
Y not observable in the first place. The common approach is
p(di | θm ) = p(xti | θm
t
). (2) to integrate over them, that is, to optimize the expected log
t=1 likelihood Ec [ln p(d, c | θ) | θ, d] instead which, according to
(6), is
2.2 Expectation Maximization Ec [ln p(d, c | θ) | θ, d]
In essence, our approach seeks to identify a model θ that max- " N
imizes the likelihood of the data. To define the likelihood of
X 1
= Ec T · M · ln √
the data under the model θ, it will be useful to introduce a set i=1
2πσ
of correspondence variables, denoted cim . Here i is the index T M
! #
of a trajectory di , and m is the index of a motion model θm . 1 XX
− 2· cim kxti − µtm k2 | θ, d . (7)
Each correspondence cim is a binary variable, that is, it is ei- 2σ t=1 m=1
ther 0 or 1. It is 1 if and only if the i-th trajectory corresponds
to the m-th motion pattern. If we think of the motion model as Since the expectation is a linear operator we can move it inside
a specific motion activity a person might engage in, cim is 1 if the expression, so that we finally get:
person was engaged in motion activity m in trajectory i.
Ec [ln p(d, c | θ) | θ, d]
N
In the sequel, we will denote the set of all correspondence vari- X 1
ables for the i-th data item by ci , that is, ci = {ci1 , . . . , ciM }. = T · M · ln √
i=1
2πσ
For any data item i, the fact that exactly one correspondence !
T M
is 1 translates into the following: 1 XX
− 2· E[cim | θ, d]kxti − µtm k2 , (8)
M
2σ t=1 m=1
X
cim = 1. (3)
m=1
where E[cim | θ, d] depends on the model θ and the data d.
p. 2
In particular, it generates a sequence of models, denoted 2.3 Monitoring Convergence and Local Maxima
θ[1] , θ[2] , . . . of increasing log likelihood. The EM-algorithm is well-known to be sensitive to local max-
ima in the search [5, 11]. In the context of identifying motion
Mathematically, the standard method is to turn (8) in a so- patterns, a typical local maximum involves situations in which
called Q-function which depends on two models, θ and θ0 : different types of trajectories are, with high probability, asso-
Q(θ0 | θ) = Ec [ln p(d, c | θ0 ) | θ, d]. (9) ciated with the same model component θm . In such cases, the
motion patterns may never develop into a clear model of peo-
In accordance with (8), this Q-function is factored as follows: ple’s motion, and specific trajectories may never be explained
well with any of the model components.
Q(θ0 | θ)
N
X 1 Luckily, such cases can be identified during the optimiza-
= T · M · ln √ tion. Our approach continuously monitors two types of oc-
i=1
2πσ
! currences:
T M
1 XX t 0t 2
− 2· E[cim | θ, d]kxi − µm k . (10)
2σ t=1 m=1 1. Low data likelihood: If a trajectory di has low likelihood
The sequence of models is then given by calculating under the model θ, this is an indication that no appropri-
ate motion pattern has yet been identified that represents
θ[j+1] = argmax Q(θ0 | θ[j] ) (11) this trajectory.
θ0
2. Low motion pattern utility: Our second criterion involves
starting with some initial model θ[0] . Whenever the Q-function
testing the utility of a motion patterns. The aim of this cri-
is continuous as in our case, the EM algorithm converges at
terion is to discover multiple model component that ba-
least to a local maximum.
sically represent the same people motion. To detect such
In particular, the optimization involves two steps: calculating cases, the total data log likelihood is calculated with and
the expectations E[cim | θ[j] , d] given the current model θ[j] , without a specific model component θm . Technically, this
and finding the new model θ[j+1] that has the maximum ex- involves executing the E step twice, once with and once
pected likelihood under these expectations. The first of these without θm . If the difference in the overall data likeli-
two steps is typically referred to as the E-step (short for: ex- hood is smaller than a pre-specified threshold, the effect
pectation step), and the latter as the M-step (short for: maxi- of removing θm from the model is negligible. This indi-
mization step). cates a case where a similar motion pattern exists and the
one at hand is a duplicate.
To calculate the expectations E[cim | θ[j] , d] we apply Bayes
rule, obeying independence assumptions between different Whenever the EM appears to have converged, our approach
data trajectories: extracts those two statistics and considers “resetting” individ-
E[cim | θ[j] , d] = p(cim | θ[j] , d) ual model components In particular, if a low data likelihood
trajectory is found, a new model component is introduced that
= p(cim | θ[j] , di )
is initialized using this very trajectory. Conversely, if a model
= ηp(di | cim , θ[j] )p(cim | θ[j] ) component with low utility is found, it is eliminated from the
= η 0 p(di | θm
[j]
), (12) model.
where the normalization constants η and η 0 ensure that the ex- In our experiments, we found this selective restarting and
pectations sum up to 1 over all m. If we combine (2) and elimination strategy extremely effective in escaping local max-
(12) exploiting the fact that the distributions are represented ima. As indicated in the experimental results section below,
by Gaussians we obtain: all our experiments converged to a model that clustered tra-
T
Y 1 t t[j] 2
jectories into categories 100% identical to those prescribed by
E[cim | θ[j] , di ] = η 0 e− 2σ2 kxi −µm k
. (13) us manually. Without this mechanism, the EM frequently got
t=1 stuck in local maxima and generated models that were signifi-
cantly less predictive of human behavior.
Finally, the M-step calculates a new model θ[j+1] by maxi-
mizing the expected likelihood. Technically, this is done by
computing for every motion pattern m and for each time step 3 Laser-based Implementation
t[j+1]
t a new mean µm of the Gaussian distribution. Thereby
we consider the expectations E[cim | θ[j] , d] computed in the The EM-based learning procedure has been implemented for
E-step: data acquired with laser-range finders. To acquire the data we
PN used three Pioneer I robots (see left image of Figure 1) which
t[j+1] E[cim | θ[j] , d]xti
µm = Pi=1 N
(14) we installed in the environments. The robots were aligned
[j]
i=1 E[cim | θ , d] so that they covered almost the whole environment. Typical
p. 3
Figure 2: Typical data sets obtained with three robots tracking a person in a home environment.
4 Experimental Results
4.1 Application of EM
In the first experiment, we applied our approach to learn a mo-
tion model for 42 trajectories recorded in a home environment.
Figure 4 shows for one experiment the expectations that were
computed in reach round of the EM given 14 possible motion
behaviors. In this particular experiment, we have exactly three
trajectories for each motion pattern. Each column in the pic-
tures contains the expectations E[ci1 | θ, d], . . . , E[ciM | θ, d]
for every trajectory di . To enhance the readability, we grouped
the examples belonging to the same motion pattern so that they
appear as blocks of three trajectories in Figure 4.
p. 4
Step 0:
Step 1:
Step 2:
80
60
Step 3: 40
20
0
0 20 40 60 80 100
length of trajectory [%]
p. 5
[4] H. Endres, W. Feiten, and G. Lawitzky. Field test of
W D a navigation system: Autonomous cleaning in supermarkets.
In Proc. of the IEEE International Conference on Robotics &
Automation (ICRA), 1998.
B [5] Z. Ghahramani and S. Roweis. Learning nonlinear
3
S 1 2
M stochastic dynamics using the generalized em algorithm. In
Advances in Neural Information Processing Systems (NIPS).
MIT Press, 1998.
[6] M. Kasper, G. Fricke, K. Stuernagel, and E. von Put-
tkamer. A behavior-based mobile robot architecture for learn-
ing from demonstration. Journal of Robotics and Automation,
Figure 7: Motion patterns and trajectory of a person. 34(2-3):153–164, 2001.
[7] S. King and C. Weiman. Helpmate autonomous mobile
robot navigation system. In Proceedings of the SPIE Confer-
ence on Mobile Robots, pages 190–198, Boston, MA, Novem-
person reaches position 3, C becomes unlikely and D becomes
ber 1990. Volume 2352.
the most probable motion behavior. This illustrates, that the
results of the prediction are useful even in situations in which [8] F. Kruse, E. und Wahl. Camera-based monitoring sys-
there are ambiguities about the actual intention of the person. tem for mobile robot guidance. In Proc. of the International
Conference on Intelligent Robots and Systems (IROS), 1998.
[9] G. Lacey and K. Dawson-Howe. The application of
robotics to a mobility aid for the elderly blind. Journal of
5 Conclusions Robotics and Autonomous Systems (RAS), 23:245–252, 1998.
[10] S. M. Lavalle, H. H. Gonzalez-Banos, G. Becker, and J.-
In this paper we presented a method for learning motion be- C. Latombe. Motion strategies for maintaining visibility of a
haviors of persons in indoor environments. To cluster similar moving target. In Proc. of the IEEE International Conference
behaviors into single motion patterns, we apply the popular on Robotics & Automation (ICRA), 1997.
expectation maximization algorithm. The output of our algo-
[11] G. McLachlan and T. Krishnan. The EM Algorithm and
rithm is a collection of motion patterns. We furthermore de-
Extensions. Wiley Series in Probability and Statistics, 1997.
scribed how to use the resulting models to predict the motions
of persons in the vicinity of the robot. [12] T. M. Mitchell. Machine Learning. McGraw-Hill, 1997.
[13] N. Roy, G. Baltus, D. Fox, F. Gemperle, J. Goetz,
Our approach has been implemented and applied to range T. Hirsch, D. Magaritis, M. Montemerlo, J. Pineau, S. J., and
data recorded with mobile robots equipped with laser sensors. S. Thrun. Towards personal service robots for the elderly. In
Special techniques allow the EM-algorithm to overcome local Proc. of the Workshop on Interactive Robotics and Entertain-
maxima in the likelihood-space which frequently occur in this ment, 2000.
application. In practical experiments we demonstrated that our [14] C. Schaeffer and T. May. Care-o-bot - a system for as-
method is able to learn typical motion behaviors of a person in sisting elderly or disabled persons in home environments. In
a domestic residence as well as in an office building. Based Assistive technology on the threshold of the new millenium.
on the resulting motion patterns our system can reliably pre- IOS Press, Amsterdam, 1999.
dict the motions of persons based on observations made by the
[15] R. Schraft and G. Schmierer. Serviceroboter. Springer
robot.
Verlag, 1998. In German.
[16] D. Schulz, W. Burgard, D. Fox, and A. Cremers. Track-
ing multiple moving targets with a mobile robot using particle
References filters and statistical data association. In Proc. of the IEEE
[1] H. Asoh, S. Hayamizu, I. Hara, Y. Motomura, S. Akaho, International Conference on Robotics & Automation (ICRA),
and T. Matsui. Socially embedded learning of office- 2001.
conversant robot Jijo-2. In Proc. of the International Joint [17] S. Tadokoro, M. Hayashi, Y. Manabe, Y. Nakami, and
Conference on Artificial Intelligence (IJCAI), 1997. T. Takamori. On motion planning of mobile robots which co-
[2] H. Bui, S. Venkatesh, and G. West. Tracking and exist and cooperate with human. In Proc. of the International
surveillance in wide-area spatial environments using the Ab- Conference on Intelligent Robots and Systems (IROS), 1995.
stract Hidden Markov Model. Intl. J. of Pattern Rec. and AI, [18] Q. Zhu. Hidden Markov model for dynamic obstacle
2001. avoidance of mobile robot navigation. IEEE Transactions on
[3] W. Burgard, A. Cremers, D. Fox, D. Hähnel, G. Lake- Robotics and Automation, 7(3), 1991.
meyer, D. Schulz, W. Steiner, and S. Thrun. Experiences with
an interactive museum tour-guide robot. Artificial Intelligence,
114(1-2), 1999.
p. 6