You are on page 1of 9

Neural Networks, Vol. 8, No. 1, pp.

125-133, 1995
~ Pergamon Copyright © 1994 Elsevier Science Ltd
Printed in the USA. All rights reserved
0893-6080/95 $9.50 + .00

0893-6080(94)E0045-M
C ONTRIBUT ED A R T I C L E

Neural Network Dynamics for Path Planning


and Obstacle Avoidance

R o Y G L A S I U S , A N D R Z E J K O M O D A , A N D STAN C . A. M . G I E L E N
University of Nijmegen
(Received 6 December 1993; accepted 28 February 1994)

Abstract---A model o f a topologically organized neural network o f a Hopfield type with nonlinear analog neurons
is shown to be very effective for path planning and obstacle avoidance. This deterministic system can rapidly provide
a proper path, from any arbitrary start position to any target position, avoiding both static and moving obstacles o f
arbitrary shape. The model assumes that an (external) input activates a target neuron, corresponding to the target
position, and specifies obstacles in the topologically ordered neural map. The path follows from the neural network
dynamics and the neural activity gradient in the topologically ordered map. The analytical results are supported by
computer simulations to illustrate the performance o f the network.

Keywords--Attractor neural networks, Dynamical systems, Topological map, Path planning, Robotics.

1. I N T R O D U C T I O N changes in the environment is becoming available con-


tinuously, because of the time needed to perform the
H u m a n m o t o r control reveals a versatility of function
planning task. Later, the idea of using an artificial po-
and economy of space that is yet beyond the reach of
tential field around each obstacle was proposed com-
robots. One of the important themes of research in the
bined with an attractive potential around the target
field of robotics is the design of a motion-planning sys-
(Kathib, 1986; Krogh & Thorpe, 1986). Unfortunately,
tem. Such a system should guide a robot or a robot
the potential functions proposed suffered from several
manipulator from an initial position to a target position,
problems, one being that of undesired local minima.
avoiding obstacles that are located between these po-
Later, several papers attempted to provide solutions to
sitions. If we take into account that the obstacles as
this problem (e.g., Barraquand, Langlois, & Latombe,
well as the target can move and that the obstacles can
1992; Barraquand & Latombe, 1990; Connolly, Burns,
have any shape, it becomes clear that this problem is
& Weiss, 1991; N e w m a n & Hogan, 1987; Warren,
not a trivial one. Planning a path for a robot in an
1989). None of these articles mentioned offers an exact
environment that is unknown and changing is an es-
potential-function-based algorithm that is guaranteed
pecially difficult problem.
to work. Recently, a new algorithm was proposed that
In the past, several authors (Barraquand & Latombe,
addressed the path-finding problem using an iterative
1989; Crowley, 1985; Kant & Zucker, 1986; Krogh,
approach under the constraints of m i n i m u m time or
1984; Lozano-Perez, 1983; Yap, 1986) have worked
m i n i m u m energy (Seshadri & Ghosh, 1993). However,
on the path-planning problem. Most work to date deals
this method will be very time consuming, especially
with static environments and used global methods,
for many, complex objects in the environment.
which can generally be viewed as a search process for
In this article we present a path-planning system
a path in a graph. However, global methods will limit
that can control robot motion in a static as well as in
the real-time capabilities of robots in a cluttered en-
a dynamic environment with a guaranteed solution,
vironment, especially when new information about
which is a shortest path in terms of the metric of the
representation, if one exists. We will assume that in-
formation about the position and shape of obstacles in
Acknowledgements:This work was partly supported by the Dutch the environment is not known beforehand and that it
Foundation for Neural Networks and by the Dutch Foundation for appears during the path planning, for example, through
Scientific Research (NWO).
Requests for reprints should be sent to Roy Glasius, Department real-time sensing. This may be rather artificial for path
of Medical Physics and Biophysics, University of Nijmegen, Geert planning in a static environment, but it is an inevitable
Grooteplein Noord 2 l, 6525 EZ Nijmegen, The Netherlands. situation for path planning in an environment in which

125
126 R. Glasius, A. Komoda, and S. C. A. M. Gielen

the positions of target and obstacles are changing con- We will only deal with the kinematics of the m o v e m e n t -
tinuously. This makes our approach distinct from other trajectory of the manipulator. We will assume that the
approaches (Dorst, Mandhyan, & Trovato, 1991; Lo- d degrees of freedom span a d-dimensional configu-
zano-Perez, 1983; Schwartz & Sharir, 1983), which ration space 6~. Each point in ~ defines a unique con-
presume an algebraic representation of a robot envi- figuration of the nonredundant robot manipulator in
ronment and which call for one-time off-line compu- workspace.
tation. In our approach we have only studied path The obstacles O~, a = 1 . . . m in workspace, are
planning, and not trajectory formation such as in Wada represented in the configuration space as the subset of
and Kawato ( 1993 ). In the latter approach, one has to those points in configuration space that cannot be
deal with forward and backward models of the system reached by the manipulator. This also includes config-
under control before trajectories can be generated from urations that are not allowed because it would require
an initial to a final point. For complex paths, such as that one of the links touches or passes through an object.
in a cluttered or changing environment, the total path Therefore, the "forbidden" regions in configuration
has to be segmented to create a sequence of trajectories. space are larger than the regions that correspond to the
Our path-planning system is based on a Hopfield positions of obstacles. In mathematical terms, the ob-
type of network with continuous neurons (Hopfield, stacle regions are defined in ~ by the points where the
1982). Analog neurons provide the possibility to make objects and the links of the robot manipulator share at
such a system a fast computing device, which can pro- least one point:
cess the massive a m o u n t of sensory data necessary to
O. = {q ~E d~l Y e ( q ) n O. 4= ~ } , (l)
move in a changing environment. In early attempts to
understand biological computation, neurons were where 97(q) is an area occupied by ~ at configuration
modeled as two-state threshold devices only. However, q in the workspace.
parallel computation with discrete neurons faces sta- Feasible paths between an initial configuration qinit
bility problems not found in sequential dynamics. Se- and a target configuration qtarg are continuous maps
quential d y n a m i c s - - w h e n neuron states are updated from a closed real interval into the free configuration,
one at a t i m e - - i m p l e m e n t e d in software on a conven- ~free, space that remains after removing the points
tional computer is simply too slow for any large network O~ from ~
application. These stability problems associated with m
parallel dynamics can be eliminated by using contin- efree = ~ o \ U ~.Oa. (2)
a-I
uous neurons (Marcus, Waugh, & Westervelt, 1990).
They also provide the possibility for hardware imple-
mentation. 2.2. Neuronal Space and Phase Space
In Section 2 we define the model and show its basic
properties. We define two types of dynamics: parallel Our path-planning system consists of a large collection
discrete- and continuous-time dynamics. We show for of identical processing units called neurons. These
both types of dynamics that the network, after receiving neurons are arranged in a d-dimensional cubic lattice
an external stimulus, evolves towards a specific state of (CL). The n u m b e r of neurons in the network is N.
neural activity that corresponds to a m i n i m u m of a They are connected only to their z nearest neighbors
Liapunov function. The path is given by the neural by excitatory and symmetric connections T o . Gener-
activity gradient. This path has the shortest length alization to other types of lattices can be obtained very
among all paths connecting initial and goal position. easily. We assume that the lattice represents a topolog-
Computer simulations illustrating the performance of ically ordered map, which can be obtained by a Ko-
the network are presented in Section 3. Section 4 sum- honen type of learning (Fritzke, 1991; Martinetz, 1993;
marizes the conclusions. Ritter, Martinetz, & Schulten, 1989). This m a p gives
a discrete topologically ordered representation of the
robot configuration space ~ (Heikonen, Koikkalainen,
2. T H E M O D E L & Oja, 1993; Morasso, Sanguineti, & Tsuji, 1993;
Vleugels, Kok, & Overmars, 1993) with nodes homo-
2.1. Work Space and Configuration Space
geneously distributed over the configuration space. We
Consider a robot manipulator Y~, equipped with sensors will call it the neuronal space Af. To each neuron in
and actuators, is operating in a subset of the real world. Af corresponds a certain subset of ~ called its receptive
This robot should have the capability of planning its field. In this discrete representation of the configuration
own motion by generating a path specified by an initial space 6° the regions U ~ ~0~ are represented by a subset
configuration, a final configuration, and by the position of nodes of the CL. We will call the nodes, which cor-
of obstacles in the workspace. We assume that the non- respond to U ~m=1 ~ , occupied nodes. Similarly, the ini-
redundant robot manipulator has d degrees of freedom, tial configuration qinit and the target configuration qta,g
which may correspond to the d joints of the robot arm. are represented by one node each. The external input
Dynamics for Path Planning and Obstacle Avoidance , 12 7

signals, which provide information about the position with G( ai ) = f ~ g - * ( x ) d x , is a global Liapunov func-
of obstacles in workspace, are supposed to clamp the tion for eqns (5) and (6). The vector a refers to the
activity of all neurons in the occupied nodes to the state of the neural network with components O-i. The
value zero corresponding to the minimal activity of the existence of a Liapunov function guarantees that the
neuron. The activity of the neuron corresponding to dynamics of the network always converges to a fixed
the target configuration node is clamped to the value point. However, in the discrete-time parallel dynamics
1. The initial configuration of the robot may correspond L is a Liapunov function if the stability condition
to any node that is not occupied.
3 < I 1/Xmi, I (8)
The activities of the neurons are characterized by
real-valued variables ai G [0, 1] and describe the state is satisfied. Here 3 is the steepest slope of the function
of the system. The set g of all possible states O-of the g ( x ) and ~krnin is the most negative eigenvalue of the
network is called the phas~ space of the system. In our matrix T. If all eigenvalues are positive then L is a
model g = [ 0, l ] N, where [ 0, 1] ~ stands for a hypercube Liapunov function for any 3. This can be understood
in ~u. if we calculate AL ~ L ( a ( t + 1)) - L(o-(t)):
1
AL = -'~ ~ Tij[ A~iAffj q- 2afit)Aai] - ~ IiAai
2.3. Dynamics t,J i

The state variables O'i, i = 1 . . . N, of the neurons can + ~ [G(ai(t + 1 ) ) - G ( a i ( t ) ) ] (9)


change due to inputs from other neurons in the network i

and due to external sensory input. The total input ui where Ao-i =- ai(t + 1) -- O-i(t). By Taylor's theorem
to the neuron i is a weighted sum of activities from we obtain
other neurons and of an external sensory input I:
N
G(~i(t + 1)) - G(tri(t))
u~(t) = ~ Toas(t) + I~ (3) = G'(~ri(t + 1))A~ri -- ½ G"(~)(A~i) 2
J
<--G'(~i(t + l))A~r~- ½(A~r~)2 rain G"(x) (10)
where the strength of the synaptic connection from node x~[0,1]

j to node i is represented by T 0. In our model the con- where ~ ~ [ O-i(t), O-i(t + 1 ) ]. Inserting eqn (10) in eqn
nections T o are excitatory ( T O > 0), symmetric ( T o = (9) and using the definition of G ( x ) we obtain
T~.i) and short range:
1
AL < --~ ~,.. [To+ ~ij/[J]AaiAaj (11)
{10 if P ( i , j ) < r t,)
T° = otherwise. (4)
where 60 is the Kronecker delta-function. Here 1/3 =
Here r is a positive number and p( i, j ) is the Euclidean minxG"(x) is the minimum curvature of the function
distance between neuron i and j in .At. G ( x ) in the closed interval [0, 1]. The right-hand side
The evolution of the configuration of this network of the inequality ( 11 ) is negative if the matrix T o. + 60/
can be defined by the discrete-time or continuous-time 3 is positive definite. A sufficient condition for matrix
dynamics. In the case of a time-discrete evolution the T ÷ 3-1I to be positive definite is given by eqn (8).
states of neurons are updated by In the case of the continuous-time dynamics, for
each finite value of 3 the dynamics (6) converges to a
a~(t + 1) = g( ~ Toa~(l) + Ii ) (5) fixed point that is a (local or global) minimum of the
J Liapunov function. By differentiation of eqn (7) and
where g ( x ) is a sigmoid function. using eqn (6) we obtain
The dynamics (5) may be parallel, sequential, or dL(a(t)) ~ OL(a(t)) d~i(t)
may be a random, asynchronous sequence. In the con- dt
.~
~ 0~ dt
tinuous-time updating, all neurons continuously and
N
simultaneously change their states. The change of ac-
= __~ (g-l(ffi) __ Ui)(O. i -- g ( u i ) ) <-- 0 (12)
tivity of the neurons is then given by the set of nonlinear i
differential equations (Grossberg, 1988 )
because g-l(o-) _ u and ~r - g ( u ) have the same sign.
dai(t)= g~ ~ Ti.ias(t) + I i ( t ) ) - ~ri(t). (6) Calculating the time derivative of eqn (3) we can also
dt ~ obtain a representation o f e q n (6) in terms of the input
variables ui:
Both types of dynamics lead to the same equilibrium
state. The function dui(t) ~
= ~ Tog(uy(t)) + I~ - u~(t). (13)
dt J
1
t(a) = -~ ~ Tijo'itr j -- ~ Iitri q- ~ a ( t r i ) (7)
t,j i i The set of eqns (13) was first proposed by Hopfield
128 R. Glasius, A. Komoda, and S. C. A. M. Gielen

(1984). There is a subtle difference though, between When all eigenvalues of the matrix T 0. are negative,
eqns (13) and (6). Hopfield assumed that the activity then L is strictly convex function. If some of them are
of neurons changes at the same time as when the neu- positive then sufficient condition for strict convexity is
rons receive input: given by

ai(t) = g ( u i ( t ) ) . (14) ~ < l/X .... (18)

In the case of dynamics (3) and (6) this assumption In the case of parallel discrete-time dynamics we have
is only valid when the system has reached a stationary to satisfy simultaneously the stability condition (8).
state. The dynamical system defined by eqn (13) to- When we choose
gether with relation (14) has the same Liapunov func-
/3<l/X, X=max{IXminl, IXmaxl}, (19)
tion as eqns (3) and (6) (Hopfleld, 1984). By direct
computation, then the Liapunov function [ eqn (7) ] is strictly convex
on ~ and at the same time condition (8) is satisfied.
dL(~r(t)) _ ~ OL(~r(t)) da~(t) The strict convexity implies that a local m i n i m u m of
~.,
dt ~ &ri dt the function (7) is a global one and is unique. In this
_ ~ [ d u i ( t ) ~ 2 ,, case the set of eqn (16) has only one solution. This
equilibrium state depends only on the position of the
obstacles and on the position of the target.
Here we have used the symmetry of TO. and the mono- Consider first the situation with static obstacles and
tonic increase of the function g. The set of eqns (13) with a static target. If the neuron in the initial config-
and the Liapunov function (7) is a particular case of uration nodejinit is active, then a path connecting initial
the more general dynamical system, the stability of position and target position is created in the following
which was studied by Cohen and Grossberg ( 1983 ). way. A new configuration for the robot manipulator is
given by the position of the neighboring neuron with
the largest activity, which serves then as the starting
2.4. Feasible Paths position for a new change. Every time when a new con-
figuration is reached a new decision is made concerning
To construct a feasible path we can use both discrete- the next configuration. This procedure is repeated until
and continuous-time dynamics given by eqns ( 5 ), (6), the neuron corresponding to the actual position of the
or ( 13 ). In our simulations we have used discrete-time robot becomes the target neuron. The created path ~P
dynamics (5) for reasons of simplicity. At the initial
passes through a sequence of points Jinit = J0, Jl . . . . .
time, t = 0, the activity of all neurons is set to zero. Jtarg = jm. The length of the path ?9 passing through
Under influence of an external input, which clamps the these sequence of points is given by
neuron at the target node to the value 1 and which
clamps all neurons in the occupied nodes to zero, the m-1

state of the system begins to change according to the D ( ~ ) = ~, P(Ji,Ji+*). (20)


i-0
dynamics of the network described by eqn (5). The
evolution of the network can be seen in the phase space In our case p(Ji, ji+l ) is equal to the lattice constant.
g as a motion of a point along the curve on which the Because the path leads from one node of the lattice to
Liapunov function decreases. This down-hill motion the neighboring node with the largest activity and ends
ends when the network reaches an equilibrium state, in the node with activity 1, the number of steps and
which is a local or global m i n i m u m of L. The network the length of the path on the grid is minimal. All other
cannot display an oscillating behavior. This is guar- paths, which pass through a sequence of points in which
anteed by condition (8). The final equilibrium states the activity does not increase monotonically or does
are solutions of the fixed point equations not increase in the optimal way (i.e., largest increment),
consist of larger number of steps and have a length larger
(16) than D (~0) because all steps have the same length. Due
~ j to the discretization of the configuration space, the path
~o has to stay on the nodes of the grid and is only an
for i = 1. . . . . N. This final equilibrium state is unique approximation of a smooth curve in 6~. The degree of
when the Liapunov function is strictly convex. If the accuracy is only limited by the grain size of the grid.
second-order partial derivatives of L exist then L is
strictly convex iffthe Hessian matrix L o =- 02 L / &ri Oaj
is positive definite. In our case 3. C O M P U T E R SIMULATIONS

1 In this section we illustrate some of the analytical results


Lij= - T i j + 6 O g,(g_~(ai) ) < - T i j + 6 o l ~ 3 . (17) of the algorithm outlined above. The range of possible
Dynamics for Path Planning and Obstacle Avoidance 129

applications includes problems classically handled by tions the configuration of the robot manipulator starts
the path-planning community, such as: controlling a to change at the m o m e n t when one of the neighboring
robot arm, car manoeuvring (Dorst et al., 1991 ), or neurons of the actual neuron is activated by the target
the piano movers problem. We have chosen four ex- neuron. This is essential when the trajectory must be
amples: an autonomous point-robot trying to find the planned in a changing environment. To understand that
center of a labyrinth; a two-link robot manipulator planning can start before a stable state is reached, we
moving in a cluttered environment; a point-robot consider the activity in the neural m a p at equilibrium.
avoiding a moving obstacle to reach the target; and a The shape of the activity in the m a p is a hill with the
point-robot following and trying to catch up with a top located at the target neuron and activity zero at the
moving target. obstacle neurons. In this activation pattern lines ofiso-
In our demonstration we used a two-dimensional activation can be defined. The structure and shape of
lattice (d = 2) with N = 2500 neurons ordered in a 50 these lines depend only on the location of the target
× 50 neural map. All four examples are two-dimen- and obstacles. Assume that target and obstacles are
sional path-planning problems. Note, however, that the fixed. At time zero only the target neuron has a nonzero
model can easily be extended to higher dimensions and activity. Initially, due to the dynamics of the neurons,
may have other types of homogeneous lattices. Accord- all neurons located on the first iso-activation line receive
ing to eqn (4), each neuron is connected bidirectionally input from the target neuron and are activated. In the
to the neurons that are lying within a distance r in the following m o m e n t they raise their activity and send in-
neural space around it. By choosing r -- 1.5 each neuron formation to the neurons on the second iso-activation
has eight neighbors (z = 8). Each simulation starts line. This will go on, and the activity of neurons on
with the activity of the neuron closest to the target po- successive iso-activation lines is raised from zero until
sition clamped to value 1. The activities of all neurons the activation landscape is stable. During this dynamics
that correspond to obstacle positions are clamped to the shape of the iso-activation lines is constant and the
the value zero. amplitude of activity decreases with increasing number
At each time step, neurons update their activations of the iso-activation line. Hence, the maximal gradient
in parallel, according to eqn (5). The changes in neuron on every position has a constant direction from the
activations stop when the network reaches an equilib- moment when the activation front reaches that location.
rium state.
The new actual neuron is the neuron with the highest
3.1. The Labyrinth
activity, chosen from the last actual neuron and its eight
neighbors. If two or more of the neighboring neurons In the first simulation we test the performance of the
have the highest activation then one of them is chosen network in the complex environment of a labyrinth.
randomly. Each configuration of a point-robot moving in a two-
Any monotonically increasing function can be used dimensional workspace is described by the two position
in our algorithm as the transfer function. In the sim- coordinates x and y. In the case of the point-robot the
ulations we chose both g ( x ) = t a n h ( ~ x ) and g ( x ) = configuration space is isomorphic to the two-dimen-
~x. To choose a value of ~ that satisfies the condition sional work space. Hence the topological neural m a p
(19), we have to estimate the largest and the smallest represents a discretization of the workspace. Each neu-
eigenvalue of the matrix T. Because each neuron is ron in it corresponds to a small unit area in the work-
connected to eight neighbors, the n u m b e r of nonzero space. Neighboring neurons correspond to neighboring
elements in each row or column is at most equal to areas. By correspondence, neurons in the neural m a p
eight. The nonzero elements are all equal to one. From represent the target position, the initial actual position,
Frobenius theorem (see Bodewig, 1959 ) it follows then and the obstacles. We simulated several situations. In
that all eigenvalues ~ ~ [ - 8 , 8 ]. In the absence of ob- each case, ifa continuous path between target and initial
stacles and when we impose periodicity of the neural position existed, the neural net found it. In the case
map, all eigenvalues are positive and the largest eigen- that more than one solution was possible, the one with
value is equal to eight. The matrix ~ T in this case is a the shortest path length was chosen. In Figure 1 we
double stochastic matrix. So we have to choose/~ ~ [0, show the neural m a p with a set of obstacle neurons
1.25 ) to guarantee the stability condition and unique- that correspond to labyrinth walls. The initial actual
ness of the equilibrium state. position is chosen outside the labyrinth at point S and
For both, the linear and t a n h ( . ) i n p u t / o u t p u t be- the target position is chosen in the center of the laby-
haviour of the neurons, we chose ~ = 0.1. The paths rinth, at point T. The obstacle neurons are represented
generated by these functions are the same. The com- by black dots and the neurons corresponding to the
putational time, however, is much shorter in the linear actual position in successive time steps by open circles.
case. Note that several paths may lead to the target position,
An optimal path can be found even before the net- but that the path generated by the network has the
work settles into an equilibrium state. In our simula- shortest length.
130 R. Glasius, A. Komoda, and S. C. A. M. Gielen

can vary from 0 to 27~ periodically. Both joints can


move simultaneously or move only one at a time. The
optimal path generated by the system is given by the
minimum joint motion of the robot arm in the work-
space.
In workspace we can choose an initial configuration
of the two-link manipulator, a target configuration, and
the obstacle positions. The shape of the obstacles in
the configuration space depends on the exact shape of
the robot arm.
In Figure 2A the workspace with the two-link ma-
nipulator is presented. The gray circles are obstacles.
The manipulator has been drawn at each time step,
illustrating the movement. In Figure 2B we show the
. . . ..“...“.......“.“......d corresponding neural space that is a discretization of
l f
f . . . . . . . ..“...“.....““ooo..“.““... f the configuration space. The black dots represent the
neurons that correspond to the forbidden configurations
of the manipulator. The open circles drawn in Figure
FIGURE 1. The labyrinth. Black dots represent neurons corre- 2B show the successive actual neurons that correspond
sponding to walls of the labyrinth. The sequence of open circles
to the manipulator configurations at every time step.
gives the path generated by the network. These dots also cor-
respond to the actual neuron at the successive time steps. The
initial position is at point S, the target position at point T.
3.3. Changing Environment

3.3.1. Moving Target. Consider a situation in which


3.2. Planar Robot
the target is moving and the robot’s task is to follow
In this example we consider a two-joint robot arm and to try to catch it. In this case the sensory infor-
moving in a two-dimensional workspace in the presence mation has to be processed continuously to update the
of obstacles. In contrast with the point-robot, all objects network for changes in the environment. The activity
have real physical dimensions. The configuration of the of neurons will change all the time in the direction of
robot arm is described by two joint angles 8, and 13~. the equilibrium configuration imposed by the instan-
We assume no constraints on the joints. In this case, taneous position of the obstacles and the target. The
we impose periodic boundary conditions: the neurons path, as in the static case, is determined by the neural
on the border of the map are connected to the neurons activity gradient. The difference with the static envi-
on the opposite border. The two-dimensional map is ronment is that the network cannot settle in an equi-
topologically equivalent to a torus and the joint angles librium state because the neuronal representation of

(4 (B)

FIGURE 2. Two link robot manipulator. (A) Representation of the workspace. Gray circles represent obstacles. The robot manipulator
is drawn in successive positions. (B) The configuration space. Black dots represent the neurons corresponding to the obstacles.
Open circles show the actual neuron at successive time steps.
Dynamics for Path Planning and Obstacle Avoidance 131

3.3.2. Moving Obstacles. In addition to avoiding static


. . . . .

obstacles, path planning must also include collision


avoidance with moving objects. To illustrate this, we
consider the case in which the point-robot can reach
the target in two ways. At the beginning of the simu-
lation one path is blocked. During the simulation, after
the robot moves along a substantial part of the path
(Figure 4A), we move the obstacle to obstruct this path
and free the other. This causes a rapid change of neural
activities and the direction of the neural activity gra-
dient. As a result, the robot reverses and reaches the
target via the other path (Figure 4B).

4. C O N C L U S I O N S

Soo!° We have proposed a model for path planning and ob-


stacle avoidance based on an analog neural network.
FIGURE 3. The pursuit. Black dots represent the target position The network is a large collection of locally and sym-
in successive time steps. Open circles show the actual position metrically connected elementary processors. The evo-
at successive time steps. lution of the network is given by parallel discrete- or
continuous-time dynamics. When the network receives
target and obstacles will change all the time. This will an external input, the neurons in the network start to
influence the direction of the gradient that will change change their activity towards a specific value that cor-
not only from node to node but also change with time responds to a minimum of the Liapunov function. The
for a given node. This gradient contains the relation direction of the motor response, the path, is determined
between neural activities and the direction of move- by the neural activity gradient.
ment. A collision-free path is generated only when the Our neural network algorithm to compute optimal
speed of the moving obstacles is smaller than the rate feasible paths is an activity wave propagation, analogous
of change of neural activities. to the Huygens principle. A distance transform method
As an example of how the network generates a path (Jarris, 1985 ) is an example of this kind of algorithm
in this case, we displaced the target by one step, after used in robotics for path planning. However, our net-
each iteration. Generally, the path made by the network work can more easily respond to sudden changes in the
is always shorter than the trajectory of the target (Figure environment, like moving obstacles and target, and with
3). This gives the robot the opportunity to catch up the possibility for hardware implementation the neural
with the target, even if both have the same speed. network is more powerful than the distance transform.

(A) (B)

L
FIGURE 4. Moving obstacle. (A) Black dots represent neurons corresponding to the labyrinth walls and the obstacle. Open circles
show the actual position of the point-robot at successive time steps from the start until the moment when the obstacle is moved.
(B) When the open pathway is blocked, the robot changes its direction towards the new open "gate."
132 R. Glasius, A. Komoda, and S. C. A. M. Gielen

O u r m e t h o d r e s e m b l e s D i j k s t r a ' s a l g o r i t h m o f find- tures. In Proceedings of the International Joint Conference on


i n g the l e n g t h s o f the shortest p a t h s f r o m a given vertex Neural Networks ( Seattle/WA ). If, 531-536.
Graf, H. P., Jackel, L. D., Howard, R. E., Straughn, B., Denker,
( t a r g e t n e u r o n ) to all t h e r e m a i n i n g vertices o f the
J. S., Hubbard, W., Tennant, D. M., & Schwartz, D. ( 1986 ). VLSI
graph. However, the t i m e t a k e n b y D i j k s t r a ' s a l g o r i t h m implementation of a neural network memory with several
o n a g r a p h w i t h N vertices is O ( N 2 ) a n d i n o u r m e t h o d hundreds of neurons. In J. S. Denker fed.), Neural Networks for
it is O ( N ) . Computing (Snowbird 1986). p. 182. New York: American In-
F o u r p o i n t s are w o r t h n o t i n g a b o u t the f u n c t i o n a l stitute of Physics.
Grossberg, S. ( 1988). Nonlinear neural networks: Principles, mech-
p r o p e r t i e s o f the n e t w o r k . First, pilot c o m p u t e r exper-
anisms, and architectures. Neural Networks, 1, 17.
i m e n t s d e m o n s t r a t e d t h a t the s y s t e m c o u l d r e m a i n Heikonen, J., Koikkalainen, P., & Oja, E. (1993). From situations
f u n c t i o n a l despite local d a m a g e s i n t h e n e t w o r k . T h e to actions: Motion behavior learning by self-organization. In Pro-
q u a l i t y o f the p a t h will be s o m e w h a t degraded, b u t it ceedings of the International Conference on Artificial Neural Net-
will still be a feasible path. Second, we w a n t to stress works (Amsterdam~The Netherlands). pp. 262-267. London:
Springer-Verlag.
the n e t w o r k ' s i n s e n s i b i l i t y to the choice o f t h e t r a n s f e r
Hopfield, J. J. (1982). Neural networks and physical systems with
f u n c t i o n . T h e s y s t e m is a b l e to p e r f o r m correctly, even emergent collective computational abilities. Proceedings ~f the
if different t r a n s f e r f u n c t i o n s are u s e d for the n e u r a l National Academy ~f Sciences of the United States of America,
m a p . T h i r d , c o n t i n u o u s n e u r o n s e l i m i n a t e oscillations 79, 2554.
related to parallel d y n a m i c s o f discrete n e u r o n s ( M a r - Hopfield, J. J. (1984). Neurons with graded response have collective
computational properties like those of two-state neurons. Pro-
cus et al., 1 9 9 0 ) . F o u r t h , the parallel a r c h i t e c t u r e a n d
ceedings of the National Academy of Sciences of the United States
d y n a m i c s o f the m o d e l give it the large speed n e e d e d of America, 81, 3088.
for r e a l - t i m e processing o f s e n s o r y i n f o r m a t i o n b e c a u s e Jarris, R. A. ( 1985). Collission-free trajectory planning using distance
all n e u r o n s s i m u l t a n e o u s l y a n d c o n t i n u o u s l y c h a n g e transforms. Mechanical Engineering Transactions of the IE Aus-
their a n a l o g states. Because the basic idea o f the n e t w o r k tralia, MEI0, 187.
Kant, K., & Zucker, S. W. (1986). Towards efficient trajectory plan-
c a n be expressed b y electrical circuits, m o d e r n tech-
ning: The path-velocity decomposition. The International Journal
nologies s h o u l d p r o v i d e the possibilities to i m p l e m e n t of Robotics Research, ~, 72-89.
the n e t w o r k with a large n u m b e r o f processing e l e m e n t s Kathib, O. (1986). Real-time obstacle avoidance for manipulators
i n h a r d w a r e . G r a f e t al. ( 1 9 8 6 ) were able to m a k e cus- and mobile robots. The International Journal ~f Robotics Research,
t o m chips w i t h N = 256 fully c o n n e c t e d u n i t s , u s i n g ~, 90-98.
Krogh, B. H. (1984). A generalized potential field approach to obstacle
2 N 2 ~ 130,000 resistors. Because i n o u r n e t w o r k all
avoidance control. In Proceedings of the IEEE International Con-
c o n n e c t i o n s are positive a n d local, the n u m b e r o f re- ~,rence on Robotics and Automation. p. 948. (Washington D.C.)
sistors r e q u i r e d will be o f the s a m e o r d e r as the n u m b e r Krogh, B. H., & Thorpe, C. E. (1986). Integrated path planning and
o f n e u r o n s i n the n e t w o r k . dynamic steering control for autonomous vehicles. In Proceedings
of the IEEE International Conference on Robotics and Automation.
pp. 1664-1669. Los Angeles: Computer Society Press of the IEEE.
REFERENCES Lozano-Perez, T. (1983). Spacial planning: a configuration space
approach. IEEE Transactions on Computers, C-32, 108-120.
Barraquand, J., Langlois, B., & Latombe, J. C. (1992). Numerical Marcus, C. M., Waugh, E R., & Westervelt, R. M. (1990). Associative
potential field techniques for robot path planning. 1EEE Trans- memory in an analog iterated-map neural network. Physical Re-
actions on Systems, Man. and Cybernetics, 22, 224-241. viewA, ,11, 3355.
Barraquand, J., & Latombe, J. C. (1989). Robot motion planning Martinetz, T. M. (1993). Competitive hebbian learning rule forms
with many degrees of freedom and dynamic constraints. In Pro- perfectly topology preserving maps. In Proceedings of the Inter-
ceedings of the 5th International Symposium on Robotics Research national Conference on Artificial Neural Networks (Amsterdam/
(Tokyo). pp. 74-83. The Netherlands). pp. 427-434. London: Springer-Verlag.
Barraquand, J., & Latombe, J. C. (1990). A monte-carlo algorithm Morasso, P., Sanguineti, V., & Tsuji, T. (1993). Neural architecture
for path-planning with many degrees of freedom. In Proceedings for robot planning. In Proceedings of the International ConJbrence
of the IEEE International Conference on Robotics and Automation on Artificial Neural Networks (Amsterdam/The Netherlands). pp.
( Cincinnati, OH). pp. 1712-1717. Los Alamos, CA: IEEE Com- 256-261. London: Springer-Verlag.
puter Society Press. Newman, W. S., & Hogan, N. (1987). High speed robot control and
Bodewig, E. (1959). In Matrix Calculus. Amsterdam: North-Holland. obstacle avoidance using dynamic potential function. In Proceed-
Cohen, M. A., & Grossberg, S. (1983). Absolute stability of global ings ~f the IEEE International Conference on Robotic~ and Au-
pattern formation and parallel memory storage by competitive tomation (Raleigh, NC). pp. 14-24. Los Angeles: Computer So-
neural networks. 1EEE Transactions on Systems, Man, and Cy- ciety Press of the IEEE.
bernetics. 13, 815. Ritter, H. J., Martinetz, T. M., & Schulten, K. J. (1989). Topology-
Connolly, C. I., Burns, J. B., & Weiss, R. ( 1991 ). Path planning using conserving maps for learning visuo-motor-coordination. Neural
Laplace's equation. In Proceedings of the IEEE International Networks, 2, 159-168.
Conference on Robotics and Automation. pp. 2102-2106. Los Schwartz, J. "E, & Sharir, M. (1983). On the "piano" movers problem.
Alamos: IEEE Computer Society Press. II General techniques for computing topological properties of real
Crowley, J. L. (1985). Navigation for an intelligent mobile robot. algebraic manifolds. Advances in Applied Mathematics, 10, 298-
IEEE Journal of Robotics and Automation, RA-I, 31. 351.
Dorst, L., Mandhyan, I., & Trovato, K. ( 1991 ). The geometrical rep- Seshadri, C., & Ghosh, A. ( 1993 ). Optimum path planning for robot
resentation of path planning problems. Robotics and Autonomous manipulators amid static and dynamic obstacles. IEEE Trans-
Systems, 7, 181. actions on Systems, Man, and Cybernetics, 23, 576-584.
Fritzke, B. ( 1991 ). Unsupervised clustering with growing cell struc- Vleugels, J. M., Kok, J. N., & Overmars, M. H. (1993). A self-or-
Dynamics for Path Planning and Obstacle Avoidance 133

ganizing neural network for robot motion planning. In Proceedings Aar~ neuron j corresponding to the target
of the International Conference on Artificial Neural Networks configuration
(Amsterdam~The Netherlands). pp. 256-261. London: Springer-
Verlag. ~min the most negative eigenvalue of matrix T
Wada, Y., & Kawato, M. ( 1993 ). A neural network model for arm L(cr) Liapunov function
trajectory formation using forward and inverse dynamics models. N number of neurons
Neural Networks, 6, 919-932. ~V neuronal space
Warren, C. W. (1989). Global path planning using artificial potential
O, obstacle a in workspace, a = 1 . . . m
fields. In Proceedings of the IEEE International Conference on
Robotics and Automation ( Scottsdale, A Z). pp. 316-321. Los
b. subset of ~ corresponding to obstacle a
Angeles: Computer Society Press of the IEEE. ~- number pi
Yap, C. K. (1986). In Algorithmic motion planning. Hillsdale, NJ: ~o path
Edbaum. q vector in configuration space
qinit initial configuration
NOMENCLATURE
qtarg target configuration
3 the steepest slope of the function g(x) p( i,j) the Euclidean distance between neuron
configuration space i and j in ./V.
~ free not occupied part of the configuration r a positive real number
space Y~ robot manipulator
CL cubic lattice set of real numbers
I~onecker delta # = [0, 1] N phase space, set of all possible states a
D(~o) length of path ~ ff system state
d dimension of the configuration space or ai(t) state of neuron i at time t
the number of degrees of freedom 01, 02 joint coordinates in configuration space
G(ai) integrated sigmoid function To synaptic connection from neuron j to i
g(x) sigmoid function t time
L external sensory input to neuron i u~(t) total input to neuron i at time t
Jinit neuron j corresponding to the initial X, ~ space coordinates in configuration space
configuration number of nearest neighbors

You might also like