You are on page 1of 6

Proceedings of the

2005 IEEE
Mcchatronics
2005, Taipei, Taiwan

International Conference on
July ]()-12,

Development of Intelligent Vision Fusion Based


Autonomous Soccer Robot
1

Chung-Hsien Kuo,

Member,

2Chun-Ming Yang, and 3 Fang-Chung Yang, Student Mcmucr, IEEE

IEEE

Graduate Institute of Medical Mechatronics, Chang Gung University, Tao-Yuan 333, Taiwan, R,O,C.,
e-mail: chkuo@mail.cgu.edu.tw
2.3 Graduate Institute of Mechanical Engineering, Chang Gung University, Tao-Yuan 333, Taiwan, R.O .c.,
e-mail: {sula.middle}@dcalab.cgu.edu.tw

This

Abstract

paper

proposes

the

hands-on

implementations of the "Formosa middle size soccer robot".


which is developed based on the intelligent vision fusion sensing
architecture. Tbe proposed autonomous soccer robot consists of
the

mecbanical

platform,

motion

control

module,

"MapCam360" omni-directional vision module, front vision


module, image processing and recognition module, investigated
target object positioning and real coordinate reconstruction,
robot path planning, subgoal definition, competition strategies,
and obstacle avoidance. Especially, in order to increase the
robustness

of

the

vIsion

sensing,

the

catadioptric

omni-directional and front vision modules are dynamically


switched according to the ball recognition results of the
omni-directional camera. The previous mentioned hardware
and software modules and components are all integrated, and
the control modes are fused based on the relationships among

the positions of the enemy robot, ball, goal, and field


boundaries. In order to increase the motility and attackability,

we design the dribbling and kicking mecbanisms. In addition.


the PC-based soccer robot controller is developed based on the
embedded Linux

platform.

The motion control uses the

Mitsubishi AC-servo motors and controllers to achieve higher


reliability. Finally, the physical robot had been finished and
tested, and we also validate the functional performance of
"Formosa Middle size soccer robot" in terms of participating
the "2004 Taiwan's
competition".

annual

middle

size

soccer

robot

Index Terms - Image servo tracking, image processing and


recognition, omni-directional vision, intelligent vision fusion,
middle size soccer robot.

I. INTRODUCTION

With the increasing developments of the artificial


intelligence (AI) and the robotics, the robot soccer
competition game was proposed to emulate the human's
soccer game. The robot soccer tournament was addressed
and fonnulated based on the real-time and practical
competition environment and problems. In general, the
soccer robot integrates the techniques of the robotics,
mechanics, control circuit design, motion controller design,
image acquisition and processing, autonomous robot
behavior modeling and programming,
intelligent
decision-making, etc. Therefore, the robot soccer tournament
is not only an interesting research topic, but also a
challenging research problem. Therefore, the robot soccer
game can be fonned as a dynamic and unpredictable
collaborative system. The RoboCup [12] and Federation of
0-7803-8998-0/05/$20.00 2005 IEEE

International Robot-soccer Association (FlRA) [9] are two


major organizations to orgamze the robot soccer
competitions.
There are several literatures discussed the visions of the
mobile robots. S.S Lin et al. [6] proposed a novel
omni-directional stereo system to capture 3D depth in real
time without additional restrictions on the operating
environment. The proposed system achieved high resolution.
It also promoted good resolution, data-flow, precision and
speed to meet omni-directional stereo robot sensor demands.
T. Schmitt et af. [7] developed and analyzed a probabilistic,
vision-based state estimation method for individual
autonomous robots. The proposed approach enabled a team
of mobile robots to estimate their joint positions in a known
environment, and the robots were capable of tracking the
positions of autonomously moving objects.
M
Dietl et al. [2] developed a new method to track
multiple and single objects using noisy and unreliable sensor
data in tenus of Kalman filtering and Markov localization
approaches. The results were compared with a simple
averaging method to emphasize their contributions. P.
Bonnin et al. [1] proposed a general methodology to
introduce behavioral and visual priori knowledge and
applied the approach in their works for the RoboCup
quadruped legged league. Finally, J. Jia et al. [4] proposed
the modular and object oriented approach for constructing
the autonomous mobile robot system. This system was
developed modularly based on the Windows operation
system. The soccer robots used the omni-vision modular for
the navigation purposes.
The middle size soccer robot is developed as a fully
autonomous mobile robot, and it has to complete the robot
soccer game based on practical concerns only by "itself'.
That is, all the tasks, including image processing, object
recognition, decision-making, path planning and robot
driving, must be done on the robol. Therefore, in addition to
develop the robot driving mechanics, the image acquisition
and software computing modules must be developed as the
kernels of the middle size soccer robot. In this paper, the
middle size soccer robot is implemented as the following
hardware and software modules.
I. Mechanical and driving modules: the mechanical
design plays an important part for the soccer robot. In this
work, we construct a strong mechanical structure as the robot
platfonn. In addition, the industry proof AC servo driving
modules are used to drive the soccer robot in a reliable
.

124

manner.

Finally, we design the dribbling and kicking

mechanism

on

the

robot

to

increase

the

attacking

spring driving mechanism. The pneumatic force is used here


to retract the kicking bar as the initial state. When the kicking
command is excited, the pneumatic force is released and then

functionality of soccer robot.


2. Image acquisition modules: the image acquisition is the

the compressed spring is desired to push the kicking bar so

major method to detect the object in the competition ground.

that the dribbled ball can be shot immediately. The shooting

To capture the overall image of the competition ground, the

force is desired depending on the spring constant and the

omni-directional

used.

mass of the kicking bar. Fig. 3 shows the computer model of

However, due to the serious image distortions of objects at

the kicking mechanism. The photo of physically assembled

vision

modules

are

generally

far distance, a front vision module is additionally used to

dribbling and kicking mechanism is shown in Fig. 4.

assist the image recognitions of the omni-directional vision


module.
3. Software computing modules: the software computing
modules are responsible for the intelligent vision fusion of
the omni-directional and front vision modules, object (e.g.,
ball, goal, enemy, field boundary) recognitions, strategy
making, path planning and motion driving. These software
modules are implemented using the C language, and these
software modules were executed on the embedded Linux
based computer. By using the embedded Linux based

Fig.l Mechanicalplatform design of Formosa middle

platform, the embedded software codes can be configured as

size soccer robot

the compact flash (CF) memory card instead of the hard disk
drive so that the system reliability can be significantly
improved during moving and unpredictable collisions of the
robot.
Finally, this paper is organized as follows. Section

II

introduces the design and implementation of the mechanical


platform. Section III illustrates the image processing and
recognitions for the intelligent vision fusion based image
acquisition
competItion

architecture.
strategies

implementations

and

Section
and

IV

subgoal

practical

describes
definition.

the
The

experimentations

discussed in Section V. Finally, the conclusions and future


works

are

proposed in Section VI.

mechanical

platform

mechanism design of Formosa middle size soccer robot

The driving of the Formosa middle size soccer robot uses


the Mitsubishi AC servomotor solution [10]. The motor is
the "HC-KFS43" model, and the molor driver is the
"MR-J2S-40A" model, as shown in Fig. 5. The driving

II. MECHANICAL PLATFORM AND DRIVING DESIGN


The

Fig.2 Dribbling

are

is

designed

based

signal of AC servomotor uses analog voltage signal (range:


on

0-10 Volts) to control the motor angular velocities. By

two-wheels driving mechanism. The industry proof AC

applying different angular velocities of two driving wheels,

servomotors are

the robot motions with different linear and angular velocities

mounted on both

sides of the robot

symmetrically. The dimension and weight are designed

can be desired as shown in (l).

according to the specification of the competition rules. In


order to achieve the stability of motion, three idle wheels are
employed.

Vx(t)

In this work, the mechanical platform is designed using

Vy (I)

the Pro! E software [11]. All the mechanical components

were produced in the machine shop of our university" The

(OB""

Pro! E mechanical platform model is designed as shown in

R cosB(w,(t) + w,(t

(I)

Fig. 1. On the other hand, we also develop the dribbling and


kicking mechanism to increase the attackability of the

R(W)tl-(O'(t

D
R(O,U) + w,(t
D
I

' (0, t (0, I>

()

{JJ, (t) (0, ()


I <

()

In (I), 8 is the robot angle corresponding to the horizontal

proposed soccer robot. The dribbling mechanism assists the


robot to keep the ball with the desired path; and the kicking

line; V xU) is the horizontal linear velocity of the robot;

mechanism is capable of increasing the impulse when

V:V (t)

is the vertical linear velocity of the robot; D is the radius of

shooting the ball. Fig. 2 shows the computer model of the

the driving wheel;

dribbling mechanism. The dribbling mechanism is designed

driving wheel;

using four rotating bars to increase the force of keeping ball.

(1)

2
R sin O(w, (I) + w,(t

cv,.(t)

mIt)

is the angular velocity of the right

is the angular velocity of the left driving

wheel; R is the distance of two driving wheels. The motion

Meanwhile, the sponges are coated on the top two rotating

driving command is transmitted via an in-lab design serial

bars to enhance the friction ofkeeping ball. Additionally, the

communication

kicking mechanism is designed using the pneumatic and

digital-to-analog

125

interface

board

moduleare

and

used

to

the

dual

control

channel
the

AC

separated into three sub-images or planes. The YUV is

servomotor drivers, as shown in Fig. 6.

different

from

RGB.

RGB

The

contain s

the

image

information of three large color channels, and the YUV deals

with one brightness or luminance channel (Y) and two col or

or

chrominance channels. When YCrCb data is being packed

as a YUV packing, the Cr component is packed as U, and the


Cb component is packed as V. The Y plane is configured as
one byte for each i mage per pixel. The Cr and Cb planes are

configured as half the width and half the height of the Y

plane (or image). Each Cr or Cb contains the information of

four neighboring pixels (a two-by-two square of the image).


For instanc e, erO belongs to Y'aa, Y'O I, Y' I0, and Y'II.
Fig.3 Kicking

mechanism design ofFonnosa middle size soccer robot

Note that the YUV and RGB can be con ve rted.

The color patterns of investigated object are recognized

using the band-pass threshold. Due to different luminance


condition

of

the

competition

ground,

the

thresholds for all investiga ted color patterns

an d

band-p a ss

obj ects are

determined before competition. To reduce the efforts on


defining the YUV band-pass thresholds, the simple formulas

are used. Initially, ten sample pixel s with YUV formats for

each investigated objects are acquired and recorded as


sequences of (Yii, Uii' Vij), where i is the index of

investigated objects such as ball, enemy, goal, and field

boundary marker ; j is the sample number for the pixels (j = I


Fig.4 Photo of assembled dribbling and kicking mechanism

to 10). Note that the relative positions of the investigated

objects

are

crucial

to

determine t he

YUV

band-pass

thresholds. In general, the extreme cases are not acceptable.

Once the YUV data of sample pixels for each investigated

objects is acquired, the mean values of the sample data are


calculated as (YMj, UM" VM,). The upper and lower bounds
of

the YUV

band-pass

thresholds

can be determined

experimentally in terms of adding the tolerance value. The


Fig.S Photo of AC servomotor and driving modules

tolerance value depends on the luminance conditions of the


competition ground, and it must consider the robustness

issues. By applying the vision fusions of the omni-directional


and front vision approach, the YUV band-pass thresholds are

individually determined.

The upper bound of the YUV band-pass thresholds is

indicated as ( YUj,

UUj, VU;);

the lower bound of the YUV

band-pass thresholds is indicated as (YLi, ULi, VLi). The


formulas

to

determine

thresholds are indicated

the

in (2) - (4).

component

band-pass

Where i is also the index

of interested objects; YMj is the mean value of comp on ent Y

i. YEj is the tolerance value of component Y for


obj ect i. The U and V c o mpo nent can also be dete rmine d in
this manner.
(2)
YM,= fy/IO
for object

Fig.6 Motion driving serial communication interface

board

III. IMAGE PROCESSING AND RECOGNITION

r=-I

YU,

A. Y-U-V Based Object Recognition

YM, + YE,

(3)

(4)

YL,= YMi - YSi

The robustness of the obj ect recognitions for specific

B. intelligent Vision Fusion Architecture

color patterns is crucial for the soccer robots. Due to the


luminance varying of th e competition ground and the

is impossible to
rec ogn ize the objects

Vision

is

the most important

sensor to detect the

distortion of the ornni-directional camera, it

investigated

use the typical RGB color format to

environment for the soccer robots. Due to large

with specified

colors.

In general, the YUV color format is

more feasible to ana lyze the images with lighting variations.


For the YUV image format

[5], three components are

objects

ground of the middle

with

specified

color

patterns

in

competition

si ze robot soccer game, the robot is

hardly to detect all objects in the competition field. Using


traditional front vision module can
126

just

inves tig ate the

objects in the front of the CCD with a particular viewing


angle. However, it cannot observe the objects except the
particular viewing angle.
In general, the omni-directional vision solution is used for
the middle size soccer robot. The omni-directional camera
can create a view screen with 360 degrees around the
observer (e.g., soccer robot). The photo of the
omni-directional camera used in this work (with model
MapCam360 model from the EeRise Corporation [8]) is
shown in Fig. 7 (a), and the omni-directional image example
is shown in Fig. 7 (b). Although the omni-directional camera
can capture 360 degrees image around the robot, the
seriously image distortion at the far locations (greater than
200 em in this work) is hardly investigated by the human or
difficultly recognized by the computer program when the
camera height is limited (80 em in this work).

primary VISIOn sensor, and the front VISIon acts as the


auxiliary vision sensor. The omni-directional camera is
mounted on the top of the soccer robot with 78 cm height
(not exceeded 80 em limitation). The front vision camera is
mounted at the front of the robot with a particular viewing
angle experimentally to investigate the ball clearly at the
possibly far distance in the competition ground.
The image fusion architecture is executed initially to
capture the images from the omni-directional camera, and
the ball (or target object) is further recognized. If the ball is
not recognized, then the ball may be either masked by the
enemy robot or far away the robot. The masking problem is
not discussed in the image recognition architecture. If the
ball cannot be recognized from the omni-directional image,
then the vision fusion controller switches to the front vision
camera automatically. At this moment, the robot rotates itself
to find the ball since the ball may be not at the front direction
of the robot. Once the ball is recognized, the robot stops
rotating and moves toward the ball straightly until the ball is
recognized by the omni-directional vision. Based on the
intelligent vision fusion architecture, the image recognition
efforts are significantly reduced.
IV. COMPETITION STRATEGY AND SUBGOAL DEFINITION

(a)
(b).
Fig.7 photo of omni-directional camera and image

Based on different image properties of the traditional


front vision and the omni-directional vision cameras, we
found their properties are complementary. Therefore, we
design the intelligent vision fusion architecture to investigate
all the objects in the competition ground completely. The
intelligent vision fusion architecture is shown in Fig. 8. Due
to the ball recognition being the most important task for the
soccer robot, the ball recognition procedure is used to
illustrate the proposed intelligent vision fusion architecture.
In this architecture, the images are fused in terms of the
strengths of individual camera properties to compensate their
weaknesses.

A. Soccer Competition Strategy

Since the proposed soccer robot is capable of dribbling


and kicking, the competition strategies are categorized as the
"with dribbling" and "without dribbling" scenarios. The
competition strategy is shown in Fig. 9. Either the "with
dribbling" or "without dribbling" model, the ball must be
recognized from the omni-directional camera instead of the
front vision camera. If the ball is far away the robot (i.e.,
cannot be recognized by the omni-directional camera), then
the proposed vision fusion approach is executed to move the
robot until the ball is recognized from the omni-directional
camera. In this manner, the soccer competition strategy can
be further executed.

DefltH."SLlbglJlllooingal
Opprro;ne DireCliOIl f from
'BaH 10 Gool wil1l1 SpeClfid o'm....-:...

Fig.9 Soccer robot competition strategy


Fig.8 Intclligent

For the "with dribbling" mode, the attack procedure is


very simple because that the distortion effects of the
omni-directional image can be ignored. Fig. 10 shows the

vision fusion architccture

In this architecture, the omni-directional vision acts as the


127

attack procedure of the "with dribbling" mode. In this case,


the robot is at position "Pos A" with direction "Dir

A".

The

ball is at position "Pos Ball". Initially, the robot rotates itself

reconstructed to calculate the Subgoal, as shown in Fig.

At

II.

the same time, the planner coordinate of the subgoal is

converted to the omni-directional image coordinate for

from "Dir A" to "Dir B" so that the ball is at front direction of

display. Note that the coordinate conversions between the

the robot, and then the robot moves straightly toward the ball.

omni-directional

When the ball is near (around 20 cm) the robot, the dribbling

coordinates uses the curve fitting approaches [3]. Based on

image

coordinates

and

the

planner

mechanism is automatically operated. Finally, the robot can

practical experimentations, the parameters of conversion

continuously keep the ball at position "Pos B". At this

equations are converged after the training of

moment, the robot rotates itself again from current direction

equations are shown in

20 data. The
(5)-(6).
Dp 0.000577 D/ - 0.16192Dl
+ 15.9966D; - 487.564
(5)
D; = 9.03xlO6 D/ - 0.00609D/
+1.48637Dp+22.1253
(6)
Where Dp indicates the real distances of planner distances

to direction "Dir C" so that the goal target is at front direction

of the robot. Then, the robot moves straightly toward the goal
target until the robot is at front of the goal with a predefined
distance ("Pos C"). Note that this position is determined in
terms of comparing the goal area pixels. Consequently, the

indi c ates the pixel distances of the

kicking mechanism is automatically operated in terms of

(in em units) and the

controlling the pneumatic solenoid value to shoot the ball.

omni-directional images. When the robot reaches "Pos B", it

D;

rotates itself until to direction "Dir C". The robot then moves

along "Dir e" to touch the ball, and further to push the ball

toward to the goal target. Consequently, the ball reaches the

goal.
P",,:R.all

Finally,

it

is noted that

the previous

described

competition strategies are executed based on the one v.s. one


robot soccer competition. The team competition can be
further modified to satisfy the practical issues. In addition,
the obstacles can be avoided in terms of changing the
subgoal dynamically based on the pre-defined rules.
V. IMPLEMENTATIONS AND PRACTICAL

Fig.IO Attack

procedure with dribbling function

EXPERIMENTAnONS

For the "without dribbling" mode, the attack procedure is


more complicated because that the distortion effects of the
omni-directional image must be considered. Fig. II shows

the attack procedure of the "without dribblin g " mode. In this


case, the robot is at position "Pos A" with direction "Dir A".

The ball is at position "Pos Ball". Because it is not easy to


keep the ball when the robot rotates with respect to a point
without the dribbling function, the soccer robot just pushes

the ball

with the desired

direction to the target. For this

reason, the "without dribbling" mode defines the subgoal for


the shooting ball purposes. The subgoal is defined at the

opposite direction of from ball to goal target with a specified


distance

(Do),

as indicated in "Pos B" in Fig. II. In this

manner, the positions of suhgoal, ball and goal target become


collinear.

The "Formosa middle size soccer robot" is physically


implemented. In order to increase the system reliability, the
embedded Linux platform is used. The proposed system is
developed on the EeRise PanoServer 3000 machine
EeRise PanoServer

[8]. The
3000 is a PC based computer with Intel

Pentium 4 2.4GHz processor. The operation system is the

Linux

2.4.20. In addition to supporting the embedded

platform, the PanoServer 3000

individual

visi on

input

is configured with for

channels

and

eight

general-purpose-input-output (GPIO) nodes to support the


soccer robot functions. The software modules are developed
using the

C language The photo of the completely


Fig. 12.

assembled soccer robot is shown in

The functions and approaches propose in this paper were


tested and verified in our laboratory. Note that the lighting
condition was poor and inconsistent during experiments. The
experimental results are proposed in this paper. Fig.

12

shows the photo of the vision fusions of the ball far away the
robot (left hand side), and the automatically switched front
vision camera (right hand side). The ball in the front vision
camera is very clear and easily recognized. In addition, the
ball and goal recognitions were also verified, and two cases
are shown in Fig. 13.

Meanwhile, the curve fittings for the conversions of the

omni-directional and planner coordinates are also practically


verified using equations
Fig. I I Attack procedure without

dribbling function

Due to the huge distortion of the omni-directional images,


the planner relative coordinates of the ball and goal target are

of

the

reconstructed

(5)-(6). Fig. 14 shows comparisons


planner

distances

from

the

omni-directional images and the real measured distances.


The reconstruction error is around

128

19.14 em

for

20 test data,

and they are acceptable for the middle size soccer robot The
reconstructed planner coordinates can be further used to

functions, intelligent vision fusion architecture,

the
(6).

black cross ("+") marks, and they are plotted using

converted planner coordinates as

indicated in equation

fitting

coordinates, competition strategies for "with dribbling" and

define the subgoals for the "without dribbling" mode, as


shown in Fig. 15. Note that the subgoal is marked using the

curve

based conversions for the omni-directional and planner


"without dribbling" modes were developed and integrated.

Especially, by introducing the intelligent vision fusion


approach, the image recognition efforts are reduced, and the
image recognition system become more robust. Finally, we
also validate the functional performance of "Formosa Middle
size soccer robot" in terms of participating the "2004
Taiwan's annual middle size soccer robot competition".

Fig.12 Assembled Fonnosa middle size soeeer robot

Fig, 16 Subgols for four eases


Fig. 13 Omni-direetional and front images

ACKNOWLEDGEMENTS

This work was supported by the Ministry of Education,


Taiwan,

R.O.C.,

under

the

2004

Student

Hand-On

Competition Project.
REFERENCES
[I]

P. Bonnin, O. Stasse, V. Huge!, P. Blazevic, "How to Introduce a Priori

V isual

and

Knowledge

Behavioral

for Autonomous and Mobile

Robots [0 Operate in Known Environments,"

Fig.14 Ball

and goal

recognitions

[2]

fEEE International
Conference on Emerging Technologies and Faclor)' Automation, Vol.
2, pp. 409-4 I 8, 200 I.
M. Dietl, 1.-5, Gutmann, B. Nebel, "Cooperative Sensing in Dynamic
Environments," IEEEIRSJ International Conference on Intelligent
Robots and Systems. Vol. 3, pp. 1706--1713,2001.

[3]

C.F.

Gerald,

P,O.

Wbeatley,

Applied

Addison-Wesley, 1999.
"""""--ITdl tii1

[4]

-rec0ller.dldJ<;t

tJ

I)

10

II

12 1.-

14

[6]

P-Ol::d

measured distances

y, Xi,

"Design and Implementation of an Open

IEEE In/emational Conference


on Rubotics and Au/umation, Vol. 2, pp, 1726-1731,2004,
A,D, Kulkarni, Artificial Neural NefworksfiJr Image Understanding,
Van Nostrand Reinhold, 1994,

16 1/ tii. lli D

Fig.15 Comparisons of the reconstructed planner distance

Chen,

and

the rcal

[7]

5,S

Lin,

R. Bajcsy, "High Resolution Catadioptric Omni-Dircetional

IEEE International Cunference on


Robotics and A u tomalion . Vol. 2, pp, 1694-1699.2003,
T, Schmitt. R. Hanck, M, Bcetz. S. Buck, B, Radig, "Cooperative
SIC reo Sensor for Robot Vision,"

Probabilistic State Estimation for Vision-Based Autonomous Mobile

IEEE Transactiuns on RobrJl!c,. and Automalion,


5, pp, 670-684, 2002.

Robots,"
Issue,

VI. CONCLUSIONS AND FUTURE WORKS

[8]
[91
[t 0]

In this paper, the "Formosa middle size soccer robot" is


introduced, and the hands-on implementations including

robot

mechanics

platform

with

dribbling

Analysis,

Autonomous Mobile Robot System,"

[5]
I

) , Jia, W,

Numeri al

and

kicking

URL:
URL:
URL:

hUp:llwww.fira,nct!

URL:

hllp:l/www.robocup,orgi

http://www.ccrise.com.tw/

http://www.mitsubishi.com/

[tl] URL: hUp:llwww.proc.coml


[12]

129

VoL 18,

You might also like