Professional Documents
Culture Documents
net/publication/224162775
Wireless Vision Based Mobile Robot Control Using Hand Gesture Recognition
Through Perceptual Color Space
CITATIONS READS
41 50
2 authors:
All content following this page was uploaded by Manigandan Muniraj on 17 May 2019.
Wireless Vision based Mobile Robot control using Hand Gesture Recognition
through Perceptual Color Space
Abstract—In this paper we have implemented a wireless vision gestures. We will focus on the recognition of static images in
based mobile robot control through hand gesture recognition which skin colour and textures detection place a vital role.
based on perceptual color space such as HIS, HSV/HSB, HSL. Skin color and textures are important cues that people use
Vision-based hand gesture recognition is an important consciously or unconsciously to infer variety of culture-
problem in the field of human-computer interaction, since related aspects about each other. In images and videos, skin
hand motions and gestures could potentially be used to interact color is an indication of the existence of humans in such
with computers in more natural ways. The robot control was media. Skin detection means detecting image pixels and
done purely based on the orientation histograms a simple and regions that contain skin-tone color. Many of those
fast algorithm on the system which would recognize static hand
approaches have been implemented to focus on a single
gestures with HSV color spaces as major parameters. The
wireless based mobile robot system using hand gestures is a
aspect of gestures, such as hand tracking, hand posture
new innovative user interface that resolves the complications of estimation or hand pose classification using uniquely colored
using numerous remote controls for various applications. gloves or markers on hands/fingers.
Based on one unified set of hand gestures, this system The desire to develop a limited set of hand gestures that
interprets the user hand gestures into pre-defined commands are distinctive has improved the processing accuracy of
to control the remote robot. The experimental results are very captured gestures from wireless pin hole camera. This also
encouraging as the system produces real-time responses and requires a less sophisticated system model for controlling the
highly accurate recognition towards various gestures under robot action from remote place using proper Decision
different lighting conditions. Making rules implemented in MATLAB software engine.
Related works are discussed in Section II and Section III
Keywords - Hand Gesture recognition, Wireless vision, Mobile describes the color spaces for Skin detection. Section IV
robot, Color spaces, Skin Detection describes about the Gesture Normalization and Section V
describes the Hand gesture recognition using COG profile.
Section VI describes the system model implementation and it
I. INTRODUCTION
are being discussed in the Section VII.
Vision-based automatic hand gesture recognition has
been a very active research topic in recent years with II. RELATED WORKS
motivating applications such as human computer interaction Early approaches to the hand gesture recognition problem
(HCI), robot control, and sign language interpretation. The in a robot control context involved the use of markers on the
general problem is quite challenging due a number of issues finger tips [1]. An associated algorithm is used to detect the
including the complicated nature of static and dynamic hand presence and colour of the markers, through which one can
gestures, complex backgrounds, and occlusions. Attacking identify which fingers are active in the gesture. The
the problem in its generality requires elaborate algorithms inconvenience of placing markers on the user’s hand makes
requiring intensive computer resources. What motivated us this an infeasible approach in practice. Recent methods use
for this work is a robot navigation problem, in which we more advanced computer vision techniques and do not
were interested in controlling a robot by pose signs given by require markers. Hand gesture recognition is performed
a human. Due to real-time operational requirements, we were through a curvature space method in [2], which involves
interested in a computationally efficient algorithm. finding the boundary contours of the hand. This is a robust
Our approach involves segmenting the hand, based on approach that is scale, translation and rotation invariant on
skin colour statistics, as well as size constraints. We then the hand pose, yet it is computationally demanding. In [3], a
find the centre of gravity (COG) of the hand region as well vision-based hand pose recognition technique using skeleton
the farthest point from the COG. Based on these pre- images is proposed, in which a multi-system camera is used
processing steps, we derive a signal that carries information to pick the centre of gravity of the hand and points with
on the activity of the fingers in the sign. Finally we identify farthest distances from the centre, providing the locations of
the sign based on that signal. We demonstrate the the finger tips, which are then used to obtain a skeleton
effectiveness of our approach on real images of hand image, and finally for gesture recognition.
96
modules of the control system. A number of hand gesture are invariant to these operations. Moment invariants offer a
images were taken and the skin segmentation and the set of features that encapsulate these properties.
subsequent normalization are shown in Figure. 3. As seen
from Figure. 3, the filter had successfully segmented the skin C. Canny edge detector
regions out of all the tested images. It was also noticeable The purpose of edge detection in general is to
that the shadow of the hand and the body did not have any significantly reduce the amount of data in an image, while
effect on the filtering process. The remaining noise and preserving the structural properties to be used for further
unfilled pixels were removed by the normalization filter image processing. In this case to extract the hand gesture
which resulted in a smooth and clear region. The resultant pose without any noise due to the lightning conditions we
effect was that the object might have been recognized as a intend to use the canny edge detector along with the
skin region as its color had been modified. In conclusion, the morphological techniques. Steps involved in canny edge
performance of the skin segmentation and normalization detection is shown in the below Figure.4
filters was firmly robust against the variance in backgrounds
and lighting conditions.
97
r= ( x i − x ) 2 + ( y i − y )2 (1)
N M N M
−
∑∑
i =1 j =1
jI (i, j )
−
∑ ∑ jI (i, j )
i =1 j =1
x= ,y = (2) Figure 6. Luminosity behind the han
A A
A number of images are used to evaluate effectiveness of
the classification system. Both the original image and the
N M skin-segmented image are used to observe and verify the
A= ∑ ∑ I (i, j )
i =1 j =1
(3) accuracy of the segmentation algorithm. Images with low
lightning condition are also tested. It is quite apparent from
Figure. 6. There is always a number of noisy spots in the
filtered images, regardless of the lightning condition. This
distortion, as expected, becomes more pronounced in low
lighting conditions. As a result, the skin-segmented image is
noisy and distorted and is likely to result in incorrect
recognition at the subsequent stages. These distortions,
however, can be removed during the gesture normalization
stage.
98
issued prior to control the robot backward, even though the hand gestures. This set of hand gestures is adequate for any
command is recognized, no action will be taken. The system electronic control system. Our research distinguishes from
was observed to be 100% accurate under normal lighting the previous attempts because of few marked differences a
conditions. The tests have firmly consolidated the hardware minimum number of gestures are used to offer higher
design and the software interface of the developed prototype. accuracy with less confusion, only low processing power is
Different hand gestures along with their gesture extractions required to process the gestures, which is useful for simple
are shown in Figure.8. it is suggested that the hand gestures consumer control devices, Very robust to lighting variations,
can easily be recognized with maximum distribution values of Real-time operation.
centroidal profile. Table I below shows the equivalent command
words for controlling the robot. IX. FUTURE WORK
These algorithms can be incorporated as part of a robot
test-bed architecture which can be used to demonstrate their
effectiveness. By processing real-time images and
communicating wirelessly with our robot, we can control the
different appliances even under complex and cluttered
backgrounds. Work is currently underway to extend our
platform by: Incorporating these algorithms for Automatic
Guided Vehicle deployment for traffic sign recognition and
AGV should be capable of taking its control action based on
the traffic sign.
ACKNOWLEDGMENT
Figure 8. Hand Gesture Pose for Various Robot action control
The authors gratefully acknowledge the following
individuals for their support: Prof.V.Bhaskaran, Dept., of
TABLE I. COMPLETE TABLE FOR COMMAND WORDS Mathematics and our colleagues for their valuable guidance
for devoting their precious time, sharing their knowledge and
S. No Robot Control Command Word to co-operation.
Operation be Transmitted
1 LEFT ‘L’
REFERENCES
2 RIGHT ‘R’
3 FRONT ‘F’ [1] J. Davis and M. Shah "Visual Gesture Recognition", IEEProc.-Vis.
4 BACK ‘B’ Image Signal Process., Vol. 141, No.2, April 1994.
5 STOP ‘S’ [2] C.-C. Chang, I.-Y Chen, and Y.-S. Huang, "Hand Pose Recognition
Table II shows the recognition results from 5 different Using Curvature Scale Space", IEEE International Conference on
Pattern Recognition, 2002.
users of each 5 input images of different hand gestures. The
[3] A. Utsumi, T. Miyasato and F. Kishino, "Multi-CameraHand Pose
numbers in the table represent each hand gesture number in Recognition System Using Skeleton Image",IEEE International
Figure.8 Proposed hand gestures recognition method can Workshop on Robot and HumanCommunication, pp. 219-224, 1995.
process the recognitions at the rate of 15 fps(frames/sec) [4] B. Martinkauppi. Face colour under varying illumination - analysis
with image size of 131x109 and PNG image format is being and applications.PhD thesis, University of Oulu, 2002.
used for better accuracy and recognition. [5] K. Schwerdt and J. Crowley. Robust face tracking using color, 2000.
[6] F. Tomaz, T. Candeias, and H. Shahbazkia. Improved automatic skin
TABLE II. RECOGNITION RATE OF HAND POSTURE WITH THE detection incolor images. In 7th Digital Image Computing:
PROPOSED ALGORITHM Techniques and Applications, pages 419–427, 2003.
[7] T. Wilhelm, H. J. Bhme, and H. M. Gross. A multi-modal system for
Left Right Back Stop Front
tracking andanalyzing faces on a mobile robot. In Robotics and
Posture
Autonomous Systems 48, pages 31–40, 2004.
Users
#1 100% 100% 97% 100% 93% [8] A. Albiol, L. Torres, and E. J. Delp. Optimum color spaces for skin
#2 100% 100% 94% 100% 98% detection. In International Conference on Image Processing, volume
1, pages 122–124, 2001.
#3 100% 100% 96% 100% 100%
#4 100% 100% 98% 100% 97% [9] R. Kjeldsen and J. Kender. Finding skin in color images. In 2nd
#5 100% 100% 95% 100% 96% International Conference on Automatic Face and Gesture
Recognition, pages 312–317, 1996.
[10] Shin, M.C., Chang, K.I., Tsap, L.V.: Does colorspace transformation
VIII. CONCLUSIONS make any difference on skin detection? In: WACV ’02: Proceedings
of the Sixth IEEEWorkshop on Applications of Computer Vision,
The system is developed to reject unintentional and Washington, DC, USA, IEEE Computer Society (2002) 275
erratic hand gestures and to supply visual feedback on the [11] Yang, M., Kriegman, D., Ahuja, N.: Detecting faces in images: A
gestures registered. This work managed to invent a set of survey. IEEE Trans. On Pattern Analysis and Machine Intelligence
gestures that are distinct from each other yet easy to (PAMI) 24(1) (2002) 34–58
recognize by the system. The accuracy of the control system [12] Shan, C., Wei, Y., Qiu, X., and Tan, T.: ‘Gesture recognition using
temporal template based trajectories’. Proc. 17th Int. Conf. Pattern
was 100% and was mainly because of the limited number of Recognition, 2004, vol. 3, pp. 954–957
99