Professional Documents
Culture Documents
Abstract—Recently, the development of nursing and welfare On the study to operate a robot by using eye tracking,
robots intended for people with disabilities and elderly has Okamoto et al. [4] measured the eyes electric potential, calcu-
been advanced. The research of the method to transmit human lated the eye position and performed an experiment to control
intentions to the robot by using the biological information of
the operator is proceeding. Particularly, in the gaze information, a robot arm. Matsumoto et al. [5] developed an intelligent
there are multiple important information including the human wheelchair with the face and gaze based interface. Wang et
psychological condition. In this research we propose a method to al. [6] performed the research that estimates the intention of
perform the locomotive control of omni-directional wheelchair a human operator in controlling the wheelchair extracting the
by a gaze instruction system based on the human intention. gaze position and characteristics and acquires the place where
The human intention is estimated by using fuzzy inference
combining the gaze map and visual saliency map. We performed he wants to go to next by using the neural networks.
the operating experiment using a gaze tracking device, and we On the other hand, the visual saliency map model that Itti
confirmed that the operator can successfully perform the moving et al. [7] proposed based on the characteristic of the retina of
control of wheelchair according to the instructions. human eyes as a method to estimate visual unconsciousness
Index Terms—Gaze Tracking, Saliency Map, Omni-directional information is well known. However, in case of estimating the
Wheelchair, Human Intention, Fuzzy Inference
intention of human using gaze information, it becomes neces-
sary to remove the unconscious information where a human
I. I NTRODUCTION
watches from conscious instructions intention information.
The research and development of nursing and welfare Therefore, at first in this study, the saliency map is extracted
robots, devices and systems intended for people with disabil- from the movement of human eyes and the scenery where he
ities and elderly have been advanced in recent years. Elec- is watching. Moreover the intention estimation map, which is
troencephalogram (EEG), electromyogram (EMG) and eye really indicated his desired place, is generated by combining
tracking are often used as the human biological information to the gaze map with the saliency map with fuzzy inference. In
operate or control robots and mechatronics machines in these the experiment, we confirmed the efficiency of the proposed
studies. For example, the operation instruction interface for method by using the gaze instruction system to control the
handicapped persons has been developed by using the non- motion of an omni-directional wheelchair.
contact gaze tracking device with a little load for human.
On the study for EEG or EMG, Tanaka et al. [1] proposed II. V ISUAL S ALIENCY M AP
a pattern matching method used EEG and performed the
operating instruction of the wheelchair robot. Maeda et al. Koch et al. [8] thought that each individual characteristic in
[2] developed a technique to convey the human instructions an image has a property to guide attention from the feature
intention to a mobile robot using EMG information of the integration theory of Treisman et al. in 1985. He encoded a
vehicle driver’s arm by using the fuzzy inference. However, single characteristic of the image and proposed a concept of
the noise based on changes of the electric potential on the the visual saliency map model to express the two dimensions
head or skin surface is easy to be carried on EEG or EMG map. Itti et al. [7] implemented this theory as a computational
and there is a problem that it takes much time to set some model in 1998 (See Fig.1).
electrodes in the measurement experiment. In Fig.1, the processing (1) shows the output process for
This study aims at estimating the human’s instruction inten- the characteristic images of color (Colors), brightness (Inten-
tion by analyzing the movement of his eyes in case of driving sity), direction (Orientations) from input image (See Fig.2).
the omni-directional wheelchair. Ogawara et al. [3] proposed a The spatial filter outputs the summation result of weighted
method that a robot estimates its next work to perform from the neighboring pixels corresponding to the input image.
eyes information of a vehicle driver by avoiding each other’s In the color information, we generate each two color feature
collisions while a robot and a human perform the collaboration maps RG(σ), BY (σ) with the relation of complementary
in the same space. color together such as red and green, blue and yellow as shown
(2)
(4)
(5)
(6)
Fig. 4. Color Feature Map Fig. 5. Intensity Feature Map Fig. 6. Orientation Feature Map
Omni-directional
Wheel Chair Fig. 8. Membership Functions Fig. 9. Membership Functions
(Saliency Map) (Gaze Map)
TABLE I
Saliency Map Gaze Map F UZZY RULE OF I NTENTION E STIMATION
Fuzzy Inference
Conscious
Unconscious
Degree s Degree g gaze \ saliency SSS SS SM SL SLL
GSS IMM ISM ILS IMS IVS
GS ILM IMM ISM ILS IMS
GM ILL ILM IMM ILS IMS
GL IML IML ILL IMM ISM
GLL IVL IVL IML ILL ILM
Intention Estimation
Intention Estimation Map Degree i SSS: Saliency minimum GSS: Gaze minimum
SS: Saliency Small GS: Gaze Small
SM: Saliency Middle GM: Gaze Middle
Fig. 7. Generation Procedure of Behavior Intention Estimation Map SL: Saliency Large GL: Gaze Large
SLL: Saliency maximum GLL: Gaze maximum
Forward Image
Displayed on Monitor
Omni-directional Wheelchair
Fig. 15. Generated Maps (Place 3) Fig. 17. Generated Maps (Place 5)
gaze point and the intention estimation point, unconscious de- or joystick control) each as processing group, the level of
gree (Saliency), conscious degree (Gaze), intention estimation significance is modified as 1/2 by the Bonferroni correction
degree (Intention) are shown in the lower part of each figures. method here. In addition, the significant differences about
Four images show the original image, the saliency map, the about the comfortability and operationability after the Bon-
gaze map, and the intention estimation map, respectively. The ferroni correction are shown in Fig.18 and 19 respectively.
human gaze point is shown with a blue circle on the original As a result of three comparison experiments, it was con-
image. firmed that the proposed method obtained the highest evalua-
In the questionnaire, 10 subjects evaluated for impres- tion in both comfortability and operationability. The evaluation
sions of three experiments: the proposed method, only gaze of the hand-operated Joystick control was the lowest, but a
control, and joystick control about the comfortability and lot of subjects felt it was unexpectedly inconvenient. When
operationability with seven points (min:1 to max:7). We used the subject operates in case of gaze control, he may turn his
t-test to inspect the significant difference in each method for own eyes to the direction where he wants to go, but, in the
the questionnaire evaluation. The calculation results of p-value case of manual operation, However, in case of joystick control
about the comfortability and operationability are shown in by hand, it is thought that he must carry out the cooperative
Table III and IV. Because we performed two times compar- behavior of joystick control and eye movement at the same
isons for the proposed method with other two methods (gaze time.
TABLE III Average
t- TEST R ESULT WITH B ONFERRONI C ORRECTION (C OMFORTABILITY) Evaluation
Value
Experiment Proposed & Gaze Proposed & Joystick
p-Value 0.4212 0.0006
Level of significance p ≥ 0.025(=0.05/2) p < 0.0025(=0.005/2)
Significant Difference n.s. ***
TABLE IV
t- TEST R ESULT WITH B ONFERRONI C ORRECTION (O PERATIONABILITY)
Experiment Proposed & Gaze Proposed & Joystick Proposed Method Gaze Control Joystick Control
p-Value 0.0051 0.0002
Level of significance p < 0.025(=0.05/2) p < 0.0025(=0.005/2) Fig. 18. Significant Difference for Comfortability
Significant Difference * ***
Average
Evaluation
Value
Finally, the gaze control was considerably better than the
joystick control because a human can intuitively operate by
the gaze point. Furthermore, we confirmed that the operation
by the proposed method is slightly superior than the gaze
control in the operationability. As the reason about this result,
we consider that human’s instruction intention was estimated
by the fuzzy inference using the saliency map and eyes
information, and the system was able to remove unconscious
eyes movements as a noise from conscious eyes information. Proposed Method Gaze Control Joystick Control