You are on page 1of 6




Mr. M. Selvaganapathy1, Mrs. N. Nishavithri2, Mr. T. Manochandar3 and Mr. G.

Assistant Professor, Department of Electronics and Communication Engineering
CK College of Engineering and Technology, Cuddalore, Tamil Nadu, India1,3,4
Mailam Engineering College, Mailam, Tamil Nadu, India2

The main aim of this project is to design a modern vehicle (Wheelchair) for the physically
challenged people using blue eye technology. In olden days, the voice oriented wheelchair is
being used by the physically disabled person to move from one place to another. Then an
advancement with voice operated wheelchair and the brain wave controlled wheelchair had taken
place. In this proposed system, the modern technology called blue eye technology is used for the
controlling of wheelchair. The main concept behind the proposed system is that the wheelchair
senses the movement of eye and locates the wheelchair movement accordingly. The proposed
system consists of eye tracking module, eye tracking computer interface, processing unit and
the wheelchair unit consisting of PMDC motor module. The outcome of this project is that one
can easily locate their wheelchair by looking at the object itself. The distance of that particular
object is being calculated initially and processed to the eye tracking computer interface and the
movement of the wheelchair is carried out.
Keywords: Blue eye technology, eye tracking computer module, Data Processing
Module, PMDC Control Unit.

The blue eye technology aims to create computational machine which have the
perceptual ability and the sensor ability. This technology uses eye tracking sensor to
locate the target at any point of extent. The term blue eye technologies can be split up
into blue and eye in which the term blue means the Bluetooth technology that enables
the powerful wireless technology and the term eye denotes the information / data
acquisition through the eyes. The blue eye technology is needed to build a prototype that
can understand the emoticons of the user.
Here the prototype used is the modern wheelchair which is for the physically
challenged people. The eye tracking computer interface is an interface that tracks the
movement of an eye and processes the input informations to the computer. The various
technologies used for locating the eye movements are emotion mouse, Manual and Gaze
Input Cascaded (MAGIC), Artificial Intelligence Speech Recognition, Simple User
Interface Tracker (SUITOR) and the eye movement sensor. The technology that is
adopted in this paper is an eye movement sensor.

The proposed methodology has various modules such as eye ball tracking module,
eye tracking computer interface, Data Processing unit, Robotic Unit and PMDC motor

Eye Movement Tracking

Eye tracking Computer


Data Processing Unit

Robotic Unit

PMDC Motor Unit

Chart 1 gives the methodology Chart for the movement of wheelchair


Figure 1 shows positioning of eye movement tracking (Center, Left and Right)
The eye movement tracking is used to determine the motion of eye balls. Based
on the motion of eye, the commands are given to the vehicle. This eye tracking module
consists of web camera which is mounted at the ECI module connected with the head of a
user. The user may wear this module when they were alive. Once the module is worn, it
is activated and starts sensing the eye motion of the user suddenly.
The camera attached with the eye movement tracking module has an additional
features like inbuilt light capability, inbuilt memory, etc. The eye tracking module is
interfaced with the computer by means of eye tracking computer interface (ECI)
module which has the driver for the eye motion tracking sensor. The ECI interface
performs various functionalities such as Motion capturing, Feature extraction and color
detection of a pixel, etc. 1


Motion Eye Motion
Capturing Detection
using Web

Bluetooth Raw Data

Module Transmission

Figure 2 shows the various blocks available in ECI Module

The ECI module has various modules such as motion capturing module, Motion
detection module, raw data transmission module and wireless module (Bluetooth). The
motion capture module initially captures the motion of the user and process it to the
human eye motion detection module. This motion detection module has specific detection
schemes based on the position of eyes (i.e.) either left, right or a side view (angled view
of an eye).
After detecting the motion of the eye, the module sends the raw data information
to the wireless module that is attached at the end of the ECI module. In Parallel, it checks
for the distance of the object from the location. Once the distance is measured, then the
start command is being forwarded; and once the wheelchair nears the targeted location,
then the wheelchair automatically sends the stop command to the PMDC unit. The
wireless module used here is a Bluetooth module; since the short distance communication
is enough to execute the ECI module2.

Danjie Zhu, Mikail Khunin and Theodore Raphan, Robust Hi Speed Binocular 3D eye movement tracking system
using two radii eye model, International Conference of IEEE Engineering in Medicine and Biology, 2006.
Mr. M. Selvaganapathy, Mrs. N. Nishavithri, Smart Wheel Chair using Neuro Sky Sensor, International Journal of
advanced Research in Computer and Communication Engineering, Vol. 4, Issue 11, November 2015, P. No. 361 366
The Data Processing unit has Bluetooth receptor which receives the Bluetooth
command sent by the Bluetooth transmitter of ECI module. The level analysis platform
checks for the direction command received from the ECI module. Once the user starts
targeting the object, the data processing unit gets the start command and once the user
nears the object, it automatically stops the wheelchair. The primary directions
concentrated in this data processing module are right or left and front or back. Once the
direction in noticed, the serial raw data is extracted and processed to the end unit called
Robotic unit.

Figure 3 shows the blocks available in Data Processing Unit

The robotic module is connected to the wheels of the wheelchair. This module has
data reception unit, ARM controller, PMDC motor unit and a display. The data reception
unit is used to receive the processed data from the previous module and processing to the
controller unit. The controller unit controls the PMDC motor as per the instructions
received from the data processing module. If the instruction received from the data
processing is left, then the ARM controller controls the PMDC motor 1 to rotate to the
left side (and PMDC motor 2 in case of right side).

Figure 4 shows the blocks available in Robotic Module

And once the user starts focusing the target object, then the robotic module
receives the command to move front, thereby commanding both the PMDC motors to
move front. And once, the targeted distance is reached, then the stop command is
instructed to the PMDC unit.
This paper proposes a modern wheelchair for the physically challenged people
controlling by the blue eye technology. The user can wear the ECI module at their head
which captures the eye motion and process it to the motion detection unit. Once, the user
targets the object present at the environment then the wheelchair automatically moves
towards the object without any primary instructions. There are six operations performed:
Start / Stop, Left / Right and Front / behind operations. Once the motion is acquisited
from the ECI module, the data processing unit processes the raw data to the robotic
module which in turn performs the movement of the wheelchair according to the request
received from the Data Processing Unit.


[1] Congcong Ma; Wenfeng Li; Raffaele Gravina; Giancarlo Fortino, Activity
recognition and monitoring for smart wheelchair users, IEEE 20th
International Conference on Computer Supported Cooperative Work in
Design (CSCWD), 2016, P. No. 664 669.
[2] Kazuto Miyawaki; Daiki Takahashi, Investigation of whole-body vibration
of passenger sitting on wheel chair and of passenger sitting
on wheelchair loaded on lifter, International Symposium on Micro Nano
Mechatronics and Human Science (MHS), 2016, P. No. 1 6.
[3] Nobuaki Kobayashi; Masahiro Nakagawa, BCI based control of
electric wheelchair, IEEE 4th Global Conference on Consumer Electronics
(GCCE), 2015, P. No. 429 430.
[4] M. Selvaganapathy and N. Nishavithri, Smart Wheel Chair using Neuro
Sky Sensor, International Journal of advanced Research in Computer and
Communication Engineering, Vol. 4, Issue 11, November 2015, P. No. 361
[5] Oral Motion Controlled Intelligent Wheelchair, SICE Annual Conference
2007, Japan.
[6] Non-Invasive, Wireless and Universal Interface for the Control of Peripheral
Devices by Means of Head Movements, Proceedings of the 2007 IEEE 10th
International Conference on Rehabilitation Robotics, June 12-15.
[7] Rafael Barea, et. al. Guidance of a Wheelchair using Electrooculography,
Electronics Department University of Alcala, SPAIN.
[8] R. C. Simpson, Smart Wheelchairs: A Literature Review, Journal of
Rehabilitation Resource Development, 2005, P. No. 423 436.
[9] L. Fehr, W. Edwin Langbein and S. B. Skaar, Adequacy of Power
Wheelchair Control Interfaces for Persons with Severe Disabilities: A Clinical
Survey, Journal of Rehabilitation Resource Development, Issue 37, No. 3,
2000, P. No. 353360.
[10] S. Stoddard, et. al. Chart book on Work and Disability in the United States,
1998, An Info Use Report, U. S. National Institute on Disability and
Rehabilitation Research, 1998.