Professional Documents
Culture Documents
Support for
Lap-tray
Array of
Sonar Sensors
Electrical
Motors
Figure 1. Wheelchair platform used for this project
Motor Control
Circuit
face where a virtual interface is projected on to a laptray
which is monitored by an overhead camera [18, 20]. Simp-
son et al. [23, 25] demonstrated the use of voice control of Figure 2. Main components of a smart wheelchair
wheelchair. Yanco [31] used eye-tracking or EOG signal for system
controlling the wheelchair in 1998. Mazo [13, 14] developed
an integrated human machine interface (HMI) that incor-
porated several types of user inputs like, breath-expulsion, Table 1. Approximate Cost for Smart Wheelchair
head movement, eye-tracking, face tracking for controlling S. No. Components Approximate Cost (USD)
the wheelchair. Min et al. [8] use force sensors to measure
shoulder and head movement which, in-turn, are used for 1 Powered Wheelchair 1200
controlling a wheelchair. Law et al. [11] developed a cap 2 Laptop with built-in 400
interface to control a wheelchair using EOG signal from webcam and micro-
eye-gaze movement and EMG signal from jaw-movement. phone
Another EMG based control wheelchair control system was
demonstrated by Felzer et al. [6] where muscle contractions 3 Sonar Sensors 200
are translated into appropriate motion commands. Quite
4 Another webcam for 30
recently, researchers have started using brain signals (EEG)
visual navigation
to control wheelchair [21, 27].
In this paper, we describe the design and development of a 5 Motor drivers and 170
smart wheelchair with a focus on minimizing the sensor and other accessories
computational requirement thereby making it affordable to
larger section of the population residing in developing coun- Total Cost 2000
tries. We avoid the use of expensive sensors like Laser, motor
encoders, gyros, accelerometers etc. and instead use ultra-
sonic sensors and webcam for navigation and path planning. structures are built in order to house sonar array and hold
The user inputs are interpreted either using a webcam or laptop. The modified structure is shown in Figure 1. The
through speech recognition, rather than using EOG, EMG wheelchair is driven by a 200 watt motor capable of carrying
or EEG signals which require sophisticated acquisition hard- a load of approximately 100 Kgs at a maximum speed of 5
ware. Our aim is to use better algorithms and software to kmph. The power supply is obtained from a 24V, 17Ah bat-
compensate for the absence of sophisticated sensors. We be- tery bank. The original system consists of a joystick which
lieve that a smart wheelchair with a set of features currently can be used for controlling the wheelchair. The wheelchair is
available with many of the existing platforms could be pro- available at a cost of approximately INR 60,000 (USD 1200)
vided at a cost of USD 2000. While lower cost will make in India. The smart wheelchair prototype is shown in Figure
this platform affordable to larger section of population in 2. It consists of the following major physical components:
developing countries, the proposed design could bring down (1) A electric wheelchair powered by battery, (2) a laptop
the overall healthcare cost also for the patients in developed with a built-in webcam and microphone, (3) motor-driver
countries who use smart wheelchair for their daily activities. system and (4) an array of sonar sensors. The approximate
The rest of this paper is organized as follows. The hard- cost of our project is about USD 2000 with break-up cost as
ware and software details of the wheelchair platform is pro- shown in Table 1.
vided in the next section. The algorithms used for develop-
ing user interfaces based on speech and vision is explained
2.2 Two-stage Control Architecture
in Section 3. The future scope of the work is provided in The schematic of a two-stage control architecture [14, 31]
Section 4 followed by conclusion in Section 5. used for our system is shown in Figure 3. At a lower level,
an Arduino Mega [1] based micro-controller system is used
to control the motor and to read the values from the sonar
2. DESCRIPTION OF THE WHEELCHAIR array. Arduino Mega has 54 digital I/O pins out of which
PLATFORM 14 can be used as PWM pins, 16 analog inputs, 4 UARTs,
a 16MHz crystal oscillator, a USB connection and a reset
2.1 The Wheelchair System button. The PWM pins are used for controlling the motor
The platform consists of an electric wheelchair obtained through a Super Hercules 9V-24V, 15A motor driver [15].
from Ostrich Mobility in India [17]. Additional support Controlling of motor require 3 pins, one for direction and
Serial Communication
two for speed. Speed is controlled by varying the PWM. Laptop Running 24V, 200W Motors
on Ubuntu
The sonar values are read through the analog input pins and
high-level system.
Since our objective was to control the wheelchair through Analog Input
a computer, we decided to get rid of the original joystick
controller that came with the wheelchair. This is unlike the Array of Sonar
Sensors
previous approaches by Simpson [25] [12], where the joystick
is an integral part of the shared-control architecture. The
joystick functionality is now provided through the laptop Figure 3. A Two-stage control architecture used
keyboard if a manual operation is desired. for controlling the wheelchair system. The low-level
micro-controller system is used for data acquisition
2.3 User Interface (input/output) and the high-level system is used for
The user can interact with the wheelchair through three decision making.
interfaces: (1) Keyboard, (2) webcam and (3) microphone.
All these three interfaces are available on the laptop. Key-
board is the replacement for the joystick which can be used In tele-operation mode, the wheelchair is controller re-
for maneuvering the wheelchair. The built-in webcam is motely over a wireless LAN connection by an operator. The
used for measuring the head-tilt of a person. In general, the operator gets the live feedback about the path from a cam-
robot is supposed to move in a straight line once started. era mounted on the wheelchair. The patient sitting on the
The wheelchair turns left or right by detecting the head-tilt wheelchair can still communicate with the operator through
through image processing. The process will be explained microphone. More details about this mode is provided in
later in the simulation section. Microphone is used for con- the next section.
trolling the wheelchair through discrete voice commands.
Specific words are mapped into different kinds of motion 2.5 Implementation: System Architecture
including ’halt’. The later two interfaces are meant for pa- The architecture of the system is shown in Figure 4. The
tients with sensori-motor disabilities who can not operate bottom-most (first) layer is the hardware which includes the
through keyboard. physical components of the wheelchair including sensors and
In order to keep the total cost low, we deliberately avoid actuators (motors). The second-layer consists of software
using other interfaces, for instance, virtual interface that necessary for interfacing with the hardware. These are the
requires additional projector and overhead camera system routines which are implemented on a micro-controller nec-
[20] [18] or EMG or EEG based interface systems [8] [11] [21]. essary for controlling the motors or reading from the sonars.
We believe that by using better algorithms, it is possible The pseudo-code of routines running on Arduino board is
to provide effective interface for a user to interact with the shown in the Table 1. Within every loop, it reads the sen-
wheelchair and in this paper, we demonstrate the working sor values sends them to laptop for processing, reads the
of speech and vision-based interfaces. decision from the serial port and then controls the motor
motion. Any other low level function can be put into this
2.4 Modes of Operation loop.
Three control modes are available for our platform: man- The third layer constitutes the Control Layer where the
ual, automatic and tele-operation. In manual mode, the higher-level decisions are taken. It reads data from the lower
operator uses one of the above three interfaces to give com- layer or from few high-level sensors like camera. Decision
mand to the wheelchair. The commands are translated into algorithms are implemented after processing the inputs ob-
the corresponding motor velocities by the micro-controller tained from these sensors. This layer can send out the de-
board. In this mode, the user is responsible for guiding the cision commands to the lower layer. The programs for this
wheelchair at all times. layer are executed on the laptop which has more computa-
In automatic mode, the goal is set by the operator and tional power compared to the micro-controller board used
the operation is carried out the wheelchair without any fur- in the lower layer. One example of a high-level task that is
ther intervention from the user. For instance, the user can performed in this layer is obstacle avoidance for which the
select a particular location visible on the screen and the pseudocode is provided in Table 2. As one can see, the sen-
robot would navigate through the environment to reach the sor readings obtained from the hardware layer is processed
destination. The goal can also be provided using voice com- and motion commands are generated by this program and
mands. For instance, “Go Home” would tell the wheelchair sent back to the hardware layer for execution.
to take the user to its pre-defined home location. Similarly, The last and upper-most layer is the User Interface Layer
a word “Explore” would tell robot to build a map of the envi- through which the user can interact with the wheelchair.
ronment while moving. During all autonomous operations, The algorithms for interpreting user inputs are also per-
the obstacle is avoided using sonar readings. formed on the laptop in our case.
robust in detecting the head outline as shown in Figure 5.
In order to compute the centroid of the head outline, it is
important to remove the background objects. Colour-based
background subtraction can not be used as the background
becomes non-static as the wheelchair moves. One simple
solution is to subtract two consecutive images to get rid of
the non-changing features of the image.
5. CONCLUSION
In this paper, we provide details of design and implemen-
tation of a smart wheelchair which can be used for assisting
patients having locomotion related disabilities. Our focus is
to keep the cost of the system low by opting for low cost
Figure 7. top: 20 utterance of a word ’left’ (x: sam- sensors and working more on algorithms. In the process, we
ple index, y: magnitude); middle: Signals obtained demonstrate the use of webcam and microphone to develop
cutting off their tails; bottom: Average Fourier user interfaces where the user can control the wheelchair us-
spectrum of the word (x: frequency (kHz), y: am- ing head-tilt or voice commands. We resort to open source
plitude) software and hardware technology platforms in order to keep
the total cost low. The smart wheelchair could be made
available at a cost as low as 2000 USD.
3.3 Tele-operation over Wi-fi
In hospitals, every disabled or paralyzed patient requires 6. ACKNOWLEDGEMENT
an attendant for carrying him/her from one place to the
This work was funded by the Ministry of Human Re-
other. This would require having a large set of manpower
source Development (MHRD), Govt. of India under the Na-
engaged in these activities leading to expensive healthcare
tional Mission on Education through Information Commu-
costs. Most of the time, a member of the family has to stay
nication Technology (NME-ICT) programme for the project
near the patient to take care of this need. The computer-
“Robotics for Education”.
controlled wheelchair could be used in a tele-operated mode
where a person sitting at one place can drive the wheelchair
remotely and take the patient from one location to another. 7. REFERENCES
The tele-operator can see the path through a camera mounted [1] Arduino Mega. An open-source electronics prototyping
on the wheelchair. The video seen through the camera is platform.
streamed over wireless LAN to ensure continuous video ex- http://arduino.cc/en/Main/arduinoBoardMega.
perience at the operator end. We use VLC video stream- [2] M. W. Brault. Americans with disabilities: 2010.
ing software on GNU/Linux (Ubuntu) platform [29] for this http:
purpose. Wireless LAN connection is available in most of //www.census.gov/prod/2012pubs/p70-131.pdf,
the hospitals these days and this approach can enable one July 2012.
operator to assist multiple patients at a given time. The [3] Census of India. Demography: Disabled population.
patient can interact with the operator through microphone http://censusindia.gov.in/Census_And_You/
and built-in webcam available on the laptop mounted on the disabled_population.aspx, 2001.
wheelchair. This can help in reducing the cost of healthcare,
[4] Demographic Profile of Older Population. World
thereby making it affordable to many more families.
Population Ageing:1950-2050, chapter 4, pages 23–27.
United Nations, Department of Economic and Social
4. FUTURE SCOPE AND DEVELOPMENT Affairs, 2001.
One of the challenges is to build a map using only cam- [5] Enable. Factsheet on persons with disabilities. http:
era without making use of encoders, gyros and accelerom- //www.un.org/disabilities/default.asp?id=18,
eters. The idea is to build a graphical map of the envi- 2012.
ronment where each place is identified as a node which is [6] T. Felzer and B. Freisleben. HaWCoS: the
connected to other nodes via links. The topological maps “hands-free” wheelchair control system. In Proc. of the
thus generated could be used for navigating in an environ- fifth international ACM conference on Assistive
ment autonomously. Microsoft Kinect [10] [9], being a low Technologies (ASSETS), pages 127–134, New York,
cost sensor, could also be used to generate 3D map of the USA, 2002.
environment. This can also be used for navigation and ob- [7] N. Grammalidis, G. Goussis, G. Troufakos, and M. G.
stacle avoidance. One can also make the wheelchair follow Strintzis. 3-d human body tracking from depth images
a human using a Kinect sensor and skeleton detection al- using analysis by synthesis. In International
(a) (b) (c)
Figure 6. Frequency Spectrum for various words spoken by the user. The x-axis shows frequency in kHz and
Y-axis shows the amplitude.
Conference on Image Processing (ICIP), volume 2, T. Kie, V. R. Kumar, J. P. Ostrowski, S. Patel, and
pages 185–188. IEEE, 2001. C. J. Taylor. Human robot interaction: Application to
[8] M. Jae-Woong, K. Lee, S.-C. Lim, and D.-S. S. Kwon. smart wheelchair. In IEEE International Conference
Human friendly interfaces of wheelchair robotic on Robotics and Automation (ICRA), volume 2, pages
system for handicapped persons. In IEEE/RSJ 3583–3588, 2002.
International Conference on Intelligent Robots and [21] B. Rebsamen, C. L. Teo, Q. Zeng, V. M. H. Ang,
Systems (IROS), volume 2, pages 1505–1510, 2002. E. Burdet, C. Guan, H. Zhang, and C. Laugier.
[9] Kinect. Xbox. http://www.xbox.com/kinect. Controlling a wheelchair indoors using thought. IEEE
[10] M. Kinect. Wikipedia. Intelligent Systems, 22(2):18–24, March-April 2007.
http://en.wikipedia.org/wiki/Kinect. [22] R. C. Simpson. Smart wheelchairs:a literature review.
[11] C. K. H. Law, M. Y. Y. Leung, Y. Xu, and S. K. Tso. Journal of Rehabilitation Research & Development,
A cap as interface for wheelchair control. In 42(4):423–436, August 2005.
IEEE/RSJ International Conference on Intelligent [23] R. C. Simpson and S. P. Levine. Adaptive shared
Robots and systems (IROS), volume 2, pages control of a smart wheelchair operated by voice
1439–1444, 2002. control. In IEEE/RSJ International Conference on
[12] S. P. Levine, D. A. Bell, L. A. Jaros, R. C. Simpson, Intelligent Robots and Systems (IROS), pages
Y. Koren, and J. Borenstein. The navchair assistive 622–626, Grenoble, September 1997.
wheelchair navigation system. IEEE Transactions on [24] R. C. Simpson, E. F. LoPresti, and R. A. Cooper.
Rehabilitation Engineering, 7(4):443–451, December How many people would benefit from a smart
1999. wheelchair? Journal of Rehabilitation Research &
[13] M. Mazo. An integral system for assisted mobility. Development, 45(1):53–72, November 2008.
IEEE Robotics and Automation Magazine, 8(1):45–56, [25] R. C. Simpson, D. Poirot, and F. Baxter. The
March 2001. haphaestus smart wheelchair system. IEEE
[14] M. Mazo, F. J. Rodriguez, J. L. Lazaro, J. Urena, Transaction on Neural Systems and Rehabilitation
J. C. Garcia, E. Santiso, P. Revenga, and J. J. Garcia. Engineering, 10(2):118–122, June 2002.
Wheelchair for physically disabled people with voice, [26] Sphinx-4. Cmu sphinx: A speech recognizer.
ultrasonic and infrared sensor control. Autonomous http://cmusphinx.sourceforge.net/sphinx4/.
Robots, 2(3):203–224, September 1995. [27] K. Tanaka, K. Matsunaga, and H. O. Wang.
[15] Nex Robotics. Super hercules 9V-24V, 15A motor Electroencephalogram-based control of an electric
driver. http: wheelchair. IEEE Transactions on Robotics,
//www.nex-robotics.com/products/motor-drivers/ 21(4):762–766, August 2005.
super-hercules-9v-24v-15a-motor-driver.html, [28] P. Viola and M. Jones. Rapid object detection using a
2012. boosted cascade of simple features. In Proceedings of
[16] OpenNI. A standard framework for 3d sensing. IEEE Conference on Computer Vision and Pattern
http://www.openni.org. Recognition (CVPR), volume 1, pages 511–518, 2001.
[17] Ostrich Mobility. Verve RX, power wheelchair. [29] VLC. Videolan streaming howto. http:
http://www.ostrichmobility.com/, 2012. //www.videolan.org/doc/streaming-howto/en/.
[18] S. P. Parikh, V. Grassi, V. R. Kumar, and [30] World Health Organization. Ageing and life course.
J. Okamoto. Incorporating user inputs in motion http://www.who.int/ageing/en/, 2013.
planning for a smart wheelchair. In IEEE [31] H. A. Yanco. Wheelesley: A robotic wheelchair
International Conference on Robotics and Automation system: Indoor navigation and user interface.
(ICRA), volume 2, pages 2043–2048, April 2004. Assistive Technology and Artificial Intelligence, LNAI,
[19] J. W. Picone. Signal modeling techniques in speech 1458:256–268, 1998.
recognition. Proceedings of the IEEE, 81(9):1215–1247,
September 1993.
[20] R. S. Rao, K. Conn, S.-H. H. Jung, J. Katupitiya,