You are on page 1of 6

Design and Implementation of a Smart Wheelchair

Amit Ranjan Trivedi, Abhash Kumar Swagat Kumar


Singh, Sundara Tejaswi Digumarti and Innovation Labs
Deepak Fulwani Tata Consultancy Services
Indian Institute of Technology Jodhpur New Delhi, India 201301
Jodhpur, Rajasthan, India 342011 swagat.kumar@tcs.com
(amitrt,abhash,tejaswi,df)@iitj.ac.in

ABSTRACT and educational opportunities, reduces dependence on care-


A smart wheelchair can restore autonomy to patients with givers and promotes feeling of self-reliance [22, 24]. Powered
sensori-motor disabilities by enabling them to move around wheelchairs are preferred over the manual wheelchairs which
freely without depending on the care givers. The objective of are physically demanding. Powered wheelchairs make use of
a smart wheelchair is to reduce user effort in controlling the electric motors which run on batteries. The wheelchairs are
wheelchair and to ensure safety during movement. In this usually controlled through a joystick. However, using a joy-
paper, our focus is to design and develop a smart wheelchair stick might not be possible for many patients who might
using inexpensive hardware and open-source software so as have sensori-motor or cognitive impairments [22]. For in-
to make it affordable to a larger section of the target popu- stance, about 10-40 percent of patients in United States
lation, particularly in developing nations. The user can con- who desired powered mobility could not be fitted with pow-
trol the wheelchair using three interfaces namely, keyboard, ered wheelchair because of sensory impairments, poor motor
a webcam and a microphone. Webcam is used to detect function or cognitive deficits [24].
head-tilt which can be used for turning the wheelchair. Mi- According to an estimate by United Nations [5], around
crophone is used for controlling the wheelchair through dis- 10% of the world’s population, or 650 million people, live
crete voice commands. The wheelchair can be operated in with disability. Eighty percent of this population live in
three modes namely, manual, automatic and tele-operation developing countries. According to US Census Bureau [2],
modes. The software and hardware architecture of the plat- there are approximately 56.7 million people living in United
form is described in detail and experiments are performed States had some kind of disability in 2010 and this accounts
to demonstrate the usability of the platform. for 18.7% of the US population. About 12.3 million people
aged 6 years and older needed assistance with one or more
activities of daily living (ADL). In India, disability consti-
Categories and Subject Descriptors tutes approximately 5% [5] (2% in 2001 [3]) of the population
B.1 [Control structures and microprogramming]: Hard- out of which at least 8% have locomotion related disabili-
ware; J.3 [Life and Medical Sciences]: Health ties. A significant proportion of this population would be
benefited from a smart wheelchair to help them with their
activities of daily living (ADL). With unprecedented rate of
General Terms ageing all over the world approximately 2 billion people will
Design,Experimentation be aged 60 and older by 2050 [4, 30]. This indicates that
the demand for some form of assistive mobility will increase
Keywords over next few decades. In [24], Simpson attempts to charac-
terize and quantify the need for smart wheelchairs. It lists
Smart wheelchair, voice control, gesture-based control out various categories of population who might benefit from
the use of smart wheelchair. It carries out a detail study
1. INTRODUCTION of various trends and projects a growing market for smart
The overall objective of this project is to restore autonomy wheelchairs across the globe.
to disabled people by helping them use a powered wheelchair A smart wheelchair consists of a standard powered wheelchair
independently. Independent mobility is recognized as an im- with an on-board computer and a collection of sensors [24,
portant factor necessary for socio-cognitive development of 25]. A smart wheelchair usually implements a shared control
an individual. Independent mobility increases vocational architecture [12, 18, 25] where the control of the wheelchair
is shared with the user while minimizing the human effort
Permission to make digital or hard copies of all or part of this work for in carrying out various tasks like target acquisition, navi-
Permission
personal or to make digital
classroom use isor hard copies
granted withoutoffee
allprovided
or part ofthatthis workare
copies for
personal gation, path planning, localization and obstacle avoidance
not madeororclassroom
distributedusefor isprofit
granted without feeadvantage
or commercial provided and
that that
copies are
etc. Readers can refer to [22] for a review on various ad-
not made or distributed for profit or commercial advantage and that copies
copies bear this notice and the full citation on the first page. Copyrights
bear this notice and the full citation on the first page. To copy otherwise, to vancements carried out in this area. Recently, the focus
for components
republish, to postofonthis workorowned
servers by otherstothan
to redistribute lists,ACM mustprior
requires be specific
honored. Abstracting with credit is permitted. To copy otherwise, or
of research is towards developing intuitive Human-machine
permission and/or a fee. interface (HMI) that would minimize the human effort in
republish,
Request to post onfrom
permission servers or to redistribute to lists,
Permission@acm.org . requires prior
specific July
AIR’13, permission and/or
04-06 2013 a fee.
Pune, India
interacting with the wheelchair. Some of the approaches
Request permissions
Copyright 2013 ACMfrom Permissions@acm.org....$15.00.
978-1-4503-2347-5/13/07 are as follows. Some researchers used a vision-based inter-
AIR '13, July 04 - 06 2013, Pune, India
http://dx.doi.org/10.1145/2506096
Copyright 2013 ACM 978-1-4503-2347-5/13/07 ...$15.00.
http://dx.doi.org/10.1145/2506095.2506140
With camera
Laptop
and Microphone

Support for
Lap-tray

Array of
Sonar Sensors
Electrical
Motors
Figure 1. Wheelchair platform used for this project
Motor Control
Circuit
face where a virtual interface is projected on to a laptray
which is monitored by an overhead camera [18, 20]. Simp-
son et al. [23, 25] demonstrated the use of voice control of Figure 2. Main components of a smart wheelchair
wheelchair. Yanco [31] used eye-tracking or EOG signal for system
controlling the wheelchair in 1998. Mazo [13, 14] developed
an integrated human machine interface (HMI) that incor-
porated several types of user inputs like, breath-expulsion, Table 1. Approximate Cost for Smart Wheelchair
head movement, eye-tracking, face tracking for controlling S. No. Components Approximate Cost (USD)
the wheelchair. Min et al. [8] use force sensors to measure
shoulder and head movement which, in-turn, are used for 1 Powered Wheelchair 1200
controlling a wheelchair. Law et al. [11] developed a cap 2 Laptop with built-in 400
interface to control a wheelchair using EOG signal from webcam and micro-
eye-gaze movement and EMG signal from jaw-movement. phone
Another EMG based control wheelchair control system was
demonstrated by Felzer et al. [6] where muscle contractions 3 Sonar Sensors 200
are translated into appropriate motion commands. Quite
4 Another webcam for 30
recently, researchers have started using brain signals (EEG)
visual navigation
to control wheelchair [21, 27].
In this paper, we describe the design and development of a 5 Motor drivers and 170
smart wheelchair with a focus on minimizing the sensor and other accessories
computational requirement thereby making it affordable to
larger section of the population residing in developing coun- Total Cost 2000
tries. We avoid the use of expensive sensors like Laser, motor
encoders, gyros, accelerometers etc. and instead use ultra-
sonic sensors and webcam for navigation and path planning. structures are built in order to house sonar array and hold
The user inputs are interpreted either using a webcam or laptop. The modified structure is shown in Figure 1. The
through speech recognition, rather than using EOG, EMG wheelchair is driven by a 200 watt motor capable of carrying
or EEG signals which require sophisticated acquisition hard- a load of approximately 100 Kgs at a maximum speed of 5
ware. Our aim is to use better algorithms and software to kmph. The power supply is obtained from a 24V, 17Ah bat-
compensate for the absence of sophisticated sensors. We be- tery bank. The original system consists of a joystick which
lieve that a smart wheelchair with a set of features currently can be used for controlling the wheelchair. The wheelchair is
available with many of the existing platforms could be pro- available at a cost of approximately INR 60,000 (USD 1200)
vided at a cost of USD 2000. While lower cost will make in India. The smart wheelchair prototype is shown in Figure
this platform affordable to larger section of population in 2. It consists of the following major physical components:
developing countries, the proposed design could bring down (1) A electric wheelchair powered by battery, (2) a laptop
the overall healthcare cost also for the patients in developed with a built-in webcam and microphone, (3) motor-driver
countries who use smart wheelchair for their daily activities. system and (4) an array of sonar sensors. The approximate
The rest of this paper is organized as follows. The hard- cost of our project is about USD 2000 with break-up cost as
ware and software details of the wheelchair platform is pro- shown in Table 1.
vided in the next section. The algorithms used for develop-
ing user interfaces based on speech and vision is explained
2.2 Two-stage Control Architecture
in Section 3. The future scope of the work is provided in The schematic of a two-stage control architecture [14, 31]
Section 4 followed by conclusion in Section 5. used for our system is shown in Figure 3. At a lower level,
an Arduino Mega [1] based micro-controller system is used
to control the motor and to read the values from the sonar
2. DESCRIPTION OF THE WHEELCHAIR array. Arduino Mega has 54 digital I/O pins out of which
PLATFORM 14 can be used as PWM pins, 16 analog inputs, 4 UARTs,
a 16MHz crystal oscillator, a USB connection and a reset
2.1 The Wheelchair System button. The PWM pins are used for controlling the motor
The platform consists of an electric wheelchair obtained through a Super Hercules 9V-24V, 15A motor driver [15].
from Ostrich Mobility in India [17]. Additional support Controlling of motor require 3 pins, one for direction and
Serial Communication
two for speed. Speed is controlled by varying the PWM. Laptop Running 24V, 200W Motors
on Ubuntu
The sonar values are read through the analog input pins and

Via USB Cable


the micro-controller board communicates with the laptop PWM Output
through a USB cable.
At a higher-level, we have a laptop running on GNU/Linux
operating system (Ubuntu) which is used for implement-
ing various complex algorithms like speech recognition or
image processing, obstacle avoidance, navigation, mapping Arduino Mega
Super Hercules
etc. While the low-level system is used for data acquisition Micro-controller
Board 9V-24V, 15A
(input/output), the decision making is carried out by the Motor Driver

high-level system.
Since our objective was to control the wheelchair through Analog Input
a computer, we decided to get rid of the original joystick
controller that came with the wheelchair. This is unlike the Array of Sonar
Sensors
previous approaches by Simpson [25] [12], where the joystick
is an integral part of the shared-control architecture. The
joystick functionality is now provided through the laptop Figure 3. A Two-stage control architecture used
keyboard if a manual operation is desired. for controlling the wheelchair system. The low-level
micro-controller system is used for data acquisition
2.3 User Interface (input/output) and the high-level system is used for
The user can interact with the wheelchair through three decision making.
interfaces: (1) Keyboard, (2) webcam and (3) microphone.
All these three interfaces are available on the laptop. Key-
board is the replacement for the joystick which can be used In tele-operation mode, the wheelchair is controller re-
for maneuvering the wheelchair. The built-in webcam is motely over a wireless LAN connection by an operator. The
used for measuring the head-tilt of a person. In general, the operator gets the live feedback about the path from a cam-
robot is supposed to move in a straight line once started. era mounted on the wheelchair. The patient sitting on the
The wheelchair turns left or right by detecting the head-tilt wheelchair can still communicate with the operator through
through image processing. The process will be explained microphone. More details about this mode is provided in
later in the simulation section. Microphone is used for con- the next section.
trolling the wheelchair through discrete voice commands.
Specific words are mapped into different kinds of motion 2.5 Implementation: System Architecture
including ’halt’. The later two interfaces are meant for pa- The architecture of the system is shown in Figure 4. The
tients with sensori-motor disabilities who can not operate bottom-most (first) layer is the hardware which includes the
through keyboard. physical components of the wheelchair including sensors and
In order to keep the total cost low, we deliberately avoid actuators (motors). The second-layer consists of software
using other interfaces, for instance, virtual interface that necessary for interfacing with the hardware. These are the
requires additional projector and overhead camera system routines which are implemented on a micro-controller nec-
[20] [18] or EMG or EEG based interface systems [8] [11] [21]. essary for controlling the motors or reading from the sonars.
We believe that by using better algorithms, it is possible The pseudo-code of routines running on Arduino board is
to provide effective interface for a user to interact with the shown in the Table 1. Within every loop, it reads the sen-
wheelchair and in this paper, we demonstrate the working sor values sends them to laptop for processing, reads the
of speech and vision-based interfaces. decision from the serial port and then controls the motor
motion. Any other low level function can be put into this
2.4 Modes of Operation loop.
Three control modes are available for our platform: man- The third layer constitutes the Control Layer where the
ual, automatic and tele-operation. In manual mode, the higher-level decisions are taken. It reads data from the lower
operator uses one of the above three interfaces to give com- layer or from few high-level sensors like camera. Decision
mand to the wheelchair. The commands are translated into algorithms are implemented after processing the inputs ob-
the corresponding motor velocities by the micro-controller tained from these sensors. This layer can send out the de-
board. In this mode, the user is responsible for guiding the cision commands to the lower layer. The programs for this
wheelchair at all times. layer are executed on the laptop which has more computa-
In automatic mode, the goal is set by the operator and tional power compared to the micro-controller board used
the operation is carried out the wheelchair without any fur- in the lower layer. One example of a high-level task that is
ther intervention from the user. For instance, the user can performed in this layer is obstacle avoidance for which the
select a particular location visible on the screen and the pseudocode is provided in Table 2. As one can see, the sen-
robot would navigate through the environment to reach the sor readings obtained from the hardware layer is processed
destination. The goal can also be provided using voice com- and motion commands are generated by this program and
mands. For instance, “Go Home” would tell the wheelchair sent back to the hardware layer for execution.
to take the user to its pre-defined home location. Similarly, The last and upper-most layer is the User Interface Layer
a word “Explore” would tell robot to build a map of the envi- through which the user can interact with the wheelchair.
ronment while moving. During all autonomous operations, The algorithms for interpreting user inputs are also per-
the obstacle is avoided using sonar readings. formed on the laptop in our case.
robust in detecting the head outline as shown in Figure 5.
In order to compute the centroid of the head outline, it is
important to remove the background objects. Colour-based
background subtraction can not be used as the background
becomes non-static as the wheelchair moves. One simple
solution is to subtract two consecutive images to get rid of
the non-changing features of the image.

Figure 4. System Architecture: Design Pyramid

Algorithm 1 Pseudocode of routines running on Arduino


1: while TRUE do
2: Read Input from Sensors.
3: Write sensor values to the serial port.
4: Read action values from Serial Port.
5: Control the motors as per action values.
6: end while
Figure 5. Detecting head-tilt using contour detec-
tion.
Algorithm 2 Pseudocode for avoiding obstacle
1: while TRUE do 3.2 Controlling the wheelchair using voice com-
2: Read sensor values from serial port.
3: Remove noise from the sensor values.
mands
4: Group sensor readings into three labels: Ls , Rs , Fs Speech recognition refers to comparing a said command to
5: Define threshold for obstacle detection for these a predefined library of words and identifying the said word.
groups: Lth , Rth , Fth We have explored quite a few software packages related to
6: Set the corresponding motion flags if they exceed speech recognition during the development of our project.
thresholds (Lm , Rm , Fm ). We initially implemented a toolkit called âĂIJSphinxkeysâĂİ
7: Decide motion command based on the state of the which is part of the CMU Sphinx package [26]. Though it
motion flags. was quite efficient, it had a large library which was redun-
8: Send the motion command to the serial port for exe- dant for our use.
cution. For this project, we also developed our own speech recog-
9: end while nition algorithm using some of the basic concepts of speech
recognition [19]. We also create our own library of words to
be used. We recorded samples of one of the authors, saying
3. ALGORITHMS FOR HUMAN ROBOT IN- 6 commands namely “left”, “right”, “front”, “down”, “up” and
“down”. The frequency spectrum of the words are shown in
TERACTION Figure 6. These were recorded at 44100Hz sampling rate.
As stated earlier, our emphasis is on using simple sensors Each word was recorded for 2 seconds and 100 recordings of
like a webcam and a microphone to control the wheelchair. each word were taken in sets of 20 each. The voice samples
We avoid using other sophisticated methods like eye-tracking were then truncated to the regions with amplitude greater
[31] [14] or using brain signals (EEG) [21] [27] or measuring than a certain threshold so that background noise might be
muscle signals (EMG) [11] [6] for user interaction. These eliminated. Then the Fourier spectrum of the data collected
methods may increase the cost of the wheelchair by a big is observed. The process of obtaining FFT for each word is
margin making it unaffordable to our targeted group. Sec- shown in Figure 7. In this figure, the first plot shows the
ondly, most of these equipments require wearing the sensor 20 utterances of a word. For example, a person says ’left’
on the body itself which might cause discomfort to the user. 20 times. Each utterance is shown in a different colour.
The X-axis shows the number of samples. Duration of each
3.1 User-input through a webcam recording is 2 second and the sampling frequency is 44100
In this work, we use a webcam to detect the head-tilt of Hz and hence there are 8.8 × 104 samples in the set. The
the operator. This information is used to turn the wheelchair Y-axis is the amplitude. In the second graph, all the utter-
either to left or right. Contour-detection is used to find the ances are made to have uniform length (number of samples)
outline of the head above the shoulder. The location of the by cutting off their edges. X-axis again shows the number
centroid of this outline is then computed. The centroid of of samples and Y-axis shows the amplitude of signals. The
this contour shifts left or right with the head movement. third plot is the average of the FFT of the 20 samples. X-
A standard face detection algorithm could also be used for axis is the frequency in Hz and Y-axis is the amplitude. This
detecting the head and then computing the centroid. We average Fourier spectrum represents each word. A training
tried using Viola-Jones face detection algorithm [28] based set was built using these FFT spectrum. A Naive Bayesian
on Haar classifier for our purpose. However, we found it classifier coupled with K-Nearest Neighbour algorithm were
quite slow and many a times, it failed to detect the face. used to recognize the words. Even though the accuracy was
In this respect, the contour detection is found to be more low, it was adequate for the task.
While one can use the standard off-the shelf solutions for gorithm [7]. OpenNI [16] is another open source software
processing speech and image signals, our objective in this pa- which could be used for developing applications for Kinect.
per is to demonstrate simple techniques which can be used In this paper, we focussed on developing the human robot
even by beginners who might not have in-depth understand- interfaces for controlling the robot in a manual mode of op-
ing of these fields. eration. In future, we aim to focus on the autonomous mode
of operation, where the robot navigates autonomously based
on pre-defined goals. The goals will be set either by using
voice commands or using vision-based gestures.
We also aim to use this wheelchair as a tele-medicine plat-
form where the diagnostics of the patient could be recorded
through an array of sensors mounted either on the wheelchair
or on the body and transmit it to a remotely located doctor.
In case of emergency, the wheelchair can raise and alarm and
inform a tele-operator for assistance.

5. CONCLUSION
In this paper, we provide details of design and implemen-
tation of a smart wheelchair which can be used for assisting
patients having locomotion related disabilities. Our focus is
to keep the cost of the system low by opting for low cost
Figure 7. top: 20 utterance of a word ’left’ (x: sam- sensors and working more on algorithms. In the process, we
ple index, y: magnitude); middle: Signals obtained demonstrate the use of webcam and microphone to develop
cutting off their tails; bottom: Average Fourier user interfaces where the user can control the wheelchair us-
spectrum of the word (x: frequency (kHz), y: am- ing head-tilt or voice commands. We resort to open source
plitude) software and hardware technology platforms in order to keep
the total cost low. The smart wheelchair could be made
available at a cost as low as 2000 USD.
3.3 Tele-operation over Wi-fi
In hospitals, every disabled or paralyzed patient requires 6. ACKNOWLEDGEMENT
an attendant for carrying him/her from one place to the
This work was funded by the Ministry of Human Re-
other. This would require having a large set of manpower
source Development (MHRD), Govt. of India under the Na-
engaged in these activities leading to expensive healthcare
tional Mission on Education through Information Commu-
costs. Most of the time, a member of the family has to stay
nication Technology (NME-ICT) programme for the project
near the patient to take care of this need. The computer-
“Robotics for Education”.
controlled wheelchair could be used in a tele-operated mode
where a person sitting at one place can drive the wheelchair
remotely and take the patient from one location to another. 7. REFERENCES
The tele-operator can see the path through a camera mounted [1] Arduino Mega. An open-source electronics prototyping
on the wheelchair. The video seen through the camera is platform.
streamed over wireless LAN to ensure continuous video ex- http://arduino.cc/en/Main/arduinoBoardMega.
perience at the operator end. We use VLC video stream- [2] M. W. Brault. Americans with disabilities: 2010.
ing software on GNU/Linux (Ubuntu) platform [29] for this http:
purpose. Wireless LAN connection is available in most of //www.census.gov/prod/2012pubs/p70-131.pdf,
the hospitals these days and this approach can enable one July 2012.
operator to assist multiple patients at a given time. The [3] Census of India. Demography: Disabled population.
patient can interact with the operator through microphone http://censusindia.gov.in/Census_And_You/
and built-in webcam available on the laptop mounted on the disabled_population.aspx, 2001.
wheelchair. This can help in reducing the cost of healthcare,
[4] Demographic Profile of Older Population. World
thereby making it affordable to many more families.
Population Ageing:1950-2050, chapter 4, pages 23–27.
United Nations, Department of Economic and Social
4. FUTURE SCOPE AND DEVELOPMENT Affairs, 2001.
One of the challenges is to build a map using only cam- [5] Enable. Factsheet on persons with disabilities. http:
era without making use of encoders, gyros and accelerom- //www.un.org/disabilities/default.asp?id=18,
eters. The idea is to build a graphical map of the envi- 2012.
ronment where each place is identified as a node which is [6] T. Felzer and B. Freisleben. HaWCoS: the
connected to other nodes via links. The topological maps “hands-free” wheelchair control system. In Proc. of the
thus generated could be used for navigating in an environ- fifth international ACM conference on Assistive
ment autonomously. Microsoft Kinect [10] [9], being a low Technologies (ASSETS), pages 127–134, New York,
cost sensor, could also be used to generate 3D map of the USA, 2002.
environment. This can also be used for navigation and ob- [7] N. Grammalidis, G. Goussis, G. Troufakos, and M. G.
stacle avoidance. One can also make the wheelchair follow Strintzis. 3-d human body tracking from depth images
a human using a Kinect sensor and skeleton detection al- using analysis by synthesis. In International
(a) (b) (c)

Figure 6. Frequency Spectrum for various words spoken by the user. The x-axis shows frequency in kHz and
Y-axis shows the amplitude.

Conference on Image Processing (ICIP), volume 2, T. Kie, V. R. Kumar, J. P. Ostrowski, S. Patel, and
pages 185–188. IEEE, 2001. C. J. Taylor. Human robot interaction: Application to
[8] M. Jae-Woong, K. Lee, S.-C. Lim, and D.-S. S. Kwon. smart wheelchair. In IEEE International Conference
Human friendly interfaces of wheelchair robotic on Robotics and Automation (ICRA), volume 2, pages
system for handicapped persons. In IEEE/RSJ 3583–3588, 2002.
International Conference on Intelligent Robots and [21] B. Rebsamen, C. L. Teo, Q. Zeng, V. M. H. Ang,
Systems (IROS), volume 2, pages 1505–1510, 2002. E. Burdet, C. Guan, H. Zhang, and C. Laugier.
[9] Kinect. Xbox. http://www.xbox.com/kinect. Controlling a wheelchair indoors using thought. IEEE
[10] M. Kinect. Wikipedia. Intelligent Systems, 22(2):18–24, March-April 2007.
http://en.wikipedia.org/wiki/Kinect. [22] R. C. Simpson. Smart wheelchairs:a literature review.
[11] C. K. H. Law, M. Y. Y. Leung, Y. Xu, and S. K. Tso. Journal of Rehabilitation Research & Development,
A cap as interface for wheelchair control. In 42(4):423–436, August 2005.
IEEE/RSJ International Conference on Intelligent [23] R. C. Simpson and S. P. Levine. Adaptive shared
Robots and systems (IROS), volume 2, pages control of a smart wheelchair operated by voice
1439–1444, 2002. control. In IEEE/RSJ International Conference on
[12] S. P. Levine, D. A. Bell, L. A. Jaros, R. C. Simpson, Intelligent Robots and Systems (IROS), pages
Y. Koren, and J. Borenstein. The navchair assistive 622–626, Grenoble, September 1997.
wheelchair navigation system. IEEE Transactions on [24] R. C. Simpson, E. F. LoPresti, and R. A. Cooper.
Rehabilitation Engineering, 7(4):443–451, December How many people would benefit from a smart
1999. wheelchair? Journal of Rehabilitation Research &
[13] M. Mazo. An integral system for assisted mobility. Development, 45(1):53–72, November 2008.
IEEE Robotics and Automation Magazine, 8(1):45–56, [25] R. C. Simpson, D. Poirot, and F. Baxter. The
March 2001. haphaestus smart wheelchair system. IEEE
[14] M. Mazo, F. J. Rodriguez, J. L. Lazaro, J. Urena, Transaction on Neural Systems and Rehabilitation
J. C. Garcia, E. Santiso, P. Revenga, and J. J. Garcia. Engineering, 10(2):118–122, June 2002.
Wheelchair for physically disabled people with voice, [26] Sphinx-4. Cmu sphinx: A speech recognizer.
ultrasonic and infrared sensor control. Autonomous http://cmusphinx.sourceforge.net/sphinx4/.
Robots, 2(3):203–224, September 1995. [27] K. Tanaka, K. Matsunaga, and H. O. Wang.
[15] Nex Robotics. Super hercules 9V-24V, 15A motor Electroencephalogram-based control of an electric
driver. http: wheelchair. IEEE Transactions on Robotics,
//www.nex-robotics.com/products/motor-drivers/ 21(4):762–766, August 2005.
super-hercules-9v-24v-15a-motor-driver.html, [28] P. Viola and M. Jones. Rapid object detection using a
2012. boosted cascade of simple features. In Proceedings of
[16] OpenNI. A standard framework for 3d sensing. IEEE Conference on Computer Vision and Pattern
http://www.openni.org. Recognition (CVPR), volume 1, pages 511–518, 2001.
[17] Ostrich Mobility. Verve RX, power wheelchair. [29] VLC. Videolan streaming howto. http:
http://www.ostrichmobility.com/, 2012. //www.videolan.org/doc/streaming-howto/en/.
[18] S. P. Parikh, V. Grassi, V. R. Kumar, and [30] World Health Organization. Ageing and life course.
J. Okamoto. Incorporating user inputs in motion http://www.who.int/ageing/en/, 2013.
planning for a smart wheelchair. In IEEE [31] H. A. Yanco. Wheelesley: A robotic wheelchair
International Conference on Robotics and Automation system: Indoor navigation and user interface.
(ICRA), volume 2, pages 2043–2048, April 2004. Assistive Technology and Artificial Intelligence, LNAI,
[19] J. W. Picone. Signal modeling techniques in speech 1458:256–268, 1998.
recognition. Proceedings of the IEEE, 81(9):1215–1247,
September 1993.
[20] R. S. Rao, K. Conn, S.-H. H. Jung, J. Katupitiya,

You might also like