Professional Documents
Culture Documents
Human-Computer Interface Controlled by The Lip: Marcelo Archajo Jos E, Member, IEEE, and Roseli de Deus Lopes
Human-Computer Interface Controlled by The Lip: Marcelo Archajo Jos E, Member, IEEE, and Roseli de Deus Lopes
IEEE JOURNAL OF BIOMEDICAL AND HEALTH INFORMATICS, VOL. 19, NO. 1, JANUARY 2015
I. INTRODUCTION
E propose a new form of interaction, a humancomputer
interface controlled by the lower lip, for users with
tetraplegia and validate it with a prototype.
While restorative treatments for spinal cord injury (SCI) [1],
[2] or invasive brainmachine interfaces [3], [4] are not available
outside research labs, some technologies [5], [6] can be used by
people with tetraplegia to interact with the world. Interaction
must be focused in the brain and muscles that users can control.
The selection of an assistive humancomputer interface requires
maximizing the flow of information and minimizing the effort
(physical and mental) to use it [6], [7].
Current alternatives include noninvasive braincomputer interfaces (BCI), eye tracking, electromyography (EMG), sipand-puff, voice commands, chin control, head control, mouth
joystick, and tongue control.
Noninvasive BCI using electroencephalography (EEG) has
two main types [6]: synchronous and asynchronous. A synchronous system [8], using P300 to control a power wheelchair,
requires 5 s, on average, to produce a highly reliable command,
too slow for a continuous control as required from a computer
pointing device (more information about EEG transfer rates can
be found in [9] and [10]). On the other hand, an asynchronous
system using sensorimotor rhythms shows the possibility to control a computer cursor [11], [12], but it will need more research
to overcome the strong performance variability [6].
In general, eye tracking can be made based on image processing [13], [14] or electro-oculography [6], [15], which uses
electrodes to monitor the eye movement. Any eye tracking system demands much user attention and errors can occur due to the
mismatch of selecting a command and the eye already changing
the position. For a pointing device, this could be acceptable, but
to control a power wheelchair, maintaining the eyes position can
be very tiring and false commands are not acceptable.
Another possibility is EMG, which uses electrodes to monitor
facial muscles to control a computer pointing device [6], [14].
An analysis of an EMG interface according to Fitts law is shown
in [16]; this interface provides discrete direction movements
(horizontal and vertical separately) not diagonal.
Sip-and-puff is an option for people with tetraplegia mainly
to control power wheelchairs [15], but this is usually difficult to
operate and usually works only with four discrete directions.
Voice commands can be useful to access some computer and
smartphone applications, but they are not adequate to direct
control pointing devices. Some research studies use voice commands to control power wheelchairs [15], [17]. A way to improve the precision of the voice recognition system in noisy environments is lip-reading (speech reading), an image processing
technology to identify speech from lip images [18].
Chin control is one of the best options currently available for
people with tetraplegia; this consists of a power wheelchair joystick adapted to be controlled by the chin [19]. Some systems
provide connection with computer as a pointing device [20],
[21]. A great advantage of the joystick is the possibility of soft
and free movements in any direction. Chin control (also applicable to head control [15], [22] and mouth joystick1 [5]) depend
on neck movements; the body must be fixed and the head must
be able to move freely. Power wheelchairs provide this condition, but vibration during the drive and body spasms (common
in spastic tetraplegia) can generate false commands. Outside the
wheelchair, the user has no control, due to the dependence of
the apparatus on the wheelchair structure.
The advantage of using tongue control [23][25] is that the
tongue is not controlled by the spinal cord; instead, it is controlled by the hypoglossal nerve directly connected to the brain.
One inconvenience of this type of interface is the hygienic issue, because people with tetraplegia depend on someone to put a
tongue piercing; for [25], it is also necessary an intraoral system.
Another restriction is that it allows only few discrete directions.
The tongue drive system (TDS) [23], [24] allows four directions (and two selection commands). Inductive tongue control
system [25] allows eight directions (and 10 sensors for other
commands).
The tongue and the mouth occupy a significant amount of
motor cortex, comparable with the hands and the fingers [26],
[27].
1 Mouth
2168-2194 2014 IEEE. Translations and content mining are permitted for academic research only. Personal use is also permitted, but republication/redistribution
requires IEEE permission. See http://www.ieee.org/publications standards/publications/rights/index.html for more information.
303
(3)
Fig. 1. (a) LCS architecture. (b) Head support. (c) Joystick support. (d) Calibration holes.
304
Fig. 2.
IEEE JOURNAL OF BIOMEDICAL AND HEALTH INFORMATICS, VOL. 19, NO. 1, JANUARY 2015
305
TABLE III
TEST DESIGN OVERVIEW
Fig. 4.
TABLE II
MULTIDIRECTIONAL DW COMBINATION
the left (or top for the vertical task) and so on. This procedure
is repeated 25 times [29] for each block.
A block is the rectangle width (W ) (pixels) and the distance
between targets (D) (pixels) combination, which defines the
index of difficulty (bits) [32]:
D
+1
(4)
ID = logz
W
We used four blocks for the task. Table I shows the combination of D and W , as well as ID; during the tests, they appear
randomly.
These values are the same as the ones for the one-directional
task of [23], but four combinations of DW that have clear
distinct ID values were chosen.
The procedure for the multidirectional circular tapping task
is to point and to select the circle indicated in red; after that,
a circle diagonally opposite will be indicated until all the 15
circles for each block have been gone through, Fig. 3(c). Here,
the data registration considers 14 steps in 15 circles.
Table II shows the combination of D and W , as well the ID;
during the tests, they appeared randomly.
These values are the same as the ones for the multidirectional
task of [23], which used just three DW combinations. Combination D = 305 W = 76 (ID = 2.33) was not used there. Here,
we prefer to maintain this combination because it has a very
distinct ID.
Many works use multidirectional circular tapping tasks, others use one-directional horizontal tapping tasks [30]. Here, we
306
IEEE JOURNAL OF BIOMEDICAL AND HEALTH INFORMATICS, VOL. 19, NO. 1, JANUARY 2015
TABLE IV
SESSIONS THROUGHPUT, AVERAGE, AND STANDARD DEVIATION (BITS/S)
Fig. 5.
The test took about 30 min per session; on the first day,
each participant spent 30 min training and 1 h in the two first
sessions; on the second day, they took 1 h and 30 min in the
three last sessions. This was done to avoid excessive mental or
physical fatigue, reducing the fatigue effects [31]. All the tests
were conducted over four weeks.
V. RESULTS AND DISCUSSION
In order to compare the three devices throughput measure, the
mean of means [30] was used. The general average throughput
and the standard deviation are shown in Fig. 5. The mouse had
4.70 bits/s for the one-directional horizontal task, 4.78 bits/s for
the one-direction vertical task and 4.95 bits/s for the multidirectional task; these values agree with many research works [23],
[30], [34] which indicates that the apparatus is adequate.
The thumb-controlled LCS achieved 4.19 bits/s for the onedirectional horizontal task, 3.97 bits/s for the vertical task and
1.8 bits/s for the multidirectional task.
The lip-controlled LCS is the main measure of this test; it had
2.59 bits/s for the one-directional horizontal task, 2.60 bits/s
for one-direction vertical task, and 1.06 bits/s for the multidirectional task. The one-directional tapping tasks show a similar throughput for horizontal and vertical movements despite
the different muscular behavior to move the joystick in these
directions. The ANOVA confirms that the difference was not
statistically significant (F1,11 = 0.015, ns). This could be due
to no significant difference in human psychomotor behavior for
the lower lip in the horizontal and vertical movements, or because the hardware achieves its limits very quickly, reducing the
muscular fine movements importance.
Table IV shows the average throughput of all the participants
for each session, the general average and standard deviation.
The lip-controlled LCS achieved throughput comparable with
other assistive technology research studies. The TDS presented
in [23] achieved for one-directional horizontal task results from
2.1 to 2.5 bits/s, vertical task from 2.2 to 2.7 bits/s and multidirectional task from 0.4 to 1.0 bits/s.
An important objective of this work is to understand how the
lower lip can be a good option to control an input device. That
is the reason for controlling the LCS with the thumb and the
lip under the same conditions. The lip-controlled LCS reaches
1.06 bits/s throughput for the multidirectional task, while the
thumb-controlled LCS had 1.80 bits/s throughput (this can be
Fig. 6.
307
Fig. 7. (a) Thumb-controlled LCS. (b) Lip-controlled LCS. (Tracing for multidirectional task for LCS of the participant #3, session 5, with D = 534 and
W = 76.)
Fig. 8.
B. Comfort Assessment
The participants answered a questionnaire defined in ISO/TS
9241-411:2012 [29] to assess the comfort in the use of the lipcontrolled LCS; the questionnaire was adapted to involve the
specific body parts used during the tests.
The assessment consists of 12 questions rated from 1 to 7,
where 7 means the most favorable answer.
The assessment in Fig. 8 shows the same verbal information
given by the participants; some fatigue on the lips (4.17) and
jaw (4.17), which affects the general comfort (3.83). The participants had no prior experience with the lip-controlled LCS,
or with the force necessary to control the joystick, or the kind
of movements likely to cause muscular discomfort. Its regular
use could promote lip muscles strength, making the device more
comfortable to use. We also observed that the participants used
the jaw as a complement to the lower lip to move the joystick
up and down.
There was practically no complaint about the neck and
shoulder.
The participants liked the operation speed (5.25) of the lipcontrolled LCS, which helped to achieve a good rate (5.25) for
the overall device operation.
A participant with orthodontic brace did not succeed in using
the LCS with the lips, and was replaced. During the tests, this
308
IEEE JOURNAL OF BIOMEDICAL AND HEALTH INFORMATICS, VOL. 19, NO. 1, JANUARY 2015
[11] G. E. Fabiani, D. J. McFarland, J. R. Wolpaw, and G. Pfurtscheller, Conversion of EEG activity into cursor movement by a braincomputer interface (BCI), IEEE Trans. Neural Syst. Rehabil. Eng., vol. 12, no. 3,
pp. 331338, Sep. 2004.
[12] D. J. McFarland, A. T. Lefkowicz, and J. R. Wolpaw, Design and operation of an EEG-based brain-computer interface with digital signal processing technology, Behavior Res. Methods, Instruments, Comput., vol. 29,
no. 3, pp. 337345, Sep. 1997.
[13] X. Zhang and I. S. Mackenzie, Evaluating eye tracking with ISO 9241
Part 9, in Proc. 12th Int. Conf. Human-Comput. Interaction, 2007,
pp. 779788.
[14] C. A. Chin, A. Barreto, and M. Adjouadi, Integration of EMG and EGT
modalities for the development of an enhanced cursor control system,
Int. J. Artif. Intell. Tools, vol. 18, no. 3, pp. 399414, Jun. 2009.
[15] M. Mazo, An integral system for assisted mobility, IEEE Robot. Autom.
Mag., vol. 8, no. 1, pp. 4656, Mar. 2001.
[16] M. R. Williams and R. F. Kirsch, Evaluation of head orientation and neck
muscle EMG signals as command inputs to a humancomputer interface
for individuals with high tetraplegia, IEEE Trans. Neural Syst. Rehabil.
Eng., vol. 16, no. 5, pp. 485496, Oct. 2008.
[17] G. Pacnik, K. Benkic, and B. Brecko, Voice operated intelligent
wheelchairVOIC, in Proc. IEEE Int. Symp. Ind. Electron., 2005, vol. 3,
pp. 12211226.
[18] Z. Yi, L. Quan-jie, L. Yan-hua, and Z. Li, Intelligent wheelchair multimodal human-machine interfaces in lip contour extraction based on
PMM, in Proc. IEEE Int. Conf. Robot. Biomimetics, Dec. 2009, pp. 2108
2113.
[19] N. Pellegrini, B. Guillon, H. Prigent, M. Pellegrini, D. Orlikovski,
J.-C. Raphael, and F. Lofaso, Optimization of power wheelchair control
for patients with severe Duchenne muscular dystrophy, Neuromuscular
Disorders, vol. 14, no. 5, pp. 297300, May 2004.
[20] H.-C. Chen, Y.-P. Liu, C.-L. Chen, and C.-Y. Chen, Design and feasibility study of an integrated pointing device apparatus for individuals with
spinal cord injury, Appl. Ergonomics, vol. 38, no. 3, pp. 275283, May
2007.
[21] H.-C. Chen, C.-L. Chen, C.-C. Lu, and C.-Y. Wu, Pointing device usage
guidelines for people with tetraplegia: A simulation and validation study
utilizing an integrated pointing device apparatus, IEEE Trans. Neural
Syst. Rehabil. Eng., vol. 17, no. 3, pp. 279286, Jun. 2009.
[22] H. V. Christensen and J. C. Garcia, Infrared non-contact head sensor, for
control of wheelchair movements, Assistive Technol.: Virtuality Reality,
vol. 16, pp. 336340, 2005.
[23] B. Yousefi, X. Huo, E. Veledar, and M. Ghovanloo, Quantitative and
comparative assessment of learning in a tongue-operated computer input
device, IEEE Trans. Inf. Technol. Biomed., vol. 15, no. 5, pp. 747757,
Sep. 2011.
[24] X. Huo and M. Ghovanloo, Using unconstrained tongue motion as an
alternative control mechanism for wheeled mobility, IEEE Trans. BioMed. Eng., vol. 56, no. 6, pp. 17191726, Jun. 2009.
[25] M. E. Lund, H. V. Christiensen, H. A. Caltenco, E. R. Lontis, B. Bentsen,
and L. N. S. Andreasen Struijk, Inductive tongue control of powered
wheelchairs, in Proc. IEEE 32nd Annu. Int. Conf. Eng. Med. Biol. Soc.,
2010, vol. 2010, pp. 33613364.
[26] X. Huo, J. Wang, and M. Ghovanloo, A magneto-inductive sensor based
wireless tongue-computer interface, IEEE Trans. Neural Syst. Rehabil.
Eng., vol. 16, no. 5, pp. 497504, Oct. 2008.
[27] E. R. Kandel, J. H. Schwartz, and T. M. Jessell, Principles of Neural Science, 4th ed. New York, NY, USA: McGraw-Hill, 2000.
[28] P. M. Fitts, The information capacity of the human motor system in
controlling the amplitude of movement, J. Exp. Psychology, vol. 47,
pp. 381391, 1954.
[29] Ergonomics of Human-System InteractionPart 411Evaluation Methods for the Design of Physical Input Devices, ISO/TS Standard 9241-411,
May 2012.
[30] R. W. Soukoreff and I. S. MacKenzie, Towards a standard for pointing
device evaluation, perspectives on 27 years of Fitts law research in HCI,
Int. J. Human-Comput. Studies, vol. 61, no. 6, pp. 751789, Dec. 2004.
[31] I. S. Mackenzie, HumanComputer Interaction: An Empirical Research
Perspective, 1st ed. San Mateo, CA, USA: Elsevier Morgan Kaufmann,
2013.
Marcelo Archanjo Jose (M07) received the graduate degree in electrical engineering from the Universidade Sao Judas Tadeu, Brazil, in 1997, and the
Masters degree in electrical engineering from the
Escola Politecnica da Universidade de Sao Paulo
(USP), Brazil, in 2008, where he is currently working toward the Ph.D. degree in electrical engineering.
In 1997, he joined Citibank and affiliates in Brazil.
In 2006, he began working at Fidelity Processadora e
Servicos. In 2008, he rejoined Citibank. Since 2009,
he has been a Researcher in NATENucleo de Aprendizagem Trabalho e Entretenimento do Laboratorio de Sistemas Integraveis da POLI USP. He is a
professional who works in research and development of hardware and software,
also heads his research teams, using his experience in personnel management,
obtained working in (Citibank and Fidelity) IT areas. He is specialized mainly in:
humancomputer interaction, computer graphics, image processing, stereo vision, stereo matching, 3-D reconstruction, computer vision, accessibility (power
wheelchair), user-friendly interfaces and games.
Mr. Jose has been a member of the Sociedade Brasileira de Computaca o
since 1999.
Roseli de Deus Lopes received the Undergraduate, Masters, Doctorate, and Postdoctorate degrees
in electrical engineering from Escola Politecnica da
Universidade de Sao Paulo, Brazil.
She is an Associate Professor at the Electronic Systems Department, School of Engineering of the University of Sao Paulo (USP). She is the Vice-Chair of
the Instrumentation Center of Interactive Technologies at the USP. She was Vice-Chair (20062008)
and Director (2008Feb. 2010) of Estaca o Ciencia, a
Center for Scientific, Technological and Cultural Dissemination of USP. She has been a Researcher at the Laboratory for Integrated
Systems since 1988, where she is a Principal Investigator of the Interactive Electronic Media Research Group (which includes research in computer graphics,
digital image processing, techniques and devices for humancomputer interaction, virtual reality, and augmented reality). She coordinates research projects in
the area of interactive electronic media, with emphasis on applications related
to education and health. She coordinates scientific dissemination initiatives and
projects aimed at identifying and developing talents in science and in engineering. She was responsible for the design and feasibility of Febrace (Brazilian
Fair of Science and Engineering), the largest national precollege science and
engineering fair in Brazil. Since 2003, she has acted as the General Coordinator
of Febrace.