You are on page 1of 2

Human-Centered, Ergonomic Wearable Device

with Computer Vision Augmented Intelligence for


VR Multimodal Human-Smart Home Object
Interaction
Ker-Jiun Wang Caroline Yan Zheng Zhi-Hong Mao
Department of Bioengineering Information Experience Design & Electrical and Computer Engineering &
University of Pittsburgh Fashion Bioengineering
Pittsburgh, USA Royal College of Art University of Pittsburgh
kew88@pitt.edu London, UK Pittsburgh, USA
yan.zheng@network.rca.ac.uk zhm4@pitt.edu

Abstract—In the future, Human-Robot Interaction should


be enabled by a compact, human-centered and ergonomic
wearable device that can merge human and machine altogether
seamlessly by constantly identifying each other’s intentions. In
this paper, we will showcase the use of an ergonomic and light-
weight wearable device that can identify human’s eye/facial
gestures with physiological signal measurements. Since human’s
intentions are usually coupled with eye movements and facial
expressions, through proper design of interactions using these
gestures, we can let people interact with the robots or smart
home objects naturally. Combined with Computer Vision object
recognition algorithms, we can allow people use very simple and
straightforward communication strategies to operate
telepresence robot and control smart home objects remotely,
Fig. 1. Our ergonomic wearable device provides flexible hardware and
totally “Hands-Free”. People can wear a VR head-mounted
software solutions. Users can customize their own design to place sensors at
display and see through the robot’s eyes (the remote camera different locations to measure different physiological signals, such that users
attached on the robot) and interact with the smart home devices can pick their favorite facial gestures and compose their own control
intuitively by simple facial gestures or blink of the eyes. It is commands to interact with different applications.
tremendous beneficial for the people with motor impairment as
an assistive tool. For the normal people without disabilities, they To address this issue, an easy and simple control interface
can also free their hands to do other tasks and operate the smart is imperatively needed. One alternative approach is to use
home devices at the same time as multimodal control strategies. Brain-Computer Interfaces (BCIs). Ideally, if we have a
Keywords—Wearable Device; Smart Home; Telepresence perfect BCI, we can directly translate brainwave signals into
Robot; Virtual Reality; Human-Computer Interaction, Human- feasible commands and operate external devices simply by
Centered Design thinking. However, many issues face the development of
BCIs, such as the low classification accuracies of brain
I. Introduction signals and the tedious human-learning procedures.
The recent development of Smart Home technologies Particularly, current non-invasive BCIs use
outlines a promising future of our society. However, such Electroencephalogram (EEG) caps where many electrodes
developments often neglect the needs of the immense must be attached on the user’s head to identify imaginary
population of the people who have motor or voice disabilities. motor commands, which brings a lot of inconvenience.
Because of the lacking of feasible control interfaces for these In this paper, we use a different approach to develop
people (Current assistive devices usually use head tracking wearable device which is comparable to BCIs. Our
control, sip-and-puff system, camera-based eye tracker, and development uses non-invasive biosensors placed on un-
wheel-mounted joysticks, etc., where users still need to apply obtrusive locations on the user’s face. People can comfortably
a lot of body/lung strength and/or extensive fine motor wear it for long hours without tiring. Under our developed
controls to operate the devices. In the meanwhile, these Machine Learning algorithms, we can identify various eye
technologies are usually attached to a wheelchair, which are movements and facial expressions with high accuracies.
not portable and confine the use to times when the users are Instead of measuring EEG alone, our approach identifies the
seated. The people with speech problems are also not able to mixture of EEG, EOG and EMG physiological signals.
use voice assistants), the disabled communities are often Through the utilization of Neuroscience and Psychology
segregated from the potential applications of such Smart studies, we can decipher situationally-aware human intentions
Home technologies to benefit their daily life. based on the identified physiological signals, and with proper
design (Design Thinking approach) of intuitive and engaging

978-1-5386-8555-6/19/$31.00 ©2019 IEEE 767


Human-Machine Interaction (HMI) strategies, we can use simple facial gestures to turn on/off the light, open/close the
very simple eye/facial gestures to generate commands with door, or switch the TV channels, etc., for instance.
the help of semi-autonomous artificial augmented intelligence, (a)
e.g., Computer Vision algorithms, to control robots, interact
with smart home and other IoTs (Internet of Things) devices
in a nature and intuitive way, totally “Hands-free”.

(b)

Fig. 2. The developed Machine Learning algorithm using Short Time Fourier
Transformation (STFT) and Convolution Neural Network (CNN). The
collected physiological signals of different channels (totally 6 channels)
within short period of time (collected by Sliding Window) will be
transformed into their corresponding power spectrums, and further sent into
CNN Deep Learning architecture to train the user’s specific physiological
gesture model. Users can arbitrarily train their favorite eye/facial gestures to
generate actionable commands to control different smart home applications.

II. Demonstrations on Control of Telepresence Drone to


Interact with Smart Home Objects Fig. 3. Strategy of telepresence drone control with Computer Vision-aided
We have set up a very engaging and intuitive interaction smart home objects interaction. (a) The concept that users can steer the drone
method to control the telepresence drone by using eye with first-person’s view seeing through the remote camera, from which they
can target each smart home device recognized by the Computer Vision
movements and facial expressions, and come up with the
algorithm and use the correspondent facial gesture control methods to
strategies to interact with the smart home objects with the
operate the device. (b) System architecture of the proposed strategy.
assistance of Computer Vision object recognition algorithms.
We use the commercially available DJI Spark Drone in the III. Discussion and Conclusion
demonstration. The real-time images captured from the
drone’s camera will be transmitted to the Virtual Reality The advancement of AI (Artificial Intelligence), Big Data,
(VR) headset and seen by the wearer. In this setup, users can IoTs (Internet of Things), Robotics, and AR/VR (Augmented
use our device to control any camera-mounted robot (e.g., Reality/Virtual Reality) outline a promising near-future of our
endoscopic camera or any camera-mounted robotic society. While people highly praise the development of these
manipulator) in the first-person’s view to directly control the technologies, however, we should not forget that such
robot using eye movements and facial expressions. For technologies should center around the human beings. Without
instance, users can use: left, right, such eye movements to human at the center of the intelligence, these enabling
steer the robot rotate in the Yaw-axis, and use “open the eyes technologies will lose their meanings, and will not be as
wider (raise the eyebrows)” or “Smile”, such facial gestures impactful as they are supposed to be. To address this issue,
to move the robot forward or backward along the depth our compact, ergonomic, human-centered interface,
direction. In such way, the people with disability can easily augmented with Machine Learning/Computer Vision aided
teleoperate the robot to explore the remote environment. intelligence, can seamlessly connect human with those
surrounding technologies. Our development outlines a bright
At the same time, the smart home objects in the remote future of HRI research that through human-centered wearable
environment will be recognized by our Deep Learning design, combed with neurophysiological understandings and
Computer Vision algorithm, modified by Transfer Learning, the assistance of proper augmented AI, we can enable an
to suit our own needs of the implemented smart home devices. intuitive Human-Machine Interface that could be potentially
When people see through the robot’s eyes (the remote camera comparable to the traditional BCI devices.
attached on the drone) with VR headset, the smart home
objects will be automatically identified, and the menus of the REFERENCES
correspondent facial gesture control methods will be [1] K.-J. Wang, L. Zhang, B. Luan, H.-W. Tung, Q. Liu, J. Wei, M. Sun, and Z.-H.
displayed alongside the identified objects. In such way, Mao. Brain-computer interface combining eye saccade two-electrode EEG
people can easily steer the drone to target any smart home signals and voice cues to improve the maneuverability of wheelchair.
Proceedings of the15th IEEE/RAS-EMBS International Conference on
device he/she wants to control, and the system will Rehabilitation Robotics (ICORR 2017), London, UK, 1073-1078,
understand which device will be selected and people can use doi:10.1109/ICORR.2017.8009392, July 2017.

768

You might also like