You are on page 1of 48

Advanced Driver Intention Inference:

Theory and Design 1st Edition Yang


Xing
Visit to download the full and correct content document:
https://ebookmass.com/product/advanced-driver-intention-inference-theory-and-desig
n-1st-edition-yang-xing/
Advanced Driver
Intention Inference
Theory and Design
YANG XING, PHD
Research Fellow
School of Mechanical and Aerospace Engineering
Nanyang Technological University
Singapore

CHEN LV
Assistant Professor
School of Mechanical and Aerospace Engineering
Nanyang Technological University
Singapore

DONGPU CAO
Canada Research Chair in Driver Cognition and Automated Driving
Department of Mechanical and Mechatronics Engineering
University of Waterloo
Canada

]
Elsevier
Radarweg 29, PO Box 211, 1000 AE Amsterdam, Netherlands
The Boulevard, Langford Lane, Kidlington, Oxford OX5 1GB, United Kingdom
50 Hampshire Street, 5th Floor, Cambridge, MA 02139, United States

Copyright Ó 2020 Elsevier Inc. All rights reserved.

No part of this publication may be reproduced or transmitted in any form or by any means, electronic or
mechanical, including photocopying, recording, or any information storage and retrieval system, without
permission in writing from the publisher. Details on how to seek permission, further information about the
Publisher’s permissions policies and our arrangements with organizations such as the Copyright Clearance
Center and the Copyright Licensing Agency, can be found at our website: www.elsevier.com/permissions.

This book and the individual contributions contained in it are protected under copyright by the Publisher (other
than as may be noted herein).

Notices
Knowledge and best practice in this field are constantly changing. As new research and experience broaden our
understanding, changes in research methods, professional practices, or medical treatment may become
necessary.

Practitioners and researchers must always rely on their own experience and knowledge in evaluating and using
any information, methods, compounds, or experiments described herein. In using such information or
methods they should be mindful of their own safety and the safety of others, including parties for whom they
have a professional responsibility.

To the fullest extent of the law, neither the Publisher nor the authors, contributors, or editors, assume any
liability for any injury and/or damage to persons or property as a matter of products liability, negligence
or otherwise, or from any use or operation of any methods, products, instructions, or ideas contained in the
material herein.

Library of Congress Cataloging-in-Publication Data


A catalog record for this book is available from the Library of Congress

British Library Cataloguing-in-Publication Data


A catalogue record for this book is available from the British Library

ISBN: 978-0-12-819113-2

For information on all Elsevier publications visit our website at


https://www.elsevier.com/books-and-journals

Publisher: Matthew Deans


Acquisitions Editor: Carrie Bolger
Editorial Project Manager: Gabriela D. Capille
Production Project Manager: Sreejith Viswanathan
Cover designer: Alan Studholme

Typeset by TNQ Technologies


List of Abbreviations

GP Gaussian Process
ABS Antilock Braking System HMM Hidden Markov Model
ACC Adaptive Cruise Control HHMM Hierarchical Hidden Markov Model
ACP Artificial Society, Computational HOG Histogram of Oriented Gradients
Experiments, and Parallel Execution HRI Human-Robot Interaction
ACT-R Adaptive Control of Thought-Rational HT Hough Transform
ADAS Advanced Driver Assistance System IDDM Intention-driven Dynamic Model
A/D Analog/Digital IMM Interactive Multiple Model
AIOHMM Autoregressive Input-Output HMM IMU Inertial Measurement Unit
ANN Artificial Neural Network IOHMM Input-Output Hidden Markov Model
AVM Around View Monitoring IPM Inverse Perspective Mapping
BCI Brain-Computer Interface JTSM Joint Time-Series Modelling
BF Bayesian Filter KRR Kernel Ridge Regression
BN Bayesian Network LADAR Laser Detection and Ranging
CAN Controller Area Network LANA Lane Finding in Another Domain
CBN Causal Bayesian Network LCII Lane Change Intention Inference
CDI Comprehensive Decision Index LDA Lane Departure Avoidance
CHMM Continuous Hidden Markov Model LDW Lane Departure Warning
CLNF Conditional Local Neural Fields LIDAR Light Detection and Ranging
CLR Constant Learning Rate LKA Lane Keeping Assistance
CPS Cyber-Physical Space LOO Leave-One-Out
CPSS Cyber-Physical-Social Space LSV Lane Sampling and Voting
CNN Convolutional Neural Network LSTM Long Short-Term Memory
CSI Channel State Information MIC Maximal Information Coefficient
DBSCAN Density-based Spatial Clustering of MHMM Modified Hidden Markov Model
Application with Noise MLR Multisteps Learning Rate
DBN Dynamic Bayesian Network NGSIM Next-Generation Simulation
DII Driver Intention Inference NFIR Nonlinear Finite Impulse Response
DWT Discrete Wavelet Transform NN Neural Networks
ED Edge Distribution OLS Orthogonal Least Squares
EEG Electroencephalograph OOB Out-of-Bag
ECG Electrocardiogram PCA Principal Component Analysis
EMG Electromyography PRM Revolutions Per Minute
EOG Electrooculography RANSAC Random Sample Consensus
ERRC Error Reduction Ratio-Causality RF Random Forest
EV Electric Vehicle RGB Red-Green-Blue
FFNN Feedforward Neural Network RGB-D Red-Green-Blue Depth
FIR Field Impedance Equalizer RMSE Root Mean Square Error
FPR False-Positive Rate RNN Recurrent Neural Network
GA Genetic Algorithm ROC Receiver Operating Characteristic
GAN Generative Adversarial Networks ROI Region of Interest
GMM Gaussian Mixture Model RVM Relevance Vector Machine
GNSS Global Navigation Satellite System SA Situation Awareness
GOLD Generic Obstacle and Lane Detection SAE Society of Automobile Engineers
GPS Global Positioning System SBL Sparse Bayesian Learning

v
vi LIST OF ABBREVIATIONS

SF Steerable Filter TS Time Sliced


SCR Skin Conductance Response TPR True-Positive Rate
SWA Side Warning Assistance TTI Time to Intersection
SVM Support Vector Machine TTC Time to Collision
SVR Support Vector Regression TTCCP Time to Critical Probability
TDV Traffic-Driver-Vehicle V2V Vehicle-to-Vehicle
THW Time Headway WHO World Health Organization
Abstract

Longitudinal and lateral control of the vehicle on the inference. Chapters 3 and 4 contain the techniques for
highway are highly interactive tasks for human drivers. traffic context perception that focus on sensor integra-
The intelligent vehicles and the advanced driver- tion, sensor fusion, and road perception. A review of
assistance systems (ADAS) need to have proper aware- lane detection techniques and its integration with a par-
ness of the traffic context as well as the driver to make allel driving framework is proposed. Next, a novel inte-
an appropriate assistant to the driver. The ADAS also grated lane detection system is designed. Chapters 5
need to understand the potential driver intent correctly and 6 discuss the driver behavior issues, which provide
since it shares the control authority with the human the driver behavior monitoring system for normal
driver. This book provides research on the driver inten- driving and secondary tasks detection. The first part is
tion inference, particular focus on the two typical based on the conventional feature selection method,
vehicle control maneuvers, namely, lane change ma- while the second part introduces an end-to-end deep
neuver and braking maneuver on highway scenarios. learning framework. Understanding the driver status
A primary motivation of this book is to propose algo- and behaviors is the preliminary task for driver inten-
rithms that can accurately model the driver intention tion inference. The design and analysis of driver braking
inference process. Driver’s intention will be recognized and lane change intention inference systems are pro-
based on the machine learning methods due to its posed in Chapters 7 and 8. Machine-learning models
good reasoning and generalizing characteristics. Sensors and time-series deep-learning models are used to esti-
in the vehicle are used to capture context traffic informa- mate the longitudinal and lateral driver intention.
tion, vehicle dynamics, and driver behavior Finally, discussions and conclusions are made in
information. Chapter 9.
This book is organized in sections and chapters,
where each chapter is corresponding to a specific topic.
Chapter 1 introduces the motivation, human intention KEYWORDS
background, and general methodological framework ADAS, Computer vision, Driver behaviors, Driver inten-
used in this book. Chapter 2 includes the literature sur- tion inference, Intelligent vehicles, Automated driving,
vey and the state-of-the-art analysis of driver intention Machine learning.

vii
CHAPTER 1

Introduction

Worldwide traffic departments have reported that more ADAS techniques such as lane departure avoidance,
than 1.2 million traffic-related injuries happen each lane keeping assistance, and side warning assistance
year. Among these traffic accidents, more than 80% can help the driver to make the right decision and
were caused by human errors [1]. The World Health Or- reduce the workload.
ganization (WHO) reported that traffic accidents each It is predicted that the shipment of ADAS in the
year cost around V518 billion worldwide and on future has great potential and can generate a huge
average, 1%e2% of the world gross domestic product amount of commercial benefit based on many automo-
[2,3]. In the past, in-vehicle passive safety systems tive market analyzers. One example is shown in Fig. 1.2
such as airbags and seat belts have played a significant according to the prediction of Grand View Research, Inc.
role in the protection of drivers and passengers. These ADAS products will show a significant increase in the
technologies have saved millions of lives. However, next 5 years. Therefore the utilization of ADAS products
they are not designed to prevent accidents from will become more accessible to the public, although this
happening but just try to minimize the injuries after can bring a series of problems. For example, the finan-
the accidents happen [4]. Therefore recent efforts have cial cost will increase. Also, as most of the automotive
been devoted to the development of safer and intelli- companies are developing their ADAS products, safety
gent systems toward the prevention of accidents. These insurance and product quality can be different from
systems are known as the Advanced Driver Assistance each other. Once the drivers are getting familiar with
Systems (ADAS). these products, they will heavily rely on these systems.
ADAS is a series of fast-developing techniques that A very famous example is the Tesla car crashes that are
are designed for improving driver safety and increasing caused by their autopilot ADAS products, as shown in
the driving experience [5]. ADAS relies on a multimodal Fig. 1.3. The autopilot products of Tesla are one of the
sensor fusion technique to integrate multiple sensors most successful commercial driver assistance and semi-
such as light detection and ranging (lidar), radar, cam- automated driving assistance system in the world. The
era, and GPS into a holistic system. The sensors’ work- product is set of intelligent computing, perception,
ing ranges are shown in Fig. 1.1. Most of the current and control units, which can significantly increase

FIG. 1.1 Distribution of Advanced Driver Assistance Systems in an advanced vehicle (deepscale.ai/adas.
html). Lidar, light detection and ranging.

Advanced Driver Intention Inference. https://doi.org/10.1016/B978-0-12-819113-2.00001-4


Copyright © 2020 Elsevier Inc. All rights reserved. 1
2 Advanced Driver Intention Inference

driving safety issues. However, even such a smart system factor, the driver. Vehicles are working in a three-
can be reported for car crashes worldwide. One of the dimensional environment with continuous driver-
most common reasons for a crash is the driver overtrust- vehicle-road interactions. Drivers are the most essential
ing the autopilot when the system is activated, which is part of this system, who control the vehicle based on the
a problem in the future. surrounding traffic context perception. Therefore allow-
The reasons why ADAS cannot be 100% trusted are ing ADAS to understand driver behaviors and follow
multifold. Currently, most of the reasons are due to driver’s intention is of importance to driver safety,
immature techniques; however, a deeper reason is that vehicle drivability, and traffic efficiency.
the driver and the automation lack mutual understand- Human driver intention inference is an ideal way to
ing. The inputs of current ADAS are mainly based only allow ADAS to obtain the ability of reasoning. The rea-
on the vehicle dynamic states and traffic context infor- sons for developing driver intention inference tech-
mation. Most of the systems ignore the most critical nique are multifold: first of all, the most important

FIGURE 1.2 Advanced Driver Assistance Systems market prediction (Grand View Research, Inc. https://
www.grandviewresearch.com). ACC, adaptive cruise control; AEB, automatic emergency braking; AFL,
adaptive front light; BSD, blind spot detection; LDWS, lane departure warning systems; TPMS, tire pressure
monitoring system.

FIGURE 1.3 A Tesla car has crashed into a parked police car in California, USA. (https://www.bbc.com/
news/technology-44300952).
CHAPTER 1 Introduction 3

and significant motivation is to improve driver and can act as the guidance to design an automatic
vehicle safety. Accordingly, two different driving sce- decision-making system.
narios require inferring the driver’s intention. The first Moreover, in terms of the level 3 automated vehicle
one is to better assess the risk in the future based on (according to the SAE International standard on the
the driver’s interesting region. The second one is to classification of automated vehicles), accurate driver
avoid making decisions that are opposite to the driver’s intention prediction enables a smoother and safer tran-
intent. For the first case, there is evidence that a large sition between the driver and the autonomous
number of accidents are caused by human error or controller [11,12]. When the level 3 automated vehicles
misbehavior, including cognitive (47%), judgment are operating in an autonomous condition, all the
(40%), and operational errors (13%) [6]. Therefore driving maneuvers are handled by the vehicle. However,
monitoring and correcting driver intention and once the vehicle cannot deal with an emergent situa-
behavior seem to be crucial in the effectiveness of a tion, it has to give the driving authority to the driver.
future ADAS. Meanwhile, the increasing use of in- This process is known as disengagement [13]. In such
vehicle devices and information systems tend to distract a case, the vehicle can assess the takeover ability of
the driver from the driving task. For the design of future the driver according to the continuously detected inten-
ADAS, it is therefore beneficial to integrate intended tion. If the driver is focusing on the driving task at that
driver behaviors from the early design stages [7,8]. moment and has an explicit intention, the vehicle can
ADAS usually automatically intervene in the vehicle warn the driver and pass the driving authority to the
dynamics and share the control authority. To ensure driver. This will make sure the transition between driver
cooperation, it is crucial that ADAS is aware of driver and controller is as smooth as possible. However, if the
intention and does not operate against the driver’s will- driver is believed to be unable to handle the driving
ingness. For example, in complex traffic conditions such task, the autonomous driving unit should help the
as intersection and roundabout, it is crucial not to inter- driver gain situation awareness as soon as possible
rupt the driver making decisions, especially not to sus- and take emergency action immediately.
pend the driver with misleading instructions. This Another essential reason to develop the driver inten-
makes it reasonable or even necessary for ADAS to tion inference system is it will contribute to the devel-
have the ability to accurately understand the driver’s opment of automated vehicles. As shown in Fig. 1.4,
driving intention. On the other hand, intention infor- each level of understanding about the driver can be
mation enables for more accurate prediction of future mapped into the corresponding intelligent level of an
trajectories, which would be beneficial to risk assess- autonomous vehicle. Comprehensive research on each
ment systems [9,10]. Driver intention inference will layer will contribute to the development of the relative
benefit the construction of the driver model, which layer in the autonomous vehicle. Driver intention

FIGURE 1.4 The evolution from the current vehicle to future autonomous vehicle.
4 Advanced Driver Intention Inference

recognition is a relative higher-level understanding of driver behavior analysis is of importance to infer


human drivers and related to the decision-making layer driver intention.
of autonomous vehicles. Modeling driver intention 4. Driver lane change intention inference algorithms:
mechanisms is critical to the construction of automated Based on the specific traffic context and driver be-
decision-making algorithms. Human drivers are the haviors, the next task is to infer driver intention
teacher of automated drivers. The automated drivers properly. The algorithms for intention inference
can learn when and how to make the decision based should have the ability to capture the long-term
on the driver’s intention knowledge. Once the human dependency between the temporal sequences.
driver becomes the passenger in the automated driving Moreover, the intention inference algorithms should
vehicles, it will be easier for the passengers to accept predict the intention as early as possible.
that the automated driving systems is such systems The driver lane change intention platform requires
that can remember the driving habit of the passenger. the integration of software and hardware systems.
Therefore a good study about when and how drivers Driver intention inference has to take the traffic context,
generate their intentions will benefit the design of the driver behaviors, or dynamic vehicle information into
decision-making module for intelligent vehicles. Based consideration, which will fuse multimodal signals and
on such a design method, the vehicles will be more mining the long-term dependency between different
similar to human drivers, which will make it much signals based on machine learning methods. In terms
easier for humans to accept these highly intelligent of the hardware system, the sensors, included in this
vehicles. book, contain RGB and RGB-D cameras and vehicle
As discussed earlier, teaching ADAS to understand navigation system. Besides, all the sensors are tested
driver intention is essential as well as challenging to and mounted on a real vehicle in this case to collect
enhance the safety of the driver-vehicle-traffic close- naturalistic data. Specifically, the traffic context such
loop system. To focus more, this book will target two as lane positions and front vehicle position will be pro-
of the most popular driving scenarios in both longitudi- cessed with image-processing methods. One web cam-
nal and lateral directions, namely, the braking and lane era is mounted inside the cabinet. The driver behavior
change maneuvers. For example, during a standard lane dynamics will be evaluated within a steady and dy-
change maneuver, the driver is expected to perform a se- namic vehicle. The RGB-D camera (Microsoft Kinect
ries of behaviors (e.g., mirror checking and turn the V2.0) will be used for the steady vehicle, while another
steering wheel). The driver’s lane change intention can web camera will be used to record the driver behavior
be inferred in an early stage by recognizing driver be- during the highway driving task. These signals are
haviors and traffic context situations. A driver lane recorded with one laptop for further processing and
change intention system facing next-generation ADAS analyzing. The algorithms used in this project are
is developed in this study. Based on this, four main ob- mainly focused on machine learning methods, which
jectives are determined: include supervised learning, unsupervised learning,
1. Driver intention process analysis: To predict driver and deep learning models. All the algorithms are writ-
lane change intention, it is vital to understand the ten in MATLAB and Cþþ.
human intention mechanism, such as how The driver’s intention inference task described in this
the intention is generated and what is the stimuli of book relies on machine learning algorithms to work in
the intention. The nature behind driver intention real time. The reasons for using machine learning can be
is the first question that needs to be answered. multifold. First, the real-time traffic context and driver
2. Traffic context perception: The driver is in the middle behavior data can be high dimensional and of large vol-
of the traffic-driver-vehicle loop. Traffic context is the ume, and very few mathematic models can deal with
input to the driver perception system, which makes such data. However, machine learning algorithms are
it act as the stimuli of the driver’s intention. There- useful for high-dimensional multimodal data process-
fore understanding the current traffic situation will ing. Second, the utilization of a machine learning algo-
benefit the intention inference system. rithm enables learning the long-term dependency
3. Understanding driver behaviors: Driver behaviors, between driver behaviors and traffic context, which
such as mirror checking, are the most important significantly increases the inference accuracy for the
clues before the driver makes a lane change. The lane change intention. Finally, it is hard to find the
driver has to perform a series of checking action to intention generation and inference pattern based on
have a full understanding of the surrounding context observation and modeling. The machine learning algo-
before he/she decides to change the lane. Therefore rithms provide an efficient way to learn knowledge
CHAPTER 1 Introduction 5

from naturalistic data. With some advanced deep Human behavior is directly influenced by intention.
learning techniques, it is even possible to achieve an Intention can be determined by the three aspects
end-to-end learning process. Although machine mentioned earlier [15]. Specifically, attitude toward the
learning algorithms are very powerful in dealing with behavior describes how much willingness does one
the tasks described in this book, they do have limita- have to take the behavior; a strong level of attitude can
tions. The major limitation of using machine learning give a strong intention of taking actions in a certain
algorithms is data collection. Data is the heart of the task. Human beliefs determine the attitude toward a
machine learning algorithms. To obtain an accurate behavior about how much outcome it brings after the
intention inference results, several experiments need behavior is taken. Second, the subjective norm reflects
to be designed and data need to be collected. Insuffi- the pressure from the surrounding social life of a human.
cient data volume will lead to overfitting and bad infer- It evaluates how much the family, friends, and the society
ence results. Besides, most of the data used in the book expect from a person to make certain behavior. Finally,
need manual labeling, which is time-consuming. perceived behavior control is developed from the self-ef-
Finally, the training and testing of machine learning al- ficacy theory given in [16]. It describes the confidence of
gorithms give rise to a higher computational burden an individual to perform the behavior. For example, if
both for the financial and temporal costs. there are two subjects with the same intention, the one
who is more confident in the task can perform better
behavior toward finishing a certain task. Based on [16],
WHAT IS HUMAN INTENTION? the planned behavior, perceived behavior control, and
This section describes the human intention mechanism the intention can be used directly to predict the behavior
based on existing studies. Driver intention is a subset of performance.
human intention that particularly occurs during Bratman also pointed out that intention is the main
driving. The human intention has been theoretically attitude that directly influences plans [17]. Also, Heinze
discussed and studied by several studies in the past [18] described a triple-level description of intentional
three decades. From the cognitive psychology perspec- behavior, namely, intentional level, activity level, and
tive, intention refers to the thoughts that one has before state level. According to Tahboub, in the human-
producing an action [14]. In the theory of reasoned ac- machine interface scope, intention recognition is gener-
tion given by Fishbein, the intention is in the center of ally defined as understanding the intention of another
the theory, which is to perform a given behavior. Three agent. More technically, it is the process of inferring an
aspects determine intentional behavior: the attitude to- agent’s intention based on its actions [19]. Elisheva pro-
ward the behavior, subjective norm, and perceived posed a cognitive model with two core components:
behavior control [15] (Fig. 1.5). intention detection and intention prediction. Intention

FIGURE 1.5 Architectural diagram of human intention [15].


6 Advanced Driver Intention Inference

detection refers to detect whether or not a sequence of ac- For example, Jang et al. [22] used the eyeball move-
tions has any underlying intention at all. Intention pre- ment pattern as the input to recognize human’s
diction, on the other hand, refers to predict and extend implicit intention. The intention recognition task
the intentional goal by a set of incomplete sequence of was viewed as a classification problem. They divided
actions [20]. In terms of driving intention inference, it implicit human intention into informational and
equals to the part of intention prediction that is navigational groups and nearest neighbor algorithms,
mentioned earlier because we assume drivers always as well as support vector machines (SVMs) to train the
have an intention during a short period (e.g., lane keep- classifier. It is also confirmed that the fixation length,
ing and following can be viewed as driver intention). fixation count, and pupil size variation were the
The inference and reasoning process make people main factors to discriminate human intention. Kang
clever and easier to take part in the social community. et al. [23] proposed a method of human implicit
A human can recognize other’s intentions based on intention recognition based on electroencephalo-
the observation and the knowledge stored in the graphic (EEG) signals. This algorithm focused on ser-
brain. However, it is a difficult task to make the intel- vice web queries. Three kinds of classification
ligent machine to infer human intention easily and methods were adopted, which were SVM, Gaussian
accurately. To some extent, only when a robot detects mixture model (GMM), and naïve Bayes. The implicit
human intention based on human observation can it human intention was classified into two types called
be viewed as an intelligent agent. Based on the study navigational and informational intentions. An eye-
by Meltzoff and Brooks [21], self-experience plays an tracking system was used to help track the subjects’
important role in making an inference of the intended eye movement in the experiment. Results showed
state of other agents. In terms of robots and intelli- that SVM gave the best classification result with
gent vehicles, self-experiences were obtained from 77.4% accuracy. Wang et al. [24] determined the user’s
learning a large amount of relevant events data. intention based on eye-tracking systems. The fuzzy
As mentioned earlier, human intention inference has inference was used to infer the user’s intention, with
been widely studied in the past decades. One of the most eye gazing frequency and gazing time as the input.
significant applications of human intention inference is The fuzzy logic controller outputs the probability of
human-robot interface design. Thousands of service ro- the user’s intention on one particular region of the
bots were designed to assist humans in completing their computer screen.
work in both daily life and a dangerous situation. The Generally speaking, the human intention inference
traditional robots were designed from a robot’s point problem contains a large amount of uncertainty and
of view rather than from a human point of view, which noise exists in the measurement device. Therefore
reduces the interaction level between humans and ro- probability-based machine learning methods are a
bots. A robot should have the ability to learn and infer powerful tool in solving this kind of problem, and it
human intentions and obtain basic reasoning intelli- has been successfully applied in many cases. In terms
gence in order to improve the efficiency of human- of human intention inference task, which is a work to
robot interaction (HRI) as well as its intelligence. A infer human mental hidden states, the hidden Markov
widely accepted method of classification of human in- model (HMM) and the dynamic Bayesian theory are
tentions in HRI scope is to classify the human intentions two very popular ways of inferring the human mental
into explicit and implicit intentions. The explicit inten- state. In Ref. [25], a living assistance system for elder
tion is much clearer than the implicit intention and and disabled people was developed. The HMM was pro-
hence easier to recognize. Explicit intention means posed based on the hand gesture information. Five basic
humans directly transmit their intention to the robot hand gestures were defined to represent human inten-
by language or directly command through the computer tion, which was come, go fetching, go away, sit down,
interface. On the contrary, implicit intention reflects and stand up. The features of hand movement data
those human mental states that cannot be communi- were extracted and converted to an observable symbol
cated to the robot. The robots have to observe and un- for HMM. The results of the experiment showed the ef-
derstand human behaviors first, and estimate the fects of the intention inference system. In Ref. [26], an
human intention based on the gained knowledge and intention recognition method for an autonomous mo-
the on-going human actions. Implicit intentions usually bile robot was developed. The HMMs were trained for
can be further separated into informational and naviga- five kinds of human actions, such as following, meeting,
tional intentions. Human implicit intention researches passing by, picking an object up, and dropping an object
have been done in various areas. off. After this stage, different models were constructed to
CHAPTER 1 Introduction 7

represent different behaviors, then the robot moved and assist his/her power. A human-will model was
autonomously and can be regarded as an observer to used to explain human intention when cooperating
recognize the human intention of the five types of ma- with the robot. A modified hidden Markov Model
neuvers and rejustify its model structure. The recognition (HMM) was proposed to infer human path intention
accuracy can reach 90%e100%. In Ref. [27], a symbiotic on the robot arm, and a filed impedance equalizer
human-robot cooperation scheme was proposed. A (FIE) was used to assist humans in merging their arm
robot was controlled wirelessly by a human to reflect a force to the robot system. The experimental result
human’s idea, and other robots were used to infer this ro- showed that the MHMM enables the intended path
bot’s intention and help it work. A vague timescale recognition in an early stage of human motion, and
method was used to help the robot infer the target ro- through the FIE, the desired impedance pattern is merged
bot’s intention with historical data rather than instanta- through the proposed assistance system.
neous information only. Given a simple task, the Kulic and Croft [30] developed a human intention
human behavior model can be constructed with fuzzy inference unit aiming to assist a controller of the HRI
automata, and the transition of the human intention is system. The intention signal was used in a planning
determined by the fuzzy rules using the qualitative and control strategy to enhance the safety and intuitive-
expression. ness of the interaction. Different physiologic data,
Rani et al. [28] aimed to study and recognize human which were blood volume pressure, skin conductance,
status from their physiologic cues by using machine chest cavity expansion/contraction, and corrugator
learning methods. Four kinds of machine learning muscle activity, were collected and featured extracted
methods were adopted, which are SVM, K-nearest before being fed into the intention recognition unit.
neighbor, Bayesian dynamic model, and regression Then a fuzzy inference mechanism was used to infer hu-
tree. The models were trained to classify five human man emotional states such as its valence and arousal
mental states (anxiety, engagement, boredom, frustra- level. The final estimation of arousal level achieved
tion, and anger) based on human physiologic cues. 94% accuracy within four subjects and the performance
Then a systematic comparison of the performance of ma- of valence estimation is 80%. The intention of an HRI
chine learning was evaluated. The results showed that monitoring system was separated into two components:
SVM gives the best classification result with 85.81% accu- attention and approval (Fig. 1.6). Human attention was
racy. In Ref. [29], a human-robot system was designed to determined by both physical and cognitive processes
reflect a human intention on the path of the robot arm focused on the robot. Human approval to robot’s

FIGURE 1.6 Architectural graph of human intention [30].


8 Advanced Driver Intention Inference

work should not only be determined based on physio- and disabled people and that have the ability to read
logic signals but also be validated physically based on both intentions. Pereira [34] proposed a novel inten-
user attention information. tion recognition method based on a causal Bayesian
In Ref. [31], the authors proposed an HRI method network and plan generator. Logic programming tech-
for a LEGO assembling task. The human upper body nique was used to compute the probability of inten-
type robot was designed to assist the human finish tions, and those with lower values were filtered out.
the assembling job by estimated human intention. This made the recognizing agent focus on the most
Eye gaze information was collected and used for inten- important events, especially when making a quick deci-
tion inference. Three kinds of potential cooperation sion. Then the plan generator generates conceivable
actions for the robot were defined, which were taking plans that can achieve the most likely intentions given
over, settlement of hesitation, and simultaneous by the first step. The plan generator guides the recog-
execution. The humanoid robot recognizes and clas- nizing process concerning hidden actions and unob-
sifies the human state into one of the three states and servable effects. In Ref. [35], an intention recognition
takes appropriate actions to help humans finish the method based on the observation of human behaviors,
assembling task. The authors introduced a human relevant goals, and current context was proposed. The
intention recognition method among human-robot author introduced an extra level between actions and
collaboration tasks [32]. The intention was recognized intention, which was called the goal level. The defini-
by using a probabilistic state machine to classify tions of intention and goal were given as goal was
explicit and implicit intentions. A human interacted something that humans want to achieve, while inten-
with a robot arm in the experiment that was executed tion was a mental state of what a human was going to
in an interactive workshop. Five explicit intentions do (see Fig. 1.7).
were introduced: picking and placing the intention of To achieve the intention recognition, an intention
an object, passing an object to the robot, placing the graph method was used. There were five elements in
object, picking and holding an object, and giving the one intention graph, namely, state, action, goal, inten-
pointed object to the human. Two implicit intentions tion, and edges. In the goal recognition step, a graph
were piling up the objects and unpiling the objects. was constructed, and the relevant goal was determined
The proposed probabilistic state machine for the robot based on the actions. Then in the intention recognition
is working efficiently on both explicit and implicit hu- part, the determined goal and user profile information
man intention recognition. from the current context were used to infer the real
Bien et al. [33] pointed out that from the point of intention. Therefore the intention inference can be
intention reading, a human-friendly interface can be viewed as the backpropagation of human behavior
classified into three classes according to the autonomy execution. In Ref. [36], the authors proposed a human
level, namely, a fully manual interface, semiautono- intention recognition method in a smart assisted living
mous interface, and fully autonomous. Given the intel- system based on a hierarchic hidden Markov model
ligence level, a human-friendly interface can be (HHMM). They used an inertial sensor mounted on
classified into two classes. In the first class, computa- one finger to collect finger gestures. Five finger gestures
tional intelligence was used and in the second class, were defined, which are come, go fetching, go away, sit
the higher level, the machine can predict human inten- down, and stand up. The final result showed that by us-
tion on the job, decide whether it can be done or not, ing HHMM, the model could achieve 0.9802, 0.8507,
and interact with a human. The author also proposed 0.9444, 0.9184, and 1.000 classification accuracy for
two different kinds of systems designed for elderly the five kinds of hand gestures, respectively. Wang

FIGURE 1.7 Human intention recognition procedure [35].


CHAPTER 1 Introduction 9

et al. [37] introduced a human intention inference sys- Finally, the measurement nodes have to be modeled
tem named the intention-driven dynamic model based on the given sensor system.
(IDDM), which is based on the extension of the One interesting point should be pointed out that
Gaussian process (GP) model and Bayes theorem. The there are some differences between the driver’s inten-
author extended an intention node in IDDM based on tion and driver’s behavior. Although there are no clear
GP and introduced the online algorithm to infer human definition and boundary between driver behavior and
intention based on movement. The authors made a suf- intention in prior research, the differences between
ficient comparison with other algorithms such as SVM these two concepts should be aware of. Driver inten-
and GP classification. The performance of the novel tion reflects a driver’s mental state, whereas driver be-
method overweighs the traditional ways showing that haviors are actions that drivers take. Driver behavior
IDDM is efficient in dealing with human intention has a much wider scope than driver intention, which
problems. is focused not only on driving behaviors but also on
In Ref. [38], a tractable probabilistic model based some other behaviors including distractions (such as
on the Bayesian network was introduced. The experi- answering phone [41]) and mental and physical fa-
ment was designed in a virtual kitchen simulation tigue (such as yawning and sleepiness [42]).
environment. The virtual human intentions are to In conclusion, driver intention is a subclass of hu-
load the dishwasher, wash dishes, cook and lay the ta- man intention and can be roughly defined as a kind
ble, get a drink from the robot, and get an object from of driver’s thoughts during the goal-oriented driving
the robot. An expert provided the mapping of the task. The driver’s intention will influence the driver’s be-
intention of actions. The user’s head and hand mo- haviors and a series set of actions will then be executed
tion were tracked, and these observable data were by the driver based on the strength of driver intention.
fed into a hidden dynamic Bayesian network
(DBN). One of the key benefits of the proposed
method is that it can derive the model directly based DRIVER INTENTION CLASSIFICATION
on expert knowledge. Bascetta et al. [39] developed a Driver intention can be classified into different cate-
human intention estimation method focusing on hu- gories based on a different perspective. It can be clas-
man intended motion trajectory and body pose pre- sified according to awareness, motivation, timescale,
diction for humans interacting with the industrial and the direction of driving. Among these, the two
robot. The proposed method enabled safe coopera- most used classifications are based on the timescale
tion between humans and robots in the industrial of the intention and on the route. In terms of time-
area even without protective fences. The robot scope classification, Michon [43] pointed out that
detected the human through a low-cost camera. Based the cognitive structure of human behavior in traffic
on the human tracking algorithm, the robot can pre- environment is a four-level hierarchic structure, which
dict the human’s intended working area (four areas contains road users, transportation consumers, social
were defined, namely, human working area, robot agents, and psychobiological organisms. Among these
area, inspection area, and cooperation area) before four levels, the road user level is directly connected
the human comes into a certain area. The intention with drivers and can be further divided into three
recognition algorithm was performed by an HMM levels: strategy, tactical, and operational levels (also
and results showed that the intention estimation al- known as control level)(Fig. 1.8). These three cogni-
gorithm could successfully predict the interaction tive levels can be viewed as three driver intention
area; 92% human intention was correctly recognized. levels based on the timescale. The strategy level de-
In Ref. [40], a generic human intention method was fines the general plane of a trip that determines the
introduced. The author analyzed the property of hu- trip route, destination, risk assessment, etc. The time
man intention and its relationship with actions. constant will be at least at the minute level or even
Then a DBN method was used because of its flexible longer. In this level the driver plans the transport
characteristics to deal with arbitrary systems or do- mobility and comfort issues, which formed a long
mains. To cover both continuous and discrete inten- timescale problem. At the tactical level, the driver
tions, a hybrid DBN method was discussed. To can make a series of short-term decisions and plans
implement the model to the robot, the possible user to control the vehicles. The time constant is much
intention and actions should be first determined; shorter than that in the strategical level, which usually
then the model parameters have to be learned from lasts for a few seconds. The hierarchic level of inten-
either expert knowledge or data-driven method. tion is shown in Fig. 1.8.
10 Advanced Driver Intention Inference

FIGURE 1.8 Three-level architecture of road users to describe driver intention [43].

Tactically planned intentional maneuvers include a according to the awareness of the current situation and
sequence of operational maneuvers to fulfill a short- information that is gathered from the control and
term goal such as turning, lane changing, and braking monitoring modules. One significant advantage of the
etc. [44]. All the control commands must meet the cognitive driver model is the incorporation of built-in
criteria from the general goal set at the strategy level. features that mimic human abilities.
Lastly, the operational level intention is the shortest Among cognition scope, Rasmussen [46] developed
among the three levels and stands for the willing to another three-level architecture model to describe hu-
remain safe and comfortable in the traffic situation man performance behavior. Three typical levels,
and the driver directly gives control signals to the namely, skill-based, rule-based, and knowledge-based
vehicle. The time constant can be a microsecond level. levels, were defined according to human performance.
Most of the existing studies have dedicated to opera- The three-level model described human performance
tional level maneuver study and the strategy route from a familiar situation to unfamiliar situations.
plan and prediction based on GPS techniques. Howev- Skill-based behavior represents human’s behavior in a
er, the main task in this book will focus on tactical familiar situation. All the activities at this level go fast,
level study, particularly lane change maneuvers; as smooth, and automated, which means a human is
the time constant of tactical maneuvers is a few sec- very confident to take any maneuver at this level and
onds, it is important to correctly identify and recognize is very familiar with the task. A more detailed example
the intention to enhance the functionality of intelli- and how this three-level architecture can be related to
gent driver assistance system. the model given by Michon is showed in Table 1.1
There are also some other ways of classification for [47]. The next level is rule-based behavior. In this level,
the cognition of driver intention. Salvucci [45] devel- humans become less familiar and confident with the
oped a driver model in Adaptive Control of situation as in the first level. All the behaviors in this
Thought-Rational cognitive architecture. Similar to the relatively familiar situation are based on the stored
three-level architecture of the road user model given rule-base in the brain. During unfamiliar situations, it
by Michon, Salvucci developed the integrated driver comes to the knowledge base level. In this level, the
model into three main components, which were con- control of the action moves to a higher conceptual level.
trol, monitoring, and decision-making modules. The behavior here is goal-oriented and clearly formu-
Particularly, the control component was similar to the lated. Humans need to analyze the environment and
operational level given by Michon, which gave the make an explicit plan to complete the task based on
charge of perception of the external world and translate the knowledge they have in their brain. The plan is usu-
any perceptual signals directly to the vehicle. The ally selected from different plans that have been made,
monitoring component keeps aware of the surrounding and the one that gives the least error should be chosen.
situation and the environment by periodically This level is similar to the strategy level mentioned
perceiving data and inferring. The decision component, earlier. When drivers are dealing with an unfamiliar
which has the same function as part of Michon’s tactical destination, they have to search the optimal route and
level, makes tactical decisions for individual maneuvers transport methods to finally complete this task.
CHAPTER 1 Introduction 11

TABLE 1.1
Examples and Relationships Between Models Given by Michon and Rasmussen [47].
Strategic Tactical/Maneuver Operational/Control
Knowledge Navigation in unfamiliar area Controlling skid Novice on first lesson
Role Choice between familiar routes Passing other vehicles Driving unfamiliar vehicles
Skill Route used for daily commute Negotiating familiar intersection Vehicle handling on curves

The classification of driver intention toward direc- the classification system. The regularized linear
tional classification, on the other hand, is quite straight- discriminant classifier was trained to identify emer-
forward. As we are considering the underground gency braking intention. Their conclusion suggested
vehicles, they can only have two directions: longitudi- that the electrophysiologic method can be used in
nal and lateral. The driver’s longitudinal behavior con- the braking assistant system. The reasons were multi-
tains braking, acceleration, deceleration, starting, and fold. First, the identifying time by using electrophysi-
lane keeping. Lateral behavior usually contains turning, ology in the real world was even earlier than that in
over-taking, and lane changing. In previous research, the simulator (can detect braking intention 237 ms
most researchers pay attention to lateral intention pre- before the braking action occurs), and second, the per-
diction because they are more complicated and more formance of intention identification method was
important than longitudinal behaviors owing to the robust even when drivers took a secondary task. Kim
interaction with other vehicles. et al. also proposed a driver braking intention method
In terms of longitudinal intention prediction, most by using EEG signals [52]. Particularly, normal
of the previous research focuses on the study on braking braking and emergency braking were defined and clas-
intention recognition. Kim et al. proposed a braking sified by the kernel ridge regression model. By
intention prediction method by using EEG and electro- analyzing EEG signals, three features were used:
myographic (EMG) signals and tried to study the neuro- event-related potentials, readiness potentials, and
driving simulation framework [48e50]. Three different event-related desynchronization. The author
braking maneuvers were defined in this paper: sharp concluded that, in general, the prediction perfor-
braking, soft braking, and no braking condition. They mance of normal braking could be much better than
applied three algorithms: kernel ridge regression, linear that of emergency braking. The reasons were the fea-
regression and combined linear regression, and classifi- tures used in the study were more suitable for normal
cation approach to make a prediction and classification braking intention analysis and the overfitting problem
of emergency braking, voluntary braking, and mixed of the regression model given by the training data.
braking. EEG signal and vehicle technical signals (i.e., Haufe used EEG and EMG as an input signal to predict
vehicle brake input) were used as the input of the infer- the driver’s emergency braking intent [53]. The signals
ence mechanism. The final result showed that the pre- after feature extraction were fed into a regularized
dicted proposed algorithm output can bring the linear discriminant analysis classifier. The simulated
minimum error with the measured value, and linear systems with EEG showed a 130 ms earlier detection
regression gave the second good result. Moreover, the than those that rely only on brake pedal signals. Kha-
result indicated that within these three methods, the liliardali proposed a driver intention prediction
prediction performance of normal braking is always model to determine whether the driver wants to go
better than that of emergency braking for the reason ahead or stop [54]. The method was to classify driver
that these features are more suitable to be used for go and stop intention based on brain-machine inter-
normal driving classification. It is considered that the face. Particularly, EEG, electrooculographic (EOG),
combined EEG features will contribute to the decrease and EMG signals from six subjects in a simulation
of prediction errors. environment were collected, and two classification
Haufe and Kim aimed to prove the ability of methods (linear and quadratic discriminant analyses)
correlated neural electrophysiology to improve the were used separately to evaluate the classification per-
prediction of emergency braking situations in the formance. The final result showed that both the algo-
real-world driving environment [51]. Vehicle states rithms could achieve above 70% accuracy for driver
as well as EEG and EMG signals were used to train intention and above 80% for brake intention.
12 Advanced Driver Intention Inference

The research mentioned earlier uses the EEG signal beyond 90%. Takahashi et al. predicted the driver decel-
as one direct measurement for a human mental state. eration intent in a downhill road [59]. A gear shift oper-
But other research tends to use surrounding context ation will be executed automatically once the driver’s
information, vehicle sensors, and driver information mental model outputs a deceleration signal. The results
from cameras and the Controller Area Network indicated that an automatic gear shift performed by the
(CAN) bus. For example, McCall and Trivedi [55] in- controller should be finished before the driver forms an
tegrated driver intention into an intelligent braking intention to shift gears. To introduce the driver’s driving
assistance system. A sparse Bayesian learning algo- experiences and personality traits, the author used Inter-
rithm was used to infer the driver’s intention of pretive Structural Modeling (ISM) method to describe
braking. Three kinds of data sources were used in these qualitative and quantitative factors. The inference
this work, which were vehicle data, surrounding model consists of two types of submodels, which can be
traffic information, and driver state. Head and foot regarded as qualitative and quantitative. The decelera-
monitoring cameras were used to capture the driver’s tion intent can be deduced based on the output of these
head and foot motion. The system collected data from two submodels.
cameras, GPS, lidar, and radar and CAN bus systems. Kumagai et al. proposed a method to predict the
To capture information about the driver’s actions, a driver’s braking intention at intersections (particularly
color camera was installed to observe the driver’s right turns) by using a DBN [60]. The data from the
head and a near-infrared camera was used to observe real vehicle, which were vehicle speed, acceleration
the driver’s feet. The system recognized driver actions pedal, and brake pedal, are collected. Two different
once the driver removed his/her foot. The results Bayesian networks, namely, HMM and switching linear
showed that the inference system could detect the dynamic model, were constructed. The results showed
driver braking intention 1e3.3 s before the braking that the inference system could predict the future
action occurs, with a relatively low false alarm rate. braking intention several seconds prior based on the
The author concluded that by adding foot cameras current observed data. The author also pointed out
to the vehicle and from the surrounding data, the that the easiest way to make stop prediction is to
inference performance could get a significant increase construct a static table that maps the observed data
of 1 s before the braking happens, while the addi- into the frequency of future stops. However, this
tional head camera showed few influences on the per- method may need more training data to get a
formance of the system. Another interesting research convincing result.
by Tran, Doshi, and Trivedi [56] predicted the driver’s In terms of lateral direction intent, most of the driver
braking intention by directly monitoring the foot intention research in a lateral direction focused on lane
gesture through cameras. They showed that driver changing and turning. As lane change analysis is the pri-
foot gesture is an important factor in vehicle control, mary object in the following chapters of this book, this
and therefore the usage of vision-based foot behavior chapter will only focus on turning intention and a few
has its advantage, which is direct and accurate. Also, types of research based on steering intention recogni-
by monitoring the foot behavior directly, it can help tion. Cheng and Trivedi [61] developed a vision-based
predict a pedal press action before it happens. A body-pose recovery and behavior recognition algo-
new vision-based framework for driver foot behavior rithms for driver assistance systems. The main task
analysis was developed using optical flow-based foot was to focus on a slow turn scenario (when the driver
tracking and HMM to characterize the instant foot meets an intersection, the driver first stops the car and
behavior. then starts to turn). It was shown that the classifier
Mabuchi and Yamada [57,58] estimated the driver's can achieve 82% and 96% prediction accuracy for the
stop and go intention at the intersection when facing left and right turns when the prediction was made 1.5
the yellow traffic lights. Two algorithms were used to s prior to the intersection. The false-alarm rate is around
predict driver intention, which were an SVM and the lo- 20%. When the decision was made 0 s before the vehi-
gistic regression model. Also, two methods were used to cles enter the intersections, the classifier performance
model vehicle behavior at the intersection depending decreased to 82% and 75% for the left and right turns,
on whether or not the driver realizes the yellow light. respectively. Meanwhile, the author examined the influ-
The experiment was designed in a real-world situation, ence of steering data and its combination with pose
and 22 subjects participated in the experiment. The final data; it was found that the classifier trained with both
results showed that with a low false-positive rate (5%), pose and steering data produced the best true-positive
the true positive rate of both the classifiers could go rates between 1- and 1.5-s decision times at better
CHAPTER 1 Introduction 13

than 85%. This means the classifier trained by head and vehicle performance at intersections, four different
hand pose, along with steering data, gives better perfor- kinds of driver intentions were defined, which were
mance than head pose only, hand pose, and steering going straight, stop at a red light, turn right, and turn
only. right but stop at a pedestrian crossing. Two particular
Windridge et al. [62] aimed to find a model that can maneuvers (going straight and turning right only)
identify intentional task hierarchies. Martin proposed were evaluated. Based on the different velocity profiles
an architecture that can deal with arbitrary combina- of different driver intentions, different probabilistic dis-
tions of subsequent maneuvers, and a varying set of tributions for the maximum acceleration parameter
available features was also investigated [63]. Right were determined. Lastly, a simple Bayesian net was
turn maneuver inference was studied as an example of used as the inference mechanism. Measurement data
the inference system. In the study, the authors were collected from four different drivers at an inner-
concluded that discriminative approaches were usually city intersection.
used to predict a single type of maneuver, while gener- Driver tactical maneuvers consist of a series of oper-
ative models were more often applied to infer multiple ational maneuvers. Some of the existing research focus
maneuvers. Ohashi et al. designed a system that recog- on the analysis of multiple tactics rather than a single
nized the turning left, right, and going straight inten- tactical task. By using machine learning theory, a single
tions by collecting data of human motions, the intention inference task can be completed with a
velocity of the vehicle, and the distance of host vehicle discriminative and generative model. However, multi-
to the intersection [64]. The intention recognition ple task inference prefers to use generative models
model was proposed using case-based learning based such as Bayesian networks and relevant inference algo-
on the fuzzy associative memory system. The final result rithms [44]. Mitrovic [68] proposed a driving event
showed that the ratio of correctly detecting left turning recognition scheme based on HMMs. Particularly, seven
is 88%, going straight is 95%, and turning right is 86%. driving events (right curve, left curve, right on the
The model was proved to be efficient and was able to roundabout, left on the roundabout, right turn, left
identify the intention at an early stage when detecting turn, and straight on the roundabout) were studied
nonlearned data. and classified by using real vehicle driving data in a
Hülnhagen predicted the turn maneuver at the normal driving environment. Multiple models were
intersection by using probabilistic finite-state ma- evaluated based on the observation sequences from
chines and fuzzy logic [65]. First, each driving ma- the training set, and the highest probability decides
neuver was modeled by a set of basic elements. Each what kind of events this sequence represents. The left-
element was specified by a set of fuzzy rules. Then a to-right HMM was selected because of its ability to bet-
Bayes approach was introduced to recognize a driving ter characterize the data than general HMM. The driving
maneuver. A training method based on the optimiza- event with the highest probability was chosen, and only
tion method was used to train the fuzzy rule based on 4 of 238 test events were wrongly classified. Although
recorded data. The final result showed that the system the system cannot always clarify between left turn and
could achieve a 73% correct rate in the 23 total events. left turn on roundabout, the system can correctly recog-
Sato and Akamatsu [66] evaluated the remaining dis- nize 234 of 238 driving events with about 98.3%
tance of a vehicle to the center of the targeted intersec- accuracy.
tion when the driver was monitored to cover the In Refs. [69 and 70], the driver assistance system
brake pedal and perform a right turn signal switch based on the human-machine interface was studied.
maneuver. It was different from most of the existing Oliver and Pentland proposed a driver behavior
research, which focused on the time and moment recognition and prediction method based on dy-
profile. The authors focused on vehicle location to namic graphic models. Particularly, a dynamic
predict driver intention. More specifically, vehicle ve- graphic model, HMM, and its extension model,
locity, relative distance to the leading and following coupled HMM, were used as the prediction controller.
vehicle, the onset locations of covering brake pedal, The primary task of these studies was to evaluate the
and turn signals were used to construct a driver struc- influence of contextual information on driver
tural equation model as the intent inference behavior recognition. Thus environmental data were
mechanism. collected and used in the paper, which was surround-
Liebner et al. [67] aimed to infer driver intention ing traffic and lane information, vehicle data, driver
based on an explicit longitudinal driver behavior model head position, and viewpoint. All the experiments
and simple Bayesian network. To evaluate the dynamic were designed in a real traffic situation. Seven driver
14 Advanced Driver Intention Inference

maneuvers were studied, including passing, changing Jingjing and Yihu [73] proposed a two-dimensional
right and left, turning right and left, starting, and HMM using the Gaussian mixture probability density
stopping. Final results indicated that the predictor as driver intention identification. The driver’s inten-
could recognize the driver’s maneuver 1 s before a sig- tion was viewed as a combination of longitudinal
nificant change in the car and contextual signals take and lateral direction intention, and six driving inten-
place. Different kinds of situations were studied to tions are studied. With a 3-s time interval of input
test the HMM performance, such as car signal only, data and eight Gaussian mixture numbers, the identi-
car and lane position, car and driver gaze, and car, fication accuracy can achieve 98.84%. The author
lane, and driver information. Final results showed believed that speed change and lane change should
that by using a combined data source, the prediction be considered at the same time rather than separately,
performance of passing and stopping could have which was not enough to reflect the actual situations.
100% accuracy, while the performances of changing Imamura developed a driver intention identification
lane right and left are much lower than the other ma- and intention label definition method based on the
neuvers (6.3% and 23.5%, respectively). It can be assumption of compliance with traffic rules [74].
concluded that the performance of only use car infor- Four subjects participated in the experiment, and the
mation will have a plateau. Some driving events, such driving behavior data were collected by using a driving
as passing and lane changing, cannot rely only on car simulator with the control of gas, brake, and steering
information. The context information is important to wheel. The GMM was used as the intention identifica-
identify turning and lane-changing maneuvers. The tion algorithm, and the expectation-maximization al-
driver’s eye gaze signal, which can reflect the driver’s gorithm was used to train GMM parameters. After
mental state, can be very useful in the prediction of training the GMM offline, the author constructs an on-
lane changing, passing and turning events, etc. line estimation system. Also, four intention labels,
Liebner inferred the driver’s intent based on an which were normal driving, deceleration, acceleration,
explicit model for the vehicle’s velocity, and an intelli- and turn left/right are defined. The final result showed
gent driver model was used to represent car-following an average estimation accuracy of 83.3%.
and turning behavior [71]. The model input data were Liu and Pentland [75] aimed to analyze the
extracted from real-world data to account for different sequence of steps within a driving action in their study
driving styles. Four kinds of driver intent were defined and to use the first steps to identify the action. The
and recognized, which were going straight, stop at the main approach was to model human behavior in a
stop line, turn right, and turn right but stop at a pedes- Markov process. The result showed that by using the
trian crossing. To construct the simple, intelligent method developed, the driver’s intention can be
driver model, the velocity profile and its derivate observed before the maneuver is performed. The
were used. The author believed driving maneuvers experiment was designed in a simulation environ-
consist of a linear sequence of actions that were ment with eight adult male subjects who drove
much suitable for the application of HMMs. He and through a citylike virtual environment. The driving
Zong used a double-layer HMM structure to predict command included stop, turn left, turn right, change
driver intention and driving behavior [72]. Specif- lanes, and passing a car. A time window method was
ically, a lower layer Gaussian HMM was designed to used to examine its effect on driver intention predic-
infer the short-term driving behavior based on the tion. In terms of a left turn prediction, the accuracy
three main data sources: brake and acceleration pedal, is between 50% and 60%. For passing, the recognition
steering wheel angle, and speed. Then, the upper accuracy is 60%e70% 1 s after the command, and the
discrete HMM can estimate one of the tactical driver in- performance of the right turn can be 60% 2 s after the
tentions (emergence braking, hill starting, braking in a command. The stopping and following can only
turn, and obstacle avoidance) based on the recognized achieve 40%e50% 2.5 s after. As can be seen, using
driving behaviors. Also, driving behaviors shortly were multiple tasks for driving intention recognition was
predicted using the likelihood maximum method. The much complex than a signal task. It has to define
driving intention was recognized online using Lab- and clear more driving maneuvers and take many
VIEW (Laboratory Virtual Instrument Engineering more experiments. Moreover, it should pay more
Workbench). The system using the proposed method attention to the generative inference algorithm, which
can achieve smooth transition of control modes be- was supposed to recognize more driving maneuvers
tween automated and manual driving. correctly and accurately.
CHAPTER 1 Introduction 15

STUDIES RELATED TO DRIVER INTENTION accurately classify the drivers’ intention into a hurry and
INFERENCE normal driving group.
There are some other research that have not directly In terms of strategic intention study, most of the
studied host driver’s tactical intention behavior but existing research focus on driving route prediction
have focused on the other aspects of driver mental state based on GPS and digital maps to estimate the driver’s
and intention relevant areas, for example, the opera- strategy intention. Rammelt [77] built two types of
tional driver modeling, strategy level programming, tactical driver models, which partly connect to the strat-
driver type classification, driver’s attention and distrac- egy level. There were two different driver models: reac-
tion analysis, and driver intention recognition of sur- tive model and planning model. Both the models
rounding vehicles. Among these, the operational and were constructed based on DBNs. In a reactive model,
strategic level intention research also have a close rela- the output probability of the intended maneuver was
tionship to the tactical driver’s intention, as they are only related to the current sensors’ input, and the plan-
from the three-level driver mental architecture and can ning model aims to find a solution to get into a zone of
interact with each other. Similarly, studies on driver’s interest. Five index values were introduced to evaluate
attention and distraction can be analyzed with the the performance of the plan, which were a prior, belief
same methods as it was used in the driver intention in success, benefit, risk, and effort. The planning model,
study because all these research were preformed to which combines a plan module and probabilistic mod-
detect and recognize the inner mental state of the driver ule, makes a natural formulation of the driving task.
by observing outer real-world variables. Driving inten- Nakajima proposed a route recommendation
tion, attention, and distraction research all belong to method for a car navigation system based on the driver’s
human cognition study, which forms a complete intention estimation [78]. The method focused on the
mental research system. Another interesting research driver’s strategic route plan. A difference-amplification
area is to infer the surrounding driver’s intention. This method was used to estimate the driver’s intention on
is also a challenging task and is important to traffic se- route based on the comparison of the newly selected
curity. Unlike host driver intention inference, surround- route and the original one given by the Dijkstra algo-
ing intention inference can only be analyzed based on rithm. The difference amplification method updated
the measurable parameters of target vehicle, such as the driver’s current route intention by recalculating the
its velocity and acceleration estimation and the sur- distance cost when the real route was different from
rounding environment. As the current technology has the original recommended route. Simmons predicted
not solved the driver’s intention inference and vehicle- driver’s every day strategic route and destination inten-
to-vehicle (V2V) data transition completely, the study tion by using the HMM [79]. Drivers’ daily driving data
on surrounding driver’s intention will still focus on were collected from the GPS and map database, and
the outer vehicle characteristics and analysis of the sur- thus their natural driving habit was recorded. The
rounding environment. method can predict the driver’s destination and route
Imamura and Takahashi et al. proposed a driver’s accurately online through the observation of GPS posi-
intention and driver style estimation method based tion. The final result showed that approximately 98%
on the HMM, considering the driver’s personality and accuracy could be achieved. The novel method was
characteristics [76]. In this paper, the authors defined designed with the assumption that the driver’s driving
four kinds of driving state, which are stopping, acceler- habits and goals were always similar during weekdays.
ation, cruise, and deceleration. Besides, the authors Data bins were used to characterize the data to construct
evaluated the different driver characteristics of driving an extended HMM.
in a hurry and driving normally by using a driver’s Driver behavior at the intersection also attracted
intention ratio index. The driver’s intention ratios of many researchers’ efforts for the reason that intersection
hurry driving were always larger than that of normal is one of the most important and challenging parts in
driving except in the group of cruise state. In the real-world driving. The traffic at the intersection can
research, the experiment was designed in driving simu- be very complex, which makes it one of the major acci-
lator environment and gas, and brake pedal signals, as dents happening areas. Therefore the development of
well as vehicle acceleration, were collected and packed ADAS, which can predict driver intention and behaviors
as the input signal. Then the HMM was trained to be at the intersection, will largely enhance driving safety
the driving state classifier. The final results showed and efficiency. Two novel classification methods were
that by considering the input data, the classifier was effi- introduced by Aoude to identify the driver types at the
cient in recognizing the different driving states and can intersection [80,81]. Specifically, a discriminative
16 Advanced Driver Intention Inference

model based on the SVM and Bayesian filter and a radar, lidar, and GPS; and vehicle dynamic sensors.
generative model based on the HMM and expectation- The merge and lane change assist system can recom-
maximization methods were evaluated separately. mend the driver when and how to make the lane change
These classifiers aim to determine whether the based on a dynamic probabilistic drivability map. Then
oncoming driver was a violator (cannot stop before the real on-road test, which is open to the public and
the stop line of the intersection) or compliant. By press, was performed to prove the efficiency of this sys-
observing 10,000 examples in a certain intersection, tem. Lastly, Doshi and Trivedi evaluated the head and
the classifiers could correctly recognize the driver’s eye gaze dynamics under various conditions leading
stop intention. Under 1 s of time-to-intersection condi- to the driver’s distraction [86]. The experiment was
tions, the SVM with a Bayesian filter classifier achieves designed in a simulation environment to make sure
85.4% accuracy and the HMM obtains 80.0% accuracy, the safety of the subject. The authors finally concluded
both can get a high-level true-positive rate with a rela- that the eye gaze and head pose information were essen-
tively low (5%) negative-positive rate. Lefèvre and Javier tial features for the recognition of driver distraction and
et al. proposed a Bayesian inference system to recognize attention states.
the driver behaviors at intersections [82,83]. They intro-
duced contextual information through a digital map to
provide informative cues for analyzing driver behaviors. CONCLUSION
Four kinds of data were used to describe the driver Driver intention inference is a very complex task that re-
behavior at the intersection, which were entrance lies heavily on knowledge from a variety of research area
road, entrance lane, turn signal, and path in intersec- such as psychology, cognition, human-machine inter-
tions. Finally, a probability density of the intended ma- face, and automotive engineering. Driver intention
neuver was inferred from these data. study is a subtask for human intention cognition. As
Another challenge in research is driver attentive level mentioned earlier, human intention can be described
and distraction monitoring, and their correlation with as a hierarchic mental architecture and will influence
driver intention inference. The driver’s attention reflects the actions that are going to occur. Therefore by
a driver’s mental state and concentration level while observing human actions, it is possible to infer human
driving, which makes the analysis very similar to driver intention with this given observation. Similar to human
intention inference. Both can be viewed as a mental intention, driver intention reflects the driver’s thoughts
inference process based on the observation of human related to the driving task and will influence the driving
outer physical behavior. McCall and Trivedi [84] action and maneuvers based on the strength of the
described a driver attention monitoring system design intention. It should also be pointed out that there are
on intelligent human-centered vehicle test bed. To some differences between driver maneuver and driver
monitor and recognize driver attention status, a variety intention. The driver’s intention here is closer to the
of data cues were captured and synchronized. For driver’s mental research and inference than driver
instance, facial expression analysis was obtained maneuver.
through the thin-plate splines algorithm. The lane posi- Driver intention can be classified based on different
tion analysis was proposed by applying one-lane criteria. For example, it can be classified according to the
tracking system. The data for the steering wheel and timescale as well as the direction of the action that is
pedal movement analyses were obtained from the going to take place. In terms of intention timescale,
vehicle CAN bus; time headway was calculated with there are three typical levels, which are strategy, tactical,
data collected via the radar system. All the cues were and operational levels. Strategy intention determines
proved to be effective in the determination of driver the driving planning level and route selection task,
attention level. Similarly, Tawari et al.introduced an ur- which is the longest task among the three. Operational
ban intelligent assistance system to deal with the com- level intention focuses on the instance and directly con-
plex and unique environment in an urban area [85]. trols the vehicle, such as steering and pedal control.
Two specific subsystems were developed, which were These maneuvers can be finished in a very short period,
driver attention guard and merge and lane change assist usually at the millisecond level. The tactical intention,
system. The intelligent driver assistance system was fully which determines the action of the next few seconds,
mounted on a real-world vehicle to collect natural data is more important in the real-world task. Most of the
while driving. The data sources covered the real-time sit- important decision during driving can be classified
uation of driver, vehicle, and surrounding environment into the tactical maneuver. Therefore the study of
based on the internal vision system; vision system; driver’s intention at this level is of importance to the
CHAPTER 1 Introduction 17

improvement of driving security. The driver will get a driver lane change focus on the state of the art of the
clearer idea about when and how to optimally execute intention inference and the challenges. Chapters 3 and
their thoughts based on the information given by the 4 cover a literature review of lane detection techniques
ADAS. and the integration methodologies for the lane detec-
It is also reasonable to classify driver intention from tion system.
the perspective of direction. Many previous types of Meanwhile, a parallel driving framework for lane
research focus on longitudinal and lateral vehicle dy- detection, namely, the parallel lane detection system
namics. Lateral dynamics are usually more complex is introduced. Chapters 5 and 6 illustrate the driver
than longitudinal dynamics. Thus driver intention in behavior recognition studies concerning the lane
the lateral direction has been the research hot spot for change maneuver prediction based on conventional
a few years. A lot of studies have focused on lane chang- machine learning and deep learning methods, respec-
ing and lane keeping, and some other lateral intention tively. Chapter 7 introduces braking intention recogni-
focus on analyzing driver behavior at intersections, tion and braking intensity prediction. Chapter 8
such as turning, stopping, or passing. On the other covers the lane change intention inference framework
hand, most of the longitudinal driver intention studies and the prediction results using naturalistic data.
focus on intelligent braking systems development. By Finally, Chapter 9 concludes the whole book and
predicting and recognizing the braking intention, the discuss the challenges as well as future works.
assistance system gives better guidance to the driver The main content of each chapter is summarized as
and improves cooperation efficiency. follows.
Apart from the direct tactical driver intention study Chapter 1: Introduction and motivation for the research
mentioned earlier, many other types of research have detailed throughout this book are proposed.
a close relationship to driver intention. Some research Chapter 2: This chapter provides studies about human
focus on driver behavior at intersections by inferring intention as well as driver intention. Driver intention
the other vehicle’s driving intention based on sensor is classified into different categories according to
fusion and vehicle communication. Moreover, to recog- different criteria. The lane change intention is also
nize the driver intention online at a strategic level is surveyed from the sensory level, algorithm level, and
another interesting aspect. Most of the research deter- the evaluation level.
mine the driver strategic intention based on GPS, digital Chapter 3: As traffic context is the major reason for lane
map information, and weekly driving task recordings. change maneuvers, one of the most important as-
From a cognitive point of view, driver attentive level pects, lane detection, is studied in this chapter. This
and distraction recognition are very similar to the driver chapter covers a sufficient survey of the lane detec-
intention study, which also can be viewed as a cognition tion techniques and the integration with other on-
process by observing the driver’s outer observable board systems.
behaviors. Chapter 4: A novel algorithm-level integrated lane
In summary, driver intention is a mental process detection framework is introduced. The integrated
during driving and can be identified by the following lane detection and evaluation system are designed to
driving actions. There are many ways to classify driver improve the robustness of the lane detection system
intention based on different criteria, and the tactical for further driver intention inference system
level intention is the most important aspect that we construction.
are concerned about. Chapter 5: Driver behaviors are the major clues for
driver intention inference because driver behaviors
carry valuable information that can reflect the
CHAPTER OUTLINES mental state of the driver. This chapter provides a
The main content of this book is divided into chapters driver behavior and secondary tasks recognition al-
according to different functional modules within the gorithm based on the driver’s head and body feature
driver intention system. The book starts with an over- selection. The feature engineering and conventional
view of the current issues related to the driver intention machine learning methods for driver behavior
inference, intelligence, and automated vehicles. This recognition are introduced.
chapter serves as a motivation to the following research Chapter 6: This chapter moves a step further in driver
and states the application area of the proposed technol- behavior monitoring, which use an end-to-end deep
ogy. Chapter 2 provides the analysis of both human learning approach to avoid complex feature extrac-
intention and driver intention. Literature surveys about tion and selection procedure.
18 Advanced Driver Intention Inference

Chapter 7: This chapter introduces an unsupervised [8] J.C. McCall, M.M. Trivedi, Human behavior based predic-
machine learning method based on braking inten- tive brake assistance, in: 2006 IEEE Intell. Veh. Symp.,
tion recognition system. The braking intention is 2006, pp. 8e12, 200.
identified and compared to two different unsuper- [9] H. Berndt, K. Dietmayer, Driver intention inference with
vehicle onboard sensors, in: 2009 IEEE Int. Conf. Veh.
vised learning methods. Then the feedforward neural
Electron. Saf., 2009, pp. 102e107.
network is used to estimate the braking intensity [10] H. Berndt, S. Wender, K. Dietmayer, Driver braking
based on different braking styles. Lastly, the feed- behavior during intersection approaches and implica-
forward neural network is used to construct an in- tions for warning strategies for driver assistant systems,
tegrated braking intention learning and braking in: 2007 IEEE Intell. Veh. Symp., 2007, pp. 245e251.
intensity estimate system. [11] A. Eriksson, N.A. Stanton, Takeover time in highly
Chapter 8: Driver intention inference algorithm automated vehicles: noncritical transitions to and
should be able to learn the dependency roles from manual control, Human Factors 59 (4) (2017)
among the temporal sequence data. In this chapter, 689e705.
a driver lane change intention inference algorithm [12] J. Nilsson, P. Falcone, J. Vinter, Safe transitions from auto-
mated to manual driving using driver controllability
based on the deep learning method is proposed.
estimation, IEEE Transactions on Intelligent Transporta-
Multimodal naturalist data that are collected on tion Systems 16 (4) (2015) 1806e1816.
highways are fed into the driver intention inference [13] C. Lv, et al., Analysis of autopilot disengagements occur-
system. The algorithm can predict driver lane ring during autonomous vehicle testing, IEEE/CAA Jour-
change intention a few seconds before the lane nal of Automatica Sinica 5 (1) (2018) 58e68.
change maneuver initiates. Then regarding the [14] P. Carruthers, The illusion of conscious will, Synthese 159
surrounding driver intention inference tasks, a (2) (2007) 197e213.
time-series trajectory prediction model is proposed [15] R. Netemeyer, M. Van Ryn, I. Ajzen, The theory of
to predict the track of the leading vehicle in real planned behavior, Orgnizational Behaviour Human De-
time. cision Processes 50 (1991) 179e211.
[16] Bandura, Toward a unifying theory of behavioral change,
Chapter 9: This chapter provides the final remarks,
Psychological Review 84 (2) (1977) 191e215.
discussions, and the future work required for driver [17] A.M. Treisman, A. Treisman, A feature-integration theory
lane change intention inference. of attention, Cognitive Psychology 12 (1980) 97e136.
[18] C. Heinze, Modelling Intention Recognition for Intelli-
gent Agent Systems, No. DSTO-RR-0286, Defence Science
REFERENCES and Technology Organisation Salisbury (Australia) Sys-
[1] L.S. Angell, J. Auflick, P.A. Austria, D.S. Kochhar, tems Sciences Lab, 2004.
L. Tijerina, W. Biever, T. Diptiman, J. Hogsett, S. Kiger, [19] K. a Tahboub, Intelligent human-machine interaction
Driver Workload Metrics Task 2 Final Report, 2006. No. based on dynamic Bayesian networks probabilistic inten-
HS-810 635. tion recognition, Journal of Intelligent and Robotic Sys-
[2] M. Peden, World Report on Road Traffic Injury tems 45 (2005) (2006) 31e52.
Prevention, World Health Organization, Geneva, 2004. [20] E. Bonchek-Dokow, Cognitive Modeling of Human
[3] Fundación Instituto Tecnológico para la Seguridad Del Intention Recognition, Dissertation, Bar Ilan University,
Automóvil (FITSA), El valor de la seguridad vial. Conocer Gonda Multidisciplinary Brain Research Center., 2011.
los costes de los accidentes de tráfico para invertir más en [21] A.N. Meltzoff, R. Brooks, Training study in social
su prevención (in Spanish), Report funded by the Spanish cognition, Developmental Psychology 44 (5) (2008)
General Directorate of Traffic, Universidad Politécnica de 1257e1265.
Madrid, Madrid, 2008. [22] Y.M. Jang, R. Mallipeddi, S. Lee, H.W. Kwak, M. Lee, Hu-
[4] T. Kowsari, Vehicular Instrumentation and Data Process- man intention recognition based on eyeball movement
ing for the Study of Driver Intent, Dissertation, The Uni- pattern and pupil size variation, Neurocomputing 128
versity of Western Ontario, 2013. (2014) 421e432.
[5] D. Geronimo, et al., Survey of pedestrian detection for [23] J.-S. Kang, U. Park, V. Gonuguntla, K.C. Veluvolu, M. Lee,
advanced driver assistance systems, IEEE Transactions Human implicit intent recognition based on the phase
on Pattern Analysis and Machine Intelligence 32 (7) synchrony of EEG signals, Pattern Recognition Letters
(2010) 1239e1258. 66 (2015) 144e152.
[6] M.G. Ortiz, Prediction of Driver Behaviour, Ph.D. [24] M. Wang, Y. Maeda, Y. Takahashi, Human intention
dissertation, Universitätsbibliothek Bielefeld, 2013. recognition via eye tracking based on fuzzy inference,
[7] J.C. McCall, Human Attention and Intent Analysis Using in: Soft Computing and Intelligent Systems (SCIS) and
Robust Visual Cues in a Bayesian Framework, 13th International Symposium on Advanced Intelligent
Dissertation, The University of California, San Diego, Systems (ISIS), 2012 Joint 6th International Conference
2006. on, IEEE, 2012.
CHAPTER 1 Introduction 19

[25] C. Zhu, W. Sun, W. Sheng, Wearable sensors based human [40] O.C. Schrempf, U.D. Hanebeck, A generic model for esti-
intention recognition in smart assisted living systems, mating user intentions in human-robot cooperation, in:
Informatica y Automatica 2008 (2008) 954e959. Proc. 2nd Int. Conf. Informatics Control. Autom. Robot.
[26] R. Kelley, A. Tavakkoli, C. King, M. Nicolescu, G. Bebis, (ICINCO 2005), vol. 3, 2005, pp. 251e256.
Understanding human intentions via hidden Markov [41] H. Alm, L. Nilsson, The effects of a mobile telephone task
models in autonomous mobile robots, in: Proc. 3rd Int. on driver behaviour in a car following situation, Accident
Conf. Hum. Robot Interact., 2008, pp. 367e374. Analysis and Prevention 27 (5) (1995) 707e715.
[27] Y. Inagaki, et al., A study of a method for intention infer- [42] Q. Ji, Z. Zhu, P. Lan, Real-time nonintrusive monitoring
ence from human’s behavior, in: Robot and Human and prediction of driver fatigue, IEEE Transactions on
Communication, 1993. Proceedings., 2nd IEEE Interna- Vehicular Technology 53 (4) (2004) 1052e1068.
tional Workshop on, IEEE, 1993. [43] J.A. Michon, A critical view of driver behavior models:
[28] P. Rani, C. Liu, N. Sarkar, E. Vanman, An empirical study what do we know, what should we do?, in: Human
of machine learning techniques for affect recognition in Behavior and Traffic Safety Springer, US, 1985,
human-robot interaction, Pattern Analysis and Applica- pp. 485e524.
tions 9 (1) (2006) 58e69. [44] A. Doshi, M.M. Trivedi, Tactical driver behavior predic-
[29] Y. Yamada, Y. Umetani, H. Daitoh, T. Sakai, Construction tion and intent inference: a review, in: IEEE Conf. Intell.
of a human/robot coexistence system based on a model of Transp. Syst. Proceedings, ITSC, 2011, pp. 1892e1897.
human will-intention and desire, International Conference [45] D.D. Salvucci, Modeling driver behavior in a cognitive
on Robotics and Automation 4 (May) (1999) 2861e2867. architecture, Human Factors 48 (2) (2006) 362e380.
[30] D. Kulic, E.A. Croft, Estimating intent for human robot [46] J. Rasmussen, Skills rules and knowledge, other distinc-
interaction, in: IEEE Int. Conf. Adv. Robot., 2003, tions in human performance models, IEEE Transactions
pp. 810e815. on Systems, Man, and Cybernetics 13 (3) (1983)
[31] K. Sakita, Flexible Cooperation between Human and 257e266.
Robot by Interpreting Human Intention from Gaze Infor- [47] A.R. Hale, J. Stoop, J. Hommels, Human error models as
mation, 2004. predictors of accident scenarios for designers in road
[32] M. Awais, D. Henrich, Human-robot collaboration by transport systems, Ergonomics 33 (10e11) (1990)
intention recognition using probabilistic state machines, 1377e1387.
in: 19th Int. Work. Robot. Alpe-Adria-Danube Reg. [48] I.-H. Kim, et al., Detection of multi-class emergency situ-
RAAD 2010 e Proc., 2010, pp. 75e80. ations during simulated driving from ERP, in: Brain-
[33] Z.Z. Bien, K.H. Park, J.W. Jung, J.H. Do, Intention reading Computer Interface (BCI), 2013 International Winter
is essential in human-friendly interfaces for the elderly Workshop on, IEEE, 2013.
and the handicapped, IEEE Transactions on Industrial [49] I.-H. Kim, J.-W. Kim, S. Haufe, S.-W. Lee, Detection of
Electronics 52 (6) (2005) 1500e1505. braking intention in diverse situations during simulated
[34] L.M. Pereira, H.T. Anh, Intention recognition via causal driving based on EEG feature combination, Journal of
bayes networks plus plan generation, in: Lect. Notes Neural Engineering 12 (1) (2014) 016001.
Comput. Sci. (Including Subser. Lect. Notes Artif. Intell. [50] J.-W. Kim, et al., Brain-computer interface for smart
Lect. Notes Bioinformatics), vol. 5816, LNAI, 2009, vehicle: detection of braking intention during simulated
pp. 138e149. driving, in: Brain-Computer Interface (BCI), 2014 Inter-
[35] S. Youn, K. Oh, Intention recognition using a graph national Winter Workshop on, IEEE, 2014.
representation, Imaging 1 (1) (2007) 2e7. [51] S. Haufe, J.-W. Kim, I.-H. Kim, A. Sonnleitner,
[36] C. Zhu, Q. Cheng, W. Sheng, Human intention recogni- M. Schrauf, G. Curio, B. Blankertz, Electrophysiology-
tion in smart assisted living systems using a hierarchical based detection of emergency braking intention in
hidden Markov model, Automation Science and Engi- real-world driving, Journal of Neural Engineering 11
neering (2008) 253e258. (5) (2014) 056011.
[37] Z. Wang, K. Mülling, M. Deisenroth, H. Ben Amor, [52] J. Kim, I. Kim, S. Lee, Decision of Braking Intensity during
D. Vogt, B. Schölkopf, J. Peters, Probabilistic movement Simulated Driving Based on Analysis of Neural Corre-
modeling for intention inference in human-robot lates, 2014, pp. 4129e4132.
interaction, The International Journal of Robotics [53] S. Haufe, M.S. Treder, M.F. Gugler, M. Sagebaum,
Research 32 (7) (2013) 841e858. G. Curio, B. Blankertz, EEG potentials predict upcoming
[38] O.C. Schrempf, D. Albrecht, U.D. Hanebeck, Tractable emergency brakings during simulated driving, Journal
probabilistic models for intention recognition based on of Neural Engineering 8 (5) (2011) 056001.
expert knowledge, in: IEEE/RSJ Int. Conf. Intell. Robot. [54] Z. Khaliliardali, R. Chavarriaga, L. Andrei Gheorghe,
Syst., 2007, pp. 1429e1434. J.D.R. Millan, Detection of anticipatory brain potentials
[39] L. Bascetta, G. Ferretti, P. Rocco, H. Ardo, H. Bruyninckx, during car driving, in: Proc. Annu. Int. Conf. IEEE Eng.
E. Demeester, E. Di Lello, Towards safe human-robot Med. Biol. Soc. EMBS, 2012, pp. 3829e3832.
interaction in robotic cells: an approach based on visual [55] J.C. McCall, M.M. Trivedi, Driver behavior and situation
tracking and intention estimation, in: 2011 IEEE/RSJ aware brake assistance for intelligent vehicles, Proceed-
Int. Conf. Intell. Robot. Syst., 2011, pp. 2971e2978. ings of the IEEE 95 (2) (2007) 374e387.
20 Advanced Driver Intention Inference

[56] C. Tran, A. Doshi, M.M. Trivedi, Modeling and prediction of intelligent driver model, in: IEEE Intell. Veh. Symp.
driver behavior by foot gesture analysis, Computer Vision Proc., 2012, pp. 1162e1167.
and Image Understanding 116 (3) (2012) 435e445. [72] L. He, C. Zong, C. Wang, Driving intention recognition
[57] R. Mabuchi, K. Yamada, “Estimation of driver’s intention and behaviour prediction based on a double-layer hidden
to stop or pass through at yellow traffic signal, Electronics Markov model, Journal of Zhejiang University - Science C
and Communications in Japan 98 (4) (2015) 35e43. 13 (3) (2012) 208e217.
[58] R. Mabuchi, K. Yamada, Study on driver-intent estima- [73] M. Jingjing, W. Yihu, Driving intentions identification
tion at yellow traffic signal by using driving simulator, based on continuous pseudo 2D hidden Markov model,
IEEE Intelligent Vehicles Symposium, Proceedings (IV) in: Intell. Comput. Technol. Autom. Int. Conf., 2012,
(2011) 95e100. pp. 629e632.
[59] H. Takahashi, K. Kuroda, A study on mental model for [74] S. Diego, Real-time Implementation of Estimation
inferring driver’s intention, Decision Control. Proceed- Method for Driver’s Intention on a Driving Simulator,
ings of 35th IEEE Conference 2 (December) (1996) 2014, pp. 1904e1909.
1789e1794. [75] A. Liu, A. Pentland, Towards real-time recognition of
[60] T. Kumagai, Y. Sakaguchi, M. Okuwa, M. Akamatsu, Pre- driver intentions, in: Proc. Conf. Intell. Transp. Syst.,
diction of driving behavior through probabilistic infer- 1997, pp. 236e241.
ence, in: Eng. Appl. Neural Networks, Eighth Int. Conf., [76] T. Imamura, T. Takahashi, Z. Zhang, T. Miyake, Estima-
no. September, 2003, pp. 8e10. tion for driver’s intentions in straight road environment
[61] S.Y. Cheng, M.M. Trivedi, Turn-intent analysis using body using hidden Markov models, in: IEEE Int. Conf. Syst.
pose for intelligent driver assistance, Pervasive Man Cybern., no. i, 2010, pp. 2971e2975.
Computing IEEE 5 (4) (2006) 28e37. [77] P. Rammelt, Planning in driver models using probabilistic
[62] D. Windridge, A. Shaukat, E. Hollnagel, Characterizing networks, in: Proceedings. 11th IEEE Int. Work. Robot
driver intention via hierarchical perception-action Hum. Interact. Commun., no. 2048, 2002, pp. 87e92.
modeling, IEEE Transactions on Human-Machine Sys- [78] S. Nakajima, D. Kitayama, Y. Sushita, K. Sumiya,
tems 43 (1) (2013) 17e31. N.P. Chandrasiri, K. Nawa, Route recommendation
[63] M. Liebner, C. Ruhhammer, F. Klanner, C. Stiller, Generic method for car navigation system based on estimation
Driver Intent Inference Based on Parametric Models, in: of driver’s intent, in: 2012 IEEE Int. Conf. Veh. Electron.
ITSC, 2013, pp. 268e275. Safety, ICVES 2012, 2012, pp. 318e323.
[64] I. Tamai, T. Yamaguchi, O. Kunihiro, Humane automo- [79] R. Simmons, B. Browning, Y.Z.Y. Zhang, V. Sadekar,
tive system using driver intention recognition, in: SICE, Learning to predict driver route and destination intent, in:
2004, pp. 4e7. 2006 IEEE Intell. Transp. Syst. Conf., 2006, pp. 127e132.
[65] T. Hülnhagen, I. Dengler, A. Tamke, T. Dang, G. Breuel, [80] G.S. Aoude, V.R. Desaraju, L.H. Stephens, S. Member,
Maneuver recognition using probabilistic finite-state ma- J.P. How, S. Member, Driver behavior classification at in-
chines and fuzzy logic, in: IEEE Intell. Veh. Symp. Proc., tersections and validation on large naturalistic data set,
2010, pp. 65e70. IEEE Transactions on Intelligent Transportation Systems
[66] T. Sato, M. Akamatsu, Influence of traffic conditions on 13 (2) (2012) 724e736.
driver behavior before making a right turn at an inter- [81] G.S. Aoude, V.R. Desaraju, L.H. Stephens, J.P. How,
section: analysis of driver behavior based on measured Behavior Classification Algorithms at Intersections and Vali-
data on an actual road, Transportation Research Part dation Using Naturalistic Data, no. Iv, 2011, pp. 601e606.
F: Traffic Psychology and Behaviour 10 (2007) [82] S. Lefèvre, J. Ibañez-Guzmán, C. Laugier, Context-based
397e413. estimation of driver intent at road intersections, in:
[67] M. Liebner, F. Klanner, M. Baumann, C. Ruhhammer, Computational Intelligence in Vehicles and Transportation
C. Stiller, Velocity-based driver intent inference at urban Systems (CIVTS), 2011 IEEE Symposium on, IEEE, 2011.
intersections in the presence of preceding vehicles, IEEE [83] S. Lefèvre, C. Laugier, J. Ibañez-guzmán, Exploiting Map
Intelligent Transportation Systems Magazine 5 (2) Information for Driver Intention Estimation at Road In-
(2013) 10e21. tersections, no. Iv, 2011, pp. 583e588.
[68] D. Mitrovic, Reliable method for driving events [84] J.C. McCall, M.M. Trivedi, Visual context capture and
recognition, IEEE Transactions on Intelligent Transporta- analysis for driver attention monitoring, in: IEEE Intell.
tion Systems 6 (2) (2005) 198e205. Transp. Syst. Conf., 2004, pp. 332e337.
[69] N. Oliver, A.P. Pentland, Graphical models for driver [85] A. Tawari, S. Sivaraman, M.M. Trivedi, T. Shannon,
behavior recognition in a SmartCar, in: Proc. IEEE Intell. M. Tippelhofer, Looking-in and looking-out vision for
Veh. Symp. 2000 (Cat. No.00TH8511), no. Mi, 2000, Urban Intelligent Assistance: estimation of driver atten-
pp. 7e12. tive state and dynamic surround for safe merging and
[70] N. Oliver, A.P. Pentland, Driver behavior recognition and braking, in: IEEE Intell. Veh. Symp. Proc., no. Iv, 2014,
prediction in a SmartCar, in: AeroSense 2000. Interna- pp. 115e120.
tional Society for Optics and Photonics, 2000. [86] A. Doshi, M.M. Trivedi, Head and eye gaze dynamics dur-
[71] M. Liebner, M. Baumann, F. Klanner, C. Stiller, Driver ing visual attention shifts in complex environments, J.
intent inference at urban intersections using the Vis. 12 (2) (2012), 9-9.
CHAPTER 2

State of the Art of Driver Lane Change


Intention Inference

DRIVER INTENTION INFERENCE Driver intention inference (DII) is an ideal way of


BACKGROUND allowing the ADAS to understand the driver. The reasons
Traffic accident statistics have shown that more than for recognizing driver intention are multifold. First, DII
80% of traffic accidents were caused by driver errors, improves driving safety. Specifically, there are two
such as misbehavior, distraction, and fatigue [1e3]. different driving scenarios that require inferring the
Various passive safety systems such as airbags and seat driver’s intention: to better assess the risk in the future
belts have played a significant role in the protection of and to avoid making decisions that are opposite to the
the driver and passengers when traffic accidents occur. driver’s intent [9]. For the first case, there is a shred of ev-
Although these technologies have saved a large number idence that a large number of accidents are caused by hu-
of lives, they are not designed to prevent traffic acci- man error, misbehavior, cognitive overload,
dents from happening but only protect the passengers misjudgment, and operational errors [2]. Monitoring
after the accident happens [4,5]. Therefore many efforts and correcting driver intention in time are crucial to the
have been devoted to the development of safer and effectiveness of ADAS. The intelligent vehicles can carry
more intelligent systems toward the prevention of acci- more high-resolution road perception sensors than a hu-
dents instead of only minimizing the impact of the ac- man driver, based on the sensor fusion technique. If the
cidents. The most successful active safety system is the vehicle can identify the driver’s intention in advance, the
advanced driver assistance systems (ADAS), which con- intelligent vehicle will pay more attention to the related
tains a series of active safety functionalities. road and analyze the possibility of such an intention
Most of the ADAS techniques such as lane departure based on the comprehensive assessment of the traffic
avoidance, lane keeping assistance (LKA), and side context. Hence, estimating driver intention prior to the
warning assistance (SWA) can assist the drivers by maneuver can largely increase driving safety. Meanwhile,
alarming in time, assisting in decision making, and the increasing usage of in-vehicle devices and other enter-
reducing driver workloads [6e8]. However, the inputs tainment devices can distract the driver from normal
of these systems usually rely on the vehicle’s dynamic driving. The driver may not sufficiently check the sur-
states such as the steering wheel angle, velocity, yaw an- rounding context before executing the control maneuver.
gles, Global Positioning System (GPS), and other traffic Hence, for the design of intelligent ADAS, it is beneficial
information. However, most of these systems ignore the to understand the driver intentions and execute correct
most important factor on the roaddthe human driver. assistance actions [10,11].
Vehicles operate in a basic three-dimensional environ- In terms of making the right decisions, ADAS inter-
ment with continuous driver-vehicle interactions. The venes with the vehicle dynamics and shares the control
three-dimensional road environment consists of road authority with the driver. The current ADAS functions
context, human driver, and vehicles. The structured or can be designed in a serial and parallel manner. For
unstructured road plays a fundamental function for example, the LKA system (LKAS) can generate feedback
drivers and other road entities. The host driver interacts control to the vehicle to slightly minimize the error be-
with other drivers by controlling the vehicle based on tween the vehicle central line and the lane middle posi-
the road context. Among the three aspects, drivers are tion. If the system does not understand the driver lane
the major component in this system, who control the change intent, conflicts between the driver and vehicle
vehicle to maintain a safe and efficient driving task. automation will occur, which can significantly decrease
Therefore allowing ADAS to understand drivers’ inten- the driver’s trust to the LKAS. Another example can be
tions and behaviors is important to driver safety, vehicle estimated in the complex traffic conditions such as at in-
drivability, and traffic efficiency. tersections and roundabouts, it is crucial not to interrupt

Advanced Driver Intention Inference. https://doi.org/10.1016/B978-0-12-819113-2.00002-6


Copyright © 2020 Elsevier Inc. All rights reserved. 21
22 Advanced Driver Intention Inference

the driver, especially not to interrupt the driver with comfortable driving experience with comfort zone,
misleading instructions. Moreover, mutual understand- the driver may try to overtake the leading vehicle that
ing between the driver and vehicle is a basic requirement has a slower speed to maintain their current speed or
for driver-vehicle collaboration. To ensure efficient coop- try to maintain a safe distance with the car behind
eration, it is important for the ADAS to be aware of the the vehicle. Sometimes lane change also happens
driver’s intention and not operate against the driver’s when the driver tries to leave a certain traffic segment.
willingness. This makes it reasonable for the ADAS to When this situation occurs, the driver will choose to
accurately understand the driver intention in real time. drive thevehicle into the adjacent lane and finish the
Furthermore, the DII system benefits from the devel- lane changing process.
opment of future automated vehicles. DII can be used When drivers initiate a lane change maneuver, they
to construct the driver model, which can act as the guid- are responsible for their own safety. A lane change ma-
ance for the design of the automated driver. Analysis of neuver requires more control actions and environment
the driver cognitive intention process based on the stim- checking behaviors than normal driving. The driver has
uli classification, control maneuver, and interaction to pay attention to the front as well as the rear situation
feedback can be used to construct a human-like of the target lane. Besides, the driver should still pay
decision-making model. Moreover, in terms of the level attention to the current context and surrounding
3 automated vehicles, as determined by the SAE Inter- vehicles around the blind spot area. The driver has to
national standard, accurate driver intention prediction decide the best moment to make a lane change maneu-
enables a smoother and safer transition between the ver based on current traffic information. Lee et al. sup-
driver and the autonomous vehicle controller posed that the driver is more likely to start a lane
[12e14]. When level 3 automated vehicles work at change when there is a gap of at least 12 m from the
automated conditions, all the driving maneuvers are front and rear vehicles or the relative velocity with the
handled by the vehicle; however, once an emergent sit- other vehicle (leading/following) is less than 22 km/h.
uation occurs, the vehicle has to disengage and give the However, the relative velocity and the distance from
driving authority to the driver. In such a case, the vehicle other vehicles are not a good measurement for lane
can determine whether the driver is ready to take over or change maneuver. The time to collision (TTC) can be a
not by assessing the driver’s intention in advance. If the better risk measurement index [120,121]. Lee deter-
driver has a reasonable driving intention, the vehicle mines that drivers will choose to change lane if the min-
should follow the driver’s maneuver and the assistance imum TTC with the front or rear vehicle is between 4
system should give fewer instructions. The application and 6 s. Wakasugi [122] suggested that the driver usually
of DII makes the transition between driver and starts a lane change when the TTC is larger than 6 s and
controller as smooth as possible. more likely to abort the lane change maneuver if TTC is
In this section, the literature review of DII will be pro- less than 10 s.
posed, and the lane change intention will be used as an Another important result given from the US traffic
exemplary scenario for further explanation. The compo- statistics is that less than half of the drivers signal
nents of this section can be summarized as follows. First, when they change the lane during their normal driving.
a state-of-the-art literature review about driver lane Besides, drivers are more likely to use a turn signal when
change intention is proposed. Then based on the lane moving into the adjacent lane rather than when coming
change intention inference (LCII) scenarios, the system back to the original lane when finishing an overtaking
architecture is categorized from different aspects. Next, [120]. Salvucci and Liu [32] revealed that 50% of the
the critical time flow of DII will be introduced, which turn signal is activated at the very beginning of a lane
leads to a comprehensive understanding of the architec- change and after 1.5e2 s this proportion increases to
ture of the intention inference system. Lastly, future 90%. This suggests two different driving patterns: using
works of DII will be proposed and discussed to benefit a turn signal to show the lane changing intent before
the development of future intelligent vehicles. performing the maneuver and using a turn signal to
indicate that the driver is performing the lane change
maneuver. They also studied the driver eye movement
LANE CHANGE MANEUVER ANALYSISdAN during the lane change process. The driver will spend
EXEMPLARY SCENARIO more time on the current lane before changing the
Lane change maneuver occurs when the host vehicle lane. As the maneuver starts, the driver moves his/her
driver is not satisfied with the current vehicle status eyes to the mirror. Then during the maneuver, the driver
or surrounding traffic context. To pursue a more directs his/her gaze to the target lane.
CHAPTER 2 State of the Art of Driver Lane Change Intention Inference 23

lane change intention will be described. According to


these limitations, the literature review of LCII will be
given in the next section.

Lane Departure Warning


Lane departure refers to the driver crossing the lane un-
aware of the surrounding context, and sometimes by
distraction, and it can be dangerous. Lane departure
warning system (LDWS) is an in-vehicle electronic sys-
tem that can alert the driver of the unintended lane de-
parture [124]. LDWS has no direct action on and
control over the vehicle chassis and steering units,
which means LDWS is just a warning system rather
than a driving assistance system like LKAS. LDWS usu-
ally uses vision sensor systems such as monocular or
stereo cameras to detect the lane boundary and vehicle
location, and then it can alert the driver if the vehicle is
coming into the dangerous situation (crossing lane) (as
shown in Fig. 2.2). LDWS is usually operated at speeds
FIG. 2.1 Lane change procedure [123].
of 56e72 km/h for passenger vehicles and 64e80 km/h
for trucks. The warning zones represent that the vehicle
According to the author in Ref. [123], lane change in is going to depart and no corrective steering force is
highway can be viewed as a seven-step procedure, applied to the vehicle. The threshold of the width of
which can be summarized as the driver using traffic in- the warning zone controls the sensitivity of the LDWS.
formation as a cue, determining the possibility of mak- Two documents guide the design of LDWS. The first
ing a lane change, monitoring the surrounding traffic, one is the ISO 17361 Lane Departure Warning Systems
rechecking the current state before lane changing, (ISO 17361:2007) (Fig. 2.3) and the second one is the
turning on signals, controlling and steering the vehicle, Lane Departure Warning Systems (FMCSA-MCRR-05-
and finishing the lane change maneuver; the process is 005). Both these systems define the LDWS in a similar
shown in Fig. 2.1. way: LDWS is an in-vehicle system that detects the
In conclusion, lane change maneuver is a complex vehicle lane departure situation through sensor systems
task during the driving process. Additional require- and alerts the driver when necessary. The driver main-
ments are needed by the drivers, such as mirror check- tains the vehicle control responsibility to make sure
ing and risk assessment. They have to be clear about the vehicle stays in the right place, and the LDWS
their surrounding traffic information, especially the does not interact with the vehicle control of the driver.
blind spot area. Sometimes driver’s distraction can According to Fig. 2.3, the status indication represents
also have an adverse effect on the lane changing maneu- the current position and state of the vehicle. There are
ver. The driver has to focus on the traffic and pay atten- three main components of the LDWS: lateral position
tion to the lane change maneuver. Hence, most of the detection and lane detection, lane departure warning
time, when the driver is going to make a lane change, and system status monitoring, and status warning sys-
the driver’s attention will be highly focused on the tem. Besides, a few other modules can help the LDWS
driving task and the distraction level is low. in making decisions, which are the suppression request,
vehicle speed reading, driver behavior, preference
reasoning, etc. If the system recognizes a proper driver
LANE CHANGE ASSISTANCE SYSTEMS intention, it may not warn the driver based on the intel-
Currently, some lane change assistance (LCA) systems ligent decision-making module.
have been developed to assist the driver in making Normally, two kinds of sensor systems are used in
lane change maneuvers and avoiding the collision. the LDWS: camera-based and infrared systems. The
This section first analyzes some recent ADAS products camera-based LDWS relies on the detection of the line
on the market such as lane departure warning, LKA, painted on the roadway and passes this information
and LCA system. Then their standard technique, advan- as well as the vehicle status (such as vehicle location, ve-
tages, disadvantages, and the necessity of introducing locity, and heading angle rate) to the processer. If the
24 Advanced Driver Intention Inference

FIG. 2.2 Detection zone definition for lane departure warning system [124].

FIG. 2.3 The lane departure warning system architecture defined in ISO 17361 [124].

data suggest the vehicle is leaving its path unintention- Lane Keeping Assistance
ally, it will alert the driver through the driver-machine LKAS can be viewed as an advanced version of LDWS for
interface. One significant drawback of camera-based the reason that it monitors the lane departure maneuver
LDWS is it relies heavily on the lane marks. If the lane and alerts the driver if necessary. Besides, it also assists
mark is not clear and the weather condition is adverse, the driver to control the vehicle by giving assist actua-
such as rainy, foggy, and ice on the carriageway, the sys- tion signals to the vehicle’s chassis and steering control
tem cannot work very well. The infrared system is usu- unit. It supports the driver in keeping to the lane by con-
ally mounted under the front bumper to detect the trolling the vehicle heading and yaw angles. LKAS is an
lane marks on the road and classifies the different reflec- extension of LDWS, which is designed to assist the
tions from the signals. When the vehicle crosses the driver in maintaining the lane position [125]. LKAS
lane, the sensor system detects a change and hence alerts uses similar sensor systems as in LDWS, in addition to
the driver. In terms of the human-machine interface the lane keeping controller and the assistance actuator
used in LDWS, there are three typical types, namely, (Fig. 2.4). The control variable for LKAS is the position
auditory warning, visual warning, and hepatic feedback. of the vehicle within the lane. LKAS normally works in a
Usually, a visual warning is suitable when the departure similar operational status as LDWS, i.e., above 65 km/h,
occurs slowly and an auditory and hepatic warnings are and focuses on the safety issues rather than on comforts.
the preferred choice when an emergency lane departure LKA can also patch with adaptive cruise system to help
happens. the driver steer and maintain the vehicle in the path.
CHAPTER 2 State of the Art of Driver Lane Change Intention Inference 25

FIG. 2.4 Functional architecture of the lane keeping assistance system [125].

The most significant difference between LKAS and collision with the approaching vehicle, and the vehicle
LDWS is whether a controller is used or not. The lane has a collision with the front overtaking car. Usually,
keeping controller calculates an optimal actuator lane change collision happens because the driver did
output based on the heading and lateral position of not see the vehicle in the blind spot area or carried
the vehicle within the lane and passes the control signal out a misbehavior during lane changing.
to the actuator unit. The actuator converts the control There is one document defined to support the LCA
signal into an action signal to support the driver. system, ISO 17387 Lane Change Decision Aid System
Currently, there are two different assist methods for (ISO 17387:2008). The standard defined in ISO
LKAS: steering input assist and corrective braking assist. 17383 supports the LCA system development for cars,
The LKAS uses the same lane detection sensor and algo- vans, and trucks. In this standard, the system definition
rithms as LDWS to identify the lane curves, vehicle po- is to warn the driver against the collision that may occur
sition, and relative status. In terms of the steering force when the driver initiates a lane change maneuver [81].
assist, when lane departure is detected, the LKAS alerts However, the systems just warn the driver through
the driver through the human-machine interface and in- human-machine interface without intervening and con-
tervenes by applying a slight countersteering angle. On trolling the vehicle. The driver should always be respon-
the other hand, the brake actuators generate the neces- sible for the safe action of the vehicle. The standard
sary yaw rate of the vehicle by controlling the brake requires the LCA to operate following a state diagram
pressure of each wheel separately. (Fig. 2.5) rather than define its functional element, as
shown in Fig. 2.5.
Lane Change Assistance There are two key aspects to be considered in LCA,
Lane change collision usually occurred when the driver which are coverage zone and the maximum closing speed
takes a lane change maneuver but neglects the adjacent of the target vehicle. The coverage zone is the entire area
vehicle in the “blind spot” or starts at an inappropriate that the sensor system monitored and covered outside
time. LCA system monitors the adjacent and rear lanes the vehicle. Two subgroups of the coverage zone are adja-
and the traffic to assist the driver when the driver is mak- cent and behind. This classification method enables the
ing a lane change maneuver. If the driver starts to LCA to be separated into three types. Type 1 covers the
change lane and the LCA detects a vehicle in the adja- adjacent region of the subject vehicle and therefore sup-
cent lane, the LCA will warn the driver to stop changing ports blind spot warning function only. Type 2 system,
the lane. Currently, this system does not have the func- on the other hand, monitors the behind area of the
tion to intervene with the driver, but in the future, LCA vehicle and only supports closing vehicle warning. Type
is expected to have lane change planning and control 3 system has a series of sensors to cover the areas both
functionality. The LCA warns the driver when he/she adjacent to and behind the vehicle and can provide blind
chooses an inappropriate time to change the lane. spot warning and closing vehicle warning function. The
Lane change collision usually happens during three sit- LCA products currently on the market can be separated
uations: the subject vehicle has a side collision with the into the blind spot monitoring system and lane change
vehicle in the lane, the subject vehicle experiences a rear warning system (monitoring the blind spot and the
26 Advanced Driver Intention Inference

FIG. 2.5 The functionality state diagram of the lane change assistance system [125].

behind area). Currently, four kinds of sensors are used in the drivers do not use turn signals in most conditions.
LCA, which are radar, camera, infrared, and ultrasonic However, some current LCA systems are activated by
sensors. Among these, radar and the camera can be the turn signals. If the drivers do not use the turn signal,
used in both blind spot and rear closing vehicle these LCA systems will fail to work. Therefore the DII al-
monitoring. gorithm is an ideal way to reduce the conflicts between
Several research have been carried out to study the driver intent and ADAS. It does not heavily rely on the
impact of LDWS and LCA on driver behavior. Portouli outer sensor system but can determine the driver’s
[126] pointed out that the usage of LDWS has no effect mental status according to inner vehicle sensors. If a
on the lane change frequency but it increases the use of driver lane change intention is estimated, it can
turn signals, which proved drivers improved their communicate with the LCA system in advance to
driving to avoid warning. It is also suggested that the make sure LCA can be executed and can improve the
driver does not drive faster because the driver feels safer driver safety more significantly. Hence, the next section
when the LDWS is activated. As LCA does not have that will give a detailed review of the research on driver lane
many uses, a few research can clearly point out how change intention.
exactly it influences the driver’s behavior. While Kiefer
and Hankey [127] suggest that the driver begins to Limitations and Emerging Requirement
look at both the side mirrors and the rearview mirror Based on the analysis of the lane change process and the
because the LCA increases the driver’s awareness of current LCA systems such as LDWS, LKAS, and LCA, it
safety. Meanwhile, no adverse effects were found on can be found that most of the current systems do not
mirror usage and lane change frequency in the have the capability to fully interact with the human
vehicle-mounted LCA. driver, as no driver behavior reasoning is used. These
The drawbacks of the current ADAS mentioned systems are mainly being activated based on the outer
earlier are that most of them rely heavily on sensor sys- traffic and road perception; however, driver behaviors
tems. If the weather condition is adverse, it is usually are a critical clue to the lane change maneuver predic-
difficult to obtain a good lane mark and around vehicle tion. If driver lane change intention can be detected in
detection, thus it can hardly provide proper assistance advance, the driver assistance system can be activated
to the driver. Moreover, as stated earlier, nearly half of in advance such that the potential driving interest
CHAPTER 2 State of the Art of Driver Lane Change Intention Inference 27

region and the blind spot region can be well monitored. the social community. Humans can recognize others’
Hence, it is necessary to design a DII unit to better intentions based on their observation and social skill
interact with the human driver. Normally the ADAS knowledge. However, it is difficult to make an intelli-
products are mainly designed to provide driving assis- gent machine, such as a smart vehicle, learn how to infer
tance in the L1 or L2 automated driving vehicles, in human intention accurately. To some extent, only when
which the driver is responsible for vehicle control. For a robot can detect human intention based on its own
higher level automated vehicles, in which the driver observation can it be viewed as an intelligent agent.
can share the control authority with the automation, Human intention inference has been widely studied
the driver intention reasoning is even more important in the past decades. One of the most significant applica-
to the automation to efficiently interact and cooperate tions of human intention inference is human-robot
with the driver. interface design [21e24]. Thousands of the service ro-
bots were designed to assist humans in completing their
works either in daily life or in a dangerous workspace.
HUMAN INTENTION MECHANISMS Traditional robots were designed from a robot’s perspec-
The human intention has been theoretically studied in tive rather than from a human’s point of view, which re-
the past two decades. From a cognitive psychology duces the interaction level between humans and robots.
perspective, intention refers to the thoughts that one To improve the efficiency of human-robot interaction
has before performing the actions [15]. Similarly, in (HRI) and the intelligence of the robot, the robot should
this study, intention (particular focus on the tactical have the ability to learn and infer human’s intention and
driving intention) is the attitude toward performing a obtain basic reasoning intelligence.
series of vehicle control maneuvers. The intention is A widely accepted method of classification for the
determined by three aspects, namely, the attitude to- human intention in HRI scope is to classify the human
ward the behavior, subjective norm, and the perceived intention into explicit and implicit intentions. Implicit
behavior control [16]. In Ref. [16], human behaviors human intention can be further separated into informa-
are found to be the response of the intention. The atti- tional and navigational according to Ref. [26]. The
tude toward the behavior describes how willing is the explicit intention is clearer than the implicit intention
human and how much effort the human puts to take and, hence, is easier to be recognized. Explicit intention
the behavior; a strong level of attitude can give a strong means humans can directly transmit their intention to
willingness of taking actions in a certain task. Second, the robot by language or direct command through the
the subjective norm reflects the pressure from the sur- computer interface, while implicit intention reflects
rounding social life of the human. Finally, the perceived the human mental state without any communication
behavior control was developed from the self-efficacy with the robot [25]. The robot has to observe and un-
theory. It describes the confidence of an individual to derstand human behavior first and then makes an esti-
perform the behavior. mation of the human intention at the right movement
Bratman [17] pointed out that intention is the main based on the knowledge base. Human intention infer-
attitude that directly influences future plans. In addi- ence problem contains a large amount of uncertainty,
tion, Heinze [18] described a triple-level description and noise exists in the measurement device. Therefore
of the intentional behavior, which contained inten- probability-based machine learning methods are
tional level, activity level, and state level. In the powerful tools in solving this kind of problem and
human-machine interface scope, according to have been successfully applied in many cases. The hu-
Ref. [19], intention recognition is the process of under- man intention inference can be mathematically
standing the intention of another agent. More techni- modeled as a process of inferring a series of human
cally, it is the process of inferring an agent’s intention mental hidden states. The hidden Markov model
based on its actions. Elisheva [20] proposed a cognitive (HMM) and the dynamic Bayesian theory are two pop-
model with two core components, which were inten- ular methods for the inference of human mental states
tion detection and intention prediction. Intention [27e29].
detection refers to detect whether a sequence of actions
has any underlying intention. Intention prediction, on
the other hand, refers to the prediction of the inten- DRIVER INTENTION CLASSIFICATION
tional goal based on a set of incomplete sequence of ac- Driver intention can be classified into different cate-
tions. The intention inference and reasoning process gories from a different perspective. For example, it can
makes people clever and enables them to take part in be classified according to the motivation, timescale,
28 Advanced Driver Intention Inference

and direction of driving. Among these, the two most operational maneuvers to fulfill the short-term goal,
straightforward ways of classifications are based on such as turning, lane changing, and braking maneuvers
the timescales of the intention and the driving [31].
direction. All the control commands must meet the criteria
from the general goal that are set at the strategic level.
Timescale-Based Driver Intention Lastly, the operational intention is the shortest one
Classification among the three levels and stands for the willing of
In terms of the timescale-based classification method, the driver to remain safe and comfortable in the traffic
Michon pointed out that the cognitive structure of hu- situation. The driver directly gives control signals to
man behavior in the traffic environment is a four-level the vehicle and the time constant is normally in milli-
hierarchic structure, which contains road user, transpor- seconds. As mentioned earlier, real-time lane change
tation consumer, social agents, and psychobiological intent inference plays a critical role in the improvement
organisms [30]. Among these, the road user level is of driving safety. In addition, continuous LCII is a rela-
directly connected with the drivers and can be further tively complex and difficult task than some other
divided into three sublevels: strategy, tactical, and oper- driving intentions. Salvucci and Liu [32] concluded
ational levels (also known as control level), as shown in that lane change not only was a control procedure but
Fig. 2.6. The three cognitive levels can be viewed as three also incorporated a set of critical aspects of driving
driver intention levels based on the timescale character- such as lower-level controls. Normally, lane change ma-
istic. Strategy level defines the general plan of a trip such neuvers will contain a series of short-term driving be-
as the trip route, destination, and risk assessment. The haviors such as the acceleration and deceleration in
time constant will be at least in minutes or even longer. the longitudinal direction and the steering wheel con-
At this moment, the driver will decide the transport trol in the lateral direction.
mobility and comfort issues, which is a long timescale There are also some other classification methods for
problem. In terms of the tactical level, in which the driver intention. Salvucci developed a driver model,
time constants are in seconds, the driver will make a namely, Adaptive Control of Thought-Rational cogni-
short-term decision and control the vehicle to negotiate tive architecture [33]. Similar to the three-level architec-
the prevailing circumstance. The tactically planned ture of road user model given by Michon, Salvucci
intentional maneuver consists of a sequence of developed the integrated driver model into three main

FIG. 2.6 Driver intention classification based on the time constant.


CHAPTER 2 State of the Art of Driver Lane Change Intention Inference 29

components, which are control, monitoring, and gesture through cameras [38,39]. They showed that
decision-making modules. The control component is the driver foot gesture plays an important role in vehicle
like the operational level given by Michon, which is control. Therefore the usage of vision-based foot
responsible for the perception of the external world tracking is more direct and accurate. Mabuchi and
and for transferring the perceptual signals directly to Yamada [40] estimated driver’s stop and go intention
the vehicle. The monitoring component keeps aware at intersections when the yellow light occurs. Takahashi
of the surrounding situation and environment by peri- and Kuroda [41] predicted the driver deceleration intent
odically perceiving data and inferring. The decision in a downhill road. Kumagai et al. [42] proposed a
component, which has the same function as part of method to predict driver’s braking intention at intersec-
Michon’s tactical level, makes tactical decisions for tions (particularly right turns) by using a dynamic
each maneuver according to the awareness of the cur- Bayesian network.
rent situation and the information gathered from the As aforementioned, direction-based intention classi-
control and monitoring modules. One significant fication is less precise than the timescale-based
advantage of the cognitive driver model is that incorpo- methods, as the driving maneuvers can be very complex
ration of the built-in features helps mimic human and may contain multiple short-stage actions. For
abilities. example, the lane change maneuvers can consist of
short-period acceleration and turn, and it is less accurate
Direction-Based Driver Intention to describe the intention as merely longitudinal or
Classification lateral.
The direction-based driver intention classification, on
the other hand, is quite straightforward. There are two Task-Based Driver Intention Classification
basic directions for the underground vehicle, which Driver tactical maneuvers consist of a series of opera-
are the longitudinal and lateral intentions. The driver’s tional maneuvers. Some of the existing studies focus
longitudinal behavior includes braking, acceleration, on the analysis of multiple tactics rather than a single
starting, lane keeping, etc. Lateral behaviors usually tactical task. The multitask-based model usually con-
contain turning, lane changing, and merging. In previ- tains both longitudinal and lateral maneuvers
ous studies, most researchers pay attention to the lateral compared with the single-task-oriented model. By using
intention prediction such as the lane change, turning, machine learning theory, a single intention inference
and overtaking maneuvers. The lateral intentions are task can be modeled with a discriminative and genera-
more complicated than the longitudinal intention tive model. However, the multitask inference model
owing to the frequent interaction with surrounding prefers to use the generative models such as Bayesian
vehicles. networks and HMM [31]. Oliver and Pentland [43] pro-
In terms of the longitudinal intention, most of the posed a driver behavior recognition and prediction
previous studies focus on braking intention recogni- method based on the dynamic graphic models. Seven
tion. Haufe et al. proposed a driver braking intention driver maneuvers, namely, passing, changing right and
prediction method using electroencephalographic left, turning right and left, starting, and stopping were
(EEG) and electromyographic (EMG) signals [34,35]. analyzed. Liebner inferred the driver’s intent based on
Khaliliardali [36] proposed a driver intention predic- an explicit model for the vehicle’s velocity, and an intel-
tion model to determine whether the driver wants to ligent driver model was used to represent the car-
go ahead or stop. The method was to classify the go following and turning behaviors [44]. Liu and Pentland
and stop intention based on the brain-machine inter- [45] aimed to analyze the patterns within a driving ac-
face. The EEG, EMG, and electrooculographic signals tion sequence. The primary approach was to model
from six subjects in the simulation environment were the human behaviors in a Markov process. Imamura
collected and two classification methods (linear and developed a driver intention identification and inten-
quadratic discriminant analyses) were used separately tion labeling method based on the assumption of
to evaluate the classification performance. McCall and compliance with traffic rules [46].
Trivedi [37] integrated driver intention into an intelli- As can be seen, multitask driving intention recogni-
gent braking assistance system. A sparse Bayesian tion is more complicated than the single-task model.
learning algorithm was used to infer the driver’s It must define and clarify several driving maneuvers
intention of braking. Trivedi et al. predicted the driver’s and need more experiments. Moreover, it relies on the
braking intention by directly monitoring the foot design of generative inference algorithms.
30 Advanced Driver Intention Inference

DRIVER INTENTION INFERENCE contains three parts, namely, the traffic context percep-
METHODOLOGIES tion unit, level 2 tactical intention unit, and level 3 con-
The Architecture of Driver Intention trol units. Specifically, in the third level, three layers are
Inference System defined. The upper layer is driver dynamics, which rep-
DII system requires multiple techniques such as vision- resents the checking and monitoring behavior of the
based perception system, data fusion and synchroniza- driver. The second interface layer will be activated
tion, and model training based on machine learning once the lane change decision is made. Finally, the con-
methods. According to the previous literature, LCII sys- trol signals are fed into the lowest vehicle control layer.
tem mainly contains the following modules: road and In the driver dynamics module, the most common dy-
traffic perception module, vehicle dynamic measure- namics are brain dynamics, which can be measured by
ment module, driver behavior recognition module, EEG, eye gaze behavior, head movement, and body
and DII module, as shown in the bounding box in movement (includes hand, body, and foot dynamics).
Fig. 2.7. From Fig. 2.9, we can clearly define the time flow of
As shown in Fig. 2.7, the traffic information is first the driver intention procedure. The driver first captures
captured by the environment perception block. Similar the traffic context and then generates the corresponding
to the previous literature that uses cameras, light detec- intention according to the traffic. Next, the driver will
tion and ranging (lidar), radar, and GPS signals to check the surrounding traffic by performing a series of
detect the surrounding traffic situation, this block will checking behaviors to make sure safety control. Once
process the current road and traffic information with the driver is confident with the lane change decision,
machine vision techniques and output the position of he/she will control the vehicle through the steering
ego-vehicle and the velocity. After the traffic context is wheel and the pedal. Finally, the vehicle responds to
detected, the relative distance and velocity between relative control behaviors and vehicle dynamic changes.
the ego-vehicle and the front vehicles can be obtained
by collecting the vehicle status information captured Inputs for Driver Intention Inference System
through the Controller Area Network (CAN) bus. The driver is in the center of the traffic-driver-vehicle
The traffic and vehicle data will be fed into the inten- (TDV) loop. The signals from the three parts of the
tion inference model, along with the driver behavior TDV loop can be used to infer human driver intentions.
signals. The driver behavioral signals include the driver The major signals that can be used as the inputs of LCII
head rotation, eye gaze, body movement, etc. Next, the system are summarized in Table 2.1. The perception of
intention inference model will calculate the probability traffic context enables the ADAS to understand the spe-
of a lane change intention based on the fused informa- cific driving environment of the human driver so that a
tion. Once the decision is made, the lane change deci- more reasonable intention inference can be detected.
sion module outputs a binary signal to indicate a lane The driver behavior information will help determine
change maneuver. After the lane change decision is acti- how long the driver has generated the intention and
vated, the interaction module models the driver’s hand how the driver checks the surrounding traffic with
and foot dynamics and the interaction with the vehicle respect to the corresponding intention. Finally, the
control interface. A taxonomy of driver intention mod- vehicle dynamic signals indicate what kind of driving
ules is depicted in Fig. 2.8. action the driver has taken to realize the intent.
The relationship between tactical intention and The input data used to infer driver mental intent and
operational intention is illustrated in Fig. 2.9. Fig. 2.9 predict driver behavior is multimodal signals. It can

FIG. 2.7 Driver lane change intention inference framework.


CHAPTER 2 State of the Art of Driver Lane Change Intention Inference 31

FIG. 2.8 Taxonomy of driver intention systems. ECG, electrocardiography; EEG, electroencephalography.

Context
Task M
Traffic
Lane Change Intention
Information
Gathering
L2 Tactical Task
Generating
Action
Sequences

Driver Dynamics
Eye Head Body
EEG
Behavior Movement Movement

Driver Vehicle Hand/Steering Foot/


Interaction Brake&Acceleration
Interface

Vehicle Dynamics
Longitudinal & Lateral
Yaw angle Path & Position
Velocity

L3 Control Task
Relationship Between Level 2 and Level 3

FIG. 2.9 Relationship between tactical intention and operational intention with respect to multimodal inputs.
EEG, electroencephalography.

contain data from the vehicle CAN bus, radar, and lidar intentions, different input data should be selected. For
system, as well as the inner driver monitoring system. example, the steering wheel angle can be a significant
Basically, there are three main kinds of sources to collect signal to reflect the vehicle lateral behavior, which is
the input data of intention inference system, which are also helpful in the prediction of driver lane change
traffic environment, vehicle parameters, and driver intention. However, in terms of longitudinal intention,
behavior information. Besides, to infer different a steering wheel angle can give limited information and
32 Advanced Driver Intention Inference

Adaptive cruise control (ACC) is a cruise system that


TABLE 2.1
can be divided into laser based and radar based. The
Common Input Signals and Sensors Used for
relevant distance between the host vehicle and the front
Driver Intention Inference.
vehicles can be detected through the radar of the ACC
Sensor system [48]. The SWA system uses at least two radars
Sources Sensor Categories that are mounted under the side mirrors to monitor
Traffic Current ego-vehicle position the rear and side vehicles. The scan area can be up to
(collected with GPS and digital map), 50 m behind the vehicle [49,50]. In Ref. [50], the au-
relative distance, velocity, and thors evaluated the impact of different sensors on the
acceleration with respect to the front prediction of driver intention. GPS, a space-based navi-
and surrounding vehicles (collected gation system, gives rough information about vehicle
with cameras, radar, or lidar) location, road type, and road geometry compared to
Vehicle CAN bus signals (including steering the abovementioned sensors. Berndt designed an
wheel angle, steering wheel velocity, HMM-based intention classifier that extracted the dis-
brake/gas pedal position, velocity, tance to the next turn, the street curvature, and street
heading angle, etc.) type from a digital map [51]. Rafael introduced an inter-
Driver Cameras (head rotation, gaze active multiple-modelebased approach to predict lane
direction, foot dynamics). EEG, EMG, change maneuvers on the highway. The system used
heart rate, etc. the GPS/IMU (inertial measurement unit) sensors to
collect the vehicle position data [52]. The advantage
CAN, Controller Area Network; EEG, electroencephalography; EMG,
electromyography; lidar, light detection and ranging. of the GPS system is that it is able to give the location
and time information in rough weather conditions
when the camera and radar system cannot work. The
the brake pedal will play an important role at this precision of GPS is usually between 3 and 10 m. How-
moment. Selecting the most important and relevant ever, it can give location and time information in all
data can increase the accuracy of the prediction rate weather conditions even in a dusty situation when the
and decrease the false alarm rate of the system. For camera and radar system cannot work. Fig. 2.10 illus-
the road-vehicle-driver system, the three parts construct trates the detection regions given by different context
a whole driving model. Each part plays its unique role sensors. In Ref. [53], McCall and Trivedi proposed a pre-
in the design of ADAS and intention inference system. liminary work for the study of a driver-centered assis-
To make the vehicle better understand the driver and tance system, which focuses on the lane change
to develop an intelligent vehicle toward automated events. A modular scalable architecture was provided
driving, information from each part should be as to capture the driver behaviors and the surrounding
detailed as possible. environment. Radar and video devices are used to
obtain the forward, rear, and side information. Mean-
Traffic context while, inner cameras were also used to monitor the
Traffic context is the major stimuli for driver intention. driver foot gesture and head movement. The work in
Traffic environment, also known as the road situation, Ref. [54] concentrated on the lane change intention pre-
is the source of stimuli for drivers to take actions during diction according to the sensor data. The data contains
their driving. A better understanding of the surrounding the lane information given by a lane tracker, the vehicle
traffic information will improve the performance of the velocity, lateral position and its derivation, and the
intention inference system. For example, lane changing steering wheel angle.
usually occurs when there is a lower vehicle in front of
the host vehicle. To keep comfortable, the driver always Vehicle dynamics
chooses to make a lane change. Nowadays, most of the Vehicle status information such as steering wheel angle,
traffic information can be obtained from ADAS. brake pedal position, and velocity can be viewed as the
There are many kinds of sensors that can be used to direct response to the control actions. This information re-
capture the surrounding traffic context, such as cameras, flects the driver’s control actions that are taken on the
radar, and lidar systems. The most popular vision-based vehicle. Hence, these signals have been widely used in
ADAS are LDW and LKA. Vision-based LDW is able to many works for driver intention identification. Vehicle
compute the distance between the host vehicle and data can be collected from the vehicle CAN bus, such as
the lane boundary, the vehicle lateral velocity and accel- velocity, acceleration, and heading angle, which can
eration, yaw angle, and road curvature, etc. [47]. reflect the driver control to the vehicle and can process a
CHAPTER 2 State of the Art of Driver Lane Change Intention Inference 33

FIG. 2.10 Traffic information detected through advanced driver assistance systems. ACC, adaptive cruise
control; SWA, side warning assistance.

large amount of data with high transfer speed. In Ref. [42], the vehicle data gives a limited contribution to the pre-
vehicle speed, acceleration pedal, and brake pedal posi- diction of driver intention, it is still an important data
tion are collected to predict the driver braking intention. source that can increase the accuracy of the intention
In terms of the lane change intent, throttle pedal posi- identification and can help predict the intent at an early
tion, brake pressure, cross-acceleration, steering wheel stage after the intended behavior is initiated.
angle, steering wheel angle velocity, yaw rate, and veloc-
ity are collected from CAN bus in Refs. [49,51]. These sig- Driver behaviors
nals are particularly useful to understand the driving With an increasing demand for ADAS, driver behavior
intent after the driver has determined to finish the inten- and driver state analysis is getting more and more pop-
tion. Schmidt and Beggiato proposed a lane change ular. Unlike the CAN bus data, driver behavioral signals,
intention recognition method based on the construction such as head and eye movement, can give an early clue
of an explicit mathematic model of the steering wheel about the driver’s intention. Many studies have evalu-
[55]. In Ref. [56], the authors proposed a driver lane ated the impact of head/eye movement on intention
change/keep intention inference method on a driving prediction. In Ref. [59], the authors applied the pupil
simulator. The input signals contained the steering wheel information as the cognitive signals for the lane change
angle, acceleration, and the relative speed of the front intent prediction. Normally, driver eye movement can
vehicle. In Ref. [57], the speed, transmission position, be classified as intention guided and nonintention
steering angle velocity, steering angle, acceleration pedal guided. Intention-guided eye movement means the
position, and lateral acceleration data were captured on a eye fixation or saccades were done on purpose, whereas
simulator for LCII. The authors in Ref. [58] proposed a the nonintention-based movement may be due to dis-
driver intention recognition method based on artificial tractions. Driver visual fixation will no longer follow
neural networks (ANNs). CAN bus data and driver gaze the driver’s attention when the driver is distracted. At
information were collected and fed into the intention this moment, eye movement cannot reflect the driver’s
model. The experiment was designed in a six-degrees- mental purpose and the intention prediction result
of-freedom dynamic driving simulator and a total of cannot be trusted [61]. In terms of intention-oriented
284 lane change and lane keeping instances were eye tracking, it can be viewed as a cognitive progress
recorded. of information gathering, which can reflect the driver’s
Although vehicle dynamic data reflect the response mental state earlier than the vehicle parameters. A
to the driving actions, it gives delayed information CAN bus dataebased driver behavior recognition
compared to the driver behavior and traffic context in- belongs to the cognitive process of action execution.
formation with respect to the intention inference. In addition, the driver’s intention at an information-
Vehicle data can reflect the driver’s intention only after gathering step is less likely to change when compared
the maneuver has been initiated. However, the vehicle with that in the action execution step [62]. Although
dynamic statusebased mathematic model has limited head/eye movement can be caused by distraction,
ability in the prediction of driver intention. Although most of the time the driver shifts the eye gaze on
34 Advanced Driver Intention Inference

purpose, which makes eye movement a useful signal for Moreover, glass, lightness, and even hairs near the eye
intention decoding and inference [63]. Doshi and Triv- can influence eye-tracking performance. According to
edi [64] evaluated the relationship between gaze these challenges, some robust algorithms for eye move-
pattern, surrounding traffic information, and driver ment detection have been proposed [68e70]. Lethaus
intention. They proposed an intention recognition et al. [71] evaluated how early the gaze information
method based on reasoning eye movement first. They can reflect the driver intent, and how many gaze features
tried to find out whether an eye movement should attri- can be used to infer the intent. In their experiment,
bute to a specific goal or irrelevant stimuli before using vehicle data was collected in a driving simulator and
eye movement to infer driver intention by fusing eye the eye data was captured by the SMI eye-tracking sys-
gaze with a saliency map. Many prior researchers have tem. Five viewing zones are defined, and ANN is used
focused on applying the eye-tracking technique to pre- to predict driver intention. They finally concluded that
dict driver intention [65e67,100,128,129], and it has a 10-s window for the eye gaze data is enough for inten-
been proved that eye movement information does tion prediction, whereas a 5-s window gives better per-
improve the intention prediction accuracy and decrease formance because the 10-s window carries more noise.
the false alarm rate. Besides, by using eye movement in- Besides, an efficient distance of 3.5e5 s before the ma-
formation, normally a driver’s intention can be recog- neuver occurs for classifying lane change left and
nized much earlier than by using vehicle parameters 2.5e3 s for lane change right detection was observed.
only. The eye-tracking system can be classified into Similar to gaze direction, head motion is another
intrusive glass type and nonintrusive camera-based sys- cognitive process for information gathering. Both eye
tem (Fig. 2.11). and head movement can be seen as a cognitive process
Eye movement is detected through cameras of information gathering. It has been proved in Ref. [94]
mounted on the vehicle dashboard or glass-type driver that head movement is a more important factor than
eye-tracking system. Still many researchers have devel- eye movement for driver intention prediction and it
oped powerful algorithms for eye movement detection has also been widely used in many past DII research
[68,130]. Many studies have paid attention to the eye- [90e92]. The true-positive rate (TRP) of using head
tracking techniques to predict driver intention movement, along with lane and vehicle information,
[65e67]. It has been proved that the eye movement in- can achieve 79.5%, whereas with only use lane and
formation does improve the intention prediction accu- vehicle information the rate is only 50%. Another inter-
racy and helps decrease the false alarm rate. A significant esting point the authors found out was that head move-
challenge to eye movement detection is eye tracking. ment information plays a more important role than eye
The eye movement is normally detected with cameras movement information in the prediction of driver
mounted on the dashboard or the wearable eye- intention. Further research is done in Ref. [94], in which
tracking system. Owing to the physical characteristics the authors claim that both eye and head movement are
of the eye (small scale and occlusion, etc.), it is not useful data for the detection of driver distraction, atten-
easy to detect the eye and track the pupil robustly. tion, and mental state inference. However, there is a

FIG. 2.11 Two different kinds of eye-tracking devices are available on the market. (A) An intrusive eye
tracking glass by Tobii and (B) a nonintrusive camera.
CHAPTER 2 State of the Art of Driver Lane Change Intention Inference 35

difference between eye and head movement in the clas- whether the ongoing driver behavior is goal oriented
sification of driver mental state. Specifically, when the or stimulus based. In addition to camera-based driver
driver is executing a mental goal-oriented task the eye- and head-tracking systems, some other driver be-
head moves early, whereas the eye stays or moves later. haviors such as the foot, hand, and body gestures
On the other hand, when outer stimuli occur, the were also recorded through a camera in some research
driver’s eye will shift first and the head moves later. [38,78,131]. A carefully selected driver behavior feature
This is an interesting research because it offers a way will improve the performance of the driver intent
to determine whether the ongoing driver behavior is predictor.
goal oriented or stimulus based.
Murphy and Trivedi [74] concluded that a variety of Electroencephalography
head pose estimation algorithm can be used to track EEG is a brain action measuring device that measures
head movement. They concluded that several head the flow of brain electric current with noninvasive elec-
pose estimation algorithms can be used to track head trodes on the scalp. It has been widely used in cognitive
movement. However, the in-vehicle head-tracking sys- neuroscience for the study of brain activities. Mean-
tem has its own problems. One significant challenge while, EEG is also an important sensor for brain-
is the online computing ability of the onboard proces- computer interface (BCI) design. EEG is sensitive to a
sor. Driver’s eye and head movements are usually small change in electric activities, which is suitable to
sampled at 30e60 Hz; for a multiple camera system, a detect a human mental state. Therefore many re-
large amount of data will influence the precision of searchers have studied the impact of introducing EEG
the driver status recognition unit. In Ref. [75], a process- on the detection of driver mental activity. A direct appli-
ing method for in-vehicle driver eye-/head-tracking sys- cation of EEG to on-vehicle BCI design is driver work-
tem was proposed aiming at handling the natural eye-/ load detection. EEG has been widely used to monitor
head-tracking system for driver distraction detection. the driver workload and other status such as drowsi-
The data processing method was able to improve the ness, happiness, sadness, mental fatigue, and abnormal
sensitivity and specificity of the eye-tracking system by conditions [77]. However, a drawback of the EEG signal
10%. Another challenge is the noise issue. Vibration is it usually contains various noise, hard to acquire, and
from the road and lightness variation issues exist in gets weak when sampled with poor quality [79]. This is
the vehicle cabinet, which brings a large amount of because the brain electric current is detected with a
noise to the captured images. Sometimes head- noninvasive method, in which the signal has to cross
tracking data will get lost because of the algorithm the brain layers, scalp, and skull. Therefore EEG
problems and head tracking usually needs reinitializing recording systems usually contain electrodes, amplifier
[76]. Besides, according to the current head-tracking al- units, A/D converter, and noise filter units. EEG signals
gorithms, those developed with a monocular camera consist of high-dimensional data with large noise,
show worse performance than the multicameras-based which leads to the use of a machine learning method
algorithms. However, using multiple cameras in the as the main solution when dealing with EEG signals.
vehicle will increase the system cost. Machine learning algorithms can be used in the feature
In Ref. [60], the authors proposed a driver head- extraction and classification unit [132]. The general EEG
tracking system for driver LCII. The head-tracking sys- processing procedure is shown in Fig. 2.12.
tem consists of six cameras. Head movement was In addition, EEG has been proved not only suitable
regarded as a more important factor than eye gaze for for driver attention and status detection but also to be
driver intention prediction [66]. Head movement was able to predict driver intention. Haufe et al. [35] have
widely adopted for DII [66,72,73]. In Ref. [66], the au- used EEG, for the first time, to design a braking assis-
thors claimed that both eye and head movements are tance system. As EEG measures brain activity, it can be
useful data for the detection of driver distraction, atten- much faster than human muscle reaction. If the braking
tion, and mental state inference. However, there is a dif- assistant system can recognize braking intention by EEG
ference between the eye and head movement for the rather than by detecting the brake pedal, it will save ex-
classification of driver mental state. Specifically, the tra time for the driver and improve the driving safety. It
head moves earlier than the eye when the driver is is thus finally proved that, by using EEG, the system can
executing a mental goal-oriented task. On the other detect a brake intention 130 ms faster than a brake
hand, when the outer stimuli occur, the driver’s eye pedal measurement only method [35]. EEG has also
will shift first and the head moves later. This is an inter- been used in driver steering intention prediction [82],
esting conclusion because it offers a way to determine with different machine learning-based classification
36 Advanced Driver Intention Inference

The impact of EEG signals on the intended inference is


worthy to be examined.
Table 2.2 illustrates previous research that used EEG
to detect driver intentions. Most of the research focus on
using EEG to recognize the driver’s operational intent
such as braking and steering. One concern is that in
the real-world driving environment, EEG signals
contain lots of noise and can be affected by head move-
ment, which makes it difficult to use. However, EEG sig-
nals are still a good measurement of driver statuses such
as attention and fatigue.

ALGORITHMS FOR DRIVER INTENTION


INFERENCE
Conventional intention inference algorithms can be
roughly divided into the following groups: mathematic
model, driver cognitive model, and the widely used ma-
chine learning models. Owing to their ability to deal
FIG. 2.12 General algorithm for electroencephalographic with a high-dimensional feature vector, machine
(EEG) signal processing [79]. learning methods are widely accepted by the LCII sys-
tem. As mentioned in Ref. [49], the intelligent vehicle
adopted 200 sensor signals. At this moment, machine
learning methods are the most suitable tools to fuse
methods, and the prediction accuracy of steering inten- the signals and construct the LCII system. The machine
tion can achieve 65%e80%. A general process of EEG learning algorithms can be divided into generative
signal processing is shown in Fig. 2.11. As shown in model and discriminative model. The generative model
prior studies, most of the EEG-based BCI is designed provides a joint probability distribution over the
to predict driver’s instant or operational level intention. observed and target values, which can generate both

TABLE 2.2
Performance of Previous Driver Intention Research Using EEG Signal.
Reference Target Inputs Algorithms Numa TRP
[102] Steering EEG Gaussian distribution 3 80 (aver)
[133] Steering EEG, steer angle, brake and Hierarchic Bayesian and 8 70 (aver)
gas pedal logistic regression
[35] Steering EEG Bayesian decision theory 5 65%
[134] Braking EEG, EMG, gas and brake Regularized linear 20 Varying with
pedal discriminant analysis time
[81] Braking EEG, EMG, brake and gas Regularized linear 15 70e90
pedal discriminant analysis
[36] Stop and go EEG, EMG, EOG, Linear and quadratic 6 0.76  0.12
steering,brake and gas pedal discriminant analysis
[135] Turning EEG, steer angle, brake and Linear discriminant analysis 7 0.69  0.16
gas pedal
[136] Alertness, EEG, driving behavior A wide variety of 6 80%e90%
drowsiness classification algorithms

EEG, electroencephalography; EMG, electromyography; EOG, electrooculography; TRP, true-positive rate.


a
Num is short for number, which indicates the number of participants in the experiment.
CHAPTER 2 State of the Art of Driver Lane Change Intention Inference 37

the inputs and outputs according to some learned hid- to identify the left and right lane change maneuvers at
den states. However, the generative model is less easy to an early stage after the lane change was started. Final re-
be trained compared with the discriminative model, as sults showed 71% and 74% recognition accuracy for the
it may require training multiple models and providing a left and right lane changes, respectively. In Ref. [85], a
model to each class based on the different probability new feature named comprehensive decision index
distribution. (CDI) was introduced.
On the contrary, the discriminative model only pro- Fuzzy logic was applied to represent the surrounding
vides the dependence of the target on the observed data. environment and the lane changing willingness of the
Discriminative models usually can be generated from driver. The overall performances of the algorithm with
the specific generative models through the Bayes rule. different input parameters (a situation, situation and
The most popular generative models include Bayesian vehicle, and situation, vehicle, and CDI) were analyzed.
networks, HMMs, Gaussian mixture models, etc., and Li et al. [86] proposed an integrated intention inference
the widely used discriminative models are support vec- algorithm based on HMM and Bayesian filter (BF) tech-
tor machine (SVM), neural networks, and linear regres- nique. A preliminary output from the HMM was further
sion. As mentioned in Ref. [31], discriminative models filtered using the BF method to make the final decision.
provide a better result on a single target problem than The HMM-BF framework achieved a recognition accu-
the generative models, whereas the generative models racy of 93.5% and 90.3% for the right and left lane
are more suitable for multitarget problems. Instead of changes, respectively. In Ref. [56], the authors proposed
these two typical methods, driver intention can also a driver lane change/keep intention inference method
be modeled based on the cognitive models and the based on a dynamic Bayesian network. A four-step
deep learning models. A taxonomy of the algorithms framework for the DII was developed and the autore-
for intention inference is shown in Fig. 2.13, and in gression was combined with an HMM to take the past
Table 2.3 the LCII results based on some articles are driver behaviors into consideration.
illustrated. In Ref. [57], the authors constructed a lane change
intention recognition method based on the continuous
Generative Model hidden Markov model (CHMM). The authors evaluated
Generative models, such as HMM, are widely used in the CHMM performance with different model struc-
previous LCII studies [55,56,59,66,83,84]. Pentland tures (3, 6, and 9 hidden states). According to the re-
and Liu [45] used the HMM to recognize seven kinds sults, the CHMM with six hidden states and 1.5 s
of driver intention. Berndt used the HMM to investigate window size (data collected between 0 and 1.5 s prior
early lane change intention [51]. The aim of the study is to the vehicle crossing the lane) gave the best

FIG. 2.13 Taxonomy of algorithms for driver intention inference system. ARHMM, autoregressive hidden
Markov model; IOHMM, input-output hidden Markov model; LSTM RNN, long short-term memory recurrent
neural network.
38 Advanced Driver Intention Inference

TABLE 2.3
Summary of Various Previous Lane Change Intention Inference Systems.
No. of Prediction
Reference Signals Algorithm Subjects Environment Performance Horizon
[57] Steering angle, CHMM 10 Simulator 100% 0.5e0.7 s After
steering force, steering
velocity
[66] Lane position, RVM 8 On road 88.51% 3 s Prior to the
CAN bus, eye lane change
and head
movement
[98] Lane position, Sparse 3 On road 90% 3 s Prior to the
CAN, and head Bayesian lane change
movement learning
[65] Eye movement Finish 17 Simulator 77% ─
questionnaire
[99] Eye movement SVM 24 On road 73.13%  1.25% ─
(pupil size Samples
variation)
[51] CAN, digital HMM 50 LCL, 50 On road 71% L, 74% R ─
map LCR
[87] CAN, distance HMM 20 On road 80%e90% ─
between
vehicles
[50] CAN, LDW, RVM 15 On road 91% 1 s Prior to the
ACC, head lane change
movement,
SWA
[43] CAN, head and HMM 70 On road 12.5% LR, 1 s Prior the
eye movement 17.6% LL maneuver
[40] CAN, lane style RVM 108 Lane On road 79.20% ─
and position, changes
head and eye
movement
[67] CAN, eye ANNs, BNs, 10 Simulator 95.8% (ANN), ─
movement NBCs 93.2% (BNs),
90.6% (NBCs)
[83] Steering angle Queuing 14 Simulator LCN 98.61%, ─
network model LCE 91.67%
[85] Steering angle, Fuzzy logic and 4 (69 Simulator 94% 1.67 s After lane
rate, HMM samples) change
comprehensive maneuver starts
decision index
[100] CAN, eye State transition 20 (8576 Simulator 80% ─
movement diagram lane
changes)
[84] Relative position SVM, RF, LR 5 Simulator 89.5% (SVM), ─
and velocity, 88.9% (RF),
heading, time to 87.2% (LR)
collision, time
headway
(continued)
CHAPTER 2 State of the Art of Driver Lane Change Intention Inference 39

TABLE 2.3
Summary of Various Previous Lane Change Intention Inference Systems.dcont'd
No. of Prediction
Reference Signals Algorithm Subjects Environment Performance Horizon
[45] Steering angle HMM 8 Simulator 88.3%  4.4% ─
and velocity,
vehicle velocity
and
acceleration
[49] CAN, ACC, RVM 15 (500 On road 80% 3 s Prior to the
SWA, LDW, samples) lane change
head movement
[101] CAN, GPS, eye Finish 22 On road ─ ─
movement questionnaire
[54] Steering angle SVM and 2 (139 On road 80% 1.3 s Prior to the
and relative lane Bayesian filter samples) lane change
position
[58] CAN, eye ANN 10 Simulator 95% (L), 85% (R) ─
movement
[102] CAN, lidar, Latent dynamic 1000 On road 90% 2 s Prior to the
radar, hand, conditional Samples lane change
head, and foot random field
dynamics
[103] CAN SVM and BN 4 Simulator 95% (LK), ─
80% (LC)
[33] CAN, eye Computational 11 Simulator 90% 1 s After
movement model based on steering
ACT-R
[86] CAN bus CHMM and 188 LCL, On road 93.5% (L), 0.5e0.7 s After
Bayesian filter 212 LCR, 90.3% (R) steering
242 LK
[94] GPS, digital LSTM-RNN 10 Drivers, On road 90.5% 3.5 s Prior to the
map, head 1180 miles lane change
movement,
CAN bus

ACC, adaptive cruise control; ACT-R, Adaptive Control of Thought-Rational; ANN, artificial neural network; BNs, Bayesian networks; CAN,
Controller Area Network; CHMM, continuous hidden Markov model; HMM, hidden Markov model; LC, lane changing; LCL, lane change left; LCR,
lane change right; LDW, lane departure warning; LK, lane keeping; LR, logistic regression; LSTM-RNN, long short-term memory-recurrent neural
network; NBCs, Naïve Bayesian classifiers; RF, random forest; RVM, relevance vector machine; SVM, support vector machine; SWA, side warning
assistance.

classification result (95.48%). The authors also From the comparison of the four scenarios, the author
concluded that the most important factors for lane pointed out that the classification performance did
change intention recognition were data representation, not show a significant increase with the additional
size of the sliding window, and the initial sets of model context information. However, they showed that the
parameters. In Ref. [87], the authors proposed a additional context information leads to a high false-
context-based highway lane change intention system positive result rate and the system performance was
based on HMM. Four different inference systems were worse than the system with vehicle state information
defined, which were vehicle state model only, front only. One possible explanation is that the HMM has
vehicle distance-based model, rear vehicle distance- limited ability to capture the context information dur-
based model, and front and rear distance-based model. ing the lane change process. Therefore a more powerful
Another random document with
no related content on Scribd:
prohibition against accepting unsolicited donations from donors in
such states who approach us with offers to donate.

International donations are gratefully accepted, but we cannot make


any statements concerning tax treatment of donations received from
outside the United States. U.S. laws alone swamp our small staff.

Please check the Project Gutenberg web pages for current donation
methods and addresses. Donations are accepted in a number of
other ways including checks, online payments and credit card
donations. To donate, please visit: www.gutenberg.org/donate.

Section 5. General Information About Project


Gutenberg™ electronic works
Professor Michael S. Hart was the originator of the Project
Gutenberg™ concept of a library of electronic works that could be
freely shared with anyone. For forty years, he produced and
distributed Project Gutenberg™ eBooks with only a loose network of
volunteer support.

Project Gutenberg™ eBooks are often created from several printed


editions, all of which are confirmed as not protected by copyright in
the U.S. unless a copyright notice is included. Thus, we do not
necessarily keep eBooks in compliance with any particular paper
edition.

Most people start at our website which has the main PG search
facility: www.gutenberg.org.

This website includes information about Project Gutenberg™,


including how to make donations to the Project Gutenberg Literary
Archive Foundation, how to help produce our new eBooks, and how
to subscribe to our email newsletter to hear about new eBooks.

You might also like