Professional Documents
Culture Documents
DOI 10.1007/s10846-011-9610-4
Received: 17 December 2010 / Accepted: 24 May 2011 / Published online: 23 July 2011
© Springer Science+Business Media B.V. 2011
Abstract Robots are increasingly autonomous in they have constraints that limit their availability
our environments, but they still must overcome due to their own work schedules. Interestingly, the
limited sensing, reasoning, and actuating capabil- study further shows that an available human is not
ities while completing services for humans. While always in close proximity to the robot. We present
some work has focused on robots that proactively an extended model that includes the availability of
request help from humans to reduce their limi- humans in the environment, and demonstrate how
tations, the work often assumes that humans are a navigation planner can incorporate this informa-
supervising the robot and always available to help. tion to plan paths that increase the likelihood that
In this work, we instead investigate the feasibility a robot can find an available helper when it needs
of asking for help from humans in the environ- one. Finally, we discuss further opportunities for
ment who benefit from its services. Unlike other the robot to adapt and learn from the occupants
human helpers that constantly monitor a robot’s over time.
progress, humans in the environment are not su-
pervisors and a robot must proactively navigate to Keywords Human–robot interaction ·
them to receive help. We contribute a study that User study · Asking for help · Planning
shows that several of our environment occupants
are willing to help our robot, but, as expected, Mathematics Subject Classifications (2010)
68T40 · 68T37
or recognizing obstacles to avoid while navigating, any given time and a robot will have to proactively
and they may not have the physical ability to navigate to seek help if it needs it. Based on these
manipulate objects. findings, we contribute a model of an environ-
To overcome these limitations, some work has ment, robot, and occupants in it and demonstrate
focused on reasoning about a robot’s current state how a navigation planner can plan paths around
and proactively requesting humans to direct the available occupants’ known locations to increase
robot’s actions during tasks [2, 16, 29, 39]. How- the likelihood that a robot can find help quickly
ever, there is always an assumption that humans when it needs it, rather than assuming there is
are available to help a robot in any location, always a human available. Finally, we discuss sev-
whether the help comes from supervisors who eral qualitative results of our study which indi-
assign tasks and have incentive for it to perform cate the need for further extending the models of
[2, 16], or from bystanders who do not benefit occupants.
from the robot’s tasks [39]. Additionally, the work
has been limited to asking for only one kind of
help (e.g., localization) at a time.
2 Related Work
In this work, we instead focus on asking for sev-
eral types of localization and manipulation help
Robots have limitations that affect their task per-
from the actual occupants of the environment and
formance. Human supervisors have traditionally
beneficiaries of a robot’s services. We argue that
been used to overcome a robot’s limitations by
robots that ask for help from occupants combine
monitoring robot progress and intervening when
the benefits of asking bystanders and supervisors
necessary, but such help can be very expensive in
in the following ways:
terms of monitoring time and cognitive load on
the helper [40]. Much recent work has focused
– Occupants should answer robots’ questions on different techniques to allow robots to reason
during tasks for other occupants with the in- about their own limitations and capabilities to
centive that the help will be reciprocated dur- proactively ask for help from supervisors, teach-
ing tasks for them. ers, and bystanders in the environment. These
– The burden of help is distributed among many different types of help have been classified in
occupants in the environment. a variety of ways [17, 20, 26, 31], but we focus
in particular on supervisors and bystanders and
However, unlike supervisors and bystanders: investigate a new classification of helpers—the in-
environment beneficiaries of robots’ tasks (i.e.,
– Occupants are not always available to help the the robot users [23, 29]).
robot.
– Occupants are spatially-situated and cannot 2.1 Supervisors
help the robot at every location in the envi-
ronment. Many different methods have been proposed for
supervisors to help including providing assistance
We first review our building environment, ro- at different levels of granularity depending on
bot, and its limitations and define several types of the robot’s capabilities [12] and participating in
help that a robot could request from occupants to mixed-initiative interactions to ensure the robot
overcome those limitations. We, then, contribute performs its task effectively [33]. Initial work in
a study that shows that, despite our busy office robot learning required teacher supervisors to
environment, several occupants are willing to help physically perform the task for the robot to im-
our robot with its limitations, but they have work itate [19, 24, 28]. Since then, teachers have been
constraints that limit their availability. Addition- able to accurately label data both when the robot
ally, there are few available occupants in our en- requests it during tasks and through corrective
vironment (and likely in many environments) at feedback after tasks are performed in order for
J Intell Robot Syst (2012) 66:205–221 207
robots to learn policies for which actions to take that bystanders in the environment can help
[2]. a robot.
Two important assumptions are made in terms
of robots employing help from supervisors: super-
2.2 Bystanders
visors have incentives to respond and they partic-
ipate in long-term interactions giving both robot
The number of bystanders is assumed to be high
and human chances to learn about each other.
compared to supervisors, distributing the burden
of help across a large number of people [4, 21, 25,
Incentives It is assumed that teachers and super-
39]. Robots can benefit from the distributed help,
visors have incentives to help their robots com-
but the requirement of a crowded environment
plete their tasks. They are either paid to monitor
requires spatial constraints on where these robots
the robot’s movements or they do so for research
can be deployed.
and teaching purposes that further the robot’s
knowledge. Additionally, while it is assumed that
teachers will provide less feedback over time, they Distributed Help Human computation (e.g., [37,
are still available to provide corrective feedback 38]) and crowd-sourcing (on websites like Ama-
or wait for questions during the entire learning zon.com’s Mechanical Turk) gather help from
process. As a result, the robot does not need a distributed set of people. Crowd-sourcing has
to take into account the number of times or been used to label pictures [37] and translate au-
frequency that it has needed help to determine dio [6, 34]. In robot domains, bystanders in busy
whether it should ask for help at the present time. environments have helped robots complete tasks
This is in contrast to other active learning methods in locations as varied as offices [4], conferences
which assume that there is a cost, such as time [25], and even on the street [39]. Because the num-
to respond or annoyance, associated with asking ber of people in these areas is so high, there is a
[1, 8, 27]. very limited possibility that any particular person
will be asked for help too frequently, limiting the
Long-Term Interaction Because supervision oc- annoyance and interruption of the requests for
curs over time, there are additional opportunities help.
for robots to take advantage of the long-term
interactions to model and learn about the human
Spatial Constraints Bystanders are more often
and vice versa. Supervisors are often assumed,
found in crowded spaces, adding spatial con-
with few exceptions (e.g., [16]), to have in-depth
straints to the robot’s deployment. While humans
knowledge about how robots work so that they
are not actively in contact with the robot all the
can help them appropriately. However, even with-
time, there is still an assumption that at least one
out this assumption, robots with knowledge of
human will help the robot shortly after it requests
supervisors should (1) take into account helper
it. Because these robots do not model humans in
expertise to determine the type of question that
the environment in terms of who will be available
the robot should ask [5, 15, 16], (2) ground or
or where they are located, nor do they proactively
familiarize the helper with the robot’s current
contact a known helper, they have little control
state to increase the likelihood of accurate re-
over the help they receive and cannot plan to
sponses [7, 30], and (3) model the helper’s in-
optimize their performance using that help. As
terruptibility or availability to answer questions
a result, their performance is dependent on the
[14, 35].
crowds in the environment.
While supervision ensures that a robot will
always be able to find help from a trusted hu-
man helper, it is an expensive requirement that 2.3 Our Work
cannot be scaled as we continue to deploy more
and more robots. To reduce the dependency on a We argue that asking for help from occupants
single supervisor, it has recently been suggested in the environment combines the benefits of
208 J Intell Robot Syst (2012) 66:205–221
supervisors in terms of incentives to answer and each human’s availability through long-term in-
long-term interactions and the benefits of by- teraction to avoid navigating to empty offices or
standers to distribute help among many peo- interrupting meetings in offices. We discuss each
ple. Additionally, seeking help from occupants in of the aspects of building occupants in turn and
offices relaxes the spatial constraints of crowded the research questions we address.
areas but requires proactive navigation to find
available occupants. We expect that because the Distributed Help Because there are many occu-
office occupants are the beneficiaries of the ro- pants in a building, the burden of helping a ro-
bot’s tasks, they will be likely to respond to bot is distributed among them. This is important
requests for help to complete tasks. However, because occupants are busy and may not be able
there has been little prior work focused on the to answer too frequently. However, it is unclear
opportunities for robots to ask for help from these whether occupants are willing to help at all given
occupants (e.g., users of the robot [23, 29]). that they have other work to do. This work will ad-
In this work, we first define the problem of dress the following questions related to occupant
seeking help from building occupants. Then, we help:
review our environment, robot, and our robot’s
– Are office occupants willing and available to
limitations that it requires human help with. We
help a robot perform its tasks?
present a study to understand the willingness and
– Are they willing to provide one type of help
availability of occupants to help a robot that per-
more than others (e.g., are they unwilling to
forms services for them. We find that few people
help with tasks that require them to leave their
are willing to help at any given time, and in order
office)?
for the robot to find these people during tasks,
– Are office occupants willing to help the robot
it will need to model where they are located and
even if they are busy in meetings or on the
modify its navigational plans to move near them.
phone?
We then contribute a model of the environment
that includes not only the robot and its robot
Incentives Because a robot would know occu-
limitations, but also available humans and demon-
pants’ office locations, it could provide services
strate that a robot can use this model to optimize
or incentives to encourage occupants to help it.
its navigation for a task to increase the likelihood
If occupants want to continue to receive services
of finding an available occupant.
and incentives from the robot (e.g., mail delivery),
they must also agree to help the robot at some
times. In this work, we aim to answer the following
3 Seeking Help from Spatially-Situated
questions relating to incentives:
Building Occupants
– Do office occupants report that they are moti-
We define occupants of buildings as having pre- vated to answer questions with incentives?
defined work spaces and conducting work which – Do they actually answer more frequently
requires that they be present over a period of when offered an incentive?
time. While occupants have similarities to both
supervisors and bystanders, they also have con- Long-Term Interaction Unlike bystanders who
straints which violate the assumptions of previous may only have to answer a particular robot’s ques-
work. In particular, they are spatially-situated in tions once, an office occupant will be asked ques-
the environment so no single occupant can help tions more often including times that may cause
a robot at every location. Additionally, the occu- interruptions to meetings and other work or other
pants are not necessarily monitoring a robot or annoyances. However, unlike supervisors, there
their email, and may miss or be late to respond to are possibly several helpers to choose from. Long-
electronic requests for help. As a result, a robot term interactions allow the robot to learn who is
will need to navigate to an occupant’s office to often available at certain times and preferences
request help. It will also need to learn and model for when and what types of help they are willing
J Intell Robot Syst (2012) 66:205–221 209
to answer. We address the following questions help it if necessary. We will answer the following
towards this interaction: question:
– Is there a novelty effect associated with will- – Can a robot use availability information to
ingness to help the robot and does willingness determine a path that increases its likelihood
to help decrease over time? of finding help when it needs it?
– Can the availability of occupants be modeled
to take into account who will be able to help? Next, we will describe our robot and environ-
– Does occupant availability change through the ment with office occupants that were used to
day and do we need to model this change? answer these questions and to understand the fea-
sibility of asking occupants for help with a variety
Navigation Finally, due to the lack of constant of different robot limitations.
supervision and asynchronicity of email, the robot
must navigate to the spatially-situated occupants
to determine their availability and to ask them for 4 Robot and Environment Domain
help if needed. Current navigational models have
included multiple layers of costs of navigation Our environment consists of one floor of an aca-
around people, such as the uncertainty in the path demic building containing 79 offices: 35 individ-
and the path distance (e.g., [9, 13, 22]), but have ual offices for faculty and administrators and 44
not included costs of who along the path is avail- offices each shared by 2–3 graduate students. Our
able to help. Intuitively, a robot should choose robot, CoBot (Fig. 1a), is capable of autonomous
short paths that also have humans available to localization and navigation in the environment as
well as dialog with humans. It has a laptop with potentially take different amounts of time to an-
the screen facing backwards, away from the di- swer and some require the occupant to leave their
rection of movement, that occupants can use to office to perform the requested help. The ques-
interact with the robot while it performs tasks tions are all spoken out loud for the occupants to
autonomously for them in the building. hear and displayed on the robot’s laptop screen,
However, like many robots today, CoBot has but we require that the occupants answer the
limitations. CoBot has high localization uncer- questions using the visual user interface on the
tainty in large open spaces (Fig. 1b—darker grey laptop as it cannot understand speech.
areas indicate more uncertainty) and also has
difficulty perceiving chairs found in common areas 4.1 Localization Help
resulting in increased navigation time as it at-
tempts to re-localize or avoid these areas entirely. When CoBot requires localization help, it can find
Additionally, CoBot does not have arms or the an open door to ask the occupants to share their
ability to manipulate objects to push chairs out of room number in the following way:
the way or pick up the mail to give to the building
I cannot determine my location. What is
occupants. While the robot can overcome some of
the room number of this office? (multiple
these challenges autonomously, inevitably, it must
choice)
ask for help from humans sometimes to resolve
each of these limitations. After speaking the question, CoBot lists three
In particular, we are interested in occupants’ predictions of the possible office numbers and
willingness to help the robot with these different an additional textbox (in case the three office
limitations when the robot proactively navigates numbers were all incorrect) for the occupant to
to their door to request the help. These requests respond with. Because building occupants should
know their own office number, CoBot’s localiza- We assume that the robot can find an office close-
tion questions should be fast for them to respond by to write the note. In this study, the office
to and do not require the occupants to leave their number of the nearby office (here 7505) changed
offices except for accessing the robot’s backwards depending on the occupants’ location to ensure
facing screen (Fig. 2a). the occupants did not have to walk too far out of
their office to deliver this message—the room was
4.2 Moving Chairs Help on average five offices away from a helper. When
CoBot navigated to offices to request help writing
Our building contains many seating areas with a note, it carried paper and pens in a tote bag for
moveable chairs that are hard for the robot to the occupants to use.
detect. Even if CoBot could detect the chairs, it
has no way of physically moving chairs that are
blocking its path. CoBot navigated to occupants’ 5 Occupant Availability Study
offices and asked them to check and move chairs
in the closest common area so that it could pass: In order to understand the feasibility of asking
My laser range finder cannot determine the occupants for different types of help, we designed
location of chair legs in the common area. a study in which CoBot visited every office on one
Can you please move chairs in the common floor of our building to ask each type of question
area to clear a path for me? at different times of day. We measure the number
of times each occupant is available in their office
While occupants can easily identify and move and willing to help for each question type as well
chairs out of CoBot’s way, this task requires that as the amount of time they spend helping the
participants leave their offices to help the robot robot. Because we were exploring the feasibility
(Fig. 2b). Help with this limitation ensures that of asking different types of questions in this study
CoBot can safely navigate through the environ- and not testing the autonomy of the robot, CoBot
ment, assuming that the occupants actually move was wizard-of-oz’d [18].
the chairs as requested. The occupants are asked
to confirm their action on the laptop user interface
so that the robot knows it is safe to continue. 5.1 Study Design and Procedure
changes by time of day but not depending on the would likely be available for the robot to ask for
type of question. Participants were equally willing help, although at any particular time there are
to help with all types of questions although some likely to be many more occupants that the robot
took much longer than others. could ask to further distribute the help.
In terms of the question types, we found that
Distributed Help In total, 130 doors were open each type of question took a very different amount
out of a combined 702 in nine trials (Fig. 3a— of time to complete but occupants helped equally
darker means the door was open more fre- with them all. Occupants took on average 30.1 sec-
quently). Occupants helped the robot 78 times out onds (s.d. 18.1) to complete localization questions,
of the 130 possible open doors. 46 offices were 55.6 seconds (s.d. 24.6) for chair questions, and
open at least once and 31 of those offices con- 88.3 seconds (s.d. 45.3) for the note writing ques-
tributed responses. Each office was open on av- tions. Despite these differences, when occupants
erage 1.8 (s.d. 1.9) times out of 9 possible and were available to help, they were willing to answer
occupants answered on average 1.1 questions any type of question. We found little difference in
(s.d. 1.7). The high standard deviation for the the response rate for each question type—57.5%
offices indicates that there were a few occupants of the localization questions, 62.5% of the chair
available almost all of the time, while many occu- questions, and 50% of the notes questions. This
pants were unavailable. Seven out of the 78 offices finding indicates that a robot would not need to
contributed 36 of the 78 responses to the robot. reduce the asking frequency of questions that take
This indicates that there is a group of people that longer to answer.
Incentives We found no statistical difference colors representing more frequent help. Interest-
in answering frequency when occupants were ingly, the few frequently available occupants were
offered gifts to when they were not, but some largely distributed around the building - especially
occupants indicated in interviews that the gift did in the areas where the robot has the most un-
affect their decision to help. In particular, we ob- certainty (Fig. 1b). A random selection of seven
served some occupants deliberately opening their offices would not necessarily result in such an even
doors when they heard the robot down the hall so distribution across the building. Every location
that they could help and possibly receive a gift. in the building was at most ten offices from an
While gifts were only offered at random times, occupant that helped the robot during many of
some participants stated in the interview that they the trials, except for the north side of the build-
would be more willing to help the robot if it ing at 2:30. Because the robot is least uncertain
offered candy more often. in the north side of the building, the robot may
be able to navigate despite the low availability.
Long Term Interaction On an average trial, 14.4 However, the distance between available occu-
(s.d. 7.5) doors were open, and occupants in those pants indicates that the robot will have to plan
offices responded to the robot’s requests for help its navigational paths to increase the likelihood of
8.7 times and refused to help or ignored the robot finding available occupants at execution time to
the remaining 5.7 times (Fig. 4). However, there is complete its tasks.
a strong effect of day on the proportion of open To summarize, based on the initial questions we
offices and available helpers. The number of asked about office occupants, we found that some
open office doors dropped each day likely due to office occupants were willing to help the robot,
occupants’ prior knowledge about when the robot although not all were. The number of people who
would be visiting. The number of occupants that did helped the robot shows that it is feasible for a
help the robot remained constant over the three robot to use office occupants as distributed help
days indicating less of a novelty effect for those through the building. Interestingly, this willing-
seven occupants who helped the robot the most. ness to help was not affected by the length of
time the question took to answer nor the incen-
Navigation Figure 5a–c show the frequency of tives the occupants received. CoBot was able to
help from each office by time of day, with darker collect data to begin to model the availability of
different offices throughout the day and could a path or state of a robot (e.g., [9, 13, 22]), little
continue to do this through long-term interac- work has modeled the states and actions that can
tions. Because the occupants that helped CoBot resolve such limitations such as human helpers
were, for the most part, distributed throughout the [29]. Robots that depend on supervisors and by-
environment, CoBot cannot assume that someone standers often assume that their human helpers
will be very close to help it and therefore must are always available (e.g., [3]). In these always-
proactively navigate to those occupants in order available cases, the robot need only determine
to seek help. when it requires help and does not take into ac-
By creating a model of occupants based on count locations of available help to plan paths for
long-term interactions with them, we will show its task.
next that it is possible for the robot to plan paths However, when there is limited and/or proba-
that minimize its own limitations and increase the bilistic availability of human helpers distributed
likelihood that occupants will be available to help through the building, we argue that a robot should
when needed. plan its paths to increase the likelihood of finding
available occupants. The shortest distance path
may result in a longer navigation time if there
6 Proactive Navigation for Help is high uncertainty along that path and no avail-
able occupants to help reduce that uncertainty.
While previous work has taken into account the There are several planning frameworks that could
robot limitations, constraints, and uncertainty of possibly be modified to incorporate help from
216 J Intell Robot Syst (2012) 66:205–221
occupants including traditional A* path planners time to complete these tasks on an edge is infinite.
with added constraints [9, 13] and POMDPs that Our planner considers the expected time to tra-
model human help [3]. verse an edge tij to include the distances and
We, first, contribute a model of the environ- velocity of the robot and its additional costs. With
ment that includes spatially-situated occupants probability nijh , the robot will experience cost cijh on
and their availabilities collected through long- eij. Otherwise, it will not experience any cost.
term interactions. We, then, demonstrate one way
to use this model for path planning, namely adapt- dij h h
tij = + nij ∗ cij (1)
ing the A* algorithm to reason about the ro- s
h∈H
bot’s capabilities and the available humans to help
reduce limitations. In order for a robot in the environment to
perform all of its tasks, we assume that the build-
6.1 Model with Spatially-Situated Occupants ing layout includes offices with occupants on the
edges. These occupants have the following con-
We define our model in terms of the building straints:
layout, robot capabilities and limitations, and oc-
cupant abilities. Our building layout includes: – availability A where aijh is the probability of an
occupant helping with type h on edge eij and
– vertices V of locations in the building, – response time R where rh is the average time
– edges E where eij is the hallway segment that it takes all occupants to respond to help h
connects vi and v j, and
– distances D where dij is the length of eij Using this help, the robot can determine its new
expected traversal time in terms of the availability
The robot navigates and performs services in the of help and time for an occupant to provide that
building with the following limitations: help. If the robot is uncertain and a human is
available, the robot can traverse the next edge
– speed s which we assume, for simplicity, is
without cijh but incurs a response cost rh . If it is
constant over all edges,
uncertain but no human is available, it traverses
– help types H that it requires to resolve
the edge with cijh .
its localization, perception, and actuation
limitations, dij h h
– need matrix N where nijh is the probability of tij = + nij aijrh + 1 − aijh cijh (2)
s
needing help h on edge eij, and h∈H
and writing notes. Instead, we chose to adapt A* Using this shortest navigation time formula-
planners, because they tion, we can show that a robot navigates to-
wards available humans when the expected time
– are commonly used on robots today,
to travel with a need for help is longer than the
– can be applied to many different types of help
expected time to travel when stopping and asking
we discussed previously and implemented
for help. Because the robot can reason about its
quickly,
limitations and the availability of the humans in
– and are easily modifiable by changing the edge
the environment, it can decide which path to take
weights to plan different shortest paths.
and when to ask for help and could easily be
We will show that our planner consolidates all implemented in our environment. Additionally, as
robot and occupant limitations and considers the expected availability of occupants in our envi-
expected navigation time to both perform au- ronment (and many others) changes, the robot can
tonomously and ask for help. replan its path to be sure someone is available to
Formally, a robot navigates along a path of help it if necessary. Next, we provide an example
edges in order to complete its tasks. We define a of the differences between these plans and show
g
path pb = E p from a beginning vb to goal vg as that the shortest path can include human help to
g g
a set of edges that connect vb to vg , and Pb = { pb } reduce the time to navigate autonomously.
as the set of all such paths. Traditionally, the path
with the shortest navigation time is the one with 6.3 Example
the globally shortest travel time:
dij Figure 6 illustrates a simple example building with
ming DistOnly( p) = min (3) six vertices and relevant distances D, time costs C
p∈Pb
g
p=E p ∈Pb s
eij ∈E p
and availability A for a robot that only needs help
A more realistic estimate of the shortest path with localization. For simplicity, we remove the
takes into account uncertainty. Specifically, in help superscripts because there is only one type
addition to distance, we calculate the expected of help. We define the need for help for each edge
time to traverse the path in terms of the time to nij to be 0.5. We evaluate which of the three non-
autonomously traverse the edges: cyclic paths from v1 to v6 a robot should take:
Using only the edge distances dij, the DistOnly Next, we discuss the opportunities to improve and
travel time estimates are DistOnly( p1 ) = 3, extend the model using results from our post-
DistOnly( p2 ) = 4, and DistOnly( p3 ) = 5 respec- study interviews.
tively. When we account for the time to relocal-
ize due to uncertainty using Eq. 4, path 2 is the
shortest: 7 Discussion and Future Work
– UncertDist( p1 ) = 11
– UncertDist( p2 ) = 10 CoBot was able to elicit responses from many
– UncertDist( p3 ) = 11 building occupants over the nine trials and collect
information about who was available, where and
In other words, if the robot were to navigate when. Next, we discuss our qualitative observa-
autonomously through this environment, it would tions and interviews with occupants after the study
be faster for it to take the second path rather than to illustrate opportunities to add further occupant
the first. constraints based on those results.
If we also include localization help based on
the availability of humans in the offices on the Interruption In designing the robot’s initial inter-
edge between v3 and v4 using Eq. 5, the third path action, we used an assumption that an open office
is the shortest. Because aij = 0 for all occupants door indicates that the occupant is interruptible.
except between edge e34 , the aij ∗ r term does not However, we found that often doors are left open
contribute to the traversal time of paths 1 and even when occupants are in meetings or on the
2 and the expected times are the same as Eq. phone. Occupants who were in meetings and did
4 (11 and 10 respectively). The expected cost of not want to help the robot either verbally tried
traversing the third path includes stopping and to send the robot away or ignored the robot until
waiting for help on e34 : it left their doorway. Surprisingly, however, many
occupants did interrupt their meetings or put their
– HelpUncert( p1 ) = 11
phone call on hold to help the robot and some
– HelpUncert( p2 ) = 10
reported that the interruptions were “well-needed
– HelpUncert( p3 ) = 9.25
breaks in their day.”
The robot is able to plan to navigate toward the Models of interruption have been used for su-
occupant for help rather than attempt to relocalize pervisors to warn them about the robot needing
while traversing another path. Notice that if c25 = help soon [35]. While it might seem obvious that
c24 = c34 = ∞, then the third path with help is the a robot in human environments should also have
only valid one. a model of interruptibility through real-time sens-
To summarize based on our original research ing, a naïve interruption model may predict that
questions, we have demonstrated that, using the occupants are not available to help when they are
locations and an availability model of the building in meetings or on the phone [14]. However, if
occupants developed through long-term interac- CoBot used this model, it would have received
tions, the robot can determine which offices to fewer responses compared to asking for help at
navigate to if it needs help. More importantly, every open door. A robot must learn, through
we have demonstrated that the robot can find the its long-term interactions, which occupants are
shortest path to its destination which increases willing to be interrupted and under which condi-
the likelihood that it will pass by many available tions (e.g., who they are speaking with, whether
offices so that it can proactively stop at those they are working on the computer) to take full
offices in case the robot needs help to overcome advantage of the occupant help.
its limitations. The robot does not need to stop
to request help if it can navigate autonomously. Deception We also assumed that participants
Our model can also be used by other planners would answer the robot’s questions accurately
such as POMDPs for localization and navigation. and completely when they agreed to help. While
J Intell Robot Syst (2012) 66:205–221 219
occupants did answer 100% of the localiza- available to help but the cost of asking someone
tion questions accurately, we found that several too frequently, and the additional constraints for
participants deceived the robot, responding that navigational planning and determining who to ask
they moved the chairs or that they had written the for help.
note when they had not. In particular, two occu- During actual deployments of a robot, how-
pants submitted blank notes and two wrote notes ever, reducing question repetition could be
with incomplete information about the package difficult. If only a single occupant is frequently
location. available in areas of frequent limitations, the ro-
It is unclear whether participants deceived the bot would have no choice but to travel in that
robot because they were told it was a study and area sometimes. In order to reduce the likelihood
not deployed for real. However, these results indi- of this happening, it could request help from an
cate that a robot must maintain some uncertainty occupant who is further away and is not asked
about whether a task was actually performed for help as often. Additionally, the robot could
for it. The robot could ask another occupant to include time in planning to vary the time of day
confirm the task was completed since it may be it would complete the task (possibly delaying its
difficult for the robot to detect deception itself, or task) if another occupant is available at another
otherwise use extra sensors to detect that the task time.
was completed (e.g., a sensor near the paper and
pens to detect if an occupant picked them up to
Supervisors and Bystanders Finally, although
write a note), or mitigate failures by apologizing
our model is based on occupants and their avail-
and requesting help from someone else [23]. If the
ability and locations, this same model can be used
robot can determine and learn through its inter-
to model supervisors and bystanders as well. Su-
actions over time which occupants are deceptive
pervisors are typically equally available on all
[10, 11], it should avoid asking for help from them
edges of the graph and can be modeled that way.
in the future by either lowering the availability of
However, if there is limited wireless communica-
the occupant or adding additional costs related to
tion range with the supervisor, for instance, the
the trustworthiness of the occupants.
availability of the supervisor can reflect the signal
strength of communication as the robot moves in
Question Repetition While most participants did
the environment. The robot could then navigate
respond to the repeated questions (each of the
towards the higher signal strength edges to ensure
three questions were repeated each day), during
that the communication is clear to request help
the interviews, occupants reported that they were
from its supervisor.
confused as to why the robot asked them to per-
Robots that depend on bystanders assume that
form the same tasks multiple days in a row. One
there is an even and high traffic flow along the
occupant reported that he wrote the incomplete
path the robot could take. Instead, the availability
notes out of frustration when he was asked to do
of bystanders in our model could reflect the traffic
the same task multiple times. This finding mirrors
flow through different areas of the graph. A robot
previous work that showed that people who are
that requests help from bystanders would then
asked for help too frequently tend to stop re-
navigate towards a path with more people. In
sponding to help requests in the future [32].
this way, it would be able to find help faster as
While this repetition of questions is an artifact
it travels rather than necessarily waiting on low-
of our study, it indicates the need for the robot
traffic edges for help.
to keep track of which occupants it has asked for
help to purposefully plan to avoid those offices
unless there is no other help available. This would
require that the robot also model the history of 8 Conclusion
questions to more heavily weigh the occupants
who have not been asked for help recently. The We have argued that robots that ask for help
robot would then need to model not only who is from occupants of the environment combine the
220 J Intell Robot Syst (2012) 66:205–221
benefits of asking bystanders with those of super- 2. Argall, B., Chernova, S., Veloso, M., Browning, B.: A
visors in the following ways: survey of robot learning from demonstration. Robot.
Auton. Syst. 57(5), 469–483 (2009)
– Occupants have incentive to answer questions 3. Armstrong-Crews, N., Veloso, M.: Oracular POMDPS:
a very special case. In: ICRA ’07: International Con-
while the robot is completing services for oth-
ference on Robotics and Automation, pp. 2477–2482
ers in order to reciprocate the help that other (2007)
occupants provide the robot when it is per- 4. Asoh, H., Hayamizu, S., Hara, I., Motomura, Y.,
forming services for them. Akaho, S., Matsui, T.: Socially embedded learning of
the office-conversant mobile robot JIJO-2. In: 15th
– The number of occupants can distribute the
International Joint Conference on Artificial Intelli-
burden of help among many more people. gence, pp. 880–885 (1997)
5. Bruemmer, D.J., Few, D.A., Boring, R.L., Marble, J.L.,
However, unlike supervisors and bystanders: Walton, M.C., Nielsen, C.W.: Shared understanding for
collaborative control. IEEE Trans. Syst. Man Cybern.
– Occupants are not always available to help the 35(4), 494–504 (2005)
robot, and this availability can be measured 6. Callison-Burch, C.: Fast, cheap, and creative: evalu-
through long-term interactions ating translation quality using Amazon’s mechanical
– Occupants are spatially-situated and the robot turk. In: EMNLP ’09: Proceedings of the 2009 Con-
ference on Empirical Methods in Natural Language
must navigate to the occupants to receive help. Processing, pp. 286–295 (2009)
7. Clark, H., Wilkes-Gibbs, D.: Referring as a Collabora-
We have shown that, in terms of distributed help, tive Process. Cognition, Manchester (1986)
several occupants were willing to help the robot 8. Cohn, D., Atlas, L., Ladner, R.: Improving generaliza-
with any of our three types of questions, irrespec- tion with active learning. Mach. Learn. 15(2), 201–221
tive of the question response time. Additionally, (1994)
9. Dautenhahn, K., Walters, M., Woods, S., Koay, K.,
we found that some occupants were even willing Nehaniv, C., Sisbot, A., Alami, R., Siméon, T.: How
to interrupt their meetings and phone calls to may I serve you?: a robot companion approaching a
help the robot. In terms of incentives, we found seated person in a helping context. In: Proceedings of
no significant different in answering frequency the ACM SIGCHI/SIGART Conference on Human–
Robot Interaction, pp. 172–179 (2006)
when gifts were offered, but some occupants that 10. Donmez, P., Carbonell, J.G.: Proactive learning:
expected gifts did open their door to help the cost-sensitive active learning with multiple imperfect
robot. We found some novelty effect of occupants oracles. In: CIKM 2008, pp. 619–628 (2008)
increasingly closing their doors over our three 11. Donmez, P., Carbonell, J.G., Schneider, J.: Efficiently
learning the accuracy of labeling sources for se-
day study, but the number of available occupants lective sampling. In: KDD ’09: Proceedings of the
remained constant. This indicates that a robot 15th ACM SIGKDD International Conference on
should be able to take advantage of occupant Knowledge Discovery and Data Mining, pp. 259–268
help over long-term interactions, and there are (2009)
12. Dorais, G.A., Bonasso, R.P., Kortenkamp, D., Pell, B.,
opportunities to learn availability based on the Schreckenghost, D.: Adjustable autonomy for human-
time of day from those interactions. We have centered autonomous systems. In: Working Notes
introduced a model of the environment that in- of the Sixteenth International Joint Conference on
cludes the learned availability information, and Artificial Intelligence Workshop on Adjustable Auton-
omy Systems, pp. 16–35 (1999)
demonstrated its use in an A* planning algo- 13. Elfes, A.: Robot navigation: integrating perception,
rithm. Finally, we discussed additional qualitative environmental constraints and task execution within
findings and opportunities to extend our model a probabilistic framework. In: Reasoning with Uncer-
of humans in the environment to include other tainty in Robotics, pp. 91–130 (1996)
14. Fogarty, J., Hudson, S.E., Atkeson, C.G., Avrahami,
parameters such as interruptibility that can be D., Forlizzi, J., Kiesler, S., Lee, J.C., Yang, J.: Predict-
learned through the long-term interactions with ing human interruptibility with sensors. ACM ToCHI
occupants. 12(1), 119–146 (2005)
15. Fong, T.W., Thorpe, C., Baur, C.: Collaboration, dia-
logue, and human-robot interaction. In: 10th Interna-
References tional Symposium of Robotics Research (2001)
16. Fong, T.W., Thorpe, C., Baur, C.: Robot, asker
1. Angluin, D.: Queries and concept learning. Mach. of questions. Robot. Auton. Syst. 42(3–4), 235–243
Learn. 2(4), 319–342 (1988) (2003)
J Intell Robot Syst (2012) 66:205–221 221
17. Goodrich, M.A., Schultz, A.C.: Human-robot interac- Joint Conference on Autonomous Agents and Multi-
tion: a survey. Found. Trends Human–Comput. Inter- agent Systems (2010)
act. 1(3), 203–275 (2007) 30. Rosenthal, S., Dey, A.K., Veloso, M.: How robots’
18. Green, P., Wei-Haas, L.: The rapid development of questions affect the accuracy of the human responses.
user interfaces: experience with the wizard of oz In: The International Symposium on Robot–Human
method. HFES Annual Meeting 29(5), 470–474 (1985) Interactive Communication, pp. 1137–1142 (2009)
19. Hayes, G., Demiris, J.: A robot controller using learn- 31. Scholtz, J.: Evaluation Methods for Human–System
ing by imitation. In: 2nd International Symposium on Performance of Intelligent Systems. Tech. Rep. 990,
Intelligent Robotic Systems, pp. 198–204 (1994) NIST Special Publication (2002)
20. Hearst, M.A.: Mixed-initiative interaction: trends and 32. Scollon, C., Kim-Prieto, C., Diener, E.: Experience
controversies. In: IEEE Intelligent Systems, pp. 14–23 sampling: promise and pitfalls, strengths and weak-
(1999) nesses. J Happiness Studies 4, 5–34 (2003)
21. Hüttenrauch, H., Eklundh, S.: To help or not to help 33. Sellner, B.P., Heger, F., Hiatt, L., Simmons, R.,
a service robot: bystander intervention as a resource Singh, S.: Coordinated multi-agent teams and sliding
in human-robot collaboration. Interact. Stud. 7(3), autonomy for large-scale assembly. In: Proceedings
455–477 (2006) of the IEEE Special Issue on Multi-Robot Systems,
22. Kirby, R., Simmons, R., Forlizzi, J.: Companion: a con- vol. 94(1), pp. 1425–1444 (2006)
straint optimizing method for person-acceptable nav- 34. Shahaf, D., Horvitz, E.: Generalized task markets
igation. In: Ro-Man ’09: Proceedings of International for human and machine computation. In: AAAI
Symposium on Robot and Human Interactive Commu- (2010)
nication, pp. 607–612 (2009) 35. Shiomi, M., Sakamoto, D., Takayuki, K., Ishi, C.T.,
23. Lee, M.K., Kielser, S., Forlizzi, J., Srinivasa, S., Ishiguro, H., Hagita, N.: A semi-autonomous com-
Rybski, P.: Gracefully mitigating breakdowns in ro- munication robot: a field trial at a train station.
botic services. In: HRI ’10: 5th ACM/IEEE Inter- In: HRI ’08: 3rd ACM/IEEE International Conference
national Conference on Human Robot Interaction, on Human Robot Interaction, pp. 303–310 (2008)
pp. 203–210 (2010) 36. Thrun, S., Beetz, M., Bennewitz, M., Burgard, W.,
24. Lockerd, A., Breazeal, C.: Tutelage and socially guided Cremers, A., Dellaert, F., Fox, D., Hahnel, D.,
robot learning. In: IEEE/RSJ International Confer- Rosenberg, C., Roy, N., et al.: Probabilistic algorithms
ence on Intelligent Robots and Systems, pp. 3475–3480 and the interactive museum tour-guide robot Minerva.
(2004) Int. J. Rob. Res. 19(11), 972–999 (2000)
25. Michalowski, M., Sabanovic, S., DiSalvo, C., Busquets, 37. von Ahn, L., Dabbish, L.: Labeling images with a
D., Hiatt, L., Melchior, N., Simmons, R.: Socially dis- computer game. ACM conference on human factors
tributed perception: grace plays social tag at AAAI in computing systems. In: CHI 2004, pp. 319–326
2005. Auton. Robots 22(4), 385–397 (2007) (2004)
26. Parasuraman, R., Sheridan, T., Wickens, C.: A model 38. von Ahn, L., Maurer, B., McMillen, C., Abraham,
for types and levels of human interaction with automa- D., Blum, M.: RECAPTCHA: human-based charac-
tion. IEEE Trans. Syst. Man Cybern. 30(3), 286–297 ter recognition via web security measures. Science,
(2000) 321(5895), 1465–1468 (2008)
27. Plutowski, M., White, H.: Selecting concise training 39. Weiss, A., Igelsböck, J., Tscheligi, M., Bauer, A.,
sets from clean data. IEEE Trans. Neural Netw. 4(2), Kühnlenz, K., Wollherr, D., Buss, M.: Robots asking
305318 (1993) for directions: the willingness of passers-by to sup-
28. Price, R.R.: Accelerating Reinforcement Learning port robots. In: HRI ’10: 5th ACM/IEEE International
Through Imitation. Ph.D. Thesis, The University of Conference on Human Robot Interaction, pp. 23–30
British Columbia (Canada) (2003). Adviser-Boutilier, (2010)
Craig 40. Yanco, H., Drury, J.L., Scholtz, J.: Beyond usability
29. Rosenthal, S., Biswas, J., Veloso, M.: An effective per- evaluation: analysis of human–robot interaction at a
sonal mobile robot agent through a symbiotic human– major robotics competition. Hum.-Comput. Interact.
robot interaction. In: AAMAS ’10: 9th International 19(1), 117–149 (2004)