Professional Documents
Culture Documents
Mercado Et Al. - 2016 - Intelligent Agent Transparency in Human-Agent Team
Mercado Et Al. - 2016 - Intelligent Agent Transparency in Human-Agent Team
support operator situation awareness (SA) of the reasons for its current behavior, and projecting
IA in its tasking environment, the agent needs to its future behaviors (Sarter & Woods, 1995). As
be transparent about its reasoning process and a result, transparency in automated systems has
projected outcomes (Chen et al., 2014; Lee & become an essential research area (Lee & See,
See, 2004). Agent transparency is the IA’s ability 2004). Although there are multiple definitions of
to communicate information to the human oper- agent transparency (Chen et al., 2014; Helldin,
ator in a clear and efficient manner, which allows 2014; Lyons & Havig, 2014), we use the defini-
the operator to develop an accurate mental tion proposed by Chen and colleagues (2014):
model of the system and its behavior, leading to “Agent transparency is the quality of an interface
calibrated trust in the system (Chen et al., 2014; pertaining to its abilities to afford an operator’s
Lee & See, 2004). comprehension about an intelligent agent’s intent,
The effect of increased agent transparency on performance, future plans, and reasoning pro-
operator workload depends on the amount of cess” (p. 2). The goal of transparency is not to
information provided and whether the informa- relay all of a system’s capabilities, behaviors, and
tion is required for the operator’s task perfor- decision-making rationale to the human operator.
mance (Lyons & Havig, 2014). If implemented Ideally, agents should relay clear and efficient
appropriately, these additional interface ele- information as succinctly as possible to the human
ments can reduce operator workload by helping operator, thus enabling the operator to maintain
the operator understand what the agent is trying proper SA of the system in its tasking environ-
to achieve and what the operator can expect to ment without becoming overloaded (Chen et al.,
happen, so the operator does not have to make 2014; Lee & See, 2004).
these connections himself or herself (Chen & In an effort to determine the essential informa-
Barnes, 2014). However, more information does tion that must be provided to the operator to sup-
not always equate to relevant and good informa- port agent transparency, we leveraged the SA-
tion. If the increased information-processing based agent transparency (SAT; Chen et al., 2014)
requirements caused by the additional informa- model. This model, based on Endsley’s (1995)
tion shown to the operator increase workload, SA model, described the information that the IA
the display may be seen as less usable and will should convey to the human in order for him or
be trusted less. her to have proper SA of the agent in its tasking
Lee and See (2004) recommend that the sys- environment. According to Endsley (1995), SA is
tem display its purpose, process, and perfor- “a[n individual’s] state of knowledge” of a
mance (3Ps) as well as respective histories. To dynamic environment (p. 36). In Endsley’s model,
avoid overwhelming the operator, the presenta- SA has three levels: perception (Level 1), com-
tion should be in a simplified form, for example, prehension (Level 2), and projection (Level 3).
integrated graphical displays and simplified text Similarly, the first level of the SAT model pro-
(Cook & Smallman, 2008; Neyedli, Hollands, & vides the operator with basic information about
Jamieson, 2011). In the present research we the IA’s current state, goals, intentions, and plan
sought to examine the level of information nec- of action; the second level provides the operator
essary to create an effective “transparent inter- with information about the IA’s reasoning process
face” that does not induce operator overload, behind the plan of action, including rationale,
specifically addressing three issues: operator capabilities, limitations, and trade-offs between
performance, operator trust in the IA, and opera- different options; and the third level provides the
tor workload. The following section explains the operator with information regarding predicted
concept of agent transparency and a framework consequences and the likelihood of a plan’s suc-
on which our experimental design was based. cess or failure (Chen et al., 2014).
The SAT model implies that incorporating
Agent Transparency and SA three levels of transparency should support the
Humans interacting with highly automated operator’s SA of an IA in its tasking environ-
systems encounter multiple challenges: under- ment, specifically the IA’s intent, reasoning,
standing the current system state, comprehending projected outcomes, and uncertainty. The opera-
Agent Transparency for Multi-UxV Management 403
tor’s SA of the IA’s intent, reasoning, projected ent interface to support human–agent teaming for
outcomes, and uncertainty should improve the management of multiple UxVs. Whereas there are
operator’s subjective trust as well as trust cali- various types of agent transparency with different
bration, which is proper reliance when the IA is levels of details, we focused on the transparency
correct and correct rejection when the IA is that would support the operator’s SA of the agent
incorrect (Lee & See, 2004). An appropriate in its tasking environment (based on the SAT
calibration of trust in an IA is especially vital in model). We hypothesized that a greater level of
high-risk situations, such as military operations, transparency in the interface should promote
where over- and undertrust can be disastrous (de effective trust calibration, which, in turn, should
Visser et al., 2012; Freedy, DeVisser, Weltman, support overall human–agent team performance.
& Coeyman, 2007; Groom & Nass, 2007; Lee & Workload is a valid concern when additional
See, 2004; Parasuraman & Riley, 1997). information is added to an interface, as it may
Although informing the operator of IA uncer- affect both trust in the system and perceived
tainty seems like a counterintuitive way to usability. We hypothesized that operator workload
improve trust calibration, understanding the lim- would decrease with increased transparency level,
itations of a system has been reported to improve because the operator would not have to speculate
trust calibration and subjective trust in the deci- on the reasoning behind the IA’s recommended
sion-making process (Dzindolet, Peterson, Pom- plans and projected outcomes. However, we also
ranky, Pierce, & Beck, 2003; Lee & See, 2004; note that increased workload is a valid concern
Wang, Jamieson, & Hollands, 2009). when additional information is added to an inter-
face, and increased workload may reduce per-
Current Study ceived usability in the system (Bevan & Macleod,
We simulated a heterogeneous multi-UxV 1994). Finally, prior research suggests that indi-
planning task, in which participants performed vidual differences factors, such as operator spatial
the role of an operator whose job was to ability, attentional control ability, video gaming
work with an intelligent planning agent. The experience, and working memory capacity
objective was to decide on which courses of (Ahmed et al., 2014), may affect human–agent
action should be carried out by the UxVs to teaming (see Chen & Barnes, 2014, for a review).
ensure mission success based on the com- The effects of these factors on operator perfor-
mander’s intent, vehicle capability, and envi- mance were also evaluated.
ronmental constraints. We utilized the Wizard
of Oz (Riek, 2012) technique to simulate both Method
current and future capabilities of the Intelligent
Participants
Multi-UxV Planner With Adaptive Collabora-
tive/Control Technologies (IMPACT) system, Participants for this experiment included
currently developed under the U.S. Department both undergraduate and graduate students from
of Defense Autonomy Research Pilot Initiative. the University of Central Florida. Thirty young
The IMPACT system combines “flexible, goal- adults (18 men, 12 women) between the ages of
oriented play-calling and human-autonomy 18 and 29 (M = 21.2, SD = 2.3) were recruited
interaction with cooperative control algorithms using an online participant pool. Participants
that provide near-optimal task assignment and were required to have normal or corrected-to-
path planning solutions as well as adaptive/ normal vision (including not being color-blind).
reactive capability” (Draper, 2013, p. 1). In They were compensated $15 per hour for their
this context, a “play” refers to a scripted plan participation, which lasted approximately 3 hr.
for coordinating multiple UxVs, which is aug-
mented by the IA that chooses the appropriate Apparatus
assets for the specific mission (Douglass, 2013; Simulator. A customized computer-based
Miller & Parasuraman, 2007). simulator, based on the U.S. Air Force Research
Our goal was to examine the level of informa- Laboratory’s FUSION interface (Spriggs, Boyer,
tion necessary to create an effective and transpar- & Bearden, 2014), was utilized in the present
404 May 2016 - Human Factors
After each training mission block, partici- and consisted of PowerPoint slides and training
pants completed an in-house usability question- missions performed using the simulator. During
naire with which they ranked how often they training, participants were informed that the IA
utilized each aspect on the decision window was not always 100% accurate. After training,
(e.g., sprocket graphic, text box). participants received 18 evaluation missions and
After each mission block, participants com- were required to perform at least 12 missions
pleted the System Usability Scale, a 10-item correctly to move on to the experimental session.
summative usability survey (Brooke, 1996). The evaluation lasted approximately 40 min. Par-
Participants’ perceived workload was evaluated ticipants were given a 5-min break, after which
with a computer-based version of the NASA the eye tracker was calibrated.
Task Load Index questionnaire (NASA-TLX; The experimental session consisted of three
Hart & Staveland, 1988). Finally, participants counterbalanced blocks of eight missions. Dur-
were asked to evaluate their trust in the IA along ing each mission, participants controlled a team
two dimensions, (a) information analysis (trust of UxVs (ground, air, and sea) in a military
in the information and analysis displayed) and perimeter defense task. Participants received
(b) decision and action selection (trust in the intel and the commander’s intent and used them
IA’s suggestions and decisions; Parasuraman, to choose one of the IA’s recommended plans.
Sheridan, & Wickens, 2000), using a modified The IA always recommended two plans: Plan A
version of the Trust Between People and Auto- (the agent’s top choice) and Plan B (the agent’s
mation questionnaire developed by Jian, Bisantz, secondary or backup choice). The IA’s top
and Drury (2000). choice was optimized for the most important
metric, which was indicated by wedge size. For
Procedure example, if the speed wedge was the largest,
After being briefed on the purpose of the study then the IA would optimize the play for speed to
and giving informed consent, participants com- complete the mission as quickly as possible. The
pleted the demographics questionnaire and the IA’s backup choice was optimized for one of the
Ishihara test. Participants then received training other two metrics (coverage or capabilities if
on the tasks, which lasted approximately 45 min speed was the most important metric).
406 May 2016 - Human Factors
Figure 4. Transparency Level 1+2 interface. Includes interface items that describe
agent rationale for recommended plan.
58) = 12.33, p < .001, η2 = .30 (Figure 6; Table F(2, 58) = 15.03, p < .001, η2 = .34 (Figure 6;
1). Post hoc comparisons indicated participants’ Table 1). Post hoc comparisons indicated that
correct IA usage rates were significantly greater participants’ correct IA rejection rates were
in Level 1+2+3 (p < .001) and Level 1+2 (p = greatest in Level 1+2+3, significantly greater
.003) compared to Level 1. than Level 1+2 (p = .04), which, in turn, is
greater than Level 1 (p = .013).
Correct IA Rejection Rates There was a marginally significant interac-
Results for correct IA rejection rates revealed tion effect for working memory capacity, F(2,
a significant main effect of transparency level, 56) = 3.07, p = .054, η2 = .01. High-WMC
408 May 2016 - Human Factors
(Figure 10). Individuals with low SpaV had longer in Level 1 (d = 1.04) than in either Level 1+2+3
fixation durations than high-SpaV individuals in (d = 0.96) or Level 1+2 (d = .82), F(1, 23) =
Level 1 (d = 0.42) and Level 1+2 (d = 0.57), but 5.54, p = .027, η2 = .19 (Figure 11).
the opposite was observed in Level 1+2+3 (d =
0.51). Trust
Pupil diameter was larger for the high-SpaO We conducted two separate between-subjects
group (M = 3.89, SD = 0.17) than for the low- ANOVAs on the Information Analysis and Deci-
SpaO group (M = 3.50, SD = 0.03) across all sion and Action Selection subscales. Only the
transparency levels. The differences were larger trust assessments for the first block were used, as
410 May 2016 - Human Factors
15
10
NASA-TLX Subscales
Table 2: Results for Trust Subscales and System Usability Scale (SUS) by Transparency Level
teaming for multirobot management. The perfor- These results, when combined with response
mance data indicated that participants’ correct time and workload data, suggest that additional
rejection accuracy increased in relation to trans- transparency elements in the display improved
parency level, whereas correct IA usage increased performance without increasing workload or
only from Level 1 to Level 1+2. The addition of response time. One alternative explanation for
reasoning information in Level 1+2 increased the improved performance with higher level of
correct IA use by 11% and correct rejection rate transparency is that the additional information
by 12%. The addition of uncertainty informa- could have caused participants to be more care-
tion (Level 1+2+3 compared with Level 1+2) ful about their decision-making process instead
improved correct IA use rate by a small amount of using the rationale or uncertainty displays. If
(2%) and correct rejection rate by 14%. this explanation is correct, we would expect
412 May 2016 - Human Factors
response times to increase with higher levels of participants’ trust in the IA. Participants may
transparency, but this result was not observed have distrusted the system and manually solved
(Table 1). Additionally, participants reported each mission without regard to the IA’s recom-
that the sprocket graphic and text tables were the mendation (Parasuraman & Riley, 1997). There-
most useful display elements for their decision fore, we used subjective trust measures to investi-
making. These data (both performance and self- gate participants’ perception of the IA’s trustwor-
report) suggest that incorporating reasoning and thiness. Results for the Suggesting or Making
uncertainty information into heterogeneous tac- Decisions subscale provide evidence that par-
tical decision making successfully allowed our ticipants trusted the IA’s recommendation more
participants to make better-calibrated decisions. when the system was more transparent. Overall,
This finding is consistent with Helldin (2014), as participants were provided with more trans-
although the performance improvement in our parency, they (correctly) rejected the agent’s
study did not come at the cost of longer response incorrect recommendation more often yet trusted
times, as Helldin reported. it more. This increase in trust may be because a
Previous research indicated that too little or more transparent agent—displaying explana-
too much trust can lead to disuse or misuse of tions of reasoning and conveying relevant uncer-
automated systems (Parasuraman & Riley, tainty—is perceived as more humanlike, intelli-
1997). In contrast, our participants who accepted gent, and trustworthy than an agent whose reason-
the IA’s correct recommendations but (correctly) ing is not evident (de Visser et al., 2012; Jian et al.,
rejected incorrect recommendations displayed 2000). Another explanation may be that in provid-
proper trust calibration for the IA’s recommen- ing the operator with reasoning and relevant
dations. Taken together, the correct IA usage and uncertainty information, the IA offers a better
correct IA rejection rates indicate that partici- understanding of the system’s limitations, thus
pants’ trust calibration increased as a function of increasing the operator’s trust.
transparency level. This result is consistent with This relationship also sheds light on Wickens
the findings of Oduor and Wiebe (2008), in and Dixon’s (2007) findings that automation
which transparency improved operator trust cal- with less than 70% reliability may not be benefi-
ibration and performance in a human–agent cial to performance. Our results suggest that
joint decision-making context. when the automation’s reasoning is transparent,
Operator performance varied with regard to the operator’s trust calibration is enhanced,
individual differences in both WMC and gaming which can result in improved subjective trust
experience. Individuals with higher WMC out- and performance even when the automation’s
performed (correct rejections) those with lower reliability is below 70%. Finally, System Usabil-
WMC during Level 1. Level 1 provided indi- ity Scale findings also show the benefit of trans-
viduals with only basic information, which in parency, as results indicate that participants’ per-
turn required individuals to process and synthe- ceived usability of the system increased as a
size that information before making a decision. function of transparency level.
Presumably, individuals with higher WMC more Prior research has shown that workload
accurately identified incorrect recommendations and response time can increase as more informa-
due to their ability to maintain the information tion is displayed to an operator due to increased
needed to make a judgment in their working information-processing demands, especially when
memory. Our findings support previous research, mental workload and effort are high (Helldin,
which indicates that individuals with low WMC 2014; Zuk & Carpendale, 2007). In contrast, our
are more likely to rely on heuristics to mitigate NASA-TLX and eye-tracking workload data as
their memory load (Quayle & Ball, 2000), but well as response time failed to show increases as
the increase in transparency of the agent inter- a function of transparency. However, when
face reduced the performance gap between low- examining individual differences, we found that
and high-WMC individuals. participants with low SpaV had longer fixation
Using operator performance as an objective durations during Level 1 and Level 1+2, which
measure of trust reveals only one aspect of the may be a result when these individuals take a
Agent Transparency for Multi-UxV Management 413
longer time to process the information on screen In contrast with Helldin (2014), our results indi-
before making a decision. This finding suggests cate that benefits of transparency based on the
that spatial abilities may have differential effects SAT model can improve performance effective-
on operator response time when agent transpar- ness without additional costs in terms of time
ency (especially uncertainty information) is and workload (Chen et al., 2014).
involved but, more importantly, identifies the Additionally, the benefits of transparency
need for further research. Additionally, in con- were also seen in the participants’ trust calibra-
trast to prior research (Rayner, 2009), we found tion, which increased as a function of transpar-
that pupil diameter was larger for individuals ency. This result led to participants’ feeling the
with high SpaO across all transparency levels. IA was trustworthy and had greater usability
However, as Rayner (2009) mentioned, the rela- when the agent was more transparent. Authors
tionship between pupil diameter and SpaO may of future research should investigate how incor-
change as an experimental task moves closer to porating uncertainty into each level of transpar-
simulating a real-world task. Our findings, in ency affects operator performance, trust, and
conjunction with those of Rayner, suggest that workload.
the relationship between pupil diameter and
SpaO may be task specific. Acknowledgments
This research was supported by the Department of
Limitations Defense Autonomy Research Pilot Initiative, under
the Intelligent Multi-UxV Planner With Adaptive
Taken together, our findings indicate that the
Collaborative/Control Technologies (IMPACT) proj-
benefits of transparency may not necessarily
ect. We wish to thank Isacc Yi, Erica Valiente, Shan
introduce potential costs in terms of speed and
Lakhmani, and Jonathan Harris for their contribution
workload. Additionally, a more transparent IA
to this project. We would also like to thank Gloria
engenders higher level of trust and perceived
Calhoun and Mark Draper for their input.
usability. However, the agent behaviors simu-
lated in the current study were based on realistic
Key Points
capabilities of the IMPACT system currently
under development, but one key function of •• Increasing agent transparency enhanced the
IMPACT was purposely removed from the overall human–agent team performance in a
simulation: operator’s manual “tweaking” of the multi-unmanned-vehicle management context.
agent’s recommended plans. Since the goal of •• Operators calibrated their trust in the agent more
the current study was to investigate the effects effectively when the agent was more transparent—
of transparency levels on operator decisions, their reliance on and compliance with the agent’s
greater control over participants’ actions (i.e., recommendation was more appropriate.
selecting from two choices rather than propos- •• The benefits of agent transparency do not neces-
ing their own solutions) was necessary in order sarily involve either a speed–accuracy trade-off or
to keep variability in participants’ responses increased workload costs.
manageable. Follow-up studies should test the •• Operators perceived a transparent agent as being
effects of transparency in a more realistic task- more trustworthy and having greater usability.
ing environment and allow the operator to mod-
ify the agent’s recommended plans. The current References
study represents our initial effort to demonstrate Ahmed, N., de Visser, E., Shaw, T., Mohamed-Ameen, A., Camp-
bell, M., & Parasuraman, R. (2014). Statistical modeling of
the potential utility of agent transparency for networked human–automation performance using working
human decision-making effectiveness. memory capacity. Ergonomics, 57, 295–318.
Bevan, N., & Macleod, M. (1994). Usability measurement in con-
text. Behaviour & Information Technology, 13, 132–145.
Conclusions Bradshaw, J. M., Feltovich, P. J., Johnson, M., Bunch, L., Breedy,
M. R., Eskridge, T. C., & Uszok, A. (2008). Coordination in
Our findings have important implications for human-agent-robot teamwork. In International Symposium on
the designing of systems to facilitate decision Collaborative Technology and Systems (pp. 467–476). Piscat-
making between the human operator and the IA. away, NJ: IEEE.
414 May 2016 - Human Factors
Brooke, J. (1996). SUS: A quick and dirty usability scale. In P. W. Hart, S., & Staveland, L. (1988). Development of NASA-TLX
Jordan, B. Thomas, B. A. Weerdmeester, & I. L. McClelland (Task Load Index): Results of empirical and theoretical
(Eds.), Usability evaluation in industry (pp. 189–194). Lon- research. In P. Hancock & N. Meshkati (Eds.), Human mental
don, UK: Taylor & Francis. workload (pp. 139–183). Amsterdam, Netherlands: Elsevier.
Chen, J. Y., Barnes, M. J., & Harper-Sciarini, M. (2011). Supervi- Hegarty, M., & Waller, D. (2004). A dissociation between mental
sory control of multiple robots: Human-performance issues and rotation and perspective-taking spatial abilities. Intelligence,
user-interface design. IEEE Transactions on Systems, Man, and 32, 175–191.
Cybernetics, Part C: Applications and Reviews, 41, 435–454. Helldin, T. (2014). Transparency for future semi-automated systems
Chen, J. Y. C., & Barnes, M. J. (2014). Human–agent teaming for (Doctoral dissertation). Örebro University, Örebro, Sweden.
multi-robot control: A review of human factors issues. IEEE Hoffman, J. E., & Subramaniam, B. (1995). The role of visual
Transactions on Human–Machine Systems, 44, 13–29. attention in saccadic eye movements. Perception & Psycho-
Chen, J. Y. C., Procci, K., Boyce, M., Wright, J., Garcia, A., & physics, 57, 787–795.
Barnes, M. (2014). Situation awareness–based agent transpar- Hwang, S. L., Yau, Y. J., Lin, Y. T., Chen, J. H., Huang, T. H., Yenn,
ency (No. ARL-TR-6905). Aberdeen Proving Ground, MD: T. C., & Hsu, C. C. (2008). Predicting work performance in
U.S. Army Research Laboratory. nuclear power plants. Safety Science, 46, 1115–1124.
Conway, A. R. A., Kane, M. J., Bunting, M. F., Hambrick, Z. D., Jian, J., Bisantz, A. M., & Drury, C. G. (2000). Foundations for
Willhelm, O., & Engle, R. (2005). Working Memory Span an empirically determined scale of trust in automated systems.
Tasks: A Methodological Review and User’s Guide. Psycho- International Journal of Cognitive Ergonomics, 4, 53–71.
nomic Bulletin & Review, 12, 769–786. Lee, J. D., & See, K. A. (2004). Trust in technology: Designing for
Cook, M., & Smallman, H. (2008). Human factors of the confir- appropriate reliance. Human Factors, 46, 50–80.
mation bias in intelligence analysis: Decision support from Levine, T. R., & Hullett, C. R. (2002). Eta squared, partial eta
graphical evidence landscapes. Human Factors, 50, 745–754. squared, and misreporting of effect size in communication
Cummings, M. L., & Mitchell, P. J. (2008). Predicting controller research. Human Communication Research, 28, 612–625.
capacity in supervisory control of multiple UAVs. IEEE Trans- Lewis, M. (2013). Human interaction with multiple remote robots.
actions on Systems, Man, and Cybernetics, Part A: Systems Reviews of Human Factors and Ergonomics, 9, 131–174.
and Humans, 38, 451–460. Linegang, M., Stoner, H. A., Patterson, M. J., Seppelt, B. D., Hoff-
de Visser, E., Krueger, F., McKnight, P., Scheid, S., Smith, M., man, J. D., Crittendon, Z. B., & Lee, J. D. (2006). Human–
Chalk, S., & Parasuraman, R. (2012). The world is not enough: automation collaboration in dynamic mission planning: A
Trust in cognitive agents. In Proceedings of the Human Factors challenge requiring an ecological approach. In Proceedings
and Ergonomics Society 56th Annual Meeting (pp. 263–267). of the Human Factors and Ergonomics Society 50th Annual
Santa Monica, CA: Human Factors and Ergonomics Society. Meeting (pp. 2482–2486). Santa Monica, CA: Human Factors
Douglass, S. (2013). Learner models in the large-scale cognitive and Ergonomics Society.
modeling initiative. In R. Sottilare, A. Graesser, X. Hu, & H. Lyons, J. B., & Havig, P. R. (2014). Transparency in a human–
Holden (Eds.), Design recommendations for adaptive intelli- machine context: Approaches for fostering shared awareness/
gent tutoring systems learner modeling (Vol. 1, pp. 111–126). intent. In R. Shumaker & S. Lackey (Eds.), Virtual, augmented
Orlando, FL: U.S. Army Research Laboratory. and mixed reality: Designing and developing virtual and
Draper, M, H. (2013). Realizing autonomy via intelligent adaptive augmented environments (pp. 181–190). Berlin, Germany:
hybrid control: Adaptable autonomy for achieving UxV RSTA Springer.
team decision superiority [Funding proposal]. Arlington, VA: Miller, C. A., & Parasuraman, R. (2007). Designing for flexible
Office of the Assistant Secretary of Defense for Research and interaction between humans and automation: Delegation inter-
Engineering. faces for supervisory control. Human Factors, 49, 57–75.
Dzindolet, M. T., Peterson, S. A., Pomranky, R. A., Pierce, L. G., Neyedli, H., Hollands, J., & Jamieson, G. (2011). Beyond identity:
& Beck, H. P. (2003). The role of trust in automation reli- Incorporating system reliability information into an automated
ance. International Journal of Human–Computer Studies, 58, combat identification system. Human Factors, 53, 338–355.
697–718. Oduor, K. F., & Wiebe, E. N. (2008, September). The effects of
Ekstrom, R. B., French, J. W., & Harman, H. H. (1976). Manual for automated decision algorithm modality and transparency on
kit of factor-referenced cognitive tests. Princeton, NJ: Educa- reported trust and task performance. In Proceedings of the
tional Testing Service. Human Factors and Ergonomics Society 52nd Annual Meeting
Endsley, M. R. (1995). Toward a theory of situation awareness in (pp. 302–306). Santa Monica, CA: Human Factors and Ergo-
dynamic systems. Human Factors, 37, 32–64. nomics Society.
Freedy, E., DeVisser, E., Weltman, G., & Coeyman, N. (2007, May). Paas, F. G., & Van Merriënboer, J. J. (1994). Instructional control
Measurement of trust in human–robot collaboration. In Proceed- of cognitive load in the training of complex cognitive tasks.
ings of IEEE International Symposium on Collaborative Technol- Educational Psychology Review, 6, 351–371.
ogies and Systems (pp. 106–114). Piscataway, NJ: IEEE. Parasuraman, R., & Riley, V. (1997). Humans and automation:
Groom, V., & Nass, C. (2007). Can robots be teammates? Bench- Use, misuse, disuse, abuse. Human Factors, 39, 230–253.
marks in human–robot teams. Interaction Studies, 8, 483–500. Parasuraman, R., Sheridan, T. B., & Wickens, C. D. (2000). A
Gugerty, L., & Brooks, J. (2004). Reference-frame misalignment model for types and levels of human interaction with automa-
and cardinal direction judgments: Group differences and strate- tion. IEEE Transactions on Systems, Man, and Cybernetics–
gies. Journal of Experimental Psychology: Applied, 10, 75–88. Part A: Systems and Humans, 30, 286–297.
Hardin, B., & Goodrich, M. A. (2009, March). On using mixed- Poole, A., & Ball, L. J. (2006). Eye tracking in HCI and usability
initiative control: A perspective for managing large-scale research. Encyclopedia of Human–Computer Interaction, 1,
robotic teams. In Proceedings of the 4th ACM/IEEE Inter- 211–219.
national Conference on Human–Robot Interaction (pp. 165– Quayle, J. D., & Ball, L. J. (2000). Working memory, metacog-
172). New York, NY: ACM. nitive uncertainty, and belief bias in syllogistic reasoning.
Agent Transparency for Multi-UxV Management 415
Quarterly Journal of Experimental Psychology: Section A, 53, Army Research Laboratory (ARL). Most recently,
1202–1223.
he was the lead for the human–robotic interaction
Rayner, K. (2009). Eye movements and attention in reading, scene
perception, and visual search. Quarterly Journal of Experi- component of Safe Operations for Unmanned Sys-
mental Psychology, 62, 1457–1506. tems for Reconnaissance in Complex Environments
Riek, L. D. (2012). Wizard of Oz studies in HRI: A systematic (SOURCE). Prior to ARL, he conducted human fac-
review and new reporting guidelines. Journal of Human–Robot
tors research (for the U.S. Navy) and served as
Interaction, 1(1).
Russell, S. J., & Norvig, P. (2009). Artificial intelligence: A mod- human factors research manager for the General
ern approach (3rd ed.) Upper Saddle River, NJ: Prentice Hall. Electric Corporation. Since coming to ARL, he has
Sarter, N. B., & Woods, D. D. (1995). How in the world did we served on a number of NATO working groups
ever get into that mode? Mode error and awareness in supervi-
sory control. Human Factors, 37, 5–19.
related to autonomous systems. He has coauthored
Spriggs, S., Boyer, J., & Bearden, G. (2014). Fusion system over- over 100 articles and coedited a book, Human–
view. Lecture given at Wright Patterson Air Force Base, OH. Robot Interactions in Future Military Operations
Wang, L., Jamieson, G. A., & Hollands, J. G. (2009). Trust and (2010). His research interests include investigations
reliance on an automated combat identification system. Human
Factors, 51, 281–291.
of risk visualization, intelligence processes, and
Wickens, C. D., & Dixon, S. R. (2007). The benefits of imperfect unmanned aerial vehicles crew systems. He was
diagnostic automation: A synthesis of the literature. Theoreti- educated at the University of Arizona (BA) and the
cal Issues in Ergonomics Science, 8, 201–212. New Mexico State University (MA).
Zuk, T., & Carpendale, S. (2007). Visualization of uncertainty and
reasoning. In A. Butz, B. Fisher, A. Krüger, P. Olivier, & S.
Owada (Eds.), Smart graphics (pp. 164–177). Berlin, Ger- Daniel Barber is an assistant research professor at
many: Springer. the University of Central Florida’s Institute for
Simulation and Training. He has extensive experi-
Joseph E. Mercado, MSC, USN, PhD, is an aerospace ence in the field of robotics and has also developed
experimental psychologist for the U.S. Navy. He works tools for synchronization, processing, and streaming
at the Naval Air Warfare Center Training Systems Divi- of data from multiple physiological sensors (e.g.,
sion in Orlando, Florida. Prior to his time in the Navy, eye tracking, electrocardiogram, and electroenceph-
he worked as a postdoctoral research fellow for the alography) within simulation and training environ-
U.S. Army Research Laboratory–Human Research and ments supporting real-time streaming adaptation to
Engineering Directorate, under Dr. Chen, at the Field user state. His current research focus is on human–
Element in Orlando, Florida. robot interaction, including multimodal communica-
tion, intuitive user interaction devices, supervisory
Michael A. Rupp received his MS degree in model- control of multiple vehicles, and adaptive systems
ing and simulation in 2012 at the University of Cen- using physiological sensors.
tral Florida, where he is pursuing his doctoral degree
in the applied experimental and human factors psy- Katelyn Procci, PhD, is a senior human perfor-
chology program. mance engineer with Cubic Corporation. She
earned her doctoral degree in human factors at the
Jessie Y. C. Chen is a senior research psychologist University of Central Florida, where she worked
with U.S. Army Research Laboratory–Human for the Institute of Simulation and Training in
Research and Engineering Directorate at the Field Orlando. During that time, she worked with Dr.
Element in Orlando, Florida. She earned her PhD in Chen’s research group and assisted with the early
applied experimental and human factors psychology stages of the IMPACT project.
from the University of Central Florida.
Michael J. Barnes is a research psychologist with the Date received: April 14, 2015
Human Research and Engineering Directorate of the Date accepted: October 26, 2015