Professional Documents
Culture Documents
DARPA’s Explainable
Artificial Intelligence Program
A
dvances in machine learning (ML) techniques promise
has led to a new wave of AI applications (for to produce AI systems that perceive, learn, decide, and
example, transportation, security, medicine, act on their own. However, they will be unable to
finance, defense) that offer tremendous explain their decisions and actions to human users. This lack
benefits but cannot explain their decisions
is especially important for the Department of Defense, whose
and actions to human users. DARPA’s
explainable artificial intelligence (XAI)
challenges require developing more intelligent, autonomous,
program endeavors to create AI systems and symbiotic systems. Explainable AI will be essential if
whose learned models and decisions users are to understand, appropriately trust, and effectively
can be understood and appropriately manage these artificially intelligent partners. To address this,
trusted by end users. Realizing this goal DARPA launched its explainable artificial intelligence (XAI)
requires methods for learning more program in May 2017. DARPA defines explainable AI as AI
explainable models, designing effective systems that can explain their rationale to a human user,
explanation interfaces, and understanding characterize their strengths and weaknesses, and convey an
the psychologic requirements for effective understanding of how they will behave in the future. Naming
explanations. The XAI developer teams are
this program explainable AI (rather than interpretable,
addressing the first two challenges by
creating ML techniques and developing comprehensible, or transparent AI, for example) reflects
principles, strategies, and human-computer DARPA’s objective to create more human-understandable AI
interaction techniques for generating effec- systems through the use of effective explanations. It also
tive explanations. Another XAI team is reflects the XAI team’s interest in the human psychology of
addressing the third challenge by summa- explanation, which draws on the vast body of research and
rizing, extending, and applying psychologic expertise in the social sciences.
theories of explanation to help the XAI
evaluator define a suitable evaluation
framework, which the developer teams
will use to test their systems. The XAI
teams completed the first of this 4-year
program in May 2018. In a series of
ongoing evaluations, the developer
teams are assessing how well their XAM
systems’ explanations improve user un-
derstanding, user trust, and user task
performance.
44 AI MAGAZINE Copyright © 2019, Association for the Advancement of Artificial Intelligence. All rights reserved. ISSN 0738-4602
Deep Learning and Security
Early AI systems were predominantly logical and Model induction refers to techniques that experi-
symbolic; they performed some form of logical in- ment with any given ML model— such as a black
ference and could provide a trace of their inference box— to infer an approximate explainable model. For
steps, which became the basis for explanation. There example, the model-agnostic explanation system of
was substantial work on making these systems more Ribeiro et al. (2016) inferred explanations by ob-
explainable, but they fell short of user needs for serving and analyzing the input-output behavior of a
comprehension (for example, simply summarizing black box model.
the inner workings of a system does not yield a suf- DARPA used these strategies to categorize a portfolio
ficient explanation) and proved too brittle against of new ML techniques and provide future practitioners
real-world complexities. with a wider range of design options covering the
Recent AI success is due largely to new ML tech- performance-explainability trade space.
niques that construct models in their internal repre-
sentations. These include support vector machines,
random forests, probabilistic graphical models, re-
inforcement learning (RL), and deep learning (DL)
XAI Concept and Approach
neural networks. Although these models exhibit high The XAI program’s goal is to create a suite of new or
performance, they are opaque. As their use has in- modified ML techniques that produce explainable
creased, so has research on explainability from the models that, when combined with effective explanation
perspectives of ML (Chakraborty et al. 2017; Ras et al. techniques, enable end users to understand, appropri-
2018) and cognitive psychology (Miller 2017). Simi- ately trust, and effectively manage the emerging gen-
larly, many XAI-related workshops have been held re- eration of AI systems. The target of XAI is an end user
cently on ML (for example, the International Conference who depends on decisions or recommendations pro-
on Machine Learning, the Conference on Neural In- duced by an AI system, or actions taken by it, and
formation Processing Systems), AI (for example, the In- therefore needs to understand the system’s rationale.
ternational Joint Conference on Artificial Intelligence), For example, an intelligence analyst who receives rec-
and HCI (for example, the Conference on Human- ommendations from a big data analytics system needs
Computer Interaction, Intelligent User Interfaces) con- to understand why it recommended certain activity for
ferences, as have special topic meetings related to XAI. further investigation. Similarly, an operator who tasks
There seems to be an inherent tension between ML an autonomous vehicle to drive a route needs to un-
performance (for example, predictive accuracy) and derstand the system’s decision-making model to ap-
explainability; often the highest-performing methods propriately use it in future missions. Figure 3 illustrates
(for example, DL) are the least explainable, and the the XAI concept: provide users with explanations that
most explainable (for example, decision trees) are the enable them to understand the system’s overall
least accurate. Figure 1 illustrates this with a notional strengths and weaknesses, convey an understanding of
graph of the performance-explainability trade-off for how it will behave in future or different situations, and
various ML techniques. perhaps permit users to correct the system’s mistakes.
When DARPA formulated the XAI program, it envi- This user-centered concept poses interrelated re-
sioned three broad strategies to improve explainability, search challenges: (1) how to produce more explain-
while maintaining a high level of learning performance, able models, (2) how to design explanation interfaces,
based on promising research at the time (figure 2): deep and (3) how to understand the psychologic require-
explanation, interpretable models, and model induction. ments for effective explanations. The first two chal-
Deep explanation refers to modified or hybrid DL lenges are being addressed by the 11 XAI research
techniques that learn more explainable features or teams, which are developing new ML techniques to
representations or that include explanation genera- produce explainable models, and new principles,
tion facilities. Several design choices might produce strategies, and HCI techniques (for example, visuali-
more explainable representations (for example, training zation, language understanding, language generation)
data selection, architectural layers, loss functions, to generate effective explanations. The third challenge
regularization, optimization techniques, training is the focus of another XAI research team that is
sequences). Researchers have used deconvolutional summarizing, extending, and applying psychologic
networks to visualize convolutional network layers, and theories of explanation.
techniques existed for associating semantic concepts The XAI program addresses two operationally rel-
with deep network nodes. Approaches for generating evant challenge problem areas (figure 4): data analytics
image captions could be extended to train a second deep (classification of events of interest in heterogeneous
network that generates explanations without explicitly multimedia data) and autonomy (decision policies
identifying the original network’s semantic features. for autonomous systems). These areas represent
Interpretable models are ML techniques that learn two important ML problem categories (supervised
more structured, interpretable, or causal models. Early learning and RL) and Department of Defense in-
examples included Bayesian rule lists (Letham et al. terests (intelligence analysis and autonomous systems).
2015), Bayesian program learning, learning models of The data analytics challenge was motivated by a
causal relationships, and use of stochastic grammars common problem: intelligence analysts are presented
to learn more interpretable structure. with decisions and recommendations from big data
SUMMER 2019 45
Deep Learning and Security
Learning Techniques
Performance vs.
Explainability
Neural Nets
Graphical
Learning Performance
Models
Deep
Ensemble
Learning Bayesian Methods
Belief Nets
SRL Random
CRFs HBNs Forests
AOGs
Statistical MLNs
Models Decision
Markov Trees
SVMs Explainability
Models
Figure 1. Learning Performance Versus Explainability Trade-Off for Several Categories of Learning Techniques.
analytics algorithms and must decide which to report as consists of two phases. Phase 1 (18 months) com-
supporting evidence in their analyses and which to menced in May 2017 and includes initial technology
pursue further. These algorithms often produce false demonstrations of XAI systems. Phase 2 (30 months)
alarms that must be pruned and are subject to concept includes a sequence of evaluations against challenge
drift. Furthermore, these algorithms often make recom- problems selected by the system developers and the
mendations that the analyst must assess to determine XAI evaluator. The first formal evaluations of XAI
whether the evidence supports or contradicts their hy- systems took place during the fall of 2018. This article
potheses. Effective explanations will help confront these describes the developer teams’ progress leading up to
issues. these evaluations, whose results were presented at an XAI
The autonomy challenge was motivated by the program meeting during the winter of 2019.
need to effectively manage AI partners. For example,
the Department of Defense seeks semiautonomous
systems to augment warfighter capabilities. Operators XAI Program
will need to understand how these behave so they can Development and Progress
determine how and when to best use them in future
missions. Effective explanations will better enable such Figure 7 summarizes the 11 XAI Technical Area 1 (TA1)
determinations. developer teams and the TA2 team [from the Florida
For both challenge problem areas, it is critical to Institute for Human and Machine Cognition (IHMC)]
measure explanation effectiveness. While it would be that is developing the psychologic model of explanation.
convenient if a learned model’s explainability could be Three TA1 teams are pursuing both challenge problem
measured automatically, an XAI system’s explanation areas (autonomy and data analytics), three are working on
effectiveness must be assessed according to how its only the former, and five are working on only the latter.
explanations aid human users. This requires human- Per the strategies described in figure 2, the TA1 teams are
in-the-loop psychologic experiments to measure the investigating a diverse range of techniques for developing
user’s satisfaction, mental model, task performance, explainable models and explanation interfaces.
and appropriate trust. DARPA formulated an initial
explanation evaluation framework that includes po- Naturalistic
tential measures of explanation effectiveness (figure 5). Decision-Making Foundations of XAI
Exploring and refining this framework is an important The objective of the IHMC team (which includes
part of the XAI program’s research agenda. researchers from MacroCognition and Michigan
The XAI program’s goal, concept, strategies, chal- Technological University) is to develop and evaluate
lenges, and evaluation framework are described in psychologically plausible models of explanation and
the program’s 2016 broad agency announcement. develop actionable concepts, methods, measures, and
Figure 6 displays the XAI program’s schedule, which metrics for explanatory reasoning. The IHMC team is
46 AI MAGAZINE
Deep Learning and Security
Learning Performance
Create a suite of Models
machine learning Deep
techniques that Learning Ensemble
Bayesian Methods
produce more
Belief Nets
explainable models,
while maintaining a SRL Random
high level of CRFs HBNs Forests
AOGs
learning Statistical MLNs
Models Decision
performance Markov Trees Explainability
SVMs Models
A1
Model
? ?
41 42 45 45
Input Hidden Output 0.5 0.5 1.0 1.0 1.0 1.0 1.0 1.0 1.0
Experiment
units units units 5 3 31 1 31 1 33 1 31
investigating the nature of explanation itself. What methodology was described in a series of essays
must happen for a person to be satisfied with an attempt (Hoffman and Klein 2017; Hoffman et al. 2017; Klein
to explain (1) the workings of a complex system and (2) 2018; Hoffman et al. 2018).
why it acted the way it did in a given situation? To Figure 8 illustrates IHMC’s model of the XAI ex-
address such questions, the team has formulated a planation process, highlighting measurement categories
naturalistic model of human explanatory reasoning and for assessing explanation effectiveness. The user re-
is providing guidance to the performer teams on ceives a recommendation or decision from an XAI
methods for evaluating the effectiveness of their XAI system, along with an explanation that could be tested
systems’ explanations. The team reviewed the relevant for goodness (versus preestablished criteria) and user
literatures in philosophy of science and specializations satisfaction. The explanation contributes to the user’s
within psychology, from which criteria were synthe- mental model of the AI system, which could be tested
sized for assessing the “goodness” of explanations. The for accuracy and comprehension. The AI system’s
team is also collecting and analyzing a corpus of cases in recommendations and the user’s mental model may
which individuals create or receive explanations of the enable, or decrease, user task performance, which
workings of complex systems. could also be measured. These processes contribute to
This team developed measures for explanation the user’s appropriate, or inappropriate, trust of the AI
goodness, a user’s mental model (for example, cor- system. The XAI evaluator is using this model to test
rectness, completeness), and user task performance. the developer teams’ XAI systems.
From this, the user can make reasonably accurate
judgments about when to trust (or doubt) the system. Evaluation
To gain this insight, the user must explore the decision- The XAI program’s independent government evalua-
making processes and performance of an XAI system, tor is the Naval Research Laboratory. For Phase 1, the
especially for boundary cases, including ways in which laboratory (with IHMC’s help) prepared an evaluation
deep neural networks (DNNs) can be spoofed. This framework for the TA1 teams to use as a template for
SUMMER 2019 47
Deep Learning and Security
48 AI MAGAZINE
Deep Learning and Security
An analyst is
Data Two trucks performing a
loading activity looking for
Recommend items of
Analytics Explainable Explanation
interest in
Model Interface
Explanation massive
Classification
multimedia
Learning Task ©Getty
data sets
Images
©Air Force Research Lab
Multimedia Data
An operator is
Autonomy ArduCopter V3.0.1 ArduCopter V3.0.1 ArduCopter V3.0.1
Actions
directing
autonomous
Explainable Explanation
Reinforcement Model Interface systems to
Explanation accomplish a
Learning Task
ArduCopter V3.0.1
click the images above for “Flight versions”
ArduCopter V3.0.1 ArduCopter V3.0.1
series of
Download firmwares PX4IO Firmware
©ArduPikot.org
Load custom firmware missions
©ArduPikot.org Beta Firmwares Pick provious firmware
©US Army
ArduPilot & SITL Simulation
analytics, DEXAI will be demonstrated using visual and how changing these factors would affect this out-
question answering (VQA) and filtering tasks (for come. Generative probabilistic models, represented in
example, using large-scale data sets such as VQA-X a probabilistic programming language, naturally ex-
and ACT-X for VQA tasks and activity recognition press causal relationships; they are well suited for this
tasks, respectively), xView, and Distinct Describable task of explaining ML systems.
Moments (Hendricks et al. 2018). CAMEL probes the internal representation of an ML
system to discover how it represents user-defined,
Causal Models to Explain Learning natural domain concepts. It then builds a causal
The goal of the Charles River Analytics (CRA) team model of their effect on the ML system’s operation by
(including researchers from the University of Massa- conducting experiments in which the domain con-
chusetts and Brown University) is to generate and cepts are systematically included or removed. CRA has
present causal explanations of ML operation, through applied this approach to DNNs for classification and
its causal models to explain learning (CAMEL) ap- RL.
proach. CAMEL explanations are presented to a user Once learned, it uses causal probabilistic models to
as narratives in an interactive, intuitive interface. infer explanations of the system’s predictions or ac-
CAMEL includes a causal probabilistic programming tions. Because inferences can be large and complex and
framework that combines representations and learning can contain many interacting components, CAMEL
methods from causal modeling (Marazopoulou et al. composes them into explanatory narratives that walk
2015) with probabilistic programming languages the user through the interactions of the major concepts
(Pfeffer 2016) to describe complex and rich phenomena. and their influence on the ML system’s output. The
CAMEL can be used to describe what an ML system did, CAMEL explanation interface, based on cognitive
how specific data characteristics influenced its outcome, systems engineering design principles and established
SUMMER 2019 49
Deep Learning and Security
Measure of Explanation
Effectiveness
User Satisfaction
Explanation Framework
• Clarity of the explanation (user rating)
Task • Utility of the explanation (user rating)
HCI techniques, allows users to understand and in- including interpretable DNNs, compositional graphical
teract with the explanatory narratives, engendering models such as and-or graphs, and models that produce
trust in automation and enabling effective user-system explanations at three levels (that is, compositionality,
teamwork. causality, and utility).
CRA is addressing both challenge problem areas. UCLA’s system includes a performer that executes
For data analytics, CAMEL has been demonstrated tasks on multimodal input data and an explainer that
using pedestrian detection (using the INRIA pedestrian explains its perception, cognitive reasoning, and de-
data set) (Harradon et al. 2018), and CRA is working cisions to a user. The performer outputs interpretable
toward activity recognition tasks (using ActivityNet). For representations in a spatial, temporal, and causal
autonomy, CAMEL has been demonstrated on the Atari parse graph (STC-PG) for three-dimensional scene
game Amidar, and CRA is working toward demon- perception (for analytics) and task planning (for au-
strating it on StarCraft II. tonomy). STC-PGs are compositional, probabilistic,
attributed, interpretable, and grounded on DNN
features from images and videos. The explainer out-
Learning and
puts an explanatory parse graph in a dialogue process
Communicating (She and Chai 2017), localizes the relevant subgraph
Explainable Representations in the STC-PG, and infers the user’s intent.
for Analytics and Autonomy The system represents explanations at three levels:
The University of California, Los Angeles (UCLA) team (1) concept compositions, represented by parse graph
(including researchers from Oregon State University and fragments that depict how information is aggregated
Michigan State University) is developing interpretable from its constituents and contexts, how decisions are
models that combine representational paradigms, made at nodes under uncertainty, and the decision’s
50 AI MAGAZINE
Deep Learning and Security
PHASE 1 PHASE 2
Technology Demonstrations Government Evaluations
Kickoff Progress Report Tech Demos Eval 1 Results Eval 2 Results Eval 3 Results
confidence levels; (2) causal and counterfactual rea- xACT enables domain experts to attach interpretable
soning, realized by extracting causal diagrams from STC- descriptions to x-features, and xDAP programmers to
PGs to predict what will happen if certain alternative annotate environment reward types and other con-
actions had been taken; and (3) utility explanations, cepts, which are automatically embedded into the
which explain why the system made certain decisions. dNNs as “annotation concepts” during learning.
UCLA is addressing both XAI challenge problem The dNN decisions can be explained via the de-
areas using a common framework of representation scriptions of relevant x-features and annotation con-
and inference. For data analytics, UCLA demonstrated cepts, which can be further understood via neural
their system using a network of video cameras for scene network saliency visualization tools. OSU is investi-
understanding and event analysis. For autonomy, gating the utility of saliency computations for
UCLA demonstrated it in scenarios using robots exe- explaining sequential decision making.
cuting tasks in physics-realistic virtual reality OSU’s explanation user interface allows users to
platforms and autonomous vehicle driving game navigate thousands of learned agent decisions and
engines. obtain visual and natural language (NL) explanations. Its
design is based on information foraging theory (IFT),
Explanation-Informed Acceptance which allows a user to efficiently drill down to the most
Testing of Deep Adaptive Programs useful explanatory information at any moment. The
assessment of rationales for learned decisions may more
Oregon State University (OSU) is developing tools for
efficiently identify flaws in the agent’s decision
explaining learned agents that perform sequential
making and improve user trust.
decision making and is identifying best principles for
OSU is addressing the autonomy challenge prob-
designing explanation user interfaces. OSU’s explain-
lem area and has demonstrated xACT in scenarios
able agent model employs explainable deep adaptive
using a custom-built real-time strategy game engine.
programs (xDAPs), which combine adaptive programs,
Pilot studies have informed the explanation user in-
deep RL, and explainability. With xDAPs, program-
terface design by characterizing how users navigate
mers can create agents by writing programs that in-
AI-agent game play and tend to explain game de-
clude choice points, which represent decisions that
cisions (Dodge et al. 2018).
are automatically optimized via deep RL through
simulator interaction. For each choice point, deep RL
attaches a trained deep decision neural network Common Ground
(dNN), which can yield high performance but is in- Learning and Explanation
herently unexplainable. The Palo Alto Research Center (PARC) team (including
After initial xDAP training, xACT trains an explana- researchers from Carnegie Mellon University, the Army
tion neural network (Qi and Li 2017) for each dNN. Cyber Institute, the University of Edinburgh, and the
These provide a sparse set of explanation features (x- University of Michigan) is developing an interactive
features) that encode properties of a dNN’s decision sensemaking system that can explain the learned ca-
logic. Such x-features, which are neural networks, are pabilities of an XAI system that controls a simulated
not initially human interpretable. To address this, unmanned aerial system.
SUMMER 2019 51
Deep Learning and Security
Explanation User
Quality Performance
Training Data
Explanation User Satisfaction
New
Explainable Explanation
Learning User’s Mental Better
Model Interface Model Performance
Process
User
Trust or Mistrust
Comprehension
Appropriate
CP Area Performer Explainable Model Explanation Interface Trust
Appropriate Use
An XAI system’s explanations should communicate system. The cognitive layer bridges human-usable
what information it uses to make decisions, whether it symbolic representations to the abstractions, composi-
understands how things work, and its goals. To tions, and generalized patterns.
address this, PARC’s common ground learning and COGLE’s explanation interfaces support performance
explanation (COGLE) and its users establish com- review, risk assessment, and training. The first provides a
mon ground about what terms to use in explana- map that traces an unmanned aerial systems’ mission
tions and their meaning. This is enabled by PARC’s actions and divides the action or decision (flight) path
introspective discourse model, which interleaves into explainable segments. The second interface’s
learning and explaining processes. tools enable users to examine and assess the sys-
Performing tasks in the natural world is challeng- tem’s competencies and make predictions about
ing for autonomous systems, requiring experience to mission performance.
create enough knowledge for reliable high perfor- COGLE will be demonstrated in ArduPilot’s Soft-
mance. COGLE employs K-models that encode an AI ware-in-the-Loop Simulator and a discretized abstract
agent’s domain-specific task knowledge. K-models orga- simulation test bed. It will be evaluated by drone op-
nize this knowledge into levels of elements, where higher erators and analysts. Competency-based evaluation will
(lower) levels model actions with longer range (local) help PARC to determine how best to develop appro-
effects. They support a competency-based framework priate domain understandable models.
that informs and guides the learning and testing of XAI
systems. Explainable Reinforcement Learning
COGLE’s multilayer architecture partitions its in- Carnegie Mellon University is creating a new disci-
formation processing into sensemaking, cognitive pline of explainable RL to enable dynamic human-
modeling, and learning. The learning layer employs machine interaction and adaptation for maximum
capacity constrained recurrent and hierarchical DNNs team performance. This effort has two goals: to de-
to produce abstractions and compositions over the velop new methods for learning inherently explain-
states and actions of unmanned aerial systems to sup- able RL policies and to develop strategies that can
port an understanding of generalized patterns. It explain existing black-box policies. For the former,
combines learned abstractions to create hierarchical, Carnegie Mellon is developing methods to improve
transparent policies that match those learned by the model learning for RL agents to capture the benefits
52 AI MAGAZINE
Deep Learning and Security
XAI Process
XAI Measurement
System Categories
Figure 8. Initial Model of the Explanation Process and Explanation Effectiveness Measurement Categories.
of model-based approaches (ability to visualize plans Carnegie Mellon is addressing the autonomy chal-
in the internal model space), while integrating lenge problem area and has demonstrated XRL in
the benefits of model-free approaches (simplicity several scenarios, including OpenAI Gym, Atari
and higher ultimate performance). These include games, autonomous vehicle simulation, mobile service
methods that incrementally add states and actions to robots, and self-improving educational software and
world models after discovering relevant latent in- games.
formation, learn models via end-to-end training
of complex model-based optimal control policies,
Deep Attention-Based
learn general DL models that directly integrate Representations for Explanation/
and exploit rigid body physics (Belbute-Peres and Explainable Generative Adversarial Networks
Kolter 2017), and learn understandable predictive SRI International’s team (including researchers from
state representations using recurrent architectures the University of Toronto, the University of Guelph,
(Hefny et al. 2018). and the University of California, San Diego) is de-
Carnegie Mellon is also developing methods that veloping an explainable ML framework for multi-
can explain the actions and plans of black-box RL modal data analytics that generates show-and-tell
agents, observed either online or from system logs. explanations with justifications of decisions accom-
This involves answering questions such as, why did panied by visualizations of input data used to generate
an agent choose a particular action? or, what train- inferences.
ing data were most responsible for this choice? To The deep attention-based representations for
achieve this, Carnegie Mellon developed techniques explanationxplainable generative adversarial networks
that generate NL descriptions of agents from behavior (DARE/X-GANS) system employs DNN architectures
logs and detect outliers or anomalies. Carnegie Mellon inspired by attentional models in visual neuroscience.
also developed improvements over traditional in- It identifies, retrieves, and presents evidence to a user as
fluence function methods in DL, allowing its XRL part of an explanation. The attentional mechanisms
system to precisely identify the portions of a train- provide a user with a means for system probing and
ing set that most influence a policy’s outcome. collaboration.
SUMMER 2019 53
Deep Learning and Security
Measure Description
ML Model performance
Various measures (on a per-challenge problem area Accuracy/performance of the ML model in its given domain (to
basis) understand whether performance improved or degraded relative to
state-of-the-art nonexplainable baselines)
Explanation Effectiveness
Explanation goodness Features of explanations assessed against criteria for explanation
goodness
Explanation satisfaction User’s subjective rating of explanation completeness, usefulness,
accuracy, and satisfaction
Mental model understanding User’s understanding of the system and the ability to predict the system’s
decisions/behavior in new situations
User task performance Success of the user performing the tasks for which the system is designed
to support
Appropriate Trust and Reliance User’s ability to know when to, and when not to, trust the system’s
recommendations and decisions
DARE/X-GANS uses generative adversarial net- interactive, explorable explanations of why it derived
works (GANs), which learn to understand data by an answer.
creating it, while learning representations with explan- The explainable question answering system
atory power. GANs are made explainable by using (EQUAS) learns explainable DNN models in which
interpretable decoders that map unsupervised clus- internal structures (for example, individual neurons)
ters onto parts-based representations. This involves have been aligned to semantic concepts (for example,
generating visual evidence, given text queries, using wheels and handlebars) (Zhou et al. 2015). This allows
text-to-parts generation (Vicol et al. 2018), the parts neural activations within the network, during a de-
being interpretable features such as human poses or cision process, to be translated to NL explanations (for
bounding boxes. This evidence is then used to example, “this object is a bicycle because it has two
search the queried visual data. wheels and handlebars”). EQUAS also uses neural
The system presents explanations of its answers visualization techniques to highlight input regions
based on visual concepts extracted from the mul- associated with neurons that most influenced its de-
timodal data inputs and knowledge base queries. cisions. To express case-based explanations, EQUAS
Given explanatory questions, it provides justifica- retains indexes and retrieves cases from its training
tions, visual evidence used for decisions, and vi- data that support its choices. Rejected alternatives are
sualizations of the system’s inner workings. This recognized and ruled out using contrastive language,
show-and-tell explanation interface ensures highly visualization, and examples. Four explanation mo-
intuitive explanations, made possible by atten- dalities map to key elements of argument construc-
tional modules that localize evidence used for tion and interactive pedagogy: didactic statements,
each visual task. Initial studies demonstrate that visualizations, cases, and rejections of alternative
such explanations substantially improve user task choices.
performance. The EQUAS explanation interface allows users to
SRI is addressing the data analytics challenge explore the explanation space populated by these
problem area and has demonstrated DARE/X-GANs explanation modes. It enables iterative and guided
using VQA and multimodal QA tasks with image and collaborative interaction, allowing users to drill down
video data sets. into the supporting evidence from each explanation
category.
Explainable Question Answering System Raytheon BBN is addressing the analytics chal-
The Raytheon BBN Technologies team (including lenge problem area and has demonstrated initial
researchers from the Georgia Institute of Technology, EQUAS capabilities on VQA tasks for images, ex-
MIT, and the University of Texas, Austin) is devel- ploring how different explanation modalities en-
oping a system that answers unrestricted NL questions able users to understand and predict the behavior of
posed by users about multimedia data and provides the underlying VQA system.
54 AI MAGAZINE
Deep Learning and Security
SUMMER 2019 55
Deep Learning and Security
Table 2. Summary of Explainable Learning System Developer Approaches and Selected Phase 1 Test Problems.
56 AI MAGAZINE
Deep Learning and Security
subset that is most representative of a model’s in- Trevor Darrell (UCB), Brian Ruttenberg and Avi Pfeffer
ference. Rutgers’ approach allows for explanation of the (CRA), Song-Chun Zhu (UCLA), Alan Fern (OSU),
inferences of any probabilistic generative and discrimina- Mark Stefik (PARC), Zico Kolter (Carnegie Mellon),
tive model, as well as influential DL models (Yang Mohamed Amer and Giedrius Burachas (SRI Interna-
and Shafto 2017). tional), Bill Ferguson (Raytheon BBN), Vibhav Gogate
Rutgers is also developing a formal theory of human- (UTD), Xia (Ben) Hu (TAMU), Patrick Shafto (Rutgers),
machine cooperation and supporting interactive and Robert Hoffman (IHMC). The authors owe a
guided explanation of complex compositional models. special thanks to Marisa Carrera for her exceptional
Common among these is a core approach of building technical support to the XAI program and her ex-
from models of human learning to foster explainability tensive editing skills.
and carefully controlled behavioral experiments to
quantify explainability.
Explanation by Bayesian teaching inputs a data set,
References
a probabilistic model, and an inference method and Belbute-Peres, F., and Kolter, J. Z. 2017. A Modular Differ-
returns a small subset of examples that best explains entiable Rigid Body Physics Engine. Paper presented at the
Neural Information Processing Systems Deep Reinforcement
the model’s inference. Experiments with unfamiliar
Learning Symposium. Long Beach, CA, December 7.
images show that explanations of inferences about
Chakraborty, S.; Tomsett, R.; Raghavendra, R.; Harborne, D.;
categories of (and specific) images increase the accu-
Alzantot, M.; Cerutti, F.; and Srivastava, M.; et al. 2017.
racy of people’s reasoning about a model (Vong et al. Interpretability of Deep Learning Models: A Survey of Re-
2018). Experiments with familiar image categories sults. Presented at the IEEE Smart World Congress 2017
show that explanations allow users to accurately Workshop: DAIS 2017 — Workshop on Distributed Ana-
calibrate trust in model predictions. lytics Infrastructure and Algorithms for Multi-Organiza-
Explanation of complex models is facilitated by tion Federations, San Francisco, CA, August 4–8. doi.org/10.
interactive guided explanations. By exploiting com- 1109/UIC-ATC.2017.8397411
positionality and cooperative modifications of ML Dodge, J.; Penney, S.; Hilderbrand, C.; Anderson, A.; and
models, Rutgers provides a generic approach to Burnett, M. 2018. How the Experts Do It: Assessing and
fostering understanding via guided exploration. In- Explaining Agent Behaviors in Real-Time Strategy Games. In
teraction occurs through an interface that exposes Proceedings of the 2018 CHI Conference on Human Factors in
Computing Systems. New York: Association for Computing
model structure and explains each component with
Machinery. doi.org/10.1145/3173574.3174136
aspects of the data. The Rutgers approach has been
demonstrated to facilitate understanding of large text Du, M.; Liu, N.; Song, Q.; and Hu, X. 2018. Towards Ex-
planation of DNN-Based Prediction and Guided Feature In-
corpora, as assessed by a human’s ability to accurately
version. In Proceedings of the 24th ACM SIGKDD International
summarize the corpus after short, guided explanations. Conference on Knowledge Discovery and Data Mining, 1358–67.
Rutgers is addressing the data analytics challenge New York: Association for Computing Machinery. doi.org/
problem area and has demonstrated its approach on 10.1145/3219819.3220099
images, text, combinations of these (for example, Gao, J.; Liu, N.; Lawley, M.; and Hu, X. 2017. An In-
VQA), and structured simulations involving temporal terpretable Classification Framework for Information Ex-
causal structure. traction from Online Healthcare Forums. Journal of
Healthcare Engineering: 2460174. doi.org/10.1155/2017/
2460174
Conclusions and Future Work Gogate, V., and Domingos, P. 2016. Probabilistic Theorem
DARPA’s XAI program is developing and evaluating a Proving. Communications of the ACM 59(7): 107–15. doi.org/10.
wide variety of new ML techniques: modified DL 1145/2936726
techniques that learn explainable features; methods Harradon, M.; Druce, J.; and Ruttenberg, B. 2018. Causal
that learn more structured, interpretable, causal Learning and Explanation of Deep Neural Networks via
models; and model induction techniques that infer an Autoencoded Activations. arXiv preprint. arXiv:1802.00541v1
explainable model from any black-box model. One year [cs.AI]. Ithaca, NY: Cornell University Library.
into the XAI program, initial technology demonstra- Hefny, A.; Marinho, Z.; Sun, W.; Srinivasa, S.; and Gordon, G.
2018. Recurrent Predictive State Policy Networks. In Pro-
tions and results indicate that these three broad strate-
ceedings of the 35th International Conference on Machine
gies merit further investigation and will provide future
Learning, 1954–63. International Machine Learning Society.
developers with design options covering the perfor-
Hendricks, L. A.; Hu, R.; Darrell, T.; and Akata, Z. 2018.
mance versus explainability trade space. The developer
Grounding Visual Explanations. Presented at the European
teams’ XAI systems are being evaluated to assess the Conference of Computer Vision (ECCV). Munich, Germany;
value of explanations that they provide, localizing the September 8–14. doi.org/10.1007/978-3-030-01216-8_17
contributions of specific techniques within this Hoffman, R.; Miller, T.; Mueller, S. T.; Klein, G.; and Clancey,
trade space. W. J. 2018. Explaining Explanation, Part 4: A Deep Dive on
Deep Nets. IEEE Intelligent Systems 33(3): 87–95. doi.org/10.
Acknowledgments 1109/MIS.2018.033001421
The authors thank the XAI development teams, Hoffman, R. R.; and Klein, G. 2017. Explaining Explanation,
specifically their principle investigators, for their Part 1: Theoretical Foundations. IEEE Intelligent Systems 32(3):
innovative research and contributions to this article: 68–73. doi.org/10.1109/MIS.2017.54
SUMMER 2019 57
Deep Learning and Security
Hoffman, R. R., Mueller, S. T.; and Klein, G. 2017. Explaining 1634–44. Stroudsburg, PA: Association for Computation
Explanation, Part 2: Empirical Foundations. IEEE Intelligent Linguistics. doi.org/10.18653/v1/P17-1150
Systems 32(4): 78–86. doi.org/10.1109/MIS.2017.3121544 Vicol, P.; Tapaswi, M.; Castrejon, L.; and Fidler, S.
Hu, R.; Andreas, J.; Rohrbach, M.; Darrell, T.; and Saenko, K. 2018.MovieGraphs: Towards Understanding Human-Cen-
2017. Learning to Reason: End-to-End Module Networks for tric Situations from Videos. In IEEE Conference on Computer
Visual Question Answering. In Proceedings of the IEEE In- Vision and Pattern Recognition. New York: IEEE. doi.org/10.
ternational Conference on Computer Vision, 804–13. New York: 1109/CVPR.2018.00895
IEEE. doi.org/10.1109/ICCV.2017.93 Vong, W.-K.; Sojitra, R.; Reyes, A.; Yang, S. C.-H.; and Shafto,
Huang, S. H.; Bhatia, K.; Abbeel, P.; and Dragan, A. 2018. P. 2018. Bayesian Teaching of Image Categories. Paper pre-
Establishing Appropriate Trust via Critical States. Pre- sented at the 40th Annual Meeting of the Cognitive Science
sented at the 13th Annual ACM/IEEE International Society (CogSci). Madison, WI, July 25–28.
Conference on Human-Robot Interaction Workshop on Yang, S. C.-H., and Shafto, P. 2017. Explainable Artificial In-
Explainable Robot Behavior. Madrid, Spain; October 1–5. telligence via Bayesian Teaching. Paper presented at the 31st
doi.org/10.1109/IROS.2018.8593649 Conference on Neural Information Processing Systems
Kim, J., and Canny, J. 2017. Interpretable Learning for Workshop on Teaching Machines, Robots and Humans. Long
Self-Driving Cars by Visualizing Causal Attention. In Beach, CA, December 9.
Proceedings of the International Conference on Computer Vi- Zhou, B.; Khosla, A.; Lapedriza, A.; Oliva, A.; and Torralba, A.
sion, 2942–50. New York: IEEE. doi.org/10.1109/ICCV. 2015. Object Detectors Emerge in Deep Scene CNNs. Paper
2017.320 presented at the International Conference on Learning Rep-
Klein, G. 2018. Explaining Explanation, Part 3: The Causal resentations. San Diego, CA, May 7–9.
Landscape. IEEE Intelligent Systems 33(2): 83–88. doi.org/10.
1109/MIS.2018.022441353
David Gunning is a program manager in DARPA’s In-
Letham, B.; Rudin, C.; McCormick, T. H.; and Madigan, D.
formation Innovation Office, as an Intergovernmental Per-
2015. Interpretable Classifiers Using Rules and Bayesian
sonnel Act from the Pacific Northwest National Labs.
Analysis: Building a Better Stroke Prediction Model. Annals of
Gunning has more than 30 years of experience in developing
Applied Statistics 9(3): 1350–71. doi.org/10.1214/15-
AI technology. In prior DARPA tours he managed the PAL
AOAS848
project that produced Siri and the CPOF project that the US
Marazopoulou, K.; Maier, M.; and Jensen, D. 2015. Learning Army adopted as its C2 system for use in Iraq and Afgha-
the Structure of Causal Models with Relational and Temporal nistan. Gunning was also a program director at PARC, a
Dependence. In Proceedings of the Thirty-First Conference on senior research manager at Vulcan, senior vice president at
Uncertainty in Artificial Intelligence, 572–81. Association for SET Corporation, vice president of Cycorp, and a senior
Uncertainty in Artificial Intelligence. scientist at the Air Force Research Labs. Gunning holds an MS
Miller, T. 2017. Explanation in Artificial Intelligence: Insights in computer science from Stanford University and an MS in
from the Social Sciences. arXiv preprint. arXiv:1706.07269v1 cognitive psychology from the University of Dayton.
[cs.AI]. Ithaca, NY: Cornell University Library.
David W. Aha is acting director of NRL’s Navy Center for
Park, D. H.; Hendricks, L. A.; Akata, Z.; Rohrbach, A.; Schiele,
Applied Research in AI in Washington, D.C. His interests
B.; Darrell, T.; and Rohrbach, M. 2018. Multimodal Expla-
include goal reasoning, XAI, case-based reasoning, and
nations: Justifying Decisions and Pointing to the Evidence.
machine learning, among other topics. He has coorganized
In Proceedings of the IEEE Conference on Computer Vision and
many events on these topics (for example, the IJCAI-18 XAI
Pattern Recognition. New York: IEEE. doi.org/10.1109/CVPR.
Workshop), launched the UCI Repository for ML Databases,
2018.00915
served as an AAAI Councilor, and leads DARPA XAI’s eval-
Pfeffer, A. 2016. Practical Probabilistic Programming.
uation team.
Greenwich, CT: Manning Publications.
Qi, Z., and Li, F. 2017. Learning Explainable Embeddings for
Deep Networks. Paper presented at the NIPS Workshop on
Interpreting, Explaining and Visualizing Deep Learning.
Long Beach, CA, December 9.
Ramanishka, V.; Das, A.; Zhang, J.; and Saenko, K. 2017. Top-
Down Visual Saliency Guided by Captions. In Proceedings of the
30th IEEE Conference on Computer Vision and Pattern Recognition,
7206–15. New York, IEEE.
Ras, G.; van Gerven, M.; and Haselager, P. 2018 Explanation
Methods in Deep Learning: Users, Values, Concerns and
Challenges. arXiv preprint. arXiv:1803.07517v2 [cs.AI].
Ithaca, NY: Cornell University Library.
Ribeiro, M. T.; Singh, S.; and Guestrin, C. 2016. “Why Should
I Trust You?”: Explaining the Predictions of Any Classifier. In
Proceedings of the 22nd ACM SIGKDD International Conference on
Knowledge Discovery and Data Mining, 1135–44. New York:
Association for Computing Machinery. doi.org/10.1145/
2939672.2939778
She, L., and Chai, J. Y. 2017. Interactive Learning for Ac-
quisition of Grounded Verb Semantics towards Human-Ro-
bot Communication. In Proceedings of the 55th Annual
Meeting of the Association for Computational Linguistics, vol. 1,
58 AI MAGAZINE