You are on page 1of 9

COVER FEATURE EXPLAINABLE ARTIFICIAL INTELLIGENCE AND MACHINE LEARNING

Toward Human–AI
Interfaces to Support
Explainability and
Causability in Medical AI
Andreas Holzinger and Heimo Müller, Medical University Graz

Our concept of causability is a measure of whether and to what


extent humans can understand a given machine explanation.
We motivate causability with a clinical case from cancer research.
We argue for using causability in medical artificial intelligence
(AI) to develop and evaluate future human–AI interfaces.

A
chieving human-level artificial intelligence inputs for the classification of skin lesions. For pretraining,
(AI) has been the ambition since the emer- they used 1.3 million images from the 2014 ImageNet chal-
gence of this field. Because of the availabil- lenge and then 130,000 clinical images with about 2,000
ity of big data and the necessary computing different diseases. The results, with an average classifica-
power, statistical machine learning, especially deep learn- tion performance of 92%, were on par with, or even better
ing, has now made tremendous progress, even in domains than, those of human dermatologists. This is a remarkable
as complex as medicine. For example, the work of the Stan- achievement, and there is no question that AI will become
ford machine learning group on dermatology1 was popu- very important for medicine.
larized in Europe as “AI is better than doctors.” The group Despite these impressive successes, we must be aware
trained a deep learning model directly from dermatolog- that these previous approaches rely on statistical mod-
ical images by using only pixels and disease labels as the el-free learning. Relying solely on statistical correlations
can be very dangerous, especially in medicine, because
Digital Object Identifier 10.1109/MC.2021.3092610
correlation must not be confused with causality, which
Date of current version: 24 September 2021 is completely missing in current AI. This is a general
This work is licensed under a Creative Commons
Attribution 4.0 License. For more information,
78 CO M P UTE R P U B LISH ED BY TH E I EEE COM P UTER SOCI E T Y see https://creativecommons.org/licenses/by/4.0/deed.ast.
problem. A specific issue is that these to sensemaking and understanding in of the derivative, where the deep neural
methods are so complex, high dimen- the context of an application domain. network is seen as a function and the
sional, nonlinear, and nonconvex that On the other hand, AI is vulnerable to explanation relies on the function’s gra-
it is difficult, if not impossible, even for even small perturbations. We empha- dient, which is available from the back-
domain experts, to understand how the size that the current approaches not propagation algorithm.8
results were obtained. Consequently, only lack robustness and generalization, Another possibility is to use decom-
these techniques are called black-box but more importantly, they are unable position methods, that is, to break up the
models. Both problems motivate us to to build causal models to support deep more complex larger parts into smaller,
make use of the expertise of a human understanding in the user.5 To make more manageable, parts, for example,
in the loop. A human expert can some- the current AI even more successful, we pixel-wise relevance propagation, lay-
times—but not always—contribute ex­­ believe that we should try to take advan- er-wise relevance propagation, or deep
perience, conceptual knowledge, and tage of the respective benefits of both Taylor decomposition. These very ver-
context understanding. statistical machine learning methods satile approaches also work on graph-
A very recent work on histopathol- and model-based approaches. More based data. 9 Other methods include
ogy2 showed that machine-generated precisely, we envision for the future deconvolution, which involves revers-
features correlate with certain mor- to interactively integrate the existing ing the effects of convolution and gen-
phological constructs and ontological implicit a priori knowledge, human erating from two functions a third func-
relationships generated by humans. experience, and conceptual understand- tion that is then the product of both, as
Such overlaps between humans and AI ing of human experts into statistical well as guided backpropagation.10
can help to solve the current problems, learning methods. All of these methods constitute an
and human experience can contribute This could result in a new hybrid excellent preprocessing step. Here we
to improve algorithm robustness and approach that fully exploits the advan- want to emphasize two issues: 1) results
explainability, which are regarded as the tages of data-driven machine learning are human interpretable when they clas-
grand challenges of current AI.3 Robust- methods and also integrates human sify objects on the basis of features that
ness is closely related to the ability to conceptual understanding, according to a human can perceive and understand
generalize, and it is no coincidence that individual needs and depending on the and 2) in the medical domain, there is
we mention robustness and interpret- problem. Such an approach will require a pressing need for a medical expert to
ability together: robustness is a ubiq- appropriate human–AI interfaces that be able to understand the causality of
uitous feature of biological systems, enable seamless interaction with ma­­ a learned representation and use it for
as we humans are. It ensures that spe- chine learning methods. However, let medical decision support. This is called
cific functions of the system are main- us first summarize the achievements etiology in medicine: the science of
tained—despite external and/or inter- of the explainable AI (XAI) community. causes and effects of pathologies.
nal perturbations. We all know that our
best machine learning algorithms are EXPLAINABILITY EXPLAINABILITY AND
very fragile and sensitive to even small The XAI community is very active in CAUSABILITY
distortions, arising from the poor data developing methods for making black- The terms interpretation and explanation
quality in the medical domain. box approaches explainable.6 The suc- are often used synonymously. Before
Consequently, a doctor in the loop4 cessful approaches focus on visualiz- we introduce the concept of causability,
will play an important role in medical ing the elements that have contributed we distinguish between interpretation,
AI, at least in the foreseeable future. to each decision, for example, through which can be defined as a mapping of
Humans are generally robust, and they heatmapping, which means highlight- an abstract statement into a domain or
can sometimes add expertise, con- ing which input parameters contribute space that the human expert can per-
ceptual understanding, and implicit most to a certain classification result.7 ceive, comprehend, and understand,
knowledge to machine learning pro- Such “mechanical explanations” can and explanation, which can be defined
cesses. Although humans make mis- be reached by using various procedures. as a collection of the features of the
takes, they can adapt quickly, impro- The simplest method works with gradi- interpretable domain or space that
vise, and exhibit a plasticity that leads ents as a multivariable generalization have contributed to a given example

OCTOBER 2021 79
EXPLAINABLE ARTIFICIAL INTELLIGENCE AND MACHINE LEARNING

to produce a statement. In an ideal 3. Causability: Causability is the when trying to achieve an explainable
world, both human and AI statements measurable extent to which medicine, this is indispensable.14
would be identical and congruent with an explanation of a statement
the ground truth, which is defined for AI to a human expert achieves a HUMAN–AI INTERFACES:
and humans equally. specified level of causal under- EFFECTIVE MAPPING OF
However, in the real world we face standing with effectiveness, EXPLAINABILITY WITH
two problems. 1) The ground truth can- efficiency, and satisfaction in CAUSABILITY
not always be well defined, especially a specified context of use. As The key to effective human–AI inter-
when making a medical diagnosis. causability is measured in terms action and, consequently, the success
2) Human (scientific) models are often of effectiveness, efficiency, and of future human–AI interfaces lies in
based on causality as an ultimate aim for (human) satisfaction related to an efficient and consistent mapping
understanding the underlying explana- causal understanding and its of explainability with causability. This
tory mechanisms, and while correla- transparency for an expert user, “mapping” (or “map metaphor”) is about
tion is accepted as a basis for decisions, it refers to a human-understand- establishing connections and relation-
it must be viewed as an intermediate able model. ships between existing areas, not about
step. This is highly relevant in the med- drawing a new map. Rather, it is about
ical domain because of the importance This is always possible for an expla- identifying the same, or at least simi-
of validity and the necessity to build nation of a human statement, as the lar, areas in two completely different
human trust and also the need to build explanation is, per se, defined related to “maps” of AI explainability and human
“AI experience.” 11 As we have men- a human model. However, to measure causability. That is why mapping is a
tioned, the most successful algorithms the causability of an explanation of a very good term. Effective and efficient
are based on probabilistic models and machine statement, it must be based on mapping is necessary, but of course it
provide only a rudimentary basis for a causal model in the sense of Pearl,12 still will not be sufficient for under-
establishing causal models. Conse- that is, to represent causal relationships standing an explanation. Whether and
quently, when we discuss the explain- and to allow inferences about those to what extent an explanation has been
ability of a machine statement, we have causal relationships from data. This is understood depends on additional fac-
to carefully distinguish among the fol- not the case for most AI algorithms, so a tors, including prior knowledge and
lowing terms: mapping between both must be defined. expectations on the human side. To fully
Here we must distinguish between the understand the importance of this map-
1. Explainability: In a technical explainable model (XAI) and an expla- ping, let us look at Figure 1.
sense, explainability highlights nation interface, which makes the results In Figure 1 we see that an explana-
decision-relevant parts of the gained in the explainable model not tion statement s can either be made by
used machine representations only usable but also useful to the expert. a human sh or a machine sm, where s is
of the algorithms and active When is it useful? It is useful when a function s = f (r, k, c) with the follow-
parts in the algorithmic model the system is able to provide causes of ing parameters: r, representations of an
that contribute either to the observed phenomena13 in a comprehen- unknown (or unobserved) fact ue related
model accuracy on the training sible and interactive manner. This is to an entity; k, preexisting knowl-
set or to a specific prediction done through, for example, a linguistic edge, which is embedded in an algo-
for one ­particular observation. description or question/answer dialogs rithm for a machine or made up for a
It does not refer to an explicit of its logical and causal relationships, human by explicit, implicit, and/or tacit
human model. enabling one to understand the causal- knowledge; and c, context, which for a
2. Usability: This term refers to the ity of the learned representations rele- machine is the technical runtime envi-
measurable extent to which a vant not only for sensemaking, but also ronment and for a human is the physical
system achieves a specified level for judging the quality of explanations. environment in which the decision was
of usability for a user with effec- This might be unnecessary for certain made (the pragmatic dimension). An
tiveness, efficiency, and satisfac- areas where we want full automation, unknown (or unobserved) fact ue rep-
tion in a specified context of use. but for the medical domain, especially resents a ground truth gt that we try to

80 COMPUTER  W W W.CO M P U T E R .O R G /CO M P U T E R


model with a machine mm or as a human human and the machine statement are Causability is the extent to which
mh. Such unobserved variables can be congruent (mh ≡ mm) and identical to an explanation of a statement to a user
found, for example, in Bayesian models the ground truth, which is defined for achieves a specified level of causal under-
or in hidden Markov models (a special machines and humans within the same standing with effectiveness, efficiency,
case of a dynamical Bayesian network). framework. The problem in the medical and satisfaction in a specified context of
In summary, we can state that caus- domain is that the ground truth is not use. As causability is measured in terms
ability as a measure of whether and to well defined and the most successful of effectiveness, efficiency, satisfaction
what extent something is understood machine learning models are based on related to causal understanding, and
consists of two parts. The first part is correlation or related concepts of sim- its transparency for a user, it refers to a
whether the human understands or ilarity and distance. All of this is prob- human-understandable model mh. This
can understand the given explanation abilistic in nature and therefore must is always possible for an explanation of
at all, and the second part is to what be considered as an intermediate step a human statement as the explanation
extent he/she can understand it, that that can only provide a basis for further is intrinsically defined related to mh. To
is, it is expressed in terms of measuring causal model building. Explainability conclude, to measure the causability of
the causability of the explanation em of in a technical sense highlights the deci- an explanation em of a machine state-
a machine statement sm under a certain sion-relevant parts of the machine rep- ment sm, either mh has to be based on a
machine model mm (see Figure 1). resentations rm and machine models causal model (which is not the case for
The central goal is that the state- mm, that is, the parts that contributed most machine learning algorithms) or
ment s is identical with the ground to model accuracy in training or to a a mapping between mm and mh must
truth gt and that a given explanation specific prediction. It is important to be defined.
of this statement is a fit to this ground emphasize that explainability does not As we can imagine, this is not triv-
truth. In an idealistic situation, both the refer to a human model mh. ial because future human–AI interfaces

kh, ch
mh rh sh eh
gt ue u
km, cm
mm rm sm em
User
Model

Ground
Truth

Unknown

Statement
Knowledge
Context

Explanation
Representation

FIGURE 1. The process of explanation. Explanations (e) by humans and machines (subscripts h and m) must be congruent with state-
ments (s) and models (m), which, in turn, are based on the ground truth (gt). Statements are a function of representations (r), knowledge
(k), and context (c). (Source: Holzinger et al. 2020,15 used with permission.)

OCTOBER 2021 81
EXPLAINABLE ARTIFICIAL INTELLIGENCE AND MACHINE LEARNING

should allow a constant feedback, in­­ spending on what. The longer a user one gradually goes “deeper” into the AI
dicating whether and to what extent looks at a particular word, sentence, system, the information flow decreases;
something has been understood, or paragraph, or data feature, the more at the same time, the semantic richness
not. In a human-to-human interaction, important the system will assume it to (SR) of potential information objects
this feedback is very much provided by be for that user. A very important aspect increases. In traditional human–com-
facial expressions. Consequently, con- lies in fusing the information from dif- puter interactions, the information
cepts of “emotional interfaces”16 will ferent sensor signals and also allowing flow is extremely asymmetrical; that
become an important part of future the metrics of importance to be changed is, much more information is shown
conversational interfaces for XAI, and in the background, thereby automati- by high-resolution displays compared
dialog systems will become important cally “bringing up” features of the great- to mouse and/or textual input—not to
for these future “explanation inter- est similarity or utility—augmented by mention other input modalities (see the
faces” where affective computing will a potential explanation of how and why. dotted line in Figure 2).
play an important role. Generally, affec- The user will similarly be able to weight
tive computing17 describes a system in the importance of the affective element A CLINICAL CASE
which the machine interacts with a via sophisticated interaction elements A clinical case of lung cancer will serve
human by observing his/her actions (face recognition, gestures, and so on). as a practical example. The survival
and varying affordances accordingly. Other tools for navigating the data will of patients with lung cancer can be
Here it is important to classify, filter, include a virtual “important features” improved by the use of immunothera-
and make use of affective computing pile (possibly on the right of the screen) pies directed against programmed cell
methods to that can be selected by gaze or mouse death 1-ligand 1 (PD-L1) and its receptor
click, which can interactively be over- PD-1. PD-L1 is a surface protein involved
1. measure the effectiveness of ridden to show any features of interest. in the inhibition of the immune response.
explainability (does the user Alternatively, the user data will be “cap- PD-L1 protein expression has emerged
really understand a given expla- tured” without any user interaction; as an effective biomarker that can pre-
nation), which can be measured that is, through a set of rich sensors, the dict which patients are more likely to
via sensors by the success rate to “machine” will get reliable information respond to immunotherapy.19 Whether
which a given explanation has about the understanding. a tumor can spread in the body depends
been understood by humans; see There are two main types of sensor on whether one’s immune system rec-
the examples in the section “A input data: 1) facial expression analy- ognizes the degenerate cells leading
Clinical Case” sis by multiple cameras and 3D sensors to the tumor as a potential danger and
2. adapt the visual communication (compare the recent developments in attacks them. The immune system
according to the mental model (a augmented reality toolkits) and 2) gaze scans all cells to distinguish the body’s
priori knowledge) of the user. patterns—which can be experimentally own cells from foreign ones. T cells
contrasted and fused with other sen- are an important component of the
Using advanced biometric technol- sor data—and they can be very useful. immune system in this process. Cancer
ogy including 2D/3D cameras and eye These data can be used to exploit affec- cells can be recognized by T cells, but
trackers, heart sensors (for electro- tive computing methods to massively they are not attacked by the immune
myography, electrocardiography, pho- improve the affordances generated by system because they are able to cam-
toplethysmography, and galvanic skin a system,18 using the fact that there are ouflage themselves by producing the
response), acoustic microphone arrays, computational models that are associ- protein PD-L1. PD-L1 is like a disguise
and other sensors (potentially machine ated with human behavior. that helps cancer cells to conceal them-
vision), an affective computing envi- In Figure 2, we outline a model for selves and remain undetected, and one
ronment may enhance the effective- the information flow between humans way to detect PD-L1 is by using immu-
ness of navigation. In particular, when and an AI system. On the interaction nohistochemistry. In Figure 3 we see
an expert user is actually reading text surface, which can be seen as a “bor- a typical part of a whole slide image
or looking at an image, the system der” between human intelligence and that is used in an ongoing validation
will “know” how much time he/she is AI, the information flow is maximal. As study (see the ethics declaration in the

82 COMPUTER  W W W.CO M P U T E R .O R G /CO M P U T E R


Human Machine
Information Flow
Displays For Example, Bits/s
Human Perception

Explainability

SR
Mental Model For Example,
SRscore, Entropy

Sensors
Machine Perception

FIGURE 2. On the upper part of vertical line, explanations are visualized and displayed to the user. Parallel to the visualization,
sensor data are captured and analyzed (see the lower part of vertical line) with less information density. Only when a rich set of
sensor inputs, for example, cameras, microphones, and movement sensors, is used can the explanation process adapt to the
prior knowledge of the user. Semantic richness (SR) can be measured via scores and entropy measures and used for
appropriate feedback.

“Acknowledgments”), focusing both on


the clinical performance and also on
the human–AI interface.
To this end, the pathologist surveys
the tissue by encircling PD-L1 stained
immune cell (IC) scored aggregates (the IC
score indicates the percentage of the area
of PD-L1 positive immune cells from the
area of vital tumor cells). The pathologist
then places them in relation to the tumor
area and can estimate the percentage rele-
vant for making a diagnosis.
In Figures 4 and 5 we can see the pro-
cess assisted by an AI algorithm. With the
results in Figure 4, the pathologist can see
from the corner of his/her eye that the
explanation of the algorithm is correct;
thus, there is congruence, as previously
explained in the model, and the patholo-
gist will be satisfied and in agreement.
FIGURE 3. The human pathologist sees the tumor above and below in the image. In the
However, in Figure 5, we see an image
middle the stroma can be seen (the tumor identification is 100% positive).
where the AI solution found the wrong

OCTOBER 2021 83
EXPLAINABLE ARTIFICIAL INTELLIGENCE AND MACHINE LEARNING

S
tumor areas. In this case, the emotion be measured very precisely by sensors tatistical machine learning is
on the pathologist’s face will be imme- and are therefore also comparable, extremely successful and has
diately visible, and the pupil widths will will show that there is no congruence made AI very popular again,
increase, which can be easily detected between the explanation provided by even in the complex application area of
by the sensors of modern eye-tracking the machine and the explanation of the medicine. Whether AI will be used or
systems. These indicators, which can human pathologist. not in this field is not up for debate; in

(a) (b)

FIGURE 4. (a) A visualization of the benign (nontumorous area) in the middle and the malignant (tumorous areas) above and below. (b)
The detected positive cells are marked. If a pathologist evaluates this, he/she intuitively sees with “one look” that the explanation of the
AI solution is correct and sufficient. In this case, there is congruence between machine and human (mh ≡ mm).

(a) (b)

FIGURE 5. (a) The same visualization method is applied, but now (b) the AI solution found the wrong tumor areas (in the middle); also the cell
detection, for both positive and negative cells, did not work. In this case we have incongruence between machine and human. (mh mm).

84 COMPUTER  W W W.CO M P U T E R .O R G /CO M P U T E R


the future, the use of medical AI will be following: How do physicians make (CY-Biobank), 824087 (European Open
practically indispensable. causal judgments? What role, if any, do Science Cloud-Life), 874662 (Human
However, the most successful AI meth- counterfactuals play in this process? Exposome Assessment Platform), and
ods are so-called black boxes, which From theory, we know that counterfac- 826078 (Feature Cloud). Parts of this
means that they are so complicated tual theories of causal judgments pre- work have received funding from the
that it is difficult or even impossible for dict that people compare what actually Austrian Research Promotion Agency
a human expert to understand how a happened with what would have hap- under grant agreement 879881 (Ecosys-
result was obtained. Because of increas- pened if the possible cause had not been tem for Pathology Diagnostics With AI
ing legal requirements,20 in the future present. Common theories also state Assistance) and by the Austrian Science
it will be imperative for results to be that people focus only on what actu- Fund, project P-32554, Explainable Arti-
made retraceable, comprehensible, and ally happened to judge the mechanism ficial Intelligence.
understandable to a human expert. linking the cause and the outcome. To
The fast-growing XAI research commu- test this, for example, we propose in the REFERENCES
nity has already developed a number future not only to record the expert’s eye 1. A. Esteva et al., “Dermatologist-level
of very successful methods of explain- movements, but also to compare them classification of skin cancer with
ability. Explainability in the technical with other circumstantial evidence rel- deep neural networks,” Nature, vol.
sense of the XAI community means evant to the decision. Here it is import- 542, no. 7639, pp. 115–118, 2017. doi:
highlighting decision-relevant parts of ant to analyze in real time, and we plan 10.1038/nature21056.
a result. However, another important to do this according to our causability 2. K. Faust et al., “Intelligent fea-
aspect is that these explanations do not model, using various methods, includ- ture engineering and ontological
refer to an explicit human model. This ing analysis of eye movements, facial mapping of brain tumour histo-
motivated us to introduce our concept expressions, and micromovements morphologies by deep learning,”
of causability. Causability is intended such as head nods. In this future work, Nature Mach. Intell., vol. 1, no. 7,
as a measure of whether and to what we plan to combine and analyze these pp. 316–321, 2019. doi: 10.1038/
extent something is understood, and it metrics. The results obtained will be s42256-019-0068-6.
basically consists of two parts. The first useful for the development of novel 3. R. Hamon, H. Junklewitz, and I.
part is whether a human understands human–AI interfaces that will benefit Sanche, Robustness and Explainability
or can understand a given explanation medical experts and also lead to further of Artificial Intelligence—From Techni-
at all, and the second part is to what contributions to the international XAI cal to Policy Solutions. Luxembourg:
extent, that is, in terms of measuring, it research community. Publications Office of the European
is understandable by a human. Conse- Union, 2020.
quently, causability will become import- ACKNOWLEDGMENTS 4. A. Holzinger, “Interactive machine
ant for the design, development, test- We are grateful for the very useful com- learning for health informatics: When
ing, and evaluation of future human–AI ments of the anonymous reviewers and do we need the human-in-the-loop?”
interfaces. Such interfaces are needed the valuable substantial comments of Brain Inf., vol. 3, no. 2, pp. 119–131,
not only for enabling a human to under- the editor. We declare that real-world 2016. doi: 10.1007/s40708-016-0042-6.
stand an explanation, but also to capac- data were used for this work, but only 5. B. M. Lake, T. D. Ullman, J. B.
itate an interaction of the human with the principles of the human–AI interface Tenenbaum, and S. J. Gershman,
the AI. This is needed because a human were investigated and used for the illus-
in the loop can (sometimes) provide trative example in the section “A Clini-
conceptual knowledge, experience, and cal Case.” We are grateful for comments DISCLAIMER
context understanding—which to date from Luca Brcic and Markus Plass for the This publication reflects only the
no AI can do. positive ethic vote of the Roche Ventana authors’ views. The European Com-
In our newly established causabil- Study. Parts of this work have received mission is not responsible for any use
ity laboratory, we currently study how funding from the European Union’s that may be made of the information
pathologists make judgments and deci- Horizon 2020 research and innovation it contains.
sions. Central questions include the program under grant agreement 857122

OCTOBER 2021 85
EXPLAINABLE ARTIFICIAL INTELLIGENCE AND MACHINE LEARNING

ABOUT THE AUTHORS


ANDREAS HOLZINGER is the head of the Human-Centered Artificial Intelli-
gence (AI) Laboratory at the Institute of Medical Informatics and Statistics at
the Medical University Graz, Graz, 8010, Austria, and a visiting professor of 14. A. Holzinger, G. Langs, H. Denk,
explainable AI at the Alberta Machine Intelligence Institute of the University K. Zatloukal, and H. Müller,
of Alberta, Canada. His research interests include human-centered AI to put “Causability and explainability of
the human-in-control of AI and align AI with human values, ensuring privacy, artificial intelligence in medi-
security and safety. Holzinger received a Ph.D. in cognitive science from Graz cine,” Wiley Interdiscipl. Rev., Data
University and a second Ph.D. in computer science from the Graz University of Mining Knowl. Discovery, vol. 9,
Technology. He is a Senior Member of IEEE. Contact him at andreas.holzinger@ no. 4, pp. 1–13, 2019. doi: 10.1002/
medunigraz.at. widm.1312.
15. A. Holzinger, A. Carrington, and
HEIMO MÜLLER is head of the Information Science and Machine Learning Labo- H. Müller, “Measuring the qual-
ratory at the Diagnostics and Research Institute of Pathology of the Diagnostics ity of explanations: The system
and Research Center for Molecular Biomedicine of the Medical University Graz, causability scale (SCS). Comparing
Graz, 8010, Austria. His research interests include human–artificial intelligence human and machine explana-
(AI) interfaces for explainable AI and machine learning, particularly for medical AI tions,” KI – Künstliche Intelligenz
in digital pathology. Müller received a Ph.D. in mathematics from the Vienna Uni- (German J. Artif. Intell.), vol. 34, no.
versity of Technology with a work on data semantics spaces. Contact him at heimo 2, pp. 193–198, 2020. doi: 10.1007/
.mueller@medunigraz.at. s13218-020-00636-z.
16. R. W. Picard, A. Wexelblat, and C. I.
Nass, “Future interfaces: Social and
emotional,” in Proc. CHI’02 Extended
“Building machines that learn and 9. A. Holzinger, B. Malle, A. Saranti, Abstracts Human Factors Comput.
think like people,” Behav. Brain Sci., and B. Pfeifer, “Towards multi-modal Syst., 2002, pp. 698–699.
vol. 40, no. e253, 2017. doi: 10.1017/ causability with graph neural net- 17. R. W. Picard, “Perceptual user
S0140525X16001837. works enabling information fusion interfaces: Affective percep-
6. A. B. Arrieta et al., “Explainable for explainable AI,” Inform. Fusion, tion,” Commun. ACM, vol. 43,
artificial intelligence (XAI): Con- vol. 71, no. 7, pp. 28–37, 2021. doi: no. 3, pp. 50–51, 2000. doi:
cepts, taxonomies, opportunities 10.1016/j.inffus.2021.01.008. 10.1145/330534.330539.
and challenges toward respon- 10. M. D. Zeiler and R. Fergus, “Visualiz- 18. S. Poria, E. Cambria, R. Bajpai,
sible AI,” Inform. Fusion, vol. 58, ing and understanding convolutional and A. Hussain, “A review of
pp. 82–115, 2020. doi: 10.1016/j. networks,” in Proc. European Conf. affective computing: From uni-
inffus.2019.12.012. Comput. Vision, 2014, pp. 818–833. doi: modal analysis to multimodal
7. S. Bach, A. Binder, K.-R. Müller, and 10.1007/978-3-319-10590-1-53. fusion,” Inform. Fusion, vol. 37,
W. Samek, “Controlling explanatory 11. F. Cabitza, A. Campagner, and C. pp. 98–125, 2017. doi: 10.1016/j.
heatmap resolution and semantics Balsano, “Bridging the ‘last mile’ gap inffus.2017.02.003.
via decomposition depth,” in Proc. between AI implementation and oper- 19. H. Yu, T. A. Boyle, C. Zhou, D.
IEEE Int. Conf. Image Process. (ICIP), ation: ‘data awareness’ that matters,” L. Rimm, and F. R. Hirsch, “PD-
2016, pp. 2271–2275. doi: 10.1109/ Ann. Transl. Med., vol. 8, no. 7, p. 501, L1 expression in lung cancer,” J.
ICIP.2016.7532763. 2020. doi: 10.21037/atm.2020.03.63. Thoracic Oncol., vol. 11, no. 7, pp.
8. G. Montavon, “Gradient-based vs. 12. J. Pearl, “The seven tools of causal 964–975, 2016. doi: 10.1016/
propagation-based explanations: An inference, with reflections on j.jtho.2016.04.014.
axiomatic comparison,” in Explainable machine learning,” Commun. ACM, 20. D. Schneeberger, K. Stoeger, and
AI: Interpreting, Explaining and Visual- vol. 62, no. 3, pp. 54–60, 2019. A. Holzinger, “The European legal
izing Deep Learning, W. Samek, doi:[10.1145/3241036. framework for medical AI,” in Proc.
G. Montavon, A. Vedaldi, L. K. Han- 13. J. Peters, D. Janzing, and B. Schölkopf, Int. Cross-Domain Conf. Machine
sen, and K.-R. Müller, Eds. Cham: Elements of Causal Inference: Founda- Learning Knowl. Extraction, Cham:
Springer International Publishing, tions and Learning Algorithms, Cam- Springer-Verlag, pp. 209–226. doi:
pp. 253–265. bridge, MA, 2017. 10.1007/978-3-030-57321-8-12.

86 COMPUTER  W W W.CO M P U T E R .O R G /CO M P U T E R

You might also like