Professional Documents
Culture Documents
Computers in Industry
journal homepage: www.elsevier.com/locate/compind
ARTICLEINFO
ABSTRACT
Article history:
Received 8 August 2018 Why do we perceive virtual assistants as something radically new? Our hypothesis is that today
Received in revised form 4 December 2018 virtual assistants are raising an expectation for natural interaction with the human. Human
Accepted 17 January 2019 interaction is by nature cognitive and collaborative. Human sciences help to flesh the ingredients of
Available online xxx such cognitive interaction. Uttering a sentence is at the same time: producing sound; making a well-
formed sentence; giving meaning; enriching a common background of beliefs and intentions; making
Keywords: something together. In this paper, we remind the basics of human cognitive communication as
Cognitive interaction
developed by human sciences, particularly philosophy of mind. We propose a definition of this way
Virtual assistant
of interacting with computer as ‘cognitive interaction’, and we summarize the main characteristics of
Cooperation
Shared intention this interaction mode into a layered model. Finally we develop case studies to illustrate concretely
Natural communication the concepts. We analyze in light of our theoretical model three approaches of conversational
systems in AI, to illustrate the different available options to implement the pragmatic dimension of
cognitive interaction. We analyze first the seminal approach of Grosz and Sidner [20], and then we
describe how the now classical approach of discourse structure developed by Asher and Lascarides
[5] could capture the pragmatic dimension of interaction with an intelligent virtual assistant. Finally,
we wonder whether a state-of-the-art chat bot framework actually implements the needed level of
cognitive interaction. The contribution of this paper is: to remind and summarize essential ideas
from other disciplines which are relevant to understand what should be the interaction with virtual
assistants should be; to explain why the cooperation with virtual assistants is something special; to
delineate the challenges we have to solve if we are to develop truly collaborative virtual assistants.
© 2019 Elsevier B.V. All rights reserved.
1. Introduction
robot interaction as a “research area that seeks to improve
interaction between robots and their users by developing
What makes virtual assistants something special lays in
cognitive models for robots and understanding human mental
human-machine interaction: virtual assistants are intended to
models for robots”. A virtual assistant as we consider it in our
interact with human users in a natural way, which is by essence
aeronautical environment differs from a robot in major
cognitive, linguistic and collaborative. In a nutshell, assisting
aspects: it would not have its own “body”, and would have
collaboratively an operator requires recognizing and representing
limited sensing and action capabilities. Nevertheless, a virtual
some of the operators cognitive attitudes related to the current
assistant for aircraft operations would also benefit from natural
task, such as intentions, assumptions and attention. To
and cognitive interaction with the operator. The handbook of
distinguish this kind of interaction from usual human-computer
Robotics identifies three sub areas to “cognitive interaction”:
interaction patterns where human cognition has no explicit role, it
human models of interaction, robots models of interaction,
is usually named cognitive interaction. Cognitive interaction is
and Interaction models that address the interaction as a joint
widely studied in the area of human robot interaction. In the
activity [35]. The work presented here would clearly lay in
“Handbook of robotics”, Multu, Roy and Sabanovic [35] define
this third category: we try to understand cognitive interaction
“cognitive human
between operators and virtual assistants as a joint activity,
shaped by verbal exchanges. Indeed, solutions for efficient
operator-assistant interaction must be inspired by what has
* Corresponding author. been designed for other applications, such as robotics. But in
E-mail addresses: denys.bernard@airbus.com (D. Bernard),
this paper, we do not intend to give solutions: our goal is to
alexandre.arnold@airbus.com (A. Arnold).
propose tracks for a technology independent specification.
https://doi.org/10.1016/j.compind.2019.01.010
Industry needs such an
0166-3615/© 2019 Elsevier B.V. All rights reserved.
D. Bernard, A. Arnold / Computers in Industry 107 (2019) 33–49
34
independent specification to evaluate, down select, specify, (A12) Let’s disconnect the control unit.
evaluate and validate the diverse frameworks, toolboxes, or (The assistant continues guiding the mechanics on how to disconnect the
methodologies provided by both the software industry and control unit)
research. Hence, we need to take a step back from technical
solutions to elaborate this neutral view. Our approach is mainly
influenced by philosophy of action and philosophy of language.
Example 1: Imaginary assistant helping a mechanics to fix
Then, this paper proposes a description of what has to be done,
a failure
while letting open the technological choices.
This example contains some interaction patterns that today
This paper is an extension of a previous conference
chatbots or Virtual Assistant frameworks would not easily
presentation
implement (except through ad-hoc hard coded behaviors and
[7] which summarized the specificity of cognitive interaction in a
poor capability to generalize beyond a few specific situations).
five layers model. In this paper, we aim at describing the
For example:
theoretical foundations in more detail, and showing diverse
example of how cognitive interaction should be deployed for
- at the beginning of the scenario, the Assistant correctly
human – virtual assistant interaction.
interprets utterance (M1)“We have STABILIZER A2 FAULT” as
Industry has good reasons to pay attention to the cognitive
an order, although it is grammatically an assertion.
dimension of human-assistant interaction: it reduces operators
- Despite utterance (M5)”In 3 h” is not a well formed sentence,
workload, improves human-system collaboration, and finally
the assistant correctly interpret it as referring to the time until
increases the quality of human task with subsequent effects on
the next aircraft flight.
safety and operational efficiency. Moreover, the urgency of
- Even more difficult, (M6) “No sorry . . . at 2PM” intends at
understanding cognitive interaction comes also from the
correcting the previous utterance: its interpretation requires
pervasive choice from providers of commercial virtual
taking into consideration the conversational context to catch the
assistants to privilege natural language as interaction mode.
mechanic’s intention to correct his previous utterance.
In this paper, we will develop further the initial conference
- When interpreting “(M13) Still faulty”, the assistant should
paper in two directions. We first introduce the interaction model
hold that not only the flight control system is still faulty
and its sources. In particular we try to clarify the role of the
(which is the explicit content of the utterance), but also that
common interaction context, in light of ideas from dynamic
the fault is the same than the initial one (otherwise, the
semantics (e.g. the recent paper of Craige Roberts [46]). Second,
human would have mentioned it to be maximally relevant
we substantiate those concepts on concrete examples. First we
[55]).
focus on two classical approaches of discourse structure in AI,
which help to understand what should be the dynamic
Although it seems very simple, our example conversation is
elaboration of the linguistic and cognitive context through the
already beyond what current Virtual Assistant development
interaction: First, the structure of discourse by Grosz and Sidner
frameworks can easily implement (note that the capabilities of
[20], then Segmented Discourse Representation Theory (SDRT)
those frameworks evolve daily, and that statement might need to
of Asher and Lascarides [25]. We apply those approaches of
be revised very soon!). Unfortunately, such context based
dialogue to the same imaginary exchange between a mechanics
interpretation processes are pervasive in operational working
and a virtual assistant, for a trouble shooting task in Aircraft
situations. The appeal to context by communicating partners is
maintenance. Finally, we analyze an existing industrial
both an opportunity – because co-workers obviously share a
framework to develop chat bots, in order to imagine how it could
common context – and an obligation for the efficiency of
be extended to fully cover all the needed layers of interaction.
collaboration. Some of those mechanisms (may be not all) must
As an introductory example, consider the following
be implemented by virtual assistants.
imaginary dialogue between a mechanics (M) and a virtual
This paper groups and summarizes a usable kit of concepts
assistant (A). The task of the mechanics is to make available for
from a diversity of disciplines interested in human verbal
flight an aircraft on which a failure occurred (the example is
communica- tion, ranging from philosophy of language to
inspired from a real failure case, although it is simplified for the
natural language understanding in AI. Those concepts help to
sake of conciseness):
specify the cognitive functions that would make a computer
(M1) We have “STABILIZER A2 FAULT”
(A1) The Actuator 2 of the stabilizer has failed. capable of operating efficiently and fluently with a human
(M2) So what? being in real operations.
(A2) In principle the aircraft can fly with actuator 1 only, but the fault might A cognitive assistant is something special because it collabo-
cause a no-go failure at next power up. rates with human operators, and conversation is a special kind of
(M3) Why?
human collaboration. Conversation is used to perform many
(A3) A short circuit in the control unit of Actuator 2 may produce a brake
effect, preventing the whole stabilizer to move other kinds of collaborative activities, in most the case
(M4) What can we do? conversation itself is not the finality of the interaction [58,61]. In
(A4) When is the next flight? operational situations such as the one depicted in the
(M5) In 3 hours.
introductory example, it might even be difficult to disentangle
(M6) No sorry . . . at 2PM
(A5) Confirm the fault, then disconnect the faulty control unit if needed
communicative actions from other activities of the common
(M7) Let’s confirm the fault task.
(A6) Is the aircraft energized? In our example dialogue, the assistant had to understand
(M8) yes elliptical phrases, to recognize a diversity of speech acts, to use
(A7) Erase the non-volatile memory (NVM) of flight control system
and interpret pronouns or indexical words, to maintain dialog
(M9) How?
(A8) On maintenance terminal, select “system test & scanning” coherence, to track mechanics expectations. Human capability
(M10) Done to act together and conversation have been considered from two
(A9) Select “Flight controls” different streams of thought in philosophy of mind. The nice
(A10) Select “Erase NVM” thing, is that those initially distinct perspectives finally converged
(M11) What’s next?
(A11) Launch the system test
towards a consistent picture where conversation is seen as a
(M12) Wait a minute . . . cooperative activity which helps agents to cooperate for other
(M13) Still faulty. purposes. Starting from the analogies between both perspectives
of action and language, we summarize the cognitive interaction
between human users and virtual assistant in a unique multi
layered description. After the philosophical foundations, the
declared intentions; intentions must persist as long as they are
paper introduces a series of examples from AI, where part of the
not realized, if we have no good reasons to cancel them; the set
ideal of cognitive interaction have been implemented or formally
of intentions one holds at a given time must be consistent.
de- scribed in an implementable formalism.
Consequently, an intention is a commitment to act: If I declare
First, we analyze two classical approaches of collaborative
my intention to make a particular action at a given time and if I
dialogue in Artificial Intelligence, and then we focus on a state
finally don't make it, I can be charged of irrationality. Because
of the art framework to develop chat bots. This analysis should
they are also commitments, intentions carry reliable information
draw a picture of where we are today on cognitive interaction
about the future behavior of the agents. It is reasonable to believe
with virtual assistants, and where could be the next research
that intentions of rational agents will be realized.
priorities.
Bratman's planning theory of agency has influenced AI since
the 90’s. It has been the case in particular for the research on BDI
2. Towards a layered model of cognitive interaction
agents, i.e. agents driven by human-like cognitive states, beliefs,
desires, intentions [39]. An approach to implement a general
A technology independent representation of cognitive interac-
system to consistently reason on intentions and beliefs has been
tion can be derived from classical accounts of human
proposed by Shoham in [52]. In the continuity of this approach
collaboration and communication. This section introduces first
[22], has elaborated the capability to handle trees of intentions.
the main cognitive notions – group agency, speech acts, and
Note that handling multilevel intentions is necessary to properly
common ground - before introducing the main model itself.
represent and reason about agent's intentions in collaborative
situations [19,20].
2.1. Joint agency vs. individual agency
In general, one cannot intend somebody else’s action
(excepted in a strong hierarchical framework). Hence, a theory of
What matters for the efficiency and the success of a virtual
collective action must explain how the intentions of a group of
assistant, is that its action plan – including communicative
agents involved in a common action are coordinated. How can
actions
individual minds shape collective actions, for example building a
– adequately meshes with operator’s action plan. Ideally, the
house, having a conversation, designing a complex system? This
operator and the virtual assistant should jointly track the goals of
question is addressed later by Bratman through the concept of
the operator. Philosophy of action provides some clues to
shared intentions [9]. A shared intention is the cognitive
understand the cognitive processes which support joint agency.
configuration of a group of partners who could say: "we are
From a classical perspective in philosophy of action, human
acting together", and act consistently with that statement.
action is explained by individual mental attitudes, which are
Bratman proposed a set of necessary conditions that define
directed towards the real world where the action is to take place
shared intention and enable collaborative behavior in a small
e.g. beliefs, hopes, intentions, desires, attention, and assumptions
group of partners. It is not necessary to expose in detail this
[49]. Such attitudes are manifestations of the general ability of
theory here, nevertheless, it’s worth reminding that a virtual
our minds to relate to the real world, or “intentionality” [48]. The
assistant must contribute to fulfilment of the user's intentions, by
assumption that action is driven by individual intentional
adjusting its plans to what the user intends, is actually doing,
attitudes holds also for group action. Hence, a central problem of
and is about to do.
philosophy of action is to explain how individual intentional
Exceptforconversationalsystemswhere intentionsare conveyed
attitudes may drive collective ones.
verbally, agent's intentions are not always easy to catch: intention
Intentional attitudes are related to particular aspects of the
recognition is an active research field particularly in human robot
real world: they have specific “contents” of different types. A
interaction (e.g. [14,23,57]). Even more, intention is not the only
belief is defined by the believed fact; the content of a desire is
cognitive attitude to be monitored by a virtual assistant. Several
a desired state of affairs, e.g. the possession of a given object;
authors, including in the field of human robot collaboration,
the content of an intention is the targeted state of affairs or an
insist on the importance of monitoring user attention (e.g
action aiming at obtaining a particular goal. Different
[34,64].).
intentional attitudes can be related to the same content: I may
To summarize, the virtual assistant must not only represent
know that the sky is blue, hope that the sky is blue, imagine
the operational situation, it must also represent the user's
that the sky is blue, or perceive that the sky is blue. All those
cognitive perspective about the task. Such requirements have
intentional attitudes have the same content that the sky is blue,
already been endorsed from diverse applications areas, ranging
but differ by the kind of relation they entertain with that fact.
from human robot interaction to conversational systems
Searle [48] distinguishes in particular two major subgroups
[2,14,19,27,57,64]). An example approach to coordinate the
depending on their “direction of fit”: beliefs succeed if their
plans of several agents is suggested in [19], where each agent,
contents match the real world; they have a “mind- to-world”
besides his own backlog of actions, manages a backlog of
direction of fit. Conversely, intentions and desires do not need
common actions that have been agreed with his partners. Most of
to match the actual world, but they succeed if the world finally
the literature agrees with Bratman on the statement that
aligns to their contents: intention and desire are said to have a
collective agency use similar planning “algorithms” than
“world-to-mind” direction of fit. Although they have the same
individual agency, but dedicated to building and maintaining
direction of fit, intentions and desires differ by their
consistent and agreed sub-plans. Less obvious is the assumption
conditions of satisfaction: a successful intention actually
that assistants rationality must be transparent: virtual assistants
causes it contents to happen. A desire is different: I can desire
must answer and act in ways which are understandable by human
that it rains tomorrow, even if I will not cause the rain
users [4], as the result of a rational planning process. The
(conversely, it sounds odd to intend that it rains tomorrow).
assistant behavior needs to be transparent and predictable
Intentions have a central role in practical reasoning, leading to
according to user's norms [61].
the elaboration of plans and the decision to take actions, as
theorized in particular by Bratman in [8] though a planning
2.2. Saying and doing
theory of agency. Making a plan consists in forming a cascade of
intentions: Main goals are progressively cascaded into lower
To coordinate their plans, collaborative agents have to
level intentions, down to instrumental intentions which match the
exchange about their current beliefs and intentions. From this
know-how of the agent. Intentions are submitted to specific
perspective, communication could be considered as an
rationality norms [8], for example, one has to act according to his
independent capability, enabling agents’ collaboration. More
consistently, communication
is itself a collaborative activity, rather than a special capability
wanting to give me your email would be the perlocutionary act.
besides joint agency. J.L. Austin has given in the 1950s a series
The interpretation of a perlocutionary act requires carrying
of conferences that inaugurated a radical change in philosophy of
inferences beyond what is said. The nature of the inference
language. Since this series of conferences published under the
needed to access the speaker perlocutionary intention is debated.
title “how to do things with words” [6], speech is no longer
Even the role of the alleged literal meaning of the utterance in
regarded only as a way of producing descriptions of the reality,
this kind of inference is questioned [42,24]. Multifaceted speech
but also as the performance of “speech acts” which purposefully
acts are pervasive in conversation in operational contexts, where
change the reality. Speech acts may succeed or not depending on
the communication needs to be parsimonious and reliable. Here
how the cognitive state of the hearer is finally affected. The
is an extract of an example conversation between pilot and co-
logical structure of speech acts parallels the structure of
pilot in a cockpit:
intentional attitudes presented above: a speech act decomposes
into content and force. The content of a speech acts represents Captain: "set speed one nine five"
the facts or objects the speech act is about. The force delineates
the purpose of the speaker when making this particular speech First Officer: "One?"
act. The same content can be pushed with different forces, for Captain: "nine five"
example "you are running" is an assertion, "run!" is a
In this exchange, First officer’s “One?” not only informs the
prescription, "are you running?" is a question but they all have
Captain that the order has been misunderstood, but also it
the same content.
requests a correction of a particular part of the order.
Speech acts purposes are cognitive and social: When giving
an Implementing the recognition and classification of multi-purpose
dialog acts is still a hard problem, but it is an active research field
order, one intends to change the intentions of the hearer, or when
(e.g. [10,24,56]). Speech acts differ from concrete action by the
making an assertion one aims primarily at changing the beliefs of
the hearer. Commitments or successful prescriptions create new so-called reflexive nature of communicative intentions [47], i.e.
obligations. A successful assertion changes the common ground, the role of speaker intention in the interpretation “algorithm”: the
i.e. what is openly known to be known among the interlocutors. speaker obtains the intended effect thanks to the recognition of
Because the common ground is the root of agents’ actions and his intention by the hearer. Recognizing speaker’s intentions is
their justification, it plays a central role in all collaboration necessary but not sufficient to ensure the success of the speech
situations, including the interaction between human operators act. In addition, the hearer has the choice to accept or not the
and virtual assistants. speech act and its subsequent cognitive effect.
The classification of possible forces in a global model is a According to this psychological perspective on speech acts
hard task, and testifies the diversity of actions that can be done (speaking is changing purposefully one’s mental state), under-
by words. For example, here are the categories proposed in [40] standing utterances is not a mere decoding process, but rather an
as a consensual high-level taxonomy of speech acts: (Fig. 1). explanation process. Interpreting an utterance is explaining the
Representative speech acts contain facts that are true or behavior of the agent by the assignment of intentions and beliefs
intended to become true. They split into performative and [41]. The central role of user communicative intentions is
assertive speech acts: an assertion says something about the partially acknowledged in today's frameworks to develop virtual
assistants or chatbots: processing a user query starts by
actual world ("it's raining"). Performative acts aim at making true
understanding his "intent". Nevertheless, as far as we know, the
their content. A declarative act makes true a social fact ("I declare
reflexive and multi- level nature of communicative intentions is
the meeting closed"). A Commitment creates obligations for the
not supported in today's implemented frameworks for developing
speaker ("I promise to go to school"). Prescriptive acts make the
chatbots, al- though research in A.I. has theorized and proposed
addressee intend a particular state of affairs or action ("Please, go
solutions in the past (see for example [2]). That limitation
to school"), provided the speaker has authority upon the hearer.
impairs the fluency of dialogue with virtual assistants, because
Finally, expressive speech acts communicate emotional states or
recognizing intention structures is necessary to understand the
social relations ("Hello","thank you").
exchange and plan the next steps.
The debate on the classification of speech acts is still alive.
But it can be objected the following to such psychological
The main difficulty is that speech acts often belong to several
approaches of conversation: in daily conversations, hearers are
categories. For example, by saying "Hello" I would not only
not permanently calculating the intentions of speakers, and
express my wish to consolidate social relations, but I also
speaker's utterances are not deliberately designed to trigger
perform the act of greeting someone: it is then a performative
precise sequences of inferences. Recent approaches of speech
act. If I say “I don’t know your email” you might understand
acts inspired by dynamic semantics tend to reduce the
that I’m requesting you to give me your address, it thus can be
psychological dimension of interpretation by understanding
interpreted as a prescription. The usual solution to this problem,
interpretation as a way of managing the conversation context
in accordance with Austin’s distinction between locutionary act,
[17,37,45,46]. Because speech acts are transparent, they can in
illocutionary act and perlocu- tionary act, is to consider that an
principle be understood through their impact on the information
utterance performs speech acts of different kinds at the same
openly shared by the interlocutors. Successful speech acts update
time. In the utterance “I don’t know your email”, the action of
and enrich the common ground. The
making you know that I don’t know your email would be an
illocutionary act, and the action of getting you
or other systems designed to interact cognitively with human speaking: when two people are carrying heavy furniture or a pilot
operators. Indeed, it is the case of some robots. For example, is executing a procedure and the co-pilot is monitoring the
consider the architecture of human-robot interaction in project execution of the procedure, or just when you are walking in a
JAST, Joint Action Science and Technology [16]. According to crowd. The agents act in a transparent way, they make sure that
the section dedicated to human robot interaction in the Handbook their gesture is understandable by the other agents and that their
of Robotics [35], this architecture is a representative example of intentions are transparent. In such case, interactions must happen
the dialog-based paradigm of human robot interaction, which also at physical, linguistic, semantic, pragmatic and cooperative
under- stands interaction as “joint action”. Starting from the levels, even if the work to be done at linguistic and semantic
bottom, the architecture presented in [16] implements the levels is obviously of different nature than in the case of spoken
physical layer through a variety of modules, specialized for the language. A recent paper [27], describes architectures for multi
different modalities available: speech recognition and synthesis, modal human-robot interaction where the knowledge base is fed
face and gaze tracking, production of facial expressions, hand from diverse directions: on top of a sensorimotor layer, the verbal
tracking. Regarding the linguistic layer, only the linguistic layer communication is managed by a specialized natural language
for speech is made explicit in the architecture (the design choice processing module, while in parallel situation assessment func-
has probably been to not consider facial expressions or hand tions feed the knowledge base with facts about the scene. Such a
movements as conventional signs, to be decoded by a specific system builds the human perspective of the scene by recognizing
module). The architecture has a dedicated semantic interpretation facts such as “the user sees X”, “the user looks at X”, or “the user
module, directly interfaced to a central decision making module, points at X”. Such a system enhance the capability of robots to
which designs outputs and actions to be done according to the act at the pragmatic level by introducing knowledge about the
current plan, the discourse history, and the world model. The shared physical environment and agents gesture into the common
semantic interpretation module includes the pragmatic background of interaction.
interpretation task: it “finds referents for noun phrases, including A major difficulty to implement cognitive interaction and
demonstrative and anaphoric pronouns” and produces particularly the pragmatic layer, is that agents have to build and
“disambiguated propositional Content”. But the update of the maintain a proper representation of the common ground. This
knowledge on operational context is distributed among the representation should include facts about the current situation,
semantic interpretation and the decision making module (indeed, some information about the cognitive attitude of the partner, and
the current updated plan is part of the interaction context). should track the interaction. The two first sections of next
The layered representation of interaction also applies to cases chapter are dedicated to two now classical approaches of
where there is no explicit communication. Think of tasks where discourse representation in AI. They both are candidate solutions
two agents coordinate their movements or thoughts without (or part of solution) to that problem.
3. Managing cognitive context: example solutions from AI
DSPs, like Discourse Segment Purposes). In the example, we
distinguish the primary intentions linked to the practical
3.1. Grosz and Sidner
objectives of the discourse segments DS0, DS1, DS2, DS3, DS4
and lower level communicative intentions contributing to fulfil
In their seminal paper Attention, Intention and the structure
the primary intentions. The intention tree could look like: (Fig.
of discourse [20], Grosz and Sidner describe the cognitive
5).
structure of discourse as composed of three subparts: (1) the
The sequence of communicative intentions corresponds to the
actual sequence of utterances (2) A tree of intentions (3) a focus
sequence of utterances in the example. The branches in the tree
stack. Even if this is not explicit in the initial approach, this
represent "dominates" relations: an intention I1 is said to
structure helps to flesh what happens at pragmatic level in a
dominate intention I2 if I2 is part of the intentions to be fulfilled in
conversation. This triple structure is a formalization of what each
order to fulfill I1.
agent has to know consistently with the other agent. Although
Not all the intentions are explicitly mentioned in the
their initial approach applies as well to monologues or texts than
conversation. For example, the mechanics does not explicitly
to conversation, the cited paper of Grosz and Sidner focuses on
require help: it is understated by his first utterance: We have
task-oriented exchanges.
“STABILIZER A2 FAULT”.
From now on, we try to illustrate this approach with our
The intention tree – as a shared representation of the
introductory example. The intentions considered here are
conversation – is built progressively along the interaction. At
typically intentions for cognitive actions, to change the
each step, each agent determines the next action to be taken and
intentions or knowledge state of the interlocutor. They represent
interprets the speech act of the partner by using the current focus
in a consistent way primary intentions linked to the task (e.g. get
space.
the mechanics to launch an auto-test of the flight control system)
The focus space is a stack of focus states, each focus state
and lower level communicative intentions s.t. getting the
representing the intentions and objects of interest introduced by
mechanics to know the operational impact of the fault.
each step of the conversation. For example, that could be the initial
The structure of the conversation parallels the upper layers of
values of the focus stack at the beginning of the conversation:
the intention tree and the structure of the procedure itself, which
(Fig. 6).
counts three major sub-sections: first to understand the fault and
The dialogue first focuses on the fault symptom, then
its consequences (DS1); then decide what to do (DS2); confirm
introduces the possible cause of the fault symptom, and then its
the fault (DS3) ; fix the fault (DS4) (Fig. 4).
impact. At each turn, the focus stack contains the focused
The structure of intentions is restricted to open intentions, i.e.
elements for each of the upper level intentions of the same
intentions to be recognized by the partners. Single open
branch in the intention tree. The focus stack is enriched with the
intentions are attached to each subpart of the conversation (they
communicative intentions and the new objects of interest
are called
introduced for the current turn.
[58]