You are on page 1of 9

AI Magazine Volume 10 Number 1 (1989) (© AAAI)

Logical and Decision-


Theoretic Methods for
Planning under Uncertainty
Curtis P. Langlotz and Edward H. Shortliffe

Decision theory and nonmonotonic logics


are formalisms that can be employed to
P lanning consists of devising a
course of action that conforms as
well as possible to a set of goals. A
not assume how or in what context
the represented knowledge will be
used (Hayes 1977). Simply stated, logi-
represent and solve problems of planning planner attempts to determine the cal inferences based on valid logical
under uncertainty. We analyze the useful- optimal action in a particular prob- statements never result in invalid
ness of these two approaches by establish- lem-solving situation. We are inter- conclusions. The consistency of first-
ing a simple correspondence between the
ested in automating decision support order logic is appealing, but it comes
two formalisms. The analysis indicates
that planning using nonmonotonic logic
for a particular set of planning prob- at a cost: All knowledge must be stat-
comprises two decision-theoretic concepts: lems distinguished by the following ed categorically; no possible occur-
probabilities (degrees of belief in planning characteristics: (1) the current situa- rences or partially achievable goals
hypotheses) and utilities (degrees of prefer- tion is not known with certainty; (2) can exist. This consistency property
ence for planning outcomes). We present the consequences of action are not prevents the use of first-order logic to
and discuss examples of the following known with certainty; and (3) the represent and solve problems of plan-
lessons from this decision-theoretic view goals of the planning process are ning under uncertainty, which inher-
of nonmonotonic reasoning: (1) decision conflicting and, therefore, are not ently involve incomplete, uncertain,
theory and nonmonotonic logics are completely satisfiable. We refer to and inconsistent information.
intended to solve different components of
problems of this type as planning In response to this limitation, sever-
the planning problem; (2) when considered
in the context of planning under uncer-
under uncertainty. Because these al researchers have devised augment-
tainty, nonmonotonic logics do not retain planning tasks entail uncertainty and ed logical systems that allow a propo-
the domain-independent characteristics of tradeoffs, a purely deductive process sition to be assigned a truth value
classical (monotonic) logic; and (3) (such as state space search [Fikes and that signifies the proposition is con-
because certain nonmonotonic program- Nilsson 1971] or skeletal plan re- sistent with the existing set of facts
ming paradigms (for example, frame-based finement [Friedland and Iwasaki (true by default) even though these
inheritance, nonmonotonic logics) are 1985]) is difficult to employ to select propositions have neither been proved
inherently problem specific, they might be the optimal plan (Langlotz et al. nor disproved (for example, Reiter
inappropriate for use in solving certain 1987). In the course of our investiga- 1980). Because these default assump-
types of planning problems. We discuss
tion of viable alternative planning tions can be withdrawn based on new
how these conclusions affect several
current AI research issues.
methodologies, we evaluated the information, it is conceivable that
applicability of two theoretical new information will cause a retrac-
0738-4602/89/$3.50 © 1989 AAAI. approaches: nonmonotonic logics and tion of an assertion and, therefore, a
decision theory. In this article, we reduction in the number of provable
establish a simple correspondence logical statements about a particular
between the two theories, describe problem. Thus, in contrast to first-
how each theory applies to the plan- order logic, the number of statements
ning task, and offer several provable from a set of default assump-
suggestions based on the strengths tions does not necessarily grow mono-
and limitations of the two approaches. tonically with the addition of new
information. Thus, these systems are
Nonmonotonic Logics called nonmonotonic logics.
To represent and solve a planning
One of the most compelling reasons problem using a nonmonotonic logic,
for the use of logic as a representation a system builder must acquire the rel-
language is that first-order logic does evant beliefs and assertions from an

SPRING 1989 39
expert and express them in the non- quences will best satisfy the goals of When a nonmonotonic logic is used
monotonic logic. The system uses a the planning task. A thorough intro- to represent and solve a planning
logical inference technique to con- duction to this theory of decision problem, unintended logical infer-
struct a proof that argues for a partic- making is provided in von Winterfeldt ences can result (Hanks and McDer-
ular course of action to satisfy the and Edwards (1986) and Holloway mott 1986). Several researchers
goals of the planning task. Because of (1979). (Shoham 1986; Lifschitz 1986) have
their ability to express inferential developed formalizations of prefer-
uncertainty as explicit default rules, ence criteria that in some situations
nonmonotonic logics are attractive Limitations of the allow only intended inferences. Some
formalisms for planning under uncer- Two Formalisms recently investigated ad hoc tech-
tainty. niques have similar goals (Cohen, et
In theory, both decision theory and al., 1987). Doyle (1985) suggests that
Decision Theory nonmonotonic logics can be used to nonmonotonic inferences can be
Decision theory is also intended to solve problems of planning under interpreted as a manifestation of pref-
formalize planning problems that uncertainty. In practice, however, erences. Shoham (1987) makes this
involve substantial uncertainty (Sav- both formalisms have weaknesses notion explicit by formally defining
age 1972), but its axiomatic basis (Cox that restrict their use in planning sys- preference logics, which implicitly
1946; von Neumann and Morgenstern tems. capture a partial order of intended
1953) is distinct from nonmonotonic inferences. He proposes this formal-
Nonmonotonic Logics ism as a generalization of other non-
logic. In contrast to nonmonotonic
logic, decision theory employs proba- By their very nature, proofs in non- monotonic logic systems and argues
bilities—continuous measures of monotonic logics can later be invali- that the various formalizations of
belief in uncertain propositions. To dated. The temporary nature of these nonmonotonic inference are simply
accommodate conflicting problem- proofs has two important conse- different methods to operationalize
solving goals, quantities called utili- quences for planning systems that use preferences for some inferences over
ties are used to denote degrees of pref- nonmonotonic logics. First, these sys- others. We agree with this point and
erence toward attaining planning tems focus on the task of constructing return to it later.
goals. A mathematical method for admissible (allowable) plans rather This view of nonmonotonic logics
combining these probabilities and than selecting the optimal plan (Char- raises questions about their practicali-
utilities is derived directly from the niak and McDermott 1985). Second, ty. We find it counterintuitive that
axiomatic basis. The solution method because not all admissible plans are preferences, which are essential to the
recommends the course of action with optimal, much of the current work on selection of intended inferences, are
the maximum expected utility score. nonmonotonic logics focuses on for- not explicitly represented in non-
Expected utility is computed as fol- malizing the contexts in which non- monotonic formalisms. In addition,
monotonic conclusions are appropri- these formalisms offer little guidance
ate, thereby precluding the construc- on how such preferences should be
m

Σ p(O |P ) * U(O )
tion of suboptimal plans. obtained and used to influence the
EU(Pi ) = j i j Because a set of logical statements design of a nonmonotonic system.
j=1 is called a theory, and because each Finally, if new preference criteria
consistent set of statements provable must be devised for new problem-
lows for m outcomes: from this theory is called an extension solving situations, this development
where EU(Pi) is the expected utility of of the theory, this task is sometimes might compromise the problem inde-
the ith plan, p(Oj|Pi) is the probability called finding minimization criteria pendence of the nonmonotonic logic.
of the jth outcome after executing the that determine the preferred exten-
Decision Theory
ith plan, and U(Oj) is the utility of the sion of a default theory (Etherington
jth outcome. To represent and solve a 1987). These criteria are represented Whereas symbolic planning method-
as logical axioms that place nonmono- ologies often concentrate on the con-
planning problem using decision theo-
tonic conclusions in order of priority. struction of plans, the decision-theo-
ry, a system builder must acquire
This prioritization is used to deter- retic for malism concentrates on
from an expert the probabilities, signi-
mine which nonmonotonic conclu- choosing the best action from a pre-
fying the likelihood of the potential
sions should be considered valid in enumerated set of plans. Thus, deci-
consequences of action, and the utili-
particular problem-solving settings. sion theory has been used to date only
ties, signifying the degree to which
For example, Poole (1986) describes as a tool to assist human planners
each potential consequence satisfies
criteria for finding the preferred exten- who have already generated a small
the problem-solving goals. The proba-
sion by giving preference to the most set of alternative plans to evaluate.
bilities and utilities are then com-
specific default theory. This technique Furthermore, decision theory has not
bined to derive an expected utility
is analogous to the notion that objects yet been used as part of a consultation
score for each plan. The plan with the
in a hierarchy should inherit proper- system, in part because of its inability
highest expected utility is considered
ties from their closest ancestor. to constrain the number of plans to
superior because its probable conse-
evaluate and in part because of its

40 AI MAGAZINE
data-hungry evaluation method. How- logics and decision theory, thereby concept of nonmonotonic reasoning.
ever, for problems limited in size, revealing the following characteristics Figure 2 shows a simple formulation
knowledge engineering time for deci- of nonmonotonic planning systems. of this problem in a nonmonotonic
sion-theoretic systems can be compa- First, planning systems that employ logic.2 Initially, two propositions exist
rable to that for rule-based systems nonmonotonic logics make implicit in the knowledge base. Statement 1
(Henrion and Cooley 1987). The use of assumptions about problem-solving expresses the fact that TWEETY is
even constrained decision-theoretic goals. Second, because goals vary known to be a bird. Statement 2
problem-solving models sometimes among problems, nonmonotonic con- expresses if we know an object is a
poses problems because the acquisi- clusions that are appropriate for one bird and if it is consistent with the
tion of probabilities and utilities is problem can be inappropriate for rest of our knowledge the object can
subject to systematic inaccuracies and another. Third, because the appropri- fly (that is, we cannot prove other-
biases (Tversky and Kahneman 1974; ateness of nonmonotonic conclusions wise), then we are willing to add to
Nutter 1987). Although the effect of can vary, the ability of nonmonotonic our knowledge base the fact the object
biases on problem-solving results can logics to represent problem-indepen- can fly. Thus, in statement 3, the con-
be estimated using sensitivity analy- dent commonsense knowledge is clusion is asserted that TWEETY can
sis (Downing 1987; Pauker and Kassir- restricted. fly, because no specific evidence
er 1980), these limitations of the We analyze nonmonotonic reason- exists to the contrary.
knowledge-acquisition process are ing using decision theory to represent Rich (1983) has observed that the
still a source of concern and a subject the decision of whether to make a default rule (2) is reasonable because
of study (Wallsten and Budescu 1983). nonmonotonic conclusion. Figure 1 it is highly likely that birds can fly.
Another practical limitation of deci- shows how a decision tree can be used Our decision-theoretic representation
sion theory is the difficulty in to represent decisions of this kind. 1 of this problem brings up a specific
explaining its inference procedure. The tree shows two choices: to assert question: How large a probability of
The mathematical nature of maximiz- a nonmonotonic conclusion or not. error would make the nonmonotonic
ing expected utility does not corre- Regardless of whether the conclusion conclusion about TWEETY unwise?
spond closely to human inference is asserted, a probability (p) exists that This question is addressed in detail in
techniques, making this for mula the conclusion is valid, and the possi- the following section.
unappealing and unconvincing as a bility (1 - p) remains that the conclu-
justification for the results of plan- sion is invalid. Preferences and
ning. A few programs have produced To facilitate an intuitive under- Default Reasoning
text explanations of quantitative prob- standing of the correspondence illus-
lem solving (Kosy and Wise 1984; trated in figure 1, we first consider the Our discussion thus far closely fol-
Langlotz, Shortliffe, and Fagan 1988), example of TWEETY the bird, which lows the analysis of several other
but this explanation process is sub- is used frequently to illustrate the researchers who acknowledge that
stantially more difficult than the anal-
ogous task for symbolic reasoning pro-
grams.
In summary, we believe that deci-
sion theory is limited in its ability to Valid Conclusion
VA (“Appropriate action”)
constrain the set of possible plans to a (p)
Conclude & act
tractable set of admissible plans. In Invalid Conclusion
contrast, nonmonotonic logics are IA (“Too aggressive”)
restricted in their ability to select the
optimal plan from the admissible
plans. In the remainder of this article,
we delineate important limitations of Valid Conclusion
IW (“Asleep at the switch”)
nonmonotonic logics with respect to (p)
the plan-selection process. Wait Invalid Conclusion
VW (“Appropriate caution”)
A Decision-Theoretic Model
of Default Reasoning
Figure 1. A Decision Tree That Represents the Decision Whether
In a previous article, we showed that
to Make a Nonmonotonic Conclusion.
decision-theoretic notions of likeli-
hood and preference underlie rule- The square nodes are decision nodes. The branches emanating from decision nodes
represent plans among which a choice must be made. The circular nodes are chance
based planning (Langlotz, Shortliffe,
nodes whose branches represent all the possible scenarios that might occur. The state-
and Fagan 1986). We now extend this ments at the right dramatize how the problem solver might perceive each scenario in
analysis by establishing a simple cor- hindsight. VW signifies valid wait, IA signifies invalid action, and so on. The proba-
respondence between nonmonotonic bility of a valid nonmonotonic conclusion is represented by p.

SPRING 1989 41
matical analysis of the decision tree
shown in figure 1. Decision theorists
(1) Bird (TWEETY ) have derived a quantity for such trees
known as the threshold probability
(2) Bird (x) and M [Fly(x)] ⊃ Fly (x) (sometimes referred to as p*) (Pauker
and Kassirer 1980). This quantity
(3) Fly (TWEETY ) denotes the probability of making a
valid conclusion below which a non-
monotonic conclusion is suboptimal.
Similarly, the quantity 1 - p* repre-
Figure 2. A Simple Proof That TWEETY Can Fly, sents the probability of error above
as Expressed in Non-monotonic Logic. which a nonmonotonic conclusion is
suboptimal. The value of p* depends
M can be interpreted as “consistent with all other known facts.” Therefore, M[Fly(x)] is
on the relative preference for each
interpreted as “the fact that x can fly is consistent with the rest of the knowledge
base.” The facts above the line are in the knowledge base initially, and the assertion
possible outcome of the problem. 4
below is proved using modus ponens. Consider the TWEETY example posed
in the context of a planning problem.
Adapted from Reiter (1980).
How would we feel if we assumed
TWEETY could fly when, in fact,
TWEETY could not? We might feel
nonmonotonic inference is related to preferable. Thus, he concludes, “One comfortable with the default rule in
notions of likelihood and probability. must maximize one’s expected utility figure 2 if we were deciding whether
Rich (1983) proposes a likelihood when selecting a nonmonotonic infer- to buy a birdbath for TWEETY, but
interpretation of default logic using ence” (Shoham 1987, p. 391). would we feel equally comfortable if
the certainty factor (CF) calculus We now explore the implications of TWEETY were a beloved pet, and we
(Shortliffe and Buchanan 1975). 3 these intuitive discussions by provid- were considering dropping him off a
Michalski and Winston (1986) pro- ing a rigorous argument to support the cliff (see figure 3)? The more serious
posed a similar framework for default claim that nonmonotonic inference the consequences, the less comfort-
reasoning called variable precision comprises not only likelihood but also able we feel making the default con-
logic. The probabilistic interpretation preference when considered in the clusion about TWEETY. This relation-
of this default logic also corresponds context of planning problems. In plan- ship holds true even when the likeli-
to the likelihood notions we describe ning, the task is to select or devise an hood TWEETY can fly is invariant.
here. appropriate action, the consequences The following medical example fur-
In contrast, we concur with other of which are vital to the decision. ther illustrates the implications of the
researchers who propose that prefer- Thus, implicit in the assertion of a relationship between the threshold
ence plays an important role in non- nonmonotonic conclusion is the probability and the utilities for the
monotonic inference. For example, belief that it is acceptable to act as if problem. Consider the following ques-
Doyle (1985) identifies parallels the assertion were true. The desirabil- tion: Does a patient with pain in the
between economic theory and non- ity of acting as if an assertion were right lower abdomen suffer from
monotonic reasoning. He proposes true depends not only on the likeli- appendicitis? This problem is illustra-
that the selection of default inferences hood of the assertion’s truth but also tive because like the birdbath exam-
can be interpreted as a reflection of on the desirability of the conse- ple, it is explicitly linked to action:
group preferences. Shoham (1987) quences of acting as if the assertion Because the consequences of untreated
makes a similar argument by distin- were true. (The concept of desirability appendicitis are extremely serious, all
guishing between first-order logics, is signified by the quoted statements patients with known appendicitis are
which make conclusions because they in figure 1). Therefore, even though immediately referred for surgery.
are valid, and nonmonotonic systems, the likelihood that a statement is true However, the surgeon cannot know
which make inferences because a rea- might not change, our willingness to with certainty before surgery whether
son exists to do so. He discusses sever- act as if it were true might vary. a patient has the condition. In fact, it
al reasons that an inference might be These changes in our willingness to is considered acceptable medical prac-
preferred. For example, some are pre- act arise because the appropriateness tice if as high as 25 percent of patients
ferred because they represent default of an assertion depends in part on our sent to surgery do not have appendici-
assumptions that facilitate communi- preferences for the potential conse- tis (Pieper, Kager, and Nasman 1982).
cation between the system and its quences of acting on this assertion. Thus, the threshold probability, p*, of
users. Similarly, inferences that are The fact that utilities dictate the appendicitis above which surgery
likely to be valid might be preferable appropriateness of nonmonotonic con- should be recommended is at least
to others; likewise, if the potential clusions is a recurring theme in our 0.75. In other words, if it is evident,
incorrectness of an inference is incon- analysis. based on the existing laboratory and
sequential, the inference might be This intuitive notion of utility examination data, that a patient has at
dependence is confirmed by a mathe- least a 75-percent chance of an appen-

42 AI MAGAZINE
dicitis, the doctor should act as if the
patient has appendicitis and, thus,
should send the patient to surgery.
Would we feel equally comfortable
assuming all birds can fly if the proba-
bility of a bird flying were only 0.75?
The answer depends on what actions
might be taken based on the assump-
tion that a bird can fly. These actions
depend, in turn, on the problem-solv-
ing situation. Unfortunately, a likeli-
hood analysis of nonmonotonic rea-
soning would judge such a default
assumption about flying birds equally
acceptable to a default assumption
about surgery.
Thus, the likelihood view of non-
monotonic reasoning must be extend-
ed to include the notion of utility.
The general principle implied by
threshold analysis is as follows: As
the consequences of an invalid non-
monotonic conclusion become seri-
ous, p* increases, and the acceptable
probability of error decreases. Thus,
two nonmonotonic conclusions with
the same probability of being invalid
might not both be appropriate if the Figure 3. But, in Fact, TWEETY Can’t Fly.
consequences of error for one are seri- The undesirable consequences of such actions underlie the appropriateness of nonmono-
ous, and those for the other are tonic conclusions.
insignificant. Consequently, we derive
the following lesson from the corre-
spondence between decision theory tonic assertion of higher priority patient to surgery).
and default reasoning: could be added that specified to refer As the number of relevant symp-
The appropriateness of a nonmono- for surgery when the patient has two toms increases, the number of neces-
tonic conclusion is, in part, dependent features that suggest appendicitis: sary logical statements increases dra-
on the (un)desirability of acting on an fever and high white–blood-cell count. matically. In the worst case, O(2n) log-
incorrect conclusion. Thus, patients would be referred to ical assertions (all combinations)
surgery unless they lacked these two must be written, where n is the num-
A Practical View indicators. This approach is oversim- ber of relevant pieces of evidence.5 On
plified because many pieces of evi- average, however, a few strong pieces
of the Two Approaches dence support (and detract from) the of evidence can dictate a decision
hypothesis that the patient has appen- regardless of the presence or absence
We showed how a single nonmono- dicitis, but it illustrates the imple- of other symptoms, thereby reducing
tonic default rule, used alone, can be mentation strategy. A set of non- the number of logical statements
suboptimal. In theory, however, non- monotonic logic statements can be required.
monotonic planning systems would constructed whose antecedents enu- The computational properties of the
contain many more logical statements merate all combinations of symptoms decision-theoretic approach are similar
that, when used in concert, produce resulting in a probability of appendici- to those of nonmonotonic logics. The
the desired behavior. We now infor- tis above the threshold, p*, and whose decision-theoretic approach entails the
mally analyze how these complex log- consequents conclude that referral to use of an inference network (Pearl
ical systems compare with their deci- surgery is the optimal plan. To view 1986) or influence diagram (Shachter
sion-theoretic counterparts. this situation from a decision-theoret- 1986) to update the probability of
Consider how the appendicitis prob- ic perspective—the “gold standard” appendicitis based on additional evi-
lem might be implemented as a com- —is to conclude surgery only if the dence. The worst case occurs when
plete system. Additional domain probability of appendicitis is greater each symptom is dependent on all oth-
knowledge would be encoded in than p*. As we showed, p* is comput- ers: O(2n) logical statements (that is,
default rules to specify those cases for ed from the undesirability of inappro- conditional probabilities) are needed in
which it is inappropriate to operate on priately referring a patient to surgery this case. If all symptoms are indepen-
a patient. For example, a nonmono- (or inappropriately not referring a dent of one another, only O(n) state-

SPRING 1989 43
ments are needed, each rule encoding achieve further modularity by separat- tion.
an update in the belief in appendicitis ing likelihood knowledge from prefer- Some researchers argue that these
based on the presence or absence of ence knowledge (that is, probabilities issues of problem dependence might
one symptom. In practice, the number from utilities). The relative modulari- be immaterial, because the effect of
of dependencies is usually intermedi- ty and explicitness of a decision-theo- the (relatively problem-independent)
ate. Thus, the decision-theoretic and retic system can simplify knowledge probabilities might far outweigh the
nonmonotonic logic approaches have base updating. To modify a decision- effect of the more problem-dependent
similar combinatoric properties when theoretic birdbath system to solve utilities. We find this argument
implemented. cliff problems, the system builder appealing. On the one hand, it advo-
need not modify the knowledge about cates distinguishing and analyzing the
TWEETY and about flying. The probabilities and utilities of a problem
Nonmonotonic Logics knowledge base only needs to be aug- to determine the applicability of a par-
and Modularity mented with the preferences for the ticular problem-solving technique. We
consequences of dropping birds off agree with this point and view it as
We define a logical assertion as modu- cliffs. one of the key lessons of our analysis.6
lar when its validity does not depend On the other hand, this argument
on the validity of other assertions. Implications for AI Research implies that nonmonotonic logics are
This definition is adapted from Heck- most appropriate when the utilities
erman and Horvitz (1987), where the As we showed, the combinatorics of are relatively unimportant (that is, in
concept is defined and discussed in the decision-theoretic and default rea- low-stakes situations when the deci-
the context of production rules. soning approaches are similar. Howev- sion maker is relatively indifferent to
We now analyze nonmonotonic rea- er, the default reasoning approach suf- the consequences of action). We also
soning and decision-theoretic reason- fers from a problem dependence that agree with this point; its frequent rep-
ing from the standpoint of modularity. complicates knowledge base mainte- etition would ensure that caution is
As we discussed, the validity of each nance. Because nonmonotonic asser- exercised when nonmonotonic logic is
nonmonotonic assertion depends on tions have been organized and ordered applied to high-stakes problem-solv-
the existence of other assertions in to reflect accurately a particular plan- ing situations.
the knowledge base. This dependency ning situation with a particular p*,
is an explicitly nonmodular property the validity of these assertions is Commonsense Knowledge
of nonmonotonic logics. For example, inherently dependent on the goals of Attempts to encode large amounts of
the likelihood that TWEETY can fly is the problem. Consequently, these problem-independent information
a property of our knowledge about nonmonotonic assertions tend to be (sometimes called commonsense
TWEETY and rarely changes as the specific to a particular planning prob- knowledge [McCarthy 1986; Lenat,
problem-solving goals change. In con- lem. We now examine the implica- Prakash, and Shepherd 1986]) are pred-
trast, as illustrated by our discussion tions of this conclusion for several icated on the assumption that the
of the birdbath and cliff examples, the current AI research issues: planning, knowledge representation used is
utility of assuming conclusions about commonsense knowledge, and analog- problem independent (because it is
TWEETY often varies among prob- ic reasoning. isomorphic to logic). As we discussed,
lem-solving situations (Keeney 1986). this property does not hold for non-
Use of Nonmonotonic
Furthermore, this knowledge is not monotonic reasoning.
explicitly represented. A nonmono- Logics for Planning Consider the implications of this
tonic reasoning system constructed The appeal of nonmonotonic logics is, observation for the CYC system
for birdbath problems would need in part, derived from the presumption (Lenat, Prakash, and Shepherd 1986),
substantial modification to ensure that they retain the same appealing which employs frame-based inheri-
adequate performance on cliff prob- characteristic as first-order logic: The tance (a form of nonmonotonic rea-
lems. The interdependence (nonmod- representation remains independent soning) to encode commonsense
ularity) among assertions makes of the problem-solving task. As we knowledge. For example, when a
updating and augmenting of the just showed, however, the validity of a knowledge engineer is encoding
knowledge base far more difficult and nonmonotonic assertion is inherently knowledge about birds in CYC, a bird
time consuming. Thus, the perfor- dependent on other assertions in the frame might be constructed contain-
mance of nonmonotonic systems is knowledge base and, thus, on the ing a can-fly slot with value true
highly dependent on the problem- problem-specific goals. This conclu- (figure 4). Several other frames, such
solving goals for which these systems sion undermines one of the frequent as hawk and seagull, might inherit
were designed. justifications for the use of nonmono- infor mation from the bird frame.
In comparison, the decision-theoret- tonic logics in the implementation of Other descendants of the bird frame,
ic approach maintains modularity automatic problem solvers (Hayes such as the frames for penguin and
when possible through the use of rep- 1977) because it shows that nonmono- ostrich, would override the can-fly
resentations that explicitly encode tonic logics have the same problem- slot with the value false. When the
dependencies if they exist (Pearl 1986; dependent features as other represen- system encounters a bird for which no
Shachter 1986). These representations tations that perform the same func- more specific frame exists, it assumes

44 AI MAGAZINE
that the bird can fly because general
bird knowledge is assumed by default.
However, there are exceptions (such BIRD
as brain-damaged birds or roast CAN FLY: YES
HAS FEATHERS: YES
turkeys). Without additional knowl-
edge, a planning system would then
act as if these invalid assumptions
were true. As we noted earlier in our
discussion of appendicitis, the conse- HAWK PENGUIN
quences of such action can sometimes EATS MICE: YES CAN-FLY: NO
be dire.7 ... ...
SEAGULL
One frequent counterargument to
...
this line of reasoning is that it ignores
other information which we might
have about the problem. More domain
knowledge might be needed, such as
frames that describe brain-damaged
birds. This additional knowledge
increases the number of problems for Figure 4. The Use of Frame-Based Inheritance
which the system’s reasoning is to Represent Knowledge about Birds.
appropriate because it lowers the Inheritance occurs along the lines between frames. Inherited knowledge can be overrid-
probability of an unanticipated excep- den by specific knowledge in the local frame.
tion. The possibility for errors still
exists, however (What about the Mal-
tese falcon?). In fact, as long as we are analogy can provide a correct solution free, might be an appropriate knowl-
not certain that a bird can fly, we can (Russell 1986). This statement, how- edge representation for use in com-
always imagine a problem-solving sit- ever, doesn’t consider the effect of monsense knowledge bases. Thus,
uation in which the consequences of utility information. For example, a decision theory represents an impor-
error are so undesirable that we still wooden table is quite analogous to a tant candidate for the implementation
do not want to make the default chair if the problem solver is looking of planning systems.
assumption a rule. Of course, all for a place to sit. However, an analogy We view our conclusions as an
exceptions not explicitly represented between the table and kindling is extension of Shoham’s in that he pro-
could be excluded from our definition especially valuable if the problem poses a logic formalism which iden-
of bird. Then the inferences would no solver needs warmth. The aptness of tifies the preference-based nature of
longer be nonmonotonic inferences; these analogies might well be related nonmonotonic inferences (Shoham
they would be logically valid. to the usefulness or utility of the pos- 1987). We provide a quantitative deci-
As we saw, a knowledge engineer sible solutions as well as to their like- sion-theoretic basis for this conclu-
who encodes knowledge in a system lihood. The utility-dependent nature sion and argue that the preference-
that employs frame-based inheritance of analogic reasoning might be poorly based nature of these nonmonotonic
must impose an estimate of the unde- understood in part because most inference systems limits their utility.
sirability of invalid conclusions about empirical studies of analogy have These limitations suggest that alter-
birds (and many other things). These examined opinions about the similari- native representations might offer
judgments can represent an average of ty of concepts essentially devoid of greater promise for the solution of
the consequences of the problem-solv- utility information, such as colors problems of planning under uncertain-
ing situations the knowledge engineer varying in lightness and saturation or ty which require explicit representa-
can anticipate, or they can represent circles varying in size (Shepard 1981). tion of uncertainty and tradeoffs.
an assessment of the utilities for one We envision decision-theoretic for-
particular problem. In either case, rea- Conclusions malisms that can be used to guide the
soning systems that might someday selection of appropriate inferences
use the knowledge base to solve par- Decision theory is designed to main- based on explicit encoding of prefer-
ticular problems can be led astray by tain an explicit separation between ences. For example, a knowledge engi-
invalid default assumptions. Thus, likelihood knowledge, which depends neer would represent commonsense
the knowledge in inheritance frames on the characteristics of the world, knowledge in a probabilistic inference
in such a knowledge base cannot truly and utility infor mation, which network (Pearl 1986; Shachter 1986;
be problem independent. depends on the problem-solving goals. Lehmann 1988). Problem-specific
This separation provides modularity preferences would be used to repre-
Analogic Reasoning and flexibility when a knowledge base sent particular planning situations
Many researchers have proposed that is adapted to a new planning situation because these preferences describe the
the value of an analogy can be judged and indicates that likelihood knowl- problem-solving goals and the trade-
in terms of the likelihood that a given edge, because it is relatively utility offs among them. This utility infor-

SPRING 1989 45
mation would be combined with the for Prospective Reasoning Systems. In Pro- Intelligence, 471–476. Menlo Park, Calif.:
relevant likelihood-based common- ceedings of the Sixth National Conference American Association for Artificial Intelli-
sense knowledge to solve the problem on Artificial Intelligence, 783–788. Menlo gence.
Park, Calif.: American Association for Holloway, C. A. 1979. Decision Making
at hand. Several researchers have
Artificial Intelligence. under Uncertainty: Models and Choices.
described instances in which the sepa-
Cox, R. 1946. Probability, Frequency, and Englewood Cliffs, N.J.: Prentice-Hall.
ration of likelihood knowledge and
Reasonable Expectation. American Journal
utility information can be incorporat- of Physics 14(1): 1–13.
Horvitz, E. J. 1988. Reasoning under Vary-
ed into the design of an expert system ing and Uncertain Resource Constraints.
Downing, K. L. 1987. Diagnostic Improve- In Proceedings of the Seventh National
(Langlotz et al. 1987; Slagle and Ham-
ment through Qualitative Sensitivity Conference on Artificial Intelligence,
burger 1985; Klein and Finin 1987;
Analysis and Aggregation. In Proceedings 111–116. Menlo Park, Calif.: American
Langlotz 1987). of the Sixth National Conference on Association for Artificial Intelligence.
Because decision theory provides a Artificial Intelligence, 789–793. Menlo Keeney, R. L. 1986. Value-Driven Expert
mechanism for combining likelihood Park, Calif.: American Association for Systems. In Proceedings of the Conference
and utility knowledge during problem Artificial Intelligence. on Multi-Attribute Decision Making via
solving, it represents an important Doyle, J. 1985. Reasoned Assumptions and O.R.-Based Expert Systems, 141–162. Uni-
tool that should be considered by Pareto Optimality. In Proceedings of the versitat Passau.
builders of planning systems. Used in Ninth International Joint Conference on Klein, D., and Finin, T. 1987. What’s in a
conjunction with nonquantitative Artificial Intelligence, 88–90. Menlo Park, Deep Model? A Characterization of Knowl-
techniques to generate candidate Calif.: International Joint Conferences on edge Depth in Intelligent Safety Systems.
plans (Wellman 1986), the decision- Artificial Intelligence. In Proceedings of the Tenth International
theoretic approach not only enhances Etherington, D. W. 1987. Formalizing Non- Joint Conference on Artificial Intelligence,
the ability of a system to select the monotonic Reasoning Systems. Artificial 559–562. Menlo Park, Calif.: International
Intelligence 31(1): 41–86. Joint Conferences on Artificial Intelli-
optimal plan when the explicit con-
sideration of tradeoffs is essential but Fikes, R. E., and Nilsson, N. J. 1971. gence.
also provides a sound basis for knowl- STRIPS: A New Approach to the Applica- Kosy, D. W., and Wise, B. P. 1984. Self-
edge engineering decisions. The plan- tion of Theorem Proving to Problem Solv- Explanatory Financial Planning Models. In
ing. Artificial Intelligence 2: 189–208. Proceedings of the Fourth National Confer-
ning problems we address indicate the
ence on Artificial Intelligence, 176–181.
continuing need to combine estab- Friedland, P. E., and Iwasaki, Y. 1985. The
Menlo Park, Calif.: American Association
lished normative techniques with the Concept and Implementation of Skeletal
for Artificial Intelligence.
symbolic reasoning and representa- Plans. Journal of Automated Reasoning
1(2): 161–208. Langlotz, C. P. 1987. Axiomatic Expert
tion techniques developed in AI. We Systems. In Proceedings of the Eleventh
believe that these combinations are Hanks, S., and McDer mott, D. 1986.
Annual Symposium on Computer Applica-
fruitful and are likely to lead to Default Reasoning, Non-Monotonic Log-
tions in Medical Care, 49–55. Washington
ics, and the Frame Problem. In Proceedings
enhanced decision support for prob- D.C.: IEEE Computer Society Press. Also
of the Fifth National Conference on
lems of planning under uncertainty. in 1989. Computer Methods and Programs
Artificial Intelligence, 328–333. Menlo
in Biomedicine. Forthcoming.
Park, Calif.: American Association for
Acknowledgments Artificial Intelligence. Langlotz, C. P., Shortliffe, E. H., and Fagan,
L. M. 1988. A Methodology for Generating
We thank Benjamin Grosof and Mark Hayes, P. J. 1977. In Defense of Logic. In
Computer-Based Explanations of Decision-
Musen for their helpful comments on ear- Proceedings of the Fifth International Joint
Theoretic Advice. Medical Decision Mak-
lier drafts of this article, and Lawrence Conference on Artificial Intelligence,
ing 8(4): 290–303.
Fagan for his comments on oral presenta- 559–565. Menlo Park, Calif.: International
tions of this work. We also thank Michael Joint Conferences on Artificial Intelli- Langlotz, C. P., Shortliffe, E. H., and Fagan,
Kahn, David Klein, Michael Wellman, gence. L. M. 1986. Using Decision Theory to Jus-
David Heckerman, and Eric Horvitz for tify Heuristics. In Proceedings of the Sixth
Heckerman, D. E. 1986. Probabilistic Inter-
many useful discussions. Support for this National Conference on Artificial Intelli-
pretations for MYCIN’s Certainty Factors.
work was provided by the National Library gence, 215–219. Menlo Park, Calif.: Ameri-
In Uncertainty in Artificial Intelligence,
of Medicine under grant no.LM-04136 and can Association for Artificial Intelligence.
eds. J. F. Lemmer, L. N. Kanal, 167–196.
the National Science Foundation under New York: North Holland. Langlotz, C. P., Fagan, L. M., Tu, S. W.,
grant no. IST83-12148. Computing facili- Sikic, B. I., and Shortliffe, E. H. 1987. A
Heckerman, D. E., and Horvitz, E. J. 1987.
ties were provided by the SUMEX-AIM Therapy Planning Architecture That Com-
On the Expressiveness of Rule-Based Sys-
resource under National Institutes of bines Decision Theory and Artificial Intel-
tems for Reasoning with Uncertainty. In
Health grant no. RR-00785 and by the ligence Techniques. Computers and
Proceedings of the Sixth National Confer-
Xerox Corporation. Biomedical Research 20(3): 279–303.
ence on Artificial Intelligence, 121–126.
Menlo Park, Calif.: American Association Lehmann, H. P. 1988. Knowledge Acquisi-
References for Artificial Intelligence. tion for Probabilistic Expert Systems. In
Proceedings of the Twelfth Annual Sympo-
Charniak, E., and McDermott, D. 1985. An Henrion, M., and Cooley, D. R. 1987. An
sium on Computer Applications in Medi-
Introduction to Artificial Intelligence. Experimental Comparison of Knowledge
cal Care, 73–77. Washington D.C.: IEEE
Reading, Mass.: Addison-Wesley. Engineering for Expert Systems and for
Computer Society Press.
Decision Analysis. In Proceedings of the
Cohen, P. R., Greenberg, M., and DeLisio, Lenat, D., Prakash, M., and Shepherd, M.
Sixth National Conference on Artificial
J. 1987. MU: A Development Environment 1986. CYC: Using Common Sense Knowl-

46 AI MAGAZINE
edge to Overcome Brittleness and Knowl- Russell, S. J. 1986. A Quantitative Analysis shop on Uncertainty in Artificial Intelli-
edge Acquisition Bottlenecks. AI Magazine of Analogy by Similarity. In Proceedings of gence, 311–318. Menlo Park, Calif.: Ameri-
6(4): 65–85. the Fifth National Conference on Artificial can Association for Artificial Intelligence.
Lifschitz, V. 1986. Pointwise Circumscrip- Intelligence, 284–288. Menlo Park, Calif.:
tion: Preliminary Report. In Proceedings of American Association for Artificial Intelli- Notes
the Fifth National Conference on Artificial gence. 1. The term decision tree has sometimes
Intelligence, 406–410. Menlo Park, Calif.: Savage, L. J. 1972. The Foundations of been used to denote a recursive structure
American Association for Artificial Intelli- Statistics. New York: Dover. that represents a sequential classification
gence. process (Quinlan 1987). We use the term in
Shachter, R. D. 1986. Intelligent Proba-
McCarthy, J. 1986. Applications of Cir- bilistic Inference. In Uncertainty in its traditional sense—to represent the pos-
cumscription to Formalizing Common Artificial Intelligence. New York: North sible scenarios that arise when we plan
Sense Knowledge. In Proceedings of the under uncertainty.
Holland. Also in 1985. Proceedings of the
AAAI Workshop on Nonmonotonic Rea- 2.We illustrate our approach using a vari-
AAAI Workshop on Uncertainty and Prob-
soning, 295–324. Menlo Park, Calif.: Amer- ant of Reiter’s default logic (Reiter 1980).
ability in Artificial Intelligence. Menlo
ican Association for Artificial Intelligence. Our basic argument also applies to other
Park, Calif.: American Association for
McCarthy, J. 1980. Circumscription—A Artificial Intelligence. formalisms for default reasoning, including
For m of Non-Monotonic Reasoning. circumscription (McCarthy 1980), McDer-
Shepard, R. N. 1981. Division 3 Presiden-
Artificial Intelligence 13 (1–2): 27–39. mott and Doyle’s nonmonotonic modal
tial Address, American Psychological
logic (McDermott and Doyle 1980), and
McDermott, D., and Doyle, J. 1980. Non- Association, Los Angeles. Also reprinted in
several other formalisms.
Monotonic Logic I. Artificial Intelligence Russell, S. J. 1986. A Quantitative Analysis
13 (1–2): 41–72. of Analogy by Similarity. In Proceedings of 3. Heckerman (1986) has subsequently
the Fifth National Conference on Artificial related the CF calculus to probability theo-
Michalski, R. S., and Winston, P. H. 1986.
Intelligence, 248–288. Menlo Park, Calif.: ry.
Variable Precision Logic. Artificial Intelli-
gence 29(2): 121–146. American Association for Artificial Intelli- 4. The derivation of p* is based on the
gence. observation that at the threshold, the deci-
Nutter, J. T. 1987. Uncertainty and Proba- sion maker is indifferent between the two
bility. In Proceedings of the Tenth Interna- Shoham, Y. 1987. Nonmonotonic Logics:
Meaning and Utility. In Proceedings of the choices. Consequently, Pauker and Kassir-
tional Joint Conference on Artificial Intel- er derive p* by mathematically expressing
ligence, 373–379. Menlo Park, Calif.: Inter- Tenth International Joint Conference on
Artificial Intelligence, 388–393. Menlo the expected utility of each choice in
national Joint Conferences on Artificial terms of the probabilities and utilities
Intelligence. Park, Calif.: International Joint Confer-
ences on Artificial Intelligence. involved, then setting these expressions
Pauker, S. G., and Kassirer, J. P. 1980. The equal to one another and solving for p.
Threshold Approach to Clinical Decision Shoham, Y. 1986. Chronological Ignorance:
Time, Nonmonotonicity, Necessity, and 5. For the purposes of this combinatoric
Making. New England Journal of Medicine analysis, we assume that evidence is con-
302(20): 1109–1117. Causal Theories. In Proceedings of the
Fifth National Conference on Artificial strained to be either present or absent.
Pearl, J. 1986. Fusion, Propagation, and 6. A formal framework for addressing these
Intelligence, 389–393. Menlo Park, Calif.:
Structuring in Belief Networks. Artificial issues was proposed by Horvitz (1988).
American Association for Artificial Intelli-
Intelligence 29(3): 241–288.
gence. 7. Shoham (1987) makes a similar point
Pieper, R., Kager, L., and Nasman, P. 1982. Shortliffe, E. H., and Buchanan, B. G. 1975. when he discusses of the chances of being
Acute Appendicitis: A Clinical Study of A Model of Inexact Reasoning in Medicine. stabbed while walking on the street.
1018 Cases of Emergency Appendectomy. Mathematical Biosciences 23: 351–379.
Acta Chirurgica Scandinavia 148: 51–62. Curtis P. Langlotz is a National Library of
Slagle, J. R., and Hamburger, H. 1985. An
Poole, D. L. 1986. On the Comparison of Medicine predoctoral fellow in Medical
Expert System for a Resource Allocation
Theories: Preferring the Most Specific Informatics at Stanford University. In June
Problem. Communications of the ACM
Explanation. In Proceedings of the Fifth of this year, he will complete an M.D. and
28(9): 994–1004.
National Conference on Artificial Intelli- a Ph.D. in Medical Information Science
Tversky, A., and Kahneman, D. 1974. Judg-
gence, 144–147. Menlo Park, Calif.: Ameri- and begin residency training in diagnostic
ment under Uncertainty: Heuristics and
can Association for Artificial Intelligence. radiology. His research interests include
Biases. Science 185: 1124-1131.
planning under uncertainty, automated
Quinlan, J. R. 1987. Generating Production von Neumann, J., and Morgenstern, O. explanation, and the integration of deci-
Rules from Decision Trees. In Proceedings 1953. Theory of Games and Economic sion theory and AI techniques.
of the Tenth International Joint Confer- Behavior. New York: Wiley.
ence on Artificial Intelligence, 304–307.
Menlo Park, Calif.: International Joint von Winterfeldt, D., and Edwards, W. 1986.
Decision Analysis and Behavioral Edward H. Shortliffe is associate professor
Conferences on Artificial Intelligence. of Medicine and Computer Science at
Research. New York: Cambridge Universi-
Reiter, R. 1980. A Logic for Default Rea- Stanford University. He is the scientific
soning. Artificial Intelligence 13 (1–2): ty Press.
director of the Medical Computer Science
81–132. Wallsten, T. S., and Budescu, D. V. 1983. Group in the Knowledge Systems Labora-
Rich, E. 1983. Default Reasoning as Likeli- Encoding Subjective Probabilities: A Psy- tory and chief of the Division of General
hood Reasoning. In Proceedings of the chological and Psychometric Review. Man- Internal Medicine at Stanford University
Third National Conference on Artificial agement Science 29(2): 151–173. School of Medicine. His research interests
Intelligence, 348–351. Menlo Park, Calif.: Wellman, M. 1986. Qualitative Probabilis- include models for evidential reasoning
American Association for Artificial Intelli- tic Networks for Planning under Uncer- and representation techniques to support
gence. tainty. In Proceedings of the AAAI Work- advanced explanation capabilities.

SPRING 1989 47

You might also like