Professional Documents
Culture Documents
This article offers an account of the nature of fallibility and its educational significance
based on recent work in cognitive science and epistemology. It explains that fallibilism is
the dominant epistemology in education and explains that the doctrine of fallibilism is an
anti-dogmatic intellectual stance or attitude. It criticizes a type of fallibility that can only
maintain its antiskepticism by introducing specialized usages for “certain” and
“possibility”.
Page 1 of 24
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). © Oxford University Press, 2018. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in
Oxford Handbooks Online for personal use (for details see Privacy Policy and Legal Notice).
Despite widespread adherence, fallibilism plays little role in education or even in contemporary
philosophy of education. A recent major—638‐page—collection in the philosophy of education
(Curren 2003) contains no reference in the index to fallibilism or to fallibility, though I expect
that the authors all endorse fallibilism. Why does fallibility or fallibilism lack bite in education?
Should it be otherwise and, if so, how?
An obvious start to answering the first question is: complete victory. The battle against
epistemologies that aspire to certainty or incorrigibility is won. No serious educational
proposal will guarantee success. Knowing that future changes will alter (p. 84) current
policies does not differentiate among prospective policies. Without contenders for
controversy, there is no visibility.
But this is only a part of the explanation: even if fallibilism has no battles to fight, it could
provide substantial guidance to policies. Democratic ideals, the authority of the teacher
in the classroom, students' use of textbooks, the centrality of mathematics in the
curriculum are hardly contestable, but still inform policy.
The major culprit is fallibility, which is the doctrine on which fallibilism rests. The
problem is that fallibility, as it is commonly understood, is too weak to justify the
significant recommendations or attitudes of fallibilism, as described above.
But how can fallibility motivate Elgin's recommendation? For the minimal inductive possibility of
error cannot be eliminated or even diminished by corrective devices. A factory is motivated to
implement a new quality‐control system only because the new system promises to lower its rate
of manufacturing defective products. It would hardly be motivated if the outcome is guaranteed
not to lower the rate.
Given the current understanding of fallibility, the response to the discovery of one's error
may just as well be the indifferent response to which we are all familiar: “I believed p, but
I was wrong. So what? I'm fallible.” The response can go even one step below
indifference: “I believed p, but it turned out false and I replaced it with the correct belief.
I learned from my mistake, so I am now more assured.”
Page 2 of 24
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). © Oxford University Press, 2018. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in
Oxford Handbooks Online for personal use (for details see Privacy Policy and Legal Notice).
Although the logical compatibility of the evidence and the denial of the inductively
reached conclusion is incontestable, the move from it to the possibility of error is not. The
possibility is supposed both to represent the logical evidence‐conclusion gap and to rule
out the conclusiveness or necessity of the evidence‐to‐conclusion relation.
Yet, we do often describe inductive inference in conclusive terms (of the necessity of the
conclusion given the premises): “John must still be in his office. I saw him go in a minute
ago and no one has come out since.” Conclusiveness characterizes knowledge—I know
John is in his office only if, given the evidence, it (p. 85) could not be otherwise. It is
contradictory to think “I know that John is in his office, but it is possible that he is not” or
even “John is in his office, but it is possible that he is not.”
These coherence‐of‐thought tests provide for a contrast with desire: of any particular
belief, I cannot think that it might be false. To do so is already to lose that belief. By
contrast, one can know that a particular desire, like the desire for a hot fudge sundae, is
bad to satisfy. Yet, there it remains.
The failed test implies that when a possibility of error remains one cannot know.
Fallibility, as the ineliminable possibility of error, is called upon to serve as a pivotal
premise for radical skepticism. The linchpin of the dreaming argument is that, compatible
with the appearances before me now of, for example, a computer screen, I cannot know
that it is a computer screen, since I cannot exclude the (way‐out) possibility that I am now
the subject of a very realistic dream. If someone affirmed, as above, that John must be in
his office and you insist that it is still possible that he is not, on a judgment that is so
obvious, you could cast into doubt most any knowledge claim.
These claims about possibility do not rest on an exotic or demanding notion of knowledge.
There is no claim that knowledge implies certainty. Urging fallibility as a reason to
replace knowledge as the central epistemological concept or goal with justified belief or
warranted assertibility does not help. Only when you have eliminated all other
possibilities are you warranted in asserting “The coffee is in the freezer.”
A related understanding of fallibility, taken by Peirce, Dewey, and Popper, is that future
inquiry may (or will) upset current conclusions.2 Godfrey‐Smith writes: “Popper placed
great emphasis on the idea that we can never be completely sure that a theory is true.
After all, Newton's physics was viewed as the best‐supported theory ever, but early in the
twentieth century it was shown to be false in several respects” (2003, p. 59; see also
Dewey 1986, chap. 2).
Page 3 of 24
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). © Oxford University Press, 2018. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in
Oxford Handbooks Online for personal use (for details see Privacy Policy and Legal Notice).
In a book that attempts to develop a Popperian philosophy of education, the authors write
in their introduction:
(p. 86)
But since we did uncover flaws in Newton's theory, which discoveries required examining
a wider variety of evidence than believed necessary at Newton's time, are we not entitled
to far greater assurance that Einstein's theory is basically correct, rather than less? I do
not doubt the heuristic value of the fate of Newton's theory for teaching a powerful
example of fallibility in scientific theorizing. Rather, I doubt whether this case should
serve as a paradigm for secure, but nevertheless fallible, beliefs. Would not much better
candidates be Descartes' “two and three added together are five” and “a square has no
more than four sides”; or, if a clearly empirical or contingent example is sought, how
about “The earth was at some time habitable” or “There are rocks” or “There is no Santa
Claus”?
These latter examples have, unlike Newton's theory, withstood the test of time, and it is
not believable that they will be discovered false in the future. We are certain (in ordinary
usage) of them—counterexamples to the ringing claim that people “can never be certain
that their ideas or beliefs are absolutely true.” If fallibility is the universally applicable
doctrine that is alleged, why can't we readily produce examples like these, where yet the
belief turned out false? These beliefs are not only painfully obvious, but they are also
implied by many other beliefs. The surrendering of Newton's theory hardly made a ripple
in our corpus of beliefs. However, if any of my mundane examples were surrendered, the
disruption would be profound.
Page 4 of 24
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). © Oxford University Press, 2018. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in
Oxford Handbooks Online for personal use (for details see Privacy Policy and Legal Notice).
2. Robust Fallibility
Is there another way to understand fallibility than the minimal constancy view, which
would support a significant educational role for the possibility of error?4 I will answer
both parts in the affirmative, though an adequate defense of the former would enter deep
epistemological waters that I cannot do here.
We want an interpretation of a thought like “I know that the coffee is in the freezer,
though I might be wrong” or an assertion like “The coffee is in the freezer, though I might
be wrong” that is consistent. The second (“might”) conjunct is not a possibility
incompatible with knowledge. Even if the evidence does conclusively establish the
conclusion, I can conceive ways, from the position I am in to judge, that I am wrong. One
may coherently think: “The coffee is in the freezer, but it seems to me, from the
(epistemic) position in which I make that judgment, that (p. 87) others are in a position to
put forward, as a warranted claim, a similar judgment, and I cannot discriminate their
position from my own.”
The view of fallibility put forth above and to be developed rejects or qualifies both horns
of this dilemma.
To pursue the epistemic import of a robust fallibility, recall Descartes' very general, yet
natural, line of thought in Meditation I to shake his confidence in even “obvious” or
necessary truths like that 2+3 = 5: “since I judge that others sometimes make mistakes
in matters that they believe they know most perfectly, may I not, in like fashion, be
deceived every time I add two and three or count the sides of a square, or perform an
even simpler operation, if that can be imagined?” (Descartes 2002, p. 461).6
Page 5 of 24
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). © Oxford University Press, 2018. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in
Oxford Handbooks Online for personal use (for details see Privacy Policy and Legal Notice).
Descartes' line of thought comes close to how we invoke fallibility to disturb those we
think of as overly confident or dogmatic. John is driving well over the speed limit. When
you tell him to slow down, he responds that he is a very good driver. You observe that
many of those who precipitated bad accidents thought of themselves as good drivers for
reasons seemingly as good as his own. Psychological studies confirm a kind of “Lake
Wobegon” effect in ratings for oneself: in a wide range of self‐judgments, (nondepressed)
subjects rate their abilities as better than others in peer groups—for example, friendlier,
less biased, better drivers, less conformist.7
Now these overratings may be so high as to generate a liability to error that is evidence
against one's believing that one is, e.g., a good driver. But only if they are so high is the
overrating incompatible with knowing that one is a good driver. (Do the high divorce
rates imply that no one can know that his or her own marital prospects are good?) But if
not, they still rationally lower one's confidence. Your pointed observation does not
directly challenge John's belief that he is a good driver. Instead, you are proposing to him
that he can easily envisage others, whom he takes to be similarly situated, who did
precipitate accidents. Consequently, John should lessen his confidence and rethink his
speeding.
However, the Cartesian line, also implicit in the Newton‐Einstein comparison, requires a
qualification. Descartes' worry arises in comparing his own confidence (p. 88) with that
of others of equal assurance, whom he takes to have been mistaken. However, first,
individuals' strength of attitude is sensitive to changes in their circumstances that do not
constitute reasons supportive or undermining of beliefs. A student becomes less confident
of his geometry proof for a final exam than merely for homework, though his reasons—the
steps in the proof—remain the same. Second, the conclusion sought of one's liability to
error requires the further assumption that one's current position does not compensate for
those errors. In earlier times, people were sure of a child's temperature by touch with the
back of one's hand or of the lines in a poem by memory. Now we have thermometers and
books or computers, respectively.
Page 6 of 24
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). © Oxford University Press, 2018. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in
Oxford Handbooks Online for personal use (for details see Privacy Policy and Legal Notice).
Robust fallibility is about our proneness to error, not just its (constant, minimal)
possibility. If there is such a proneness, there should be evidence for it. If so, that
evidence is worth study by educators to evaluate curricula and by students to appreciate
their own errors, including the sources of those errors (poor understanding, confusion,
hasty judgment, forgetfulness, false assumptions, distraction…). Studying that evidence
should contribute to the students' self‐understanding and to weakening that proneness.8
The proposal, consonant with appreciation of our robust fallibility, is then for the study of
errors, relying on research on reasoning as evidence for our proneness to error.
A typical answer is: “As the season commences a player will, I think, become less
motivated to impress people with a powerful bat—he is taking a sort of (p. 89) ho‐hum
attitude about it” (compare to Kahneman and Tversky 1982). The favored response is
grounded in chance—more specifically, the law of large numbers and regression. Earlier
in the season, with far fewer times at bat, deviation from a well‐established average is
more likely. With the much greater number of times at bat by the end of the season, the
chance influences on deviations from average performance are more likely to balance out
to smaller deviations from that average. Batting averages that reflect greater times at bat
are less subject to the influence of chance and so provide better evidence “as to how well
a batter hits” (Jepson, Krantz, and Nisbett 1983, p. 500). In general, the experimenters
found that “Most of these answers [of subjects] presumed, either tacitly or explicitly, that
the sample was adequate and showed no recognition of the uncertain or probabilistic
nature of events of the kind presented in the problems” (p. 496).
A related example is to explain the alleged sophomore slump, whereby outstanding rookie
baseball players do substantially worse their second year. Subjects give similar answers—
for example, success spoiled them—in response to which Nisbett and Ross (1980)
comment: “by chance alone, some mediocre athletes will perform exceptionally well in
their first year but perform less well in subsequent years” (p. 164). The explanations that
subjects offer—for example, that the batter became too cocky—fit common patterns:
performers who become conceited owing to their success in a skilled, demanding activity
relax their efforts and then do worse.
The weaknesses in reasoning revealed by these studies should fuel, I assume, efforts to
promote an education that seeks to develop an informed and critical citizenry. Although I
will provide further and varied illustration of such weaknesses, I cannot undertake to
Page 7 of 24
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). © Oxford University Press, 2018. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in
Oxford Handbooks Online for personal use (for details see Privacy Policy and Legal Notice).
establish this assumption here, since it requires showing that many faulty judgments in
public policy rely for their support on distortions of evidence, argument, and reasoning,
which exploit the weaknesses that the above studies and their neighbors expose.
Distortions in argument can anticipate success at persuasion because our curiosity and
what we accept as explanations to satisfy that curiosity can be turned on us, as the
baseball examples suggest. Another captivating example is in Michael Schermer's column
in Scientific American (“Mr. Skeptic Goes to Esalen”): “Once it become known that Mr.
Skeptic was there [at Esalen Institute], for example, I heard one after another ‘How do
you explain this?’ story, mostly involving angels, aliens, and the usual paranormal
fare” (p. 38). The request for an explanation is really a challenge: if Schermer cannot
supply a better explanation, he is rationally bound to endorse the one offered. In
accepting the legitimacy of the “why” question, we bias ourselves against explaining‐
away some of the alleged surprising phenomena with a chance account: “Look, there is
no ‘why’ to it. It's just chance.”
Another set of studies is from elementary physics (mechanics), and I quote from Howard
Gardner's (1991) presentation:
The goal of the game is to give instructions to the dynaturtle such that it will hit a
target and do so with minimum speed at impact.…
So described, the game sounds simple enough… Yet nearly everyone at both levels
of expertise fails dismally. The reason, briefly, is that success at the game requires
an understanding and application of Newton's laws of motion. To succeed, the
player must be able to take into account the direction in which and the speed with
which the dynaturtle has already been moving. Whatever their formal training,
however, players of this game reveal themselves to be dyed‐in‐the‐wool
Aristotelians. They assume that, so long as they aim the dynaturtle directly at the
target, they will succeed, and they are mystified when the KICK does not result in
the desired collision. (pp. 152–53)
Gardner concludes that “This is not a case of simple ignorance of the principle under
investigation; many of the students know—and can state—the laws on which they would be
drawing. Nor is it a case of factual errors; the students are not being asked whether the sun is a
star, or a dolphin is a fish” (p. 155).
These two illustrations can be marshaled against another instance of the above false
dilemma that made robust fallibility itself appear impossible. The dilemma in this
configuration concludes that robust fallibility cannot be of educational import:
Page 8 of 24
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). © Oxford University Press, 2018. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in
Oxford Handbooks Online for personal use (for details see Privacy Policy and Legal Notice).
Propensities to err are either very weak, in which case, though they may apply
generally, they provide neither direction for policy nor sufficient prospective
benefits to warrant investment; or else they are substantial, in which case they
will be too variable across individuals and domains to serve as a basis for general
(uniform) policies.
Against this false dilemma, I offer the empirically based illustrations above, which stand for
many others, of widespread difficulties that most any student would confront. The fallacies
derive from entrenched and heavily used ways of thinking. They are evidence of a substantial
proneness to error, and yet, contrary to the dilemma, they hold generally without implying a
radical skepticism.
The faults that the research uncovers have been construed as impediments to learning,
across the liberal arts curriculum. In presenting these problems in The Unschooled Mind
(also his The Disciplined Mind), Gardner means to show:
The two previous illustrations are, however, initially problematic as evidence for a robust
fallibility. For fallibility teaches its sharpest lesson when it applies to judgments or beliefs
that are warranted, which does not hold in these cases (participants are wrong about how
to explain the baseball discrepancies or how to kick the dynaturtle). The lesson of
fallibility is close by, though: One can reasonably think one's judgment warranted—and it
not be. But fallacious patterns of reasoning can even be involved in judgments that are
warranted. Those who failed with di Sessa's dynaturtle test can nevertheless succeed at a
wide range of judgments of motion. Students who commit well‐known fallacies with
conditionals, like affirming the consequent, can be warranted in judging valid that “if
today is Tuesday, I have class. So since today is Tuesday, I do have class” and invalid that
“if I do not practice, I will not make the varsity. So if I do practice, I will make it.”
Page 9 of 24
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). © Oxford University Press, 2018. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in
Oxford Handbooks Online for personal use (for details see Privacy Policy and Legal Notice).
The proneness to error of fallacious patterns of reasoning need not reveal itself in a
substantial actual rate of error. To the contrary, these patterns may not only guide
successful action, but they will be reinforced, within our narrow confines of experience,
by the success of those actions. The weaknesses are uncovered only by going beyond
those confines or by contriving tests where failures are more vivid and precise, as with
the dynaturtle experiments.
It is not only the world that may be misleadingly cooperative as far as exposing limits on
patterns of reasoning but also pervasive human activities. In the baseball cases, the
exclusion of the chance account is reinforced by ordinary conversation. To ask why there
is the discrepancy between early and late batting averages, or freshman and sophomore
batting performance, is to presuppose that the discrepancy calls for explanation and so it
is not a chance phenomenon. Since hearers normally accommodate to a speaker's
presupposition, we are conversationally disposed to place a chance account off the table,
as assumed above. (There are, of course, other reinforcers to this nonchance,
deterministic explanation for the discrepancy in batting performance.)
The standard pattern of learning from experience is that—for example—you are trying to
get to the 50th Street stop from Chambers Street. You take the first uptown train to arrive
on the line with a 50th Street stop—the #2. But the #2 does not stop at 50th Street,
proceeding from Times Square (42nd Street) directly to the 72nd Street stop. Realizing
the error, you check the map to discover that 50th Street is a local stop and the #2 is an
express train. You lose your original belief that any train on the line with your destination
station and heading in its direction will stop at the (p. 92) station. You replace that belief
with a qualified one to take account of the difference between express and local stations.
This pattern can be represented as:
of belief, where the arrow stands for some immediate (e.g., intentional, causal, epistemic)
connection. (This is roughly the pattern taken as central in Ohlsson 1996.)
The cases that we are concerned with are those in which the first three steps, especially
the middle two, are not evident in normal experience. There are no simple actions or
predictions that could yield clear falsifying observations. In short, you must endorse your
own reasoning. The reasoning by which you draw a conclusion thereby also provides you
with sufficient reason neither to question that reasoning nor to conceptualize various
outcomes as falsifying. The explanation of the discrepancy in batting averages as due to
the batters' becoming too cocky and cutting down on practice will deter me from putting
Page 10 of 24
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). © Oxford University Press, 2018. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in
Oxford Handbooks Online for personal use (for details see Privacy Policy and Legal Notice).
that explanation to the test or to take real notice of whether it holds up over time and
across players. To the contrary, that explanation satisfies me. So it undermines motivation
to pursue further analysis, while rendering salient those cases that appear confirmatory.
Contrast this self‐protection with reasoning from desire to action. Given my desire to eat
a hot fudge sundae, I reason practically to what I might do to obtain and eat one. The
reasoning does not nullify the reasons I have to criticize the desire. Or, to recall an earlier
example, if I conclude that the coffee is in the freezer by elimination, but I do not find it
there, the guiding belief that there is no further location to search is noticeably
jeopardized.
(p. 93)
Students, even young ones, already have vast stores of knowledge that regularly awake
them in conversation to highly implausible claims, regardless of the arguments or
evidence alleged for them. You cannot wildly fool many of the people much of the time.
Page 11 of 24
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). © Oxford University Press, 2018. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in
Oxford Handbooks Online for personal use (for details see Privacy Policy and Legal Notice).
Unfortunately, this optimism works only to expose the strikingly implausible. Optimism
sinks rapidly for claims and conclusions that require reflection to draw out the
implausibility. The example I turn to is another of Gardner's illustrations of widespread
and deep‐seated misconceptions. Gardner writes:
If one is told that there are six times as many students as professors and that
there are ten professors, nearly everyone can compute the number of students.…
But when students are asked to write out a formula that captures the relevant
proportion, using S for students, and P for professors, the majority of college
students fail.… The majority of college students write the formula 6S = P.(1991, p.
160)
Gardner comments: “Yet this formula would yield the astonishing conclusion that if there are 60
students, there would be 360 professors!” (p. 160).
Gardner's counterexample is a knock‐down objection to the solution. Students, who
answer wrongly, already have the competence to recognize that their solution is baldly
mistaken. But, of course, Gardner can only offer the example to illustrate misconceptions
because the students do endorse their answer. They fail to activate, or deactivate, their
own powerful critical resources.
The students who give the incredible answer find no discomfort in their reasoning. They
do not recognize a reason to check their answer, let alone to check it in a novel way,
which is another surfacing of the problem of self‐protective reasoning. Bad reasoning
does not come so marked, and whether reasoning is good or bad, it provides grounds for
a judgment, which are thereby grounds not to doubt that judgment. In cases like the
algebra one, as well as the earlier ones, the (p. 94) great lesson of fallibility is to be found
in discovering that one's own error is obstructed by the reasoning that yields that error.
Page 12 of 24
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). © Oxford University Press, 2018. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in
Oxford Handbooks Online for personal use (for details see Privacy Policy and Legal Notice).
An addition that the study of errors does require for the higher grades is the study of
reasoning, focusing on probability and statistics; explicitness of steps (assumptions) in
proofs and arguments; and the appreciation of correlative concepts of chance, causality,
randomness, proof, consistency, implication, and explanation. A more demanding
suggestion is that more of science be taught historically with original texts, as part of a
larger place for the history of ideas.10 Real errors, modifications, and dead‐ends become
apparent, yet placed in a web of beliefs where those errors emerge sensibly.11
The study of errors has a number of dimensions, which I will list and illustrate from our
examples so far, with further ones to be introduced. The lessons to be drawn next will be
in sharper focus if the errors illustrated are contrasted with simple errors of, say,
spelling, calculation, memory, or formulation, even if each is evidence of fallibility.
The discovery that you are wrong in the two main illustrations is very surprising. Subjects
confidently believed p to be true, but it turned out to be false. The surprise is second‐
order: I come to believe that what I believed, or thought I knew, was wrong. (Adler 2008)
Surprise, and appreciation of fallibility, cannot arise in an area of recognized ignorance,
but only where one thinks one knows. The surprise brings salience and focus to the
outcome or observation, as contrary to one's expectation and to one's poor
understanding.
As already emphasized, in the main two illustrations there were numerous, ordinary
experiences that presented contrary evidence, which the students or subjects missed. In
watching lots of baseball games, for example, subjects had the opportunity to discover by
observation alone that their account is mistaken.
In any of these examples, once an error is detected in a belief, the belief is surrendered,
as in the subway case. But that detection requires seeing the event under a description of
it as a counterinstance, which ordinarily cannot be counted on—an aspect of the self‐
protection problem. The falsifying description in the motion case is compelled by the
Page 13 of 24
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). © Oxford University Press, 2018. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in
Oxford Handbooks Online for personal use (for details see Privacy Policy and Legal Notice).
computer simulation, which eliminates all variables but the ones essential to the
Newtonian processes.
Still, even the contrivances of the computer environment do not wholly eliminate the
latitude to explain‐away the observation. In di Sessa's (1982) study, described earlier by
Gardner, the reaction of the students, when they observe their dynaturtle miss its target,
is that there is cheating going on—the software is deceiving them. It takes reassurance
and further testing to convince them otherwise. In the cases that we have presented,
normal experience does not even yield the formation of the specific beliefs. Instead, it
yields only clusters of assumptions, vaguely related beliefs, and patterns of thought. In
the baseball examples, the beliefs that immediately yield the judgments are partly
induced by the experimenter's question. Previous to that question, subjects are simply not
sufficiently focused on the comparison of batting averages to even notice the discrepancy.
4.3. Dependence
The students' discovery of their error with little latitude for evasion required external
guidance, as by teachers (or computer simulations). It cannot, except by accident or
happenstance, derive from normal experience—a crucial premise to argue for the study of
errors in the liberal arts curriculum.
The teacher's role is not just to guide students to a better understanding of the relevant
subject and to help them to develop knowledge of it, but also to bring students to
recognize their own misconceptions. An important lesson of the study is of the extensive
dependence on others, which is also a lesson in humility, a facet of fallibilism. The
correlative lesson is the importance of discerning good authorities or experts, and
respecting a division of (epistemic) labor among one's peers in social inquiries.
Each of these cases, and this is typical of the reasoning studies, involve a conflict
between our ordinary understanding and a more abstract, less contextualized (p. 96)
understanding. The real “why” of a chance (regression) account in either baseball case is
in discord with natural ways of thinking about such phenomena—discrepancies from a
trend are to be explained by a change of circumstance. But the chance account denies
that there is any real discrepancy. (“Natural” does not imply “inevitable.” We each have
explained‐away discrepant phenomena and coincidences by saying “It's just chance.”)
Page 14 of 24
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). © Oxford University Press, 2018. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in
Oxford Handbooks Online for personal use (for details see Privacy Policy and Legal Notice).
Appreciation of this minimal realism also prepares students to admit ignorance, which is
especially valuable when confronted, as Schermer was at the Esalen Institute, with those
who offer a ready account for seemingly puzzling phenomena. In addition to the humble
“I do not know,” intellectual courage is called for to maintain this discomforting
puzzlement and to resist accepting the account offered, merely because it is the only one
available.
In the reasoning studies, the correct and discomforting alternative patterns of reasoning
or analysis are not merely less preferred to the ones that students and subjects rely upon.
These alternatives are not generated. The simple thought‐experiment for constructing a
blatant counterexample is not even attempted in Gardner's algebra example. In the
baseball cases, randomness or chance does not even appear to subjects as a candidate
account of the discrepancy in averages. (Studies on individual differences in reasoning
show that when arguments that back subjects' conclusions are explicitly compared to
arguments for the preferred answer (e.g., to factor into the motion problem the resolution
of vectors), subjects shift, mainly favoring the latter; Stanovich 1999.)
The next example, an extended one from history, illustrates the importance of
appreciating the full set of options in correcting for recurrent sources of human (p. 97)
fallibility: a narrowing of the set of options either by not noticing genuine alternatives, as
has preoccupied discussion so far, or by treating them as off‐the‐table. The example is
particularly distressing for the grave importance of the events and the false restrictions
that many experts and leaders imposed, despite extensive time and resources for
informed deliberation.
In his book Humanity: A Moral History of the Twentieth Century, Glover (1999) develops
a contrast between the leadership of 1914, involved in World War I, and that involved in
the Cuban Missile Crisis of 1962. To enormously simplify a comparison that runs over five
Page 15 of 24
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). © Oxford University Press, 2018. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in
Oxford Handbooks Online for personal use (for details see Privacy Policy and Legal Notice).
chapters, the crucial contrast is in the responses to what Glover refers to as “traps,”
which position leaders so that there appears no exit to a highly risky course of action.
Even when they desperately did not want a major war, the leaders of the nations involved
in World War I and their advisers took themselves to be so bound by treaties, political
pressures, and sensitivities of national pride, as well as caught in miscommunication,
misunderstanding, and haste, that they entered the war almost as an act of resignation.
Glover writes: “The political and psychological traps combined with this inadequate mode
of thinking to cause the statesmen collectively to bring about the disaster. It was a
disaster unwanted, certainly by most of them, and probably by all of them” (p. 199).
The inadequate mode of thinking that Glover's “this” points to is social Darwinism, which
viewed conflicts among nations as inevitable, as another leader's apology for the war
affirms: “It is in accord with this great principle [‘mankind's struggle for existence’] that
the catastrophe of the world war came about inevitably and irresistibly as the result of
the motive forces in the lives of states and peoples, like a thunderstorm which must by its
nature discharge itself” (Glover 1999, p. 195).
Glover's contrast with the peaceful resolution of the Cuban Missile Crisis of 1962 is that
when devastating costs are a serious expectation from a course of action, leaders must
rethink the starting presupposition that there are no other alternatives. In the Cuban
Missile Crisis, leaders recognized a substantial probability that if the current trajectory
continued, the U.S.–Russia confrontation would escalate to nuclear war. Both Kennedy
and Khruschchev took this as a gamble that should be refused. They both, though, appear
trapped, especially by public opinion and national pride. However, “The other side of the
crisis is the way the two governments managed to slip out of the trap. They did not pull
tight the knot of war. In this they succeeded, but they nearly failed” (Glover 1999, p. 212).
Neither Kennedy nor Kruschchev acted in haste, despite intense political and military
pressure to act. They took time to reflect (on the origins of World War I, specifically), one
benefit of which was that the anger “was replaced by fear” (p. 214). There was a change
in “tone of voice” and something approximating explicit acknowledgment of their
respective traps, so that “they can then turn escape from the trap into a shared
project” (p. 215). As in 1914, the thinking in 1962 was “limited by dangerously crude
assumptions,” including influential hawks who thought that the idea of escalation to
nuclear war was “preposterous” (p. 218).
By contrast, “The doves placed much emphasis on uncertainty and human fallibility” (p.
218). They found unacceptable even a small risk of nuclear war, and (p. 98) they favored
their emotional reaction and “great imaginative awareness” (of the potential devastating
outcomes) over the calculations of rational choice theory. Ultimately, both leaders
generated and chose alternatives that, because of negative consequences, were
previously outside the set of feasible options that had defined their respective traps:
Page 16 of 24
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). © Oxford University Press, 2018. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in
Oxford Handbooks Online for personal use (for details see Privacy Policy and Legal Notice).
Kruschchev had to make what was widely seen as a humiliating climbdown. And
Kennedy, in order to defend his reputation for toughness, had to deceive the public
about the existence of the deal.…
Through climbdown and deception they avoided war. After the crisis, Krushchev
wrote to Kennedy that “we had to step over our pride, both you and [me], in order
to reach this agreement.” This stepping over pride was what the statesmen of
1914 had not been able to do to their sense of national honour. The humiliating
climbdown, the necessary deception, and stepping over one's pride: they should
each have their honoured place in a modern account of the political virtues.
(Glover 1999, pp. 222–23)
In the closing section, Glover draws an important lesson from taking fallibility seriously: “At a
deeper level people are trapped by the limitations of their view of the world” (p. 230).13 And
further trapped by those limitations obscuring recognition of them as limitations.
The main reason that recognition and self‐protection are not serious barriers to self‐
correction for science is that science's formulation and testing of hypotheses is ongoing,
requiring no special motivation. In that ongoing testing, the role of previously accepted
hypotheses in developing new, testable predictions routinely places those (former)
hypotheses at risk without any reason to doubt them. Scientists are varied in ideas and
commitments, and they act free of (p. 99) others' control. They are motivated to perform
original and creative work, including the generation of new hypotheses and the criticism
of previous claims, not only by internal drive and curiosity but also by an extensive
reward system (Kitcher 1993).
Page 17 of 24
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). © Oxford University Press, 2018. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in
Oxford Handbooks Online for personal use (for details see Privacy Policy and Legal Notice).
Given that scientific hypotheses and theories yield observable predictions only as heavily
mediated by further assumptions, why can we, as individuals, not also expect that our
errors will be “sooner or later detected?” Why do we need to rely much more on design
and intervention? What are the limits or obstacles (without contrivance) to our
approximating science as a self‐corrective system—besides the obvious differences, such
as our pursuing a wide range of inquiries as individuals, compared to a huge, trained
professional community devoted to specialized inquiries?
One broad and positive implication for education is that students need to be taught to
generate (imagine) alternatives and other potential grounds of self‐criticism, and to
impose mechanisms of critical control that are most effective to the extent that they are
independent of the beliefs or judgments to which their findings are potentially contrary.
Teachers already impose independent mechanisms toward self‐correction. For example,
mathematics teachers have students check their answers; English teachers ask or require
students to hand in revisions of their papers. Computer programming is practically
impossible without numerous, easy, corrective simulations—successive approximation.
The additional need in teaching is for explicit mention of the connection to fallibility and a
suitable common vocabulary for facilitating a focus on error. Few mathematics and
English teachers encourage, I expect, reflection on the basis for their recommendations:
even when students are sure their work is impeccable, they should check and revise. The
meta‐lesson of fallibility is not so conceptualized and presented. Fallibility appreciation is
a meta‐cognitive lesson in monitoring your own understanding and progress; planning
your steps in learning; and generating, and heightening access to, self‐knowledge of
weaknesses, biases, and past errors—and when (and from whom) to seek more
information and assistance.
Page 18 of 24
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). © Oxford University Press, 2018. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in
Oxford Handbooks Online for personal use (for details see Privacy Policy and Legal Notice).
In routine cases, when these intellectual virtues are applied automatically, rather than
calling upon deliberation they are expected to provide guidance beyond the school
setting. Well‐acquired habits are reliable and economical. Cognitive science teaches that
learning is facilitated by ease of usage and the lessening of burdens on thought.
Given our limited time and resources, a crucial role for good judgment (“practical
wisdom”) in guiding the application of these intellectual virtues is a high degree of
selectivity. Greater efforts are invested as the potential cost of error or the importance of
being correct is greater.14 We are more open‐minded in those domains and to those
questions that we believe are less well established—those domains in which we are likely
to be more robustly fallible. It is worse than pointless to be open‐minded about whether
the sun will rise tomorrow, even granted that it is (inductively) possible. Intellectual
virtues, especially open‐mindedness, can easily be overwhelmed, if you are
indiscriminately open‐minded, since time and attention are very limited resources (Adler
2004).
However, these intellectual virtues are not as promising applied to problems of benefit,
reward, or self‐interest. These virtues involve rendering oneself vulnerable to criticism or
counterevidence and to tolerating new, discomforting assumptions, as arise by generating
alternatives. They involve positioning oneself to invest time, resources, and effort into
activities and reflection whose most justifying outcome will be to unsettle beliefs that, by
implication, deny that the investment will pay off.
Intellectual honesty involves admitting the influences of others' work, and so downplaying
one's originality. It imposes burdens of explicitness about assumptions and precision in
formulating claims that foreclose “plausible deniability.” In writing an opinionated essay,
pleasure is with style, which dictates varying terms and other devices of enjoyable
reading. But clarity of argument is enhanced by stating only what cogency requires and
so leads to homogenizing terms. When confidence in a particular belief is diminished, a
self‐critical person attends to his belief from a detached perspective on himself, which is
neither natural nor in accord with inclination.
The conceptual connections among belief, fallibility, correction, and the intellectual
virtues are links to reward and benefit. Fallibility, as proneness to error, is in conflict with
belief's aim of truth. Even more motivating practically than this internal or conceptual
demand of belief for truth is the external value of truth: your actions are expected to
largely succeed only if the beliefs guiding them are true. If I think the name of the person
approaching me at a party is “Bill” and my wife reminds me that it is “Tony,” I gladly
surrender my belief, acquire hers, and look forward to greeting him without
embarrassment with, “Hi, Tony.”
Page 19 of 24
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). © Oxford University Press, 2018. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in
Oxford Handbooks Online for personal use (for details see Privacy Policy and Legal Notice).
well. By drawing the conceptual connections of belief, truth, and fallibility with these
virtues, their acquisition is shown to be ultimately consonant with inclination or pleasure.
Fallibility thus provides an internal motive for students to acquire those virtues, as well as
stimulating their manifest external motive of wanting their actions to succeed.
The self‐understanding as to why students should conform to the virtues because of their
conceptual ties to fallibility and the claims of belief may not be as effective, practically, as
are rewards or threats (tests, grades). Nevertheless, understanding the conceptual ties
contributes to the stability and resilience of the virtues across diverse settings,
particularly when the dictates of these virtues conflict with self‐interest. It is harder to
negotiate away the demands of intellectual honesty when one grasps the arguments for it
—arguments that reveal the depth of one's commitment.
The initial examples implicate the students' own self‐understandings. They are, then,
gripping evidence of a significant fallibility in reasoning, rather than a lapse in memory or
patience. The description is essentially first‐person: “I did not really know, though I
thought I did,” representing an intrinsic reason to learn and to correct one's error. The
study of error engages students' interests easily because it addresses their own
preconceptions and misconceptions, challenging them at the borders of their
understanding.15
Students will enjoy, or be stimulated by, the surprises and the challenges to their self‐
understanding. This interest, of course, threatens to be a detour away from disciplinary
studies and toward self‐absorption. But the risk is not severe, for reasons already
mentioned. The misconceptions and faulty reasoning are common, and common as
impediments to learning. The tendencies that research on reasoning and the study of
error uncover, however, are human tendencies, not isolated or idiosyncratic ones.
References
Adler, J. E. (2002a). Belief's Own Ethics. Cambridge, MA: MIT Press.
Bransford, J. D., A. L. Brown, and Cocking, R. R. (2000). How People Learn: Brain, Mind,
Experience, and School, expanded ed. Committee on Developments in the Science of
Learning. Washington, DC: National Academy Press.
Blackwell.
Page 20 of 24
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). © Oxford University Press, 2018. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in
Oxford Handbooks Online for personal use (for details see Privacy Policy and Legal Notice).
Dewey, J. (1986). Logic: The Theory of Inquiry. Vol. 12: 1938. The Later Works, 1925–
1953. Carbondale, IL: Southern Illinois University Press.
Elga, A. (2005). “On Overrating Oneself.… and Knowing it.” Philosophical Studies 123:
115–24.
Evnine, S. (2001). “Learning from One's Mistakes: Epistemic Modesty and the Nature of
Belief.” Pacific Philosophical Quarterly 82: 157–77.
Feris, T. (2002). “The Whole Shebang.” American Educator 26 (Fall): 20–23, 43–45.
Glover, J. (1999). Humanity: A Moral History of the Twentieth Century. New Haven, CT:
Yale University Press.
Kahneman, D., and A. Tversky. (1982). “On the Psychology of Prediction.” In Judgment
under Uncertainty: Heuristics and Biases, eds. D. Kahneman, D. Slovic, and A. Tversky
(pp. 48–68). Cambridge: Cambridge University Press.
Kuhn, D. (1991). The Skills of Argument. New York: Cambridge University Press.
Leite, A. (in press). “Fallibilism.” In Blackwell's Companion to Epistemology, 2nd ed., eds.
E. Sosa and M. Steup. Oxford: Blackwell.
Nisbett, R., and L. Ross. (1980). Human Inference: Strategies and Shortcomings of Social
Judgment. Englewood Cliff, NJ: Prentice‐Hall.
Page 21 of 24
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). © Oxford University Press, 2018. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in
Oxford Handbooks Online for personal use (for details see Privacy Policy and Legal Notice).
Ohlsson, S. (1996). “Learning from Performance Errors.” Psychological Review 103: 241–
62.
Schermer, Michael. (2005). “Mr. Skeptic Goes to Esalen.” Scientific American, December,
p. 38.
Swartz, Ronald N., Henry J. Perkinson, and Stephanie G. Edgerton, eds. (1980).
Knowledge and Fallibilism: Essays on Improving Education. New York: New York
University Press.
Notes:
(2.) Some identify fallibilists as those who accept the assumption, crucial to the “preface
paradox,” that we should each believe that “Some of my beliefs are false.” For arguments
against this assumption, see my 2002a, chap. 7.
(3.) Popper's (e.g., 1959) philosophy of science is a bad model for a substantive role for
fallibility, since he denied that there could be (inductive) learning from experience.
However, in his concern with seeking out and examining errors, as well as problem
solving, his is a useful view, which inspires Swartz, Perkinson, and Edgerton 1980.
(4.) The proposal to come is related to the second‐order view that is advocated by Evnine
2001 and Adler 2002a, chap. 11, and 2004.
(5.) Throughout I am concerned only with Type I error when one accepts a false
hypothesis. One commits a Type II error when one rejects a true hypothesis (or, we can
add, does not accept a true hypothesis, presented to him). The omission is not merely for
brevity: 1. As with self‐interest, our natural disposition is toward acceptance (gullibility,
hasty generalization …), so the main worries are over false acceptance or error. 2.
James's two objectives—seek truth; avoid error—are not equal. The latter is an absolute
demand—once error is even suspected, the belief is erased. However, on the “seek truth”
aim, a person may agree that the evidence points toward a conclusion, but hesitate to
accept it because he is very cautious. 3. Since a type II error, where one remains ignorant
Page 22 of 24
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). © Oxford University Press, 2018. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in
Oxford Handbooks Online for personal use (for details see Privacy Policy and Legal Notice).
or forgets one's rejection, will involve not believing that one knows, Type I errors are
bound to be more costly, in general.
(6.) The analysis to follow depends on a distinction between belief and confidence,
elaborated in Adler 2002a, chap. 10.
(7.) For references and insightful reflections, see Elga 2005. Below, I deny, however,
Elga's thesis that the ratings are always evidence of self‐overrating.
(8.) Objection: If we study errors, we study what is wrong. But right ideas are always
more valuable, all other things equal. So, given limited time, the course of study should
deal only with what is correct—that is, what we believe now.Reply: But labeling errors as
such does amout to teaching truths. Moreover, the objection assumes that students will
learn better only if they study just what we regard as true or known, given limited time.
But why accept this, particularly when taking account of students' quality of thought
outside and subsequent to school? We do not, after all, think that the structure of
knowledge in a discipline is the same as the structure of how it should be taught.
(9.) Although transfer is limited in that way, meta‐cognitive practices and strategies do
show some transfer. For discussion and citations, see Bransford, Brown, and Cocking
2000.
(10.) Here, as elsewhere, I am not treating specialized training—for example, the training
of future scientists. On science education, see Siegel 1988, chap. 6.
(11.) Another addition is that of much more use of argumentation in disciplinary studies;
see Kuhn 2005, part III.
(13.) Cross‐cultural contrasts are an excellent way to teach these limits and the fallibility
of taking one's perspective to be universal.
(14.) Rescher ( 2007) distinguishes between the gravity or seriousness of error and its
extent, or “how wide off the mark a mistake happens to be” (p. 4).
(15.) The relevance of such engagement to learning is cited in Bransford, Brown, and
Cocking ( 2000, pp. 14–15). For further examples, see their chap. 7.
Jonathan E. Adler
Page 23 of 24
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). © Oxford University Press, 2018. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in
Oxford Handbooks Online for personal use (for details see Privacy Policy and Legal Notice).
Inquiry: Classic and Contemporary Readings (2007), and co‐editor (with Lance Rips)
of Reasoning: Studies of Human Inference and Its Foundation (2008).
Page 24 of 24
PRINTED FROM OXFORD HANDBOOKS ONLINE (www.oxfordhandbooks.com). © Oxford University Press, 2018. All Rights
Reserved. Under the terms of the licence agreement, an individual user may print out a PDF of a single chapter of a title in
Oxford Handbooks Online for personal use (for details see Privacy Policy and Legal Notice).