Professional Documents
Culture Documents
net/publication/324502214
CITATION READS
1 192
2 authors, including:
SEE PROFILE
Some of the authors of this publication are also working on these related projects:
All content following this page was uploaded by Ali Muhammad Rushdi on 13 April 2018.
Authors’ contributions
This work was carried out in collaboration between the both authors. Author RAR envisioned and
designed the study, performed the analyses, stressed the medical context for the fallacies considered,
and solved the detailed examples. Author AMR managed the literature search and wrote the
preliminary manuscript. Both authors read and approved the final manuscript.
Article Information
DOI: 10.9734/JAMMR/2018/40784
Editor(s):
(1) Angelo Giardino, Professor, Texas Children’s Hospital, Houston, Texas and Pediatrics, Baylor College of Medicine,
Houston, TX, USA.
Reviewers:
(1) Maria A. Ivanchuk, Bucovinian State Medical University, Ukraine.
(2) Grienggrai Rajchakit, Maejo University, Thailand.
(3) Jeffrey Clymer, USA.
Complete Peer review History: http://www.sciencedomain.org/review-history/24107
th
Received 16 January 2018
Accepted 4th April 2018
Review Article th
Published 11 April 2018
ABSTRACT
This paper presents and explores the most frequent and most significant fallacies about probability
in the medical fields. We allude to simple mathematical representations and derivations as well as to
demonstrative calculations to expose these fallacies, and to suggest possible remedies for them.
We pay a special attention to the evaluation of the posterior probability of disease given a positive
test. Besides exposing fallacies that jeopardize such an evaluation, we offer an approximate method
to achieve it under justified typical assumptions, and we present an exact method for it via the
normalized two-by-two contingency matrix. Our tutorial exposition herein might hopefully be helpful
for our intended audience in the medical community to avoid the detrimental effects of probabilistic
fallacies. As an offshoot, the pedagogical nature of the paper might allow probability educators to
utilize it in helping their students to learn by unraveling their private misconceptions about
probability.
_____________________________________________________________________________________________________
2
Rushdi and Rushdi; JAMMR, 26(1): 1-21, 2018; Article no.JAMMR.40784
with the solution of general probability problems “conditional probability” is included. Any equation
in this paper. We restrict ourselves to simple we present herein that is not generally true will
well-posed probability problems of medical be identified as such (in an admittedly harsh way)
context that have already been known for some by labeling it as “WRONG.”
time and have well-established correct solutions
in the literature, but could be mishandled by 2. MULTIPLICATION AND ADDITION
medical students and practitioners due to some FALLACIES
inherent misconceptions or as a result of lack of
adequate training. The Multiplication and Addition Fallacies for two
general events A and B amount to
Due to space limitations, we restrict our mathematically expressing the probabilities of the
discussion to fallacies encountered frequently in intersection and union of these two events (see
the medical circles. We have to leave out many Appendix A) as the product and sum of their
fallacies that are well-known in other walks of life probabilities, namely
and other scientific disciplines. These include
(albeit not restricted to) the Equi-probability ( ∩ ) = ( ) ( ), (WRONG) (1)
Fallacy [58], the Gambler's Fallacy [33], the
Fallacy of the Impartial Expert [24], the Efficient ( ∪ ) = ( ) + ( ). (WRONG) (2)
Breach Fallacy [32], the Individualization Fallacy
[45], the Association Fallacy [46], the Defense- The wide-spread prevalence of these fallacies is
Attorney Fallacy [31,51], the Uniqueness Fallacy perhaps mainly due to the way probability is
[45], and the Disjunction Fallacy [59,60]. introduced in pre-college education. In fact, these
fallacies are appealing because they simply
The organization of the rest of this paper is as replace set operations in the event domain by
follows. Section 2 explores the Multiplication and their arithmetic counterparts in the probability
Addition Fallacies, and as an offshoot, comments domain. Another possible reason for the
on basic probability formulas. Section 3 is the popularity of these fallacies is the inadvertent
main contribution of this paper. Besides exposing neglect or disregard of the conditions under
the Inverse Fallacy, it presents a method to which they become valid. Equation (1) is correct
estimate P(Disease|positive test) approximately provided the events A and B are (statistically)
under typical assumptions, and also offers independent, while Equation (2) is exactly correct
another method to evaluate this probability when the events A and B are mutually exclusive
exactly via the normalized two-by-two ( ∩ = ∅). In particular, the addition formula (2)
contingency table. Section 3 also presents three is correct if A and B are primitive outcomes or
illustrative examples, which are computational in singletons, i.e., events comprising individual
nature and medical in context. Moreover, Section points of the pertinent sample space. Moreover,
3 reflects on certain comments available in the Equation (2) is approximately correct when the
literature on the validity of the pertaining model events and are independent and the
itself. Section 4 reviews the concept of an event probabilities ( ) and ( ) are particularly very
being favorable to another and discusses the small. The correct versions for (1) and (2) are the
Fallacy of the Favorable Event. Section 5 elementary formulas [1,61].
investigates the Conditional-Marginal Fallacy and
discusses the relations among conditional and ( ∩ ) = ( | ) ( ) = ( ) ( | ) (3)
marginal probabilities. Section 6 investigates
Simpson's Paradox through it is not really a ( ∪ ) = ( )+ ( )− ( ∩ ) (4)
fallacy as such, but, being a paradox, it shares
the problematic nature of a fallacy. Section 7 Equation (3) asserts that the fundamental
demonstrates the Conjunction Fallacy via a concept of conditional probability is unavoidable,
medical example while Section 8 discusses the indeed, since statistical independence is not
Appeal-to-Probability Fallacy. Section 9 always guaranteed. Equation (3) provides
illustrates the drastic effects of the Base-Rate definitions for the conditional probabilities ( | )
Fallacy by considering its effect on one of the and ( | ) provided ( ) ≠ 0, and ( ) ≠ 0,
examples of Section 3. Section 10 covers the respectively. The probability ( ∩ ) can be
Representative-Sampling Fallacy. Section 11 neglected (considered almost 0) in (4) so as to
adds some useful observations, while Section 12 approximate (4) by (2) when the events and
concludes the paper. To make the paper self- are independent and ( ) and ( ) are very
contained, an appendix (Appendix A) on small. However, this same probability cannot be
3
Rushdi and Rushdi; JAMMR, 26(1): 1-21, 2018; Article no.JAMMR.40784
neglected in (3), even when ( ) and ( ) are P(A) (true probability of disease presence or
extremely small since such an action leads to a such a probability according to a gold standard)
catastrophic error (100% relative error), or to a to Perceived or Apparent Prevalence P(B)
fallacy of its own called the Rare-Event Fallacy. (probability of disease presence according to the
test). The Perceived Prevalence P(B) is given by
3. THE INVERSE FALLACY the Total Probability Formula [1, 61] as
| ℎ ( )≈ ( )+ | (11)
( | )=
(5) The approximation (11) does not violate the
probability axiom { ( ) ⩽ 1}, since both P(A) and
is the same as the test Sensitivity . | are known to be small compared to 1.
However, the probability | (called the False
ℎ | Positive Rate, ) (albeit small) could be
( | )=
significantly larger than P(A). Hence, the
(6)
perceived P(B) is greater (or even much greater
than) the true prevalence P(A). This makes the
Since the former probability is typically
ratio in (7) definitely smaller (usually much
substantially smaller than the latter one, this
smaller) than 1. In other words, ( | ) > ( | )
fallacy has grave consequences, as it means
(typically ( | ) ≫ ( | )). This means that in
misinterpreting false positive test results (which
are already bad and alarming besides being many cases, a test PPV is (significantly) smaller
than its Sensitivity, and should not be mistaken
misleading) to make them even more disturbing
as being equal to it. Under typical mild
and threatening.
assumptions, we can assess the PPV ( | )
approximately through a combination of (7) and
The two conditional probabilities ( | ) and
( | ) are not related by the equality relation (11) as
fallaciously assumed, but are related by Bayes' ( )
formula expressed by Equation (3). Therefore, ( | )≈ ( | ) (12)
( ) |
the ratio of these two conditional probabilities is
equal to the ratio of the unconditional or marginal ( )
≈ (12a)
probabilities, namely ( ) |
With our earlier designation of A and B as “Assume you conduct breast cancer screening
{Disease is present} and {Test says disease is using mammography in a certain region. You
present}, the ratio in (7) is not 1 as the fallacy know the following information about the women
demands, but it is the ratio of True Prevalence in this region:
4
Rushdi and Rushdi; JAMMR, 26(1): 1-21, 2018; Article no.JAMMR.40784
(a) The probability that a woman has breast answer of 0.10 in (c) while 47% and 13% of them
cancer is 1% (True Prevalence) grossly overestimated the answer as 0.90 and
(b) If a woman has breast cancer, the probability 0.81, respectively, perhaps falling victim to (or at
that she truly tests positive is 90% (Sensitivity) least being influenced by) the Inversion Fallacy.
(c) If a woman does not have breast cancer, the Only 19% of the respondents underestimated the
probability that she nevertheless tests positive is answer.
9% (False-Positive Rate)
Another example reported by Eddy [29] runs as
A woman tests positive. She wants to know from follows:
you whether that means that she has breast
cancer for sure, or what the chances are. What is “The prior probability, P(ca), 'the physician's
the best answer?” subjective probability', that the breast mass is
malignant is assumed to be 1%. To decide
In this problem, we identify the given information whether to perform a biopsy or not, the physician
in our notation as: orders a mammogram and receives a report that
in the radiologist's opinion the lesion is malignant.
(a) ( ) = 0.01 (True Prevalence) (13a) This is new information and the actions taken will
depend on the physician's new estimate of the
(b) ( | ) = 0.90 (Sensitivity or TPR) (13b) probability that the patient has cancer. This
estimate also depends on what the physician will
(c) | = 0.09 (FPR) (13c) find about the accuracy of mammography. This
accuracy is expressed by two figures: sensitivity,
and recognize the required unknown in this or true-positive rate P(+ | ca), and specificity, or
problem as the PPV or ( | ). We observe that true-negative rate P( - | benign). They are
the assumptions we made above are all valid, respectively 79.2% and 90.4%.”
namely
We choose to give a detailed analysis of this
( ) = 0.01 ≪ 1, example via the normalized contingency table of
1. = 0.99 ≃ 1 (14a)
Fig. 1, which summarizes our earlier findings in
[21-23]. Substituting for the symbolic notation in
2. ( | ) = 0.90 ≃ 1 (14b) Fig. 1, we produce a complete solution for the
aforementioned example in Fig. 2. The results
3. | = 0.09 is small but is (much) obtained indicate that the particularly-required
larger than ( ), (14c) result of P(cancer | positive test), or in our
notation = P(j=+1|i=+1) = ( | ) is
Our approximate answer in (11) is 0.076923 or approximately 7.7%. According to
Eddy [29], most physicians interviewed estimated
( ) ≃ 0.01 + 0.09 = 0.10, (15) this posterior probability to be about 75%, i.e.,
almost ten times larger. He attributed this to the
while the exact answer computed by Rushdi, et Inverse Fallacy, which led them to believe that
al. [23] (for the same problem) is 0.0981. the required probability is approximately equal to
Correspondingly, our approximate answer in its transpose of P(Test positive|cancer) = P(i=+1|
(12a) is j=+1) which is given as 0.792.
0.01 Boumans [62-64] criticizes the above findings, by
( | )≈ = 0.1 (16)
0.01+0.09 questioning the validity of the model on which
they are based. He constructs a different double-
In an experiment conducted by Gigerenzer, et al. threshold model that ultimately justifies why
[2], 160 gynecologists were requested to choose physicians tend to give high estimates for the
the best value for ( | ) among four given posterior probability P(cancer|positive test). It
values seems that this tendency among physicians is
excused on the grounds that two wrongs (a
(a) 0.81, (b) 0.90, (c) 0.10, (d) 0.01 wrong model and a wrong method of calculations)
can possibly make one right. Boumans [62-64]
where incorrect answers were spaced about one asserts that decision making in real-life situations
order of magnitude away from the best answer. is different from decision making in a laboratory
Only 21% of the gynecologists found the best controlled experiment. “A model of a decision
5
Rushdi and Rushdi; JAMMR, 26(1): 1-21,, 2018; Article no.JAMMR.40784
no.
( = +1) ( = −1) 1
= =1 −
=
Fig. 2. Complete solution of the second example of Sec. 2 with the aid of the normalized
contingency table introduced in Fig. 1. Initially known entries are highlighted in red
Fig. 3. Complete solution of the third example in Sec. 2 via a normalized contingency table.
Given data is shown in red. The assumption of FNR=0 (TPR=1)
(TPR=1) was added (to the original
formulation by Casscells et al.. [7]) by subsequent authors such as Westbury [4] and Sloman et
al. [65]
6
Rushdi and Rushdi; JAMMR, 26(1): 1-21, 2018; Article no.JAMMR.40784
differences among statisticians and physicians is stating that “Boys more at risk on bicycles” is
highly urged. Further exploration of this subject is cited [42] to be based on the report that “among
definitely warranted, albeit it seems somewhat children involved in bicycle accidents the majority
beyond the scope of this paper. were boys.” The writer(s) of the headline, in fact,
observed that
We close this section with a third example due to
Casscells, et al. [7]. This example reads as P(Boys|bicycle accident) is “large” (18)
follows:
and went on to conclude that
“If a test to detect a disease whose prevalence is
1/1000 has a false positive rate of 5 per cent, ( | )>
what is the chance that the person found to have ( ) (19)
a positive result actually has the disease,
assuming that you know nothing about the
The statement (19) is only possibly unwarranted,
person's symptoms or signs?”
i.e., it is not necessarily false, but the fact is that
it cannot be logically inferred from (18).
Fig. 3 indicates immediately that the data given is
not complete and must be supplemented by
The concept of “favorableness” discussed in this
something else. The solution given in Fig. 3 is
section is also involved in Simpson's Paradox [25,
based on neglecting the False Negative Rate
42]. In general, Simpson's paradox describes a
(FNR), and is the solution intended by those who
phenomenon in which a trend appears in
posed the problem. However, the missing
individual groups of data, but disappears or
information in the example defeats the purpose
reverses when these groups are combined, or
of the experiment done by Casscells, et al. [7].
amalgamated [25,26,28,35,42,48]. We will
The physicians who resorted during that
discuss this reversal-upon-amalgamation
experiment to the Inverse Fallacy are only
paradox in Sec. 6.
partially to be blamed, since they were forced to
seek a means to fill in an inadvertent and 5. THE CONDITIONAL-MARGINAL
unnecessary gap. We noted that subsequent
FALLACY
authors who referred to the problem of Casscells,
et al. [7] augmented the original formulation In the Conditional-Marginal Fallacy, the
above by adding {FNR = 0} [4], or equivalently, conditional probability ( | ) is mistaken for the
{TPR = 1} [65]. marginal unconditional probability ( ), or,
equivalently (according to Eq. (7)), the inverse
4. FALLACY OF THE FAVORABLE conditional probability ( | ) is equivocated with
EVENT ( ). We note that this is generally fallacious
unless the two events A and B are (statistically)
An event A is called favorable to another B [25, independent. In fact, the very definition of
42] when the occurrence of A implies an increase independence of event A from event B is the
in the chances B occurs, i.e., requirement that ( | ) be equal to ( ).
Similarly, independence of event B from event A
( | )> ( ) (17) is the requirement that ( | ) be equal to ( ).
These two definitions are equivalent, and hence,
The Fallacy of the Favorable Event is to infer we do not need to refer to independence of one
from the fact that the conditional probability event from another, but to independence
( | ) is “large” that the conditional (conditioned) between the two events. Any of the following
event A is favorable to the conditioning event B. equivalent twelve relations can be used to
This fallacy is so problematic that it is not even denote (statistical) independence between
amenable to precise mathematical description, events A and B, and can be used to
since one does not really know how “large” is mathematically deduce any of the other relations
“large.”
( )= ( | )= | (20a)
Krämer and Gigerenzer [42] cite many examples
in which this fallacy occurs in various contexts, P(B) = P(B|A) = P B|A (20b)
and suggest that it is “possibly the most frequent
logical error that is found in the interpretation of P A = P A|B = P A|B (20c)
statistical information.” A newspaper headline
7
Rushdi and Rushdi; JAMMR, 26(1): 1-21, 2018; Article no.JAMMR.40784
8
Rushdi and Rushdi; JAMMR, 26(1): 1-21, 2018; Article no.JAMMR.40784
9
Rushdi and Rushdi; JAMMR, 26(1): 1-21, 2018; Article no.JAMMR.40784
(a)
(b)
(c)
Fig. 4. Representation of events A and B via area-proportional Venn diagrams (The areas
allotted to an event is proportional to its probability)
10
Rushdi and Rushdi; JAMMR, 26(1): 1-21, 2018; Article no.JAMMR.40784
Table 1. Possible cases of favorableness between two events and their complements
Table 2. Listing of famous fallacies pertaining to probability and comparing fallacious formulas
to correct one
( ∩ ) = ( ) ( ),
A and B are statistically independent
Addition Fallacy ( ∪ ) = ( ) + ( ), ( ∪ )
A and B are general = ( )+ ( )− ( ∩ )
( ∪ ) = ( ) + ( ),
A and B are mutually exclusive
Inverse Fallacy ( | )= ( | ) ( )
( | )= ( | )
( )
Conditional-Marginal Fallacy ( | )= ( ) ( | )= ( )
A and B are general A and B are statistically independent
Conjunction Fallacy ( ∩ )> ( ) ( ∩ )⩽ ( ), ( )
or
( ∩ )> ( )
The Appeal-to-Probability { ( ) > 0} ⇒ { ( ) > 0} ⇒ { ( ) ∈ (0,1]}
Fallacy { ( ) = 1} trivially
The Base-Rate Neglect ( )= = 0.5 P(A) is not necessarily equal to
Typically For medical applications when
P(A) denotes disease prevalence
( ) ≪ 1, ≃1
11
Rushdi and Rushdi; JAMMR, 26(1): 1-21, 2018; Article no.JAMMR.40784
Fig. 5. Utilization of the Karnaugh map as a convenient and natural sample space to
demonstrate Simpson's Paradox. The full sample space in (a) is compacted via additive
elimination [21] to the reduced one in (b)
Fig. 6. Complete solution of the second example of Sec. 2 under the Base-Rate Fallacy
ignoring the prior knowledge of true prevalence. An extremely exaggerated value of 89.2% for
the PPV is obtained
The common cold was chosen for the vignette the potential to improve students’ decision
above, since entering medical students have little making.
or no clinical experience but are assumed at one
point or another to have suffered themselves In passing, we note that while most people tend
from the common cold and would have some to overestimate a conjunctive probability [68]. a
knowledge of typical and atypical symptoms. majority of people are also more likely to
Runny nose is widely known to be a common underestimate a disjunctive probability, which is
symptom; diarrhea is not [68]. A violation of the a phenomenon referred to as the Disjunction
conjunction rule (i.e., Conjunction Fallacy) was Fallacy [59, 60].
recorded if diarrhea was assigned a lower
probability than the combination of runny nose 8. THE APPEAL-TO-PROBABILITY
and diarrhea, regardless of the absolute FALLACY
assigned probability value or the values recorded
The Appeal-to-Probability Fallacy (sometimes
for the other options [68]. In the exercise, the
called the Appeal-to-Possibility Fallacy) equates
mean estimate of the probability of diarrhea was
a probable event to a certain one, i.e., it asserts
17.2%. The mean estimate of the probability of
that if A is not the impossible event, then it is the
the combination of runny nose and diarrhea was
certain event
31.6%. Overall, 47.8% of the students violated
the conjunction rule by assigning a higher { ≠ ∅} ⇒ { = } (WRONG) (27a)
probability to runny nose and diarrhea than to or
diarrhea alone [68]. The moral of the study in [68] { ( ) > 0} ⇒ { ( ) = 1} (WRONG) (27b)
is that teaching medical students about the
Conjunction Fallacy and other biases in This fallacy is perhaps committed by patients
assessment of probability has, in theory at least, rather than physicians. It is particularly
12
Rushdi and Rushdi; JAMMR, 26(1): 1-21, 2018; Article no.JAMMR.40784
misleading when its fallacious reasoning is drawn from a population is highly representative
preceded by some sort of wild guessing. A of the population, i.e., similar to the population in
meticulous, doubting and suspicious person all its “essential” characteristics [75]. This leads
believes for sure that he/she definitely has a to the expectation that any two samples drawn
certain disease when he/she senses or imagines from a particular population to be more similar to
having some of its symptoms. Having the one another and to the population than sampling
disease might (probably) be the case but cannot theory predicts, at least for small samples. In
be taken for granted, and should not be assumed fact, the law of large numbers guarantees that
as a matter of fact. That is basically why a patient very large samples will indeed be highly
should seek medical help, consultation, and representative of the population from which they
diagnosis, and why a physician should be well- are drawn. The aforementioned intuitions about
trained to meet the expectations and needs of random sampling appear to follow an alleged law
the patients. The essence of the diagnosis of small numbers [75, 76], which asserts that the
process is to employ scientific methodology to law of large numbers applies also to small
attribute correctly-observed symptoms to their numbers (through a presumed self-corrective
genuine causes. tendency).
13
Rushdi and Rushdi; JAMMR, 26(1): 1-21, 2018; Article no.JAMMR.40784
fallacies herein to medical subject matter, and test characteristics only and do not stress
restricted our framework to elementary (even enough the need to know the status of the
simplistic) mathematics, and tailored our patient. Therefore, they are being replaced [101-
exposition to address a medical audience. Out of 105] by the following more explicit forms, in
the extensive multitude of existing fallacies, we which both test properties and patient status are
strived to cover a sample of the most frequently- specified.
encountered (and hopefully, the most
representative). We hope our work might have SnNOUT : If Sensitivity is high, a Negative test
some modest contribution towards the ultimate will rule the disorder OUT.
desirable goal of minimization, elimination, (For a highly-sensitive test, a positive test result
removal, suppression, and eradication of is not very helpful, but a negative result is useful
fallacious reasoning. Without sincere correcting for asserting disorder absence).
and remedial efforts, perpetuated and
unchallenged fallacies may proliferate so as to SpPIN : If Specificity is high, a Positive test will
comprise a dominant portion of applicable rule the disorder IN.
knowledge. (For a highly-specific test, a negative test result
is not very helpful, but a positive result is useful
Our strategy to confront fallacies herein is simply for asserting disorder presence).
to unravel them from the point of view of
elementary mathematics. There is a long history The assertion that: “If a test has high Sensitivity,
of research challenging such fallacies [91-96]. In a Negative result helps rule out the disease”
particular, we note that Arkes [94] counts might be mathematically understood as follows. If
influence of preconceived notions among five a person actually does have the disease
impediments to accurate clinical judgment, and { ( = +1) = 1} , we would expect a highly-
discusses possible ways to minimize their impact. sensitive test { ( = +1| = +1) ≃ 1} to be
positive with high probability { ( = +1) ≃ 1} .
In passing, we discuss some other problematic Therefore, when a highly-sensitive test is
notions and abbreviated rules of diagnostic negative, we can confidently assume disease
testing, which are claimed even to be absence (rule out the disease). Likewise, we
counterintuitive and misleading or to suffer from interpret the assertion: “If a test has a high
definitional arbitrariness. It is desirable that a Specificity, a Positive result helps rule in the
diagnostic test possess high values for both disease” mathematically as follows. If a person
Sensitivity and Specificity. Sensitivity is the actually does not have the disease { ( = −1) =
probability of a positive test, given the presence 1} , we would with high probability expect a
of the disease { ( = +1| = +1)} , while highly-specific test { ( = −1| = −1) ≃ 1} to be
specificity is the probability of a negative test, negative { ( = −1) ≃ 1} . Therefore, when a
given the absence of the disease { ( = −1| = highly-specific test is positive, we can confidently
−1}. The natural inclination among many people assume disease presence (rule in the disease).
is to think that a highly-sensitive test is effective
at identifying persons who have the disease, and 12. SUMMARY AND CONCLUSIONS
that a highly-specific test is effective at identifying
those without the disease. By contrast, a highly- We have studied probabilistic fallacies in
sensitive test is effective at ruling out the disease medicine using simple mathematical
(when it is really absent), while a highly-specific representations and derivations. We summarize
test is effective at ruling in the disease (when it is our results in Table 2, which shows the wrong
really present). The following acronyms are used proposition of each fallacy as well as an
as mnemonics to help remember the appropriate correction for it. The study made
aforementioned fact [97-100] herein should hopefully be of significant help to
medical students and medical practitioners alike.
SnOUT : If Sensitivity is high, a negative test will It might ensure that they acquire the necessary
rule the disorder OUT. knowledge of elementary probability, but it does
not demand that they gain too much knowledge
SpIN : If Specificity is high, a positive test will that might distract them from their genuine (vital
rule the disorder IN. and critical) subject matter. It also attempts to
remedy the notorious and grave ramifications of
These two mnemonics might be sometimes probabilistic fallacies residing as permanent
misleading since they seem to be concerned with misconceptions in their “private” knowledge
14
Rushdi and Rushdi; JAMMR, 26(1): 1-21, 2018; Article no.JAMMR.40784
15
Rushdi and Rushdi; JAMMR, 26(1): 1-21, 2018; Article no.JAMMR.40784
unpublished master thesis, department of 32. Friedmann D. The efficient breach fallacy.
pediatrics, faculty of medicine at Kasr Al- The Journal of Legal Studies. 1989;18(1):
Ainy, Cairo University, Cairo, Egypt; 2017. 1-24.
Available:https://www.researchgate.net/pro 33. Clotfelter CT, Cook PJ. The gambler's
file/Rufaidah_Rushdi/contributions fallacy in lottery play. Management
21. Rushdi RA, Rushdi AM. Karnaugh-map Science. 1993;39(12):1521-1525.
utility in medical studies: The case of Fetal 34. Falk R, Greenbaum CW. Significance tests
Malnutrition. International Journal of die hard: The amazing persistence of a
Mathematical, Engineering and probabilistic misconception. Theory &
Management Sciences (IJMEMS). 2018; Psychology. 1995;5(1):75-98.
3(3):220-244. 35. Julious SA, Mullee MA. Confounding and
Available:www.ijmems.in/ijmems— Simpson’s paradox. British Medical Journal,
volumes.html 1994;209(6967):1480–1481.
22. Rushdi AM, Talmees FA. An exposition of 36. Evett IW. Avoiding the transposed
the eight basic measures in diagnostic conditional. Science and Justice.
testing using several pedagogical tools. 1995;35(3):127-131.
Journal of Advances in Mathematics and 37. Balding DJ. Interpreting DNA evidence:
Computer Science. 2018;26(3):1-17. Can probability theory help? In Statistical
23. Rushdi RA, Rushdi AM, Talmees FA. science in the courtroom. Springer, New
Novel pedagogical methods for York, NY. 2000;51-70.
conditional-probability computations in 38. Jowett C. Lies, damned lies, and DNA
medical disciplines, Journal of Advances in statistics: DNA match testing, Bayes'
Medicine and Medical Research. Theorem, and the criminal courts.
2018;25(10):1-15. Medicine, Science and the Law.
24. Diamond BL. The fallacy of the impartial 2001;41(3):194-205.
expert. Archives of Criminal 39. Pfannkuch M, Seber GA, Wild CJ.
Psychodynamics. 1959;3(2): 221-236. Probability with less pain. Teaching
25. Blyth CR. Simpson's paradox and mutually Statistics. 2002;24(1):24-30.
favorable events. Journal of the American 40. Villejoubert G, Mandel DR. The inverse
Statistical Association. 1973;68(343):746- fallacy: An account of deviations from
746. Bayes’ theorem and the additivity principle.
26. Gardner M. Fabric of inductive logic, and Memory & cognition. 2002;30(2):171-178.
some probability paradoxes. Scientific 41. Leung WC. The prosecutor's fallacy—A
American. 1976;234(3):119-122. pitfall in interpreting probabilities in forensic
27. Fryback DG. Bayes' theorem and evidence. Medicine, Science and the Law.
conditional non-independence of data in 2002;42(1):44-50.
medical diagnosis. Computers and 42. Krämer W, Gigerenzer G. How to confuse
Biomedical Research. 1978;11(5):423-434. with statistics or: The use and misuse of
28. Wagner CH. Simpson's paradox in real life. conditional probabilities. Statistical Science.
The American Statistician. 1982;36(1):46- 2005;20(3):223-230.
48. 43. Batanero C, Sanchez E. What is the nature
29. Eddy DM. Probabilistic reasoning in clinical of high school students' conceptions and
medicine: Problems and opportunities. In D. misconceptions about probability? In
Kahneman, P. Slovic and A. Tversky (Eds.) Exploring Probability in school Springer,
Judgment under uncertainty: Heuristics Boston, MA. 2005;241-266
and biases. New York: Cambridge 44. Kuklin B. Probability misestimates in
University Press. 1982;249–267. medical care. Arkansas Law Review. 2006;
30. Tversky A, Kahneman D. Extensional 59:527-554.
versus intuitive reasoning: The conjunction 45. Saks MJ, Koehler JJ. The individualization
fallacy in probability judgment. fallacy in forensic science evidence.
Psychological Review. 1983;90(4):293-315. Vanderbilt Law Review. 2008;61(1):199-
31. Thompson WC, Schumann EL. 219.
Interpretation of statistical evidence in 46. De Macedo C. Guilt by statistical
criminal trials: The prosecutor's fallacy and association: Revisiting the prosecutor's
the defense attorney's fallacy. Law and fallacy and the interrogator's fallacy. The
Human Behavior. 1987;11(3):167-187. Journal of Philosophy. 2008;105(6):320-
332.
16
Rushdi and Rushdi; JAMMR, 26(1): 1-21, 2018; Article no.JAMMR.40784
17
Rushdi and Rushdi; JAMMR, 26(1): 1-21, 2018; Article no.JAMMR.40784
75. Tversky A, Kahneman D. Belief in the law testing. Quality of Life Research. 2014;
of small numbers. Psychological Bulletin. 23(1):5-7.
1971;76(2):105-110. 90. Greenland S, Senn SJ, Rothman KJ,
76. Rabin M. Inference by believers in the law Carlin JB, Poole C, Goodman SN, Altman
of small numbers. The Quarterly Journal of DG. Statistical tests, P values, confidence
Economics. 2002;117(3):775-816. intervals, and power: A guide to
77. Rozeboom WW. The fallacy of the null- misinterpretations. European Journal of
hypothesis significance test. Psychological Epidemiology. 2016;31(4):337-350.
Bulletin. 1960;57(5):416-428. 91. Cheever DW. The value and the fallacy of
78. Cohen J. The earth is round (p<. 05). statistics in the observation of disease.
American Psychologist. 1994;49(12):997- The Boston Medical and Surgical Journal.
1003. 1861; 63(24):476-483.
79. Kirk RE. Practical significance: A concept 92. Ingle DJ. Fallacies and errors in the
whose time has come. Educational and wonderlands of biology, medicine, and
Psychological Measurement. 1996;56(5): Lewis Carroll. Perspectives in Biology and
746-759. Medicine. 1972;15(2):254-283.
80. Goodman SN. Toward evidence-based 93. Christensen-Szalanski JJ, Bushyhead JB.
medical statistics. 1: The P value fallacy. Physicians' use of probabilistic information
Annals of Internal Medicine. 1999;130(12): in a real clinical setting. Journal of
995-1004. Experimental Psychology: Human
81. Sterne JA, Smith GD. Sifting the Perception and Performance. 1981;7(4):
evidence—what's wrong with significance 928-935.
tests? Physical Therapy. 2001;81(8):1464- 94. Arkes HR. Impediments to accurate clinical
1469. judgment and possible ways to minimize
82. Dixon P. The p-value fallacy and how to their impact. Journal of Consulting and
avoid it. Canadian Journal of Experimental Clinical Psychology. 1981;49(3):323-330.
Psychology / Revue Canadienne de
95. Gigerenzer G. How to make cognitive
Psychologie Expérimentale. 2003;57(3):
illusions disappear: Beyond “heuristics and
189-202.
biases”. European Review of Social
83. Fidler F, Geoff C, Mark B, Neil T. Statistical
Psychology. 1991;2(1):83-115.
reform in medicine, psychology and
ecology. The Journal of Socio-Economics. 96. Croskerry P. The importance of cognitive
2004;33(5):615-630. errors in diagnosis and strategies to
84. Moyé LA. Statistical reasoning in medicine: minimize them. Academic Medicine. 2003;
The intuitive P-value primer. Springer 78(8):775-780.
Science & Business Media; 2006. 97. Davidson M. The interpretation of
85. Wagenmakers EJ. A practical solution to diagnostic tests: A primer for
the pervasive problems of p values. physiotherapists. Australian Journal of
Psychonomic Bulletin & Review. 2007; Physiotherapy. 2002;48(3):227-232.
14(5):779-804. 98. Suss R. Sensitivity and specificity: alien
86. Goodman S. A dirty dozen: Twelve p-value edition: A light-hearted look at statistics.
misconceptions. In Seminars in Canadian Family Physician. 2007;53(10):
Hematology, Elsevier. 2008;45(3):135-140. 1743-1744.
87. Li XR, Li XB. Common fallacies in applying 99. Hayden SR. Basic statistics: Assessing the
hypothesis testing. In Information Fusion, impact of a diagnostic test; choosing a
2008 11th International Conference on (pp. gold standard, sensitivity, specificity, PPV,
1-8). IEEE; 2008. NPV, and likelihood ratios. Doing
88. Biau DJ, Jolles BM, Porcher R. P value Research in Emergency and Acute Care:
and the theory of hypothesis testing: An Making Order out of Chaos. 2015;205-222.
explanation for new researchers. Clinical 100. Williams KA, Harrild D, Williams DN.
Orthopaedics and Related Research. 2010; Statistics in nuclear cardiology. Annals of
468(3):885-892. Nuclear Cardiology. 2016;2(1):174-177.
89. Verdam MG, Oort FJ, Sprangers MA. 101. Pewsner D, Battaglia M, Minder C, Marx A,
Significance, truth and proof of p values: Bucher HC, Egger M. Ruling a diagnosis in
Reminders about common misconceptions or out with “SpPIn” and “SnNOut”: A note
regarding null hypothesis significance of caution. BMJ. 2004;329(7459):209-213.
18
Rushdi and Rushdi; JAMMR, 26(1): 1-21, 2018; Article no.JAMMR.40784
102. Aslan D, Sandberg S. Simple statistics in “probability”. Computers & Education. 2012;
diagnostic tests. Journal of Medical 58(3):931-941.
Biochemistry. 2007;26(4):309-313. 112. Da Costa NC. Pragmatic probability.
103. Baudrillard J, Peat J, Barton B. Categorical Erkenntnis. 1986;25(2):141-162.
and continuous variables: Diagnostic 113. Paolella MS. Fundamental probability: A
statistics. Medical Statistics: A Guide to computational approach. Wiley, New York,
Data Analysis and Critical Appraisal. 2008; NY. USA; 2006.
278-295. 114. Shafer G. Conditional probability.
104. Hegedus EJ, Stern B. Beyond SpPIN International Statistical Review/Revue
and SnNOUT: Considerations with Internationale de Statistique. 1985;53(3):
dichotomous tests during assessment of 261-275.
diagnostic accuracy. Journal of Manual & 115. Tomlinson S, Quinn R. Understanding
Manipulative Therapy. 2009;17(1):1E-5E. conditional probability. Teaching Statistics.
105. Denegar CR, Cordova ML. Application of 1997;19(1):2-7.
statistics in establishing diagnostic
116. Utts J. What educated citizens should
certainty. Journal of Athletic Training. 2012;
know about statistics and probability. The
47(2):233-236.
American Statistician. 2003;57(2):74-79.
106. Garfield J, Ahlgren A. Difficulties in
learning basic concepts in probability and 117. Carlton M. Pedigrees, prizes, and
statistics: Implications for research. Journal prisoners: The misuse of conditional
for Research in Mathematics Education. probability. Journal of Statistics Education.
1988;19(1):44-63. 2005;13(2).
107. Lord TR. 101 reasons for using Available:http://ww2.amstat.org/publication
cooperative learning in biology teaching. s/jse/v13n2/carlton.html
The American Biology Teacher. 2001; 118. Ancker JS. The language of conditional
63(1):30-38. probability. Journal of Statistics Education.
108. Prince M. Does active learning work? A 2006;14(2).
review of the research. Journal of Available:http://ww2.amstat.org/publication
Engineering Education. 2004;93(3):223- s/jse/v14n2/ancker.html
231. 119. Manage AB, Scariano SM. A classroom
109. Modell H, Michael J, Wenderoth MP. note on: student misconceptions regarding
Helping the learner to learn: The role of probabilistic independence vs. mutual
uncovering misconceptions. The American exclusivity. Mathematics and Computer
Biology Teacher. 2005;67(1):20-26. Education. 2010;44(1):14-21.
110. Michael J. Where's the evidence that 120. Böcherer-Linder K, Eichler A, Vogel M.
active learning works? Advances in Understanding conditional probability
Physiology Education. 2006;30(4):159-167. through visualization. In Proceedings of
111. Gürbüz R, Birgin O. The effect of the International Conference Turning Data
computer-assisted teaching on remedying into Knowledge: New Opportunities for
misconceptions: The case of the subject Statistics Education. 2015;14-23.
19
Rushdi and Rushdi; JAMMR, 26(1): 1-21, 2018; Article no.JAMMR.40784
There are three important and differing interpretations of probability, namely: the empirical, the logical,
and the subjectivistic. Despite this disagreement on the meaning of probability, there is a widespread
agreement on the basic axioms of the probability calculus and its mathematical structure [112]. We
use the empirical (frequentistic or common-sense) interpretation, and base our probability notions on
a probability “sample space” which constitutes the set of all possible (equally-likely) outcomes or
primitive events of the underlying random “experiment.” We describe events as subsets of the sample
space, and hence get N = 2 events for a sample space of outcomes or sample points [1, 113].
Since events are sets, we can also describe events via elementary set operations (complementation,
intersection, union, and set difference), namely:
The complement of a set is the set of all elements of the universal set that are not elements of .
The intersection ∩ of two sets and is the set that contains all elements of that also belong to
(or equivalently, all elements of that also belong to ), but no other elements. It is also the set of
all elements of the universal set that are not elements of ∪ .
The union ∪ of two sets and is the set of elements which are in alone, in alone, or in
both and . It is also the set of all elements of the universal set that are not elements of ∩ .
The difference − of two sets and is the set that contains all elements of that are not
elements of . It is also the set of elements in ∩ .
Since the outcomes in a sample space are equally likely, the probability ( ) of an event is defined
as the number of outcomes constituting (favoring ) divided by the total number of outcomes in the
sample space. The concept of conditional probability of event given event is based on the notion
that event replaces the universal set as a certain event. Hence, this conditional probability is given
by [1,6,9,21-23,113-120]
( | )= ( ∩ )⁄ ( ), ( )≠0 (A1)
We utilize the Venn diagram and the Karnaugh map of Fig. A1 to visualize the concept of conditional
probability. In this figure, ( | ) can be interpreted as the ratio of two areas, provided the diagram
and map be drawn as area-proportional. Table A1 lists the values ( | ) for some special cases.
Understanding these special cases is helpful for grasping the basic concepts of (and consolidating
knowledge about) conditional probability. The case = { } is typically used as a readily-
comprehensible starting point for the introduction of conditional probability. Its aggregation over a
multitude of sample points comprising immediately produces the general definition (A1).
20
Rushdi and Rushdi; JAMMR, 26(1): 1-21, 2018; Article no.JAMMR.40784
Fig. A1. Definition of the conditional probability ( | ) as the black area common to and ,
divided by the blue area of (both the Venn diagram and the Karnaugh map are assumed to
be area-proportional)
_________________________________________________________________________________
© 2018 Rushdi and Rushdi; This is an Open Access article distributed under the terms of the Creative Commons Attribution
License (http://creativecommons.org/licenses/by/4.0), which permits unrestricted use, distribution, and reproduction in any
medium, provided the original work is properly cited.
Peer-review history:
The peer review history for this paper can be accessed here:
http://www.sciencedomain.org/review-history/24107
21