formal fallacy. If presented with related base rate
information (i.e. generic, general information) and specific
information (information pertaining only to a certain case), the mind
tends to ignore the former and focus on the latter.
The base rate fallacy, also called base rate neglect or base rate bias, is a
One type of base rate fallacy is the false positive paradox, where false positive
tests are more probable than true positive tests, occurring when the
overall population has a low incidence of a condition and the incidence
rate is lower than the false positive rate. The probability of a
positive test result is determined not only by the accuracy of the test
but by the characteristics of the sampled population. When the incidence, the proportion of those who have a given condition, is lower than the test's false positive rate, even tests that have a very low chance of giving a false positive in an individual case will give more false than true positives overall.
So, in a society with very few infected people—fewer proportionately
than the test gives false positives—there will actually be more who test
positive for a disease incorrectly and don't have it than those who
test positive accurately and do. The paradox has surprised many.
It is especially counter-intuitive when interpreting a positive result in a test on a low-incidence population after having dealt with positive results drawn from a high-incidence population. If the false positive rate of the test is higher than the proportion of the new
population with the condition, then a test administrator whose
experience has been drawn from testing in a high-incidence population
may conclude from experience
that a positive test result usually indicates a positive subject, when
in fact a false positive is far more likely to have occurred.
Examples
Example 1: HIV
High-incidence population
Number of people |
Infected | Uninfected | Total |
---|---|---|---|
Test positive |
400 (true positive) |
30 (false positive) |
430 |
Test negative |
0 (false negative) |
570 (true negative) |
570 |
Total | 400 | 600 | 1000 |
Imagine running an HIV test on population A of 1000 persons, in which 40% are infected. The test has a false positive rate of 5% (0.05) and no false negative rate. The expected outcome of the 1000 tests on population A would be:
- Infected and test indicates disease (true positive)
- 1000 × 40/100 = 400 people would receive a true positive
- Uninfected and test indicates disease (false positive)
- 1000 × 100 – 40/100 × 0.05 = 30 people would receive a false positive
- The remaining 570 tests are correctly negative.
So, in population A, a person receiving a positive test could be over 93% confident (400/30 + 400) that it correctly indicates infection.
Low-incidence population
Number of people |
Infected | Uninfected | Total |
---|---|---|---|
Test positive |
20 (true positive) |
49 (false positive) |
69 |
Test negative |
0 (false negative) |
931 (true negative) |
931 |
Total | 20 | 980 | 1000 |
Now consider the same test applied to population B, in which only 2% is infected. The expected outcome of 1000 tests on population B would be:
- Infected and test indicates disease (true positive)
- 1000 × 2/100 = 20 people would receive a true positive
- Uninfected and test indicates disease (false positive)
- 1000 × 100 – 2/100 × 0.05 = 49 people would receive a false positive
- The remaining 931 (= 1000 - (49 + 20)) tests are correctly negative.
In population B, only 20 of the 69 total people with a
positive test result are actually infected. So, the probability of
actually being infected after one is told that one is infected is only
29% (20/20 + 49) for a test that otherwise appears to be "95% accurate".
A tester with experience of group A might find it a paradox that in group B, a result that had usually correctly indicated infection is now usually a false positive. The confusion of the posterior probability of infection with the prior probability of receiving a false positive is a natural error after receiving a life-threatening test result.
Example 2: Drunk drivers
A group of police officers have breathalyzers displaying false drunkenness in 5% of the cases in which the driver is sober. However, the breathalyzers never fail to detect a truly drunk person. One in a thousand drivers is driving drunk. Suppose the police officers then stop a driver at random, and force the driver to take a breathalyzer test. It indicates that the driver is drunk. We assume you don't know anything else about him or her. How high is the probability he or she really is drunk?
Many would answer as high as 95%, but the correct probability is about 2%.
An explanation for this is as follows: on average, for every 1,000 drivers tested,
- 1 driver is drunk, and it is 100% certain that for that driver there is a true positive test result, so there is 1 true positive test result
- 999 drivers are not drunk, and among those drivers there are 5% false positive test results, so there are 49.95 false positive test results
Therefore, the probability that one of the drivers among the 1 + 49.95 = 50.95 positive test results really is drunk is .
The validity of this result does, however, hinge on the validity
of the initial assumption that the police officer stopped the driver
truly at random, and not because of bad driving. If that or another
non-arbitrary reason for stopping the driver was present, then the
calculation also involves the probability of a drunk driver driving
competently and a non-drunk driver driving (in-)competently.
More formally, the same probability of roughly 0.02 can be established using Bayes's theorem.
The goal is to find the probability that the driver is drunk given that
the breathalyzer indicated he/she is drunk, which can be represented as
where D means that the breathalyzer indicates that the driver is drunk. Bayes's theorem tells us that
We were told the following in the first paragraph:
- and
As you can see from the formula, one needs p(D) for Bayes' theorem, which one can compute from the preceding values using the law of total probability:
which gives
Plugging these numbers into Bayes' theorem, one finds that
Example 3: Terrorist identification
In
a city of 1 million inhabitants let there be 100 terrorists and 999,900
non-terrorists. To simplify the example, it is assumed that all people
present in the city are inhabitants. Thus, the base rate probability of a
randomly selected inhabitant of the city being a terrorist is 0.0001,
and the base rate probability of that same inhabitant being a
non-terrorist is 0.9999. In an attempt to catch the terrorists, the city
installs an alarm system with a surveillance camera and automatic
facial recognition software.
The software has two failure rates of 1%:
- The false negative rate: If the camera scans a terrorist, a bell will ring 99% of the time, and it will fail to ring 1% of the time.
- The false positive rate: If the camera scans a non-terrorist, a bell will not ring 99% of the time, but it will ring 1% of the time.
Suppose now that an inhabitant triggers the alarm. What is the chance
that the person is a terrorist? In other words, what is P(T | B), the
probability that a terrorist has been detected given the ringing of the
bell? Someone making the 'base rate fallacy' would infer that there is a
99% chance that the detected person is a terrorist. Although the
inference seems to make sense, it is actually bad reasoning, and a
calculation below will show that the chances he/she is a terrorist are
actually near 1%, not near 99%.
The fallacy arises from confusing the natures of two different
failure rates. The 'number of non-bells per 100 terrorists' and the
'number of non-terrorists per 100 bells' are unrelated quantities. One
does not necessarily equal the other, and they don't even have to be
almost equal. To show this, consider what happens if an identical alarm
system were set up in a second city with no terrorists at all. As in the
first city, the alarm sounds for 1 out of every 100 non-terrorist
inhabitants detected, but unlike in the first city, the alarm never
sounds for a terrorist. Therefore, 100% of all occasions of the alarm
sounding are for non-terrorists, but a false negative rate cannot even
be calculated. The 'number of non-terrorists per 100 bells' in that city
is 100, yet P(T | B) = 0%. There is zero chance that a terrorist has
been detected given the ringing of the bell.
Imagine that the first city's entire population of one million
people pass in front of the camera. About 99 of the 100 terrorists will
trigger the alarm—and so will about 9,999 of the 999,900 non-terrorists.
Therefore, about 10,098 people will trigger the alarm, among which
about 99 will be terrorists. So, the probability that a person
triggering the alarm actually is a terrorist, is only about 99 in
10,098, which is less than 1%, and very, very far below our initial
guess of 99%.
The base rate fallacy is so misleading in this example because
there are many more non-terrorists than terrorists, and the number of
false positives (non-terrorists scanned as terrorists) is so much larger
than the true positives (the real number of terrorists).
Findings in psychology
In
experiments, people have been found to prefer individuating information
over general information when the former is available.
In some experiments, students were asked to estimate the grade point averages
(GPAs) of hypothetical students. When given relevant statistics about
GPA distribution, students tended to ignore them if given descriptive
information about the particular student even if the new descriptive
information was obviously of little or no relevance to school
performance. This finding has been used to argue that interviews are an unnecessary part of the college admissions process because interviewers are unable to pick successful candidates better than basic statistics.
Psychologists Daniel Kahneman and Amos Tversky attempted to explain this finding in terms of a simple rule or "heuristic" called representativeness.
They argued that many judgments relating to likelihood, or to cause and
effect, are based on how representative one thing is of another, or of a
category. Kahneman considers base rate neglect to be a specific form of extension neglect. Richard Nisbett has argued that some attributional biases like the fundamental attribution error
are instances of the base rate fallacy: people do not use the
"consensus information" (the "base rate") about how others behaved in
similar situations and instead prefer simpler dispositional attributions.
There is considerable debate in psychology on the conditions under which people do or do not appreciate base rate information.
Researchers in the heuristics-and-biases program have stressed
empirical findings showing that people tend to ignore base rates and
make inferences that violate certain norms of probabilistic reasoning,
such as Bayes' theorem. The conclusion drawn from this line of research was that human probabilistic thinking is fundamentally flawed and error-prone.
Other researchers have emphasized the link between cognitive processes
and information formats, arguing that such conclusions are not generally
warranted.
Consider again Example 2 from above. The required inference is to
estimate the (posterior) probability that a (randomly picked) driver is
drunk, given that the breathalyzer test is positive. Formally, this
probability can be calculated using Bayes' theorem,
as shown above. However, there are different ways of presenting the
relevant information. Consider the following, formally equivalent
variant of the problem:
- 1 out of 1000 drivers are driving drunk. The breathalyzers never fail to detect a truly drunk person. For 50 out of the 999 drivers who are not drunk the breathalyzer falsely displays drunkness. Suppose the policemen then stop a driver at random, and force them to take a breathalyzer test. It indicates that he or she is drunk. We assume you don't know anything else about him or her. How high is the probability he or she really is drunk?
In this case, the relevant numerical information—p(drunk), p(D | drunk), p(D | sober)—is presented in terms of natural frequencies with respect to a certain reference class.
Empirical studies show that people's inferences correspond more closely
to Bayes' rule when information is presented this way, helping to
overcome base-rate neglect in laypeople and experts. As a consequence, organizations like the Cochrane Collaboration recommend using this kind of format for communicating health statistics.
Teaching people to translate these kinds of Bayesian reasoning problems
into natural frequency formats is more effective than merely teaching
them to plug probabilities (or percentages) into Bayes' theorem.
It has also been shown that graphical representations of natural
frequencies (e.g., icon arrays) help people to make better inferences.
Why are natural frequency formats helpful? One important reason
is that this information format facilitates the required inference
because it simplifies the necessary calculations. This can be seen when
using an alternative way of computing the required probability p(drunk|D):
where N(drunk ∩ D) denotes the number of drivers that are drunk and get a positive breathalyzer result, and N(D)
denotes the total number of cases with a positive breathalyzer result.
The equivalence of this equation to the above one follows from the
axioms of probability theory, according to which N(drunk ∩ D) = N × p (D | drunk) × p
(drunk). Importantly, although this equation is formally equivalent to
Bayes' rule, it is not psychologically equivalent. Using natural
frequencies simplifies the inference because the required mathematical
operation can be performed on natural numbers, instead of normalized
fractions (i.e., probabilities), because it makes the high number of
false positives more transparent, and because natural frequencies
exhibit a "nested-set structure".
It is important to note that not any kind of frequency format facilitates Bayesian reasoning. Natural frequencies refer to frequency information that results from natural sampling,
which preserves base rate information (e.g., number of drunken drivers
when taking a random sample of drivers). This is different from systematic sampling,
in which base rates are fixed a priori (e.g., in scientific
experiments). In the latter case it is not possible to infer the
posterior probability p (drunk | positive test) from comparing
the number of drivers who are drunk and test positive compared to the
total number of people who get a positive breathalyzer result, because
base rate information is not preserved and must be explicitly
re-introduced using Bayes' theorem.