3.4. The basis of epistemic exceptionalism: The overconfidence feedback loop

К оглавлению1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 
17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 
34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 
51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 
68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 
85 86 87 88 89 90 91 92 93 94 

Let’s recap briefly. We aren’t especially good at detecting the properties

that covary with the target property we want to predict—especially when

we have strong background opinions and when the informational situation

is complex. We aren’t especially good at recalling or attending to lots

of different avenues of information. And often, the feedback we get about

the quality of our judgments or behavior is unrepresentative (and we don’t

know it) or incomplete (and we don’t see that this is a serious problem).

As a result, it is not surprising that we aren’t especially reliable in our

judgments about complex social phenomena.

Against this background, the sluggish reception SPRs have received in

the disciplines whose business it is to predict and diagnose is particularly

puzzling. (Resistance to the use of SPRs is particularly strong when it

comes to making social predictions. SPRs have found easier acceptance in

non-psychiatric medical diagnosis.) In the face of a half century of studies

showing the superiority of SPRs, many experts still base judgments on

subjective impressions and unmonitored evaluation of the evidence. Resistance

to the SPR findings runs deep and typically comes as a kind of

epistemic exceptionalism. Those who resist the SPR findings take their

reasoning powers to be exceptional, and so they defect from the judgments

of SPRs when they find what they take to be exceptions to it. They are

typically quite willing to admit that in the long run, SPRs will be right more

often than human experts. But their (over)confidence in their subjective

powers of reflection leads them to deny that we should believe the SPR’s

prediction in this particular case.

We suspect that epistemic exceptionalism, which we suggest has led

to the sluggish reception of SPRs, is the result of two facts about people.

When it comes to prediction, we find the success of SPRs hard to believe,

The Amazing Success of Statistical Prediction Rules 43

and we find our lack of success hard to believe. The reason we find our

own lack of success hard to believe is that most of the failures of our

predictive capacities are hidden from us. We don’t see what’s gone wrong.

We don’t detect the right covariations, but we think we do. We can’t attend

to the relevant complexities, but we think we have. We aren’t getting

representative feedback on our predictions, but we think we are. As a result,

we tend to be overconfident about the power of our subjective reasoning

faculties and about the reliability of our predictions (Trout 2002).

Our faith in the reliability of our subjective powers of reasoning bolsters

our (over)confidence in our judgments; and our (over)confident judgments

bolster our belief in the reliability of our subjective faculties (Arkes

1991; Sieck and Arkes [unpublished manuscript]). Let’s focus on each side

of this overconfidence feedback loop.

The first side of the overconfidence feedback loop consists in overconfidence

in our judgments. This overconfidence leads too often to defection

from a successful SPR. That we fall victim to an overconfidence

bias is one of the most robust findings in contemporary psychology:

[A] large majority of the general public thinks that they are more intelligent,

more fair-minded, less prejudiced, and more skilled behind the wheel of an

automobile than the average person. . . . A survey of one million high school

seniors found that 70% thought they were above average in leadership

ability, and only 2% thought they were below average. In terms of ability to

get along with others, all students thought they were above average, 60%

thought they were in the top 10%, and 25% thought they were in the top

1%! Lest one think that such inflated self-assessments occur only in the

minds of callow high-school students, it should be pointed out that a survey

of university professors found that 94% thought they were better at their

jobs than their average colleague. (Gilovich 1991, 77)

The overconfidence bias goes far beyond our inflated self-assessments. For

example, Fischhoff, Slovic, and Lichtenstein (1977) asked subjects to indicate

the most frequent cause of death in the U.S. and to estimate their

confidence that their choice was correct (in terms of ‘‘odds’’). When

subjects set the odds of their answer’s correctness at 100:1, they were

correct only 73% of the time. Remarkably, even when they were so certain

as to set the odds between 10,000:1 and 1,000,000:1, they were correct only

between 85% and 90% of the time. It is important to note that the overconfidence

effect is systematic (it is highly replicable and survives changes

in task and setting) and directional (the effect is in the direction of

over rather than underconfidence). But overconfidence is eliminated or

44 Epistemology and the Psychology of Human Judgment

reversed when the questions are very easy. This phenomenon is known as

the difficulty (or hard-easy) effect (Lichtenstein and Fischhoff 1977).

The second side of the overconfidence feedback loop consists of our

overconfidence in the reliability of our subjective reasoning faculties. We

are naturally disposed to exaggerate the powers of our subjective faculties.

A very prominent example that we have already discussed is the interview

effect. When gatekeepers avail themselves of unstructured interviews, they

actually degrade the reliability of their predictions. Although the interview

effect is one of the most robust findings in psychology, highly educated

people ignore its obvious practical implication. We suspect that this occurs

because of our confidence in our subjective ability to ‘‘read’’ people.

We suppose that our insight into human nature is so powerful that we

can plumb the depths of a human being in a 45-minute interview—unlike

the lesser lights who were hoodwinked in the SPR studies. As we have said,

a major reason our (over)confidence survives is because we typically don’t

get systematic feedback about the quality of our judgments (e.g., we can’t

compare the long-term outcomes of our actual decisions against the

decisions we would have made if we hadn’t interviewed the candidates).

To put this in practical terms, the process by which most working philosophers

were hired was seriously and, at the time, demonstrably flawed.

This will be of no comfort to our colleagues, employed or unemployed.

We expect, however, that the unemployed will find it considerably less

surprising.

Let’s recap briefly. We aren’t especially good at detecting the properties

that covary with the target property we want to predict—especially when

we have strong background opinions and when the informational situation

is complex. We aren’t especially good at recalling or attending to lots

of different avenues of information. And often, the feedback we get about

the quality of our judgments or behavior is unrepresentative (and we don’t

know it) or incomplete (and we don’t see that this is a serious problem).

As a result, it is not surprising that we aren’t especially reliable in our

judgments about complex social phenomena.

Against this background, the sluggish reception SPRs have received in

the disciplines whose business it is to predict and diagnose is particularly

puzzling. (Resistance to the use of SPRs is particularly strong when it

comes to making social predictions. SPRs have found easier acceptance in

non-psychiatric medical diagnosis.) In the face of a half century of studies

showing the superiority of SPRs, many experts still base judgments on

subjective impressions and unmonitored evaluation of the evidence. Resistance

to the SPR findings runs deep and typically comes as a kind of

epistemic exceptionalism. Those who resist the SPR findings take their

reasoning powers to be exceptional, and so they defect from the judgments

of SPRs when they find what they take to be exceptions to it. They are

typically quite willing to admit that in the long run, SPRs will be right more

often than human experts. But their (over)confidence in their subjective

powers of reflection leads them to deny that we should believe the SPR’s

prediction in this particular case.

We suspect that epistemic exceptionalism, which we suggest has led

to the sluggish reception of SPRs, is the result of two facts about people.

When it comes to prediction, we find the success of SPRs hard to believe,

The Amazing Success of Statistical Prediction Rules 43

and we find our lack of success hard to believe. The reason we find our

own lack of success hard to believe is that most of the failures of our

predictive capacities are hidden from us. We don’t see what’s gone wrong.

We don’t detect the right covariations, but we think we do. We can’t attend

to the relevant complexities, but we think we have. We aren’t getting

representative feedback on our predictions, but we think we are. As a result,

we tend to be overconfident about the power of our subjective reasoning

faculties and about the reliability of our predictions (Trout 2002).

Our faith in the reliability of our subjective powers of reasoning bolsters

our (over)confidence in our judgments; and our (over)confident judgments

bolster our belief in the reliability of our subjective faculties (Arkes

1991; Sieck and Arkes [unpublished manuscript]). Let’s focus on each side

of this overconfidence feedback loop.

The first side of the overconfidence feedback loop consists in overconfidence

in our judgments. This overconfidence leads too often to defection

from a successful SPR. That we fall victim to an overconfidence

bias is one of the most robust findings in contemporary psychology:

[A] large majority of the general public thinks that they are more intelligent,

more fair-minded, less prejudiced, and more skilled behind the wheel of an

automobile than the average person. . . . A survey of one million high school

seniors found that 70% thought they were above average in leadership

ability, and only 2% thought they were below average. In terms of ability to

get along with others, all students thought they were above average, 60%

thought they were in the top 10%, and 25% thought they were in the top

1%! Lest one think that such inflated self-assessments occur only in the

minds of callow high-school students, it should be pointed out that a survey

of university professors found that 94% thought they were better at their

jobs than their average colleague. (Gilovich 1991, 77)

The overconfidence bias goes far beyond our inflated self-assessments. For

example, Fischhoff, Slovic, and Lichtenstein (1977) asked subjects to indicate

the most frequent cause of death in the U.S. and to estimate their

confidence that their choice was correct (in terms of ‘‘odds’’). When

subjects set the odds of their answer’s correctness at 100:1, they were

correct only 73% of the time. Remarkably, even when they were so certain

as to set the odds between 10,000:1 and 1,000,000:1, they were correct only

between 85% and 90% of the time. It is important to note that the overconfidence

effect is systematic (it is highly replicable and survives changes

in task and setting) and directional (the effect is in the direction of

over rather than underconfidence). But overconfidence is eliminated or

44 Epistemology and the Psychology of Human Judgment

reversed when the questions are very easy. This phenomenon is known as

the difficulty (or hard-easy) effect (Lichtenstein and Fischhoff 1977).

The second side of the overconfidence feedback loop consists of our

overconfidence in the reliability of our subjective reasoning faculties. We

are naturally disposed to exaggerate the powers of our subjective faculties.

A very prominent example that we have already discussed is the interview

effect. When gatekeepers avail themselves of unstructured interviews, they

actually degrade the reliability of their predictions. Although the interview

effect is one of the most robust findings in psychology, highly educated

people ignore its obvious practical implication. We suspect that this occurs

because of our confidence in our subjective ability to ‘‘read’’ people.

We suppose that our insight into human nature is so powerful that we

can plumb the depths of a human being in a 45-minute interview—unlike

the lesser lights who were hoodwinked in the SPR studies. As we have said,

a major reason our (over)confidence survives is because we typically don’t

get systematic feedback about the quality of our judgments (e.g., we can’t

compare the long-term outcomes of our actual decisions against the

decisions we would have made if we hadn’t interviewed the candidates).

To put this in practical terms, the process by which most working philosophers

were hired was seriously and, at the time, demonstrably flawed.

This will be of no comfort to our colleagues, employed or unemployed.

We expect, however, that the unemployed will find it considerably less

surprising.