2. Overcoming overconfidence

К оглавлению1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 
17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 
34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 
51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 
68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 
85 86 87 88 89 90 91 92 93 94 

If error is the constant companion of inquiry, so is overconfidence. People

are systematically prone to giving more credence to their beliefs than they

deserve. The literature demonstrating overconfidence is very large. To cite

one representative example mentioned earlier (chapter 2, section 3.4),

Fischhoff, Solvic, and Lichtenstein (1977) asked subjects to specify the

most frequent cause of death in the U.S. and to estimate their confidence

that their choice was correct. It turns out that when subjects set the odds

of their answer’s correctness at 100:1, they were correct 73% of the time;

even when they were so sure as to set the odds between 10,000:1 and

1,000,000:1, they were correct only between 85% and 90% of the time. The

overconfidence effect is systematic (it is highly replicable and survives

changes in task and setting) and directional (the effect is in the direction of

overconfidence rather than underconfidence—except when it comes to

objectively easy problems [Lichtenstein and Fischhoff 1980].) Given these

findings, one might hope that expert training reduces or eliminates

overconfidence. This does not appear to be so. Physicists, economists, and

demographers have all been observed to suffer from this bias, even when

reasoning about the content of their special discipline (Henrion and

Fischoff 1986). In fact, it might be useful for readers of this book to know

that people who score better on SATs are also more prone to overconfidence

(Stanovich 1999, 120). Overconfidence is not the mere result of

individual differences in personality or of clinical delusions of grandeur. It

is the normal consequence of routine cognitive activity.

Overconfidence in highly significant reasoning problems can lead to

fatally inaccurate judgments—for example, that guilt has been proven beyond

a reasonable doubt. There have been 110 people who were convicted

of murder who have been found to be not guilty on the basis of DNA

evidence. Most of these errors have been found to be the result of mistaken

eyewitness testimony and the overconfidence of eyewitnesses and juries in

the reliability of such testimony (Wells, Olson, and Charman 2002).

Are there any reliable strategies for counteracting this overconfidence?

Arkes, Dawes, and Christensen (1986) found that accuracy incentives did

not improve accuracy of judgments (also see Arkes 1991). Lord et al.

(1984) found that simple motivational declarations do not reduce bias

either. So, for example, offering subjects $1 for each correct answer or

urging subjects to be unbiased do not reduce overconfidence. There are

some ways to reduce bias, but they involve introducing elements that are

unlike what people tend to face in daily life. For example, when subjects

are made to feel accountable for their judgments (by being told that their

answers will be discussed after the session), they assimilated evidence in a

less biased way. Further, overconfidence can be eliminated when subjects

are persistently exposed to a rigorous schedule of specially prepared feedback

(Arkes, Christensen, Lai, and Blumer 1987). These sorts of findings

are important for thinking about what sorts of procedures to implement

when deliberations occur in stable social settings (such as a jury or a policy

panel). These sorts of calibration exercises can prompt groups of people to

be less overconfident than each individual, reasoning alone, might have

been. Indeed, it may be that controlled environments, where debiasing

decision-making procedures can be implemented, offer our best hope of

overcoming overconfidence.

The most prominent of the individual, subjective debiasing methods

is the consider-the-opposite strategy. One of the groundbreaking studies on

debiasing argued that people ‘‘have a blind spot for opposite possibilities’’

when making social and policy judgments (Lord, Lepper, and Preston,

1984). When subjects were asked to generate pros and cons for a judgment

they had made, Koriat, Lichtenstein, and Fischhoff (1980) demonstrated

that overconfidence bias was reduced. Indeed, they found that it was the

generation of opposing reasons that did all of the bias-reducing work. The

most effective version of the consider-the-opposite strategy involves a simple

rule: ‘‘Stop to consider why your judgment might be wrong’’ (Plous

1993, 228). Here are some beliefs that we might hold with undue certainty.

If S believes that New York State is the largest state on the Eastern Seaboard,

S might consider that he might be wrong because New York gets a

lot more press than South Atlantic states. If S believes that Los Angeles

is west of Reno, he might ask himself whether he might not have a clear

sense of the orientation of the American landmass. If S believes that a

defendant is guilty beyond a reasonable doubt, he might consider that he

might be mistaken because of his confusion over the DNA evidence.

While we are somewhat hopeful about the ability of the consider-theopposite

strategy to reduce our overconfidence, there are a number of issues

that need to be resolved before we can embrace it wholeheartedly.One serious

worry is whether the consider-the-opposite strategy can be exported to a

setting in which it can be usefully implemented. This is not a criticism of the

ecological validity of the experiments; there is no question that if you could

get people in natural settings to perform the same experimental debiasing

task, overconfidence would be reduced. The question instead is whether, as

people go through their daily lives, they will have the discipline, motivation,

and concentration required to implement the consider-the-opposite strategy.

This kind of worry drops out of our view quite naturally. Strategic

reliabilismis a view that explicitly recognizes our psychological limitations. If

it is part of our psychological functioning that we tend to ignore or undervalue

corrective strategies, this is something we must take seriously.

But even if we can implement the consider-the-opposite strategy to

good effect, there is another serious worry: What is its range to be? After all, it would surely be a mistake for us to adopt the strategy for every belief

we entertain. Such a strategy would turn us into unhappy, neurotic nebbishes,

impossible to get along with. (One of us is reminded of the student

who was obviously having serious life troubles and explained that it was

because he had really taken Descartes’ Meditations to heart.) Recall our

discussion in chapter 3 of the Four Ways to better reasoning. We noted

that there could be occasions when one could become a better reasoner by

replacing a reliable but expensive strategy with another that was less reliable

but less expensive. Replacing a wide ranging consider-the-opposite

strategy with a consider-the-opposite strategy with a narrower range might

be an example of this. If a consider-the-opposite strategy is a good idea, it

is surely one whose range is to be limited to highly significant problems in

which short-term reliability is very important.

If error is the constant companion of inquiry, so is overconfidence. People

are systematically prone to giving more credence to their beliefs than they

deserve. The literature demonstrating overconfidence is very large. To cite

one representative example mentioned earlier (chapter 2, section 3.4),

Fischhoff, Solvic, and Lichtenstein (1977) asked subjects to specify the

most frequent cause of death in the U.S. and to estimate their confidence

that their choice was correct. It turns out that when subjects set the odds

of their answer’s correctness at 100:1, they were correct 73% of the time;

even when they were so sure as to set the odds between 10,000:1 and

1,000,000:1, they were correct only between 85% and 90% of the time. The

overconfidence effect is systematic (it is highly replicable and survives

changes in task and setting) and directional (the effect is in the direction of

overconfidence rather than underconfidence—except when it comes to

objectively easy problems [Lichtenstein and Fischhoff 1980].) Given these

findings, one might hope that expert training reduces or eliminates

overconfidence. This does not appear to be so. Physicists, economists, and

demographers have all been observed to suffer from this bias, even when

reasoning about the content of their special discipline (Henrion and

Fischoff 1986). In fact, it might be useful for readers of this book to know

that people who score better on SATs are also more prone to overconfidence

(Stanovich 1999, 120). Overconfidence is not the mere result of

individual differences in personality or of clinical delusions of grandeur. It

is the normal consequence of routine cognitive activity.

Overconfidence in highly significant reasoning problems can lead to

fatally inaccurate judgments—for example, that guilt has been proven beyond

a reasonable doubt. There have been 110 people who were convicted

of murder who have been found to be not guilty on the basis of DNA

evidence. Most of these errors have been found to be the result of mistaken

eyewitness testimony and the overconfidence of eyewitnesses and juries in

the reliability of such testimony (Wells, Olson, and Charman 2002).

Are there any reliable strategies for counteracting this overconfidence?

Arkes, Dawes, and Christensen (1986) found that accuracy incentives did

not improve accuracy of judgments (also see Arkes 1991). Lord et al.

(1984) found that simple motivational declarations do not reduce bias

either. So, for example, offering subjects $1 for each correct answer or

urging subjects to be unbiased do not reduce overconfidence. There are

some ways to reduce bias, but they involve introducing elements that are

unlike what people tend to face in daily life. For example, when subjects

are made to feel accountable for their judgments (by being told that their

answers will be discussed after the session), they assimilated evidence in a

less biased way. Further, overconfidence can be eliminated when subjects

are persistently exposed to a rigorous schedule of specially prepared feedback

(Arkes, Christensen, Lai, and Blumer 1987). These sorts of findings

are important for thinking about what sorts of procedures to implement

when deliberations occur in stable social settings (such as a jury or a policy

panel). These sorts of calibration exercises can prompt groups of people to

be less overconfident than each individual, reasoning alone, might have

been. Indeed, it may be that controlled environments, where debiasing

decision-making procedures can be implemented, offer our best hope of

overcoming overconfidence.

The most prominent of the individual, subjective debiasing methods

is the consider-the-opposite strategy. One of the groundbreaking studies on

debiasing argued that people ‘‘have a blind spot for opposite possibilities’’

when making social and policy judgments (Lord, Lepper, and Preston,

1984). When subjects were asked to generate pros and cons for a judgment

they had made, Koriat, Lichtenstein, and Fischhoff (1980) demonstrated

that overconfidence bias was reduced. Indeed, they found that it was the

generation of opposing reasons that did all of the bias-reducing work. The

most effective version of the consider-the-opposite strategy involves a simple

rule: ‘‘Stop to consider why your judgment might be wrong’’ (Plous

1993, 228). Here are some beliefs that we might hold with undue certainty.

If S believes that New York State is the largest state on the Eastern Seaboard,

S might consider that he might be wrong because New York gets a

lot more press than South Atlantic states. If S believes that Los Angeles

is west of Reno, he might ask himself whether he might not have a clear

sense of the orientation of the American landmass. If S believes that a

defendant is guilty beyond a reasonable doubt, he might consider that he

might be mistaken because of his confusion over the DNA evidence.

While we are somewhat hopeful about the ability of the consider-theopposite

strategy to reduce our overconfidence, there are a number of issues

that need to be resolved before we can embrace it wholeheartedly.One serious

worry is whether the consider-the-opposite strategy can be exported to a

setting in which it can be usefully implemented. This is not a criticism of the

ecological validity of the experiments; there is no question that if you could

get people in natural settings to perform the same experimental debiasing

task, overconfidence would be reduced. The question instead is whether, as

people go through their daily lives, they will have the discipline, motivation,

and concentration required to implement the consider-the-opposite strategy.

This kind of worry drops out of our view quite naturally. Strategic

reliabilismis a view that explicitly recognizes our psychological limitations. If

it is part of our psychological functioning that we tend to ignore or undervalue

corrective strategies, this is something we must take seriously.

But even if we can implement the consider-the-opposite strategy to

good effect, there is another serious worry: What is its range to be? After all, it would surely be a mistake for us to adopt the strategy for every belief

we entertain. Such a strategy would turn us into unhappy, neurotic nebbishes,

impossible to get along with. (One of us is reminded of the student

who was obviously having serious life troubles and explained that it was

because he had really taken Descartes’ Meditations to heart.) Recall our

discussion in chapter 3 of the Four Ways to better reasoning. We noted

that there could be occasions when one could become a better reasoner by

replacing a reliable but expensive strategy with another that was less reliable

but less expensive. Replacing a wide ranging consider-the-opposite

strategy with a consider-the-opposite strategy with a narrower range might

be an example of this. If a consider-the-opposite strategy is a good idea, it

is surely one whose range is to be limited to highly significant problems in

which short-term reliability is very important.