3.3. Lack of reliable feedback

К оглавлению1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 
17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 
34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 
51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 
68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 
85 86 87 88 89 90 91 92 93 94 

Even if we don’t start off making complex social judgments in a reliable

fashion, we can at least hope to improve our judgments by receiving and

acting on accurate feedback. If we can determine that a depressingly large

number of our past judgments were mistaken (or even that a well-defined

class of past predictions was mistaken), perhaps we can modify our reasoning

strategies and so judge more accurately. (The fact that a person

might have made such modifications might lead him to discount the

pessimism we seem to be insisting upon here.) Unfortunately, there are a

number of quite natural phenomena that keep us from getting accurate

feedback on our past judgments and behaviors.

For many irrevocable decisions we make, the feedback we receive on

our judgments is almost inevitably incomplete. Consider the grizzled

philosopher who has played a major role in hiring a number of junior

colleagues and who takes the interviews very seriously. Given the nature of

the job market in philosophy, it’s quite likely that his junior colleagues are,

by and large, a pretty impressive lot. Given this feedback, he is likely to

think quite highly of his ability to identify in interviews good young

philosophers. The problem here is that the grizzled philosopher doesn’t

know whether his predictions would have turned out better or worse

without the interviews. (And even if he did, it’s unlikely he would have a

large enough sample size to draw a reasonable conclusion.) Simply put,

most gatekeepers don’t have control groups to test the effectiveness of

their reasoning strategies. After all, the set of junior colleagues who would

have been hired without interviews (the control group) might have been

even more terrific than his actual set of junior colleagues. The problem is

The Amazing Success of Statistical Prediction Rules 41

not just that most gatekeepers don’t have control groups—that is often a

practical inevitability. The problem is that they don’t recognize that this is

a serious problem. Most gatekeepers should probably have much more

diffidence concerning their powers of prediction—especially in a job market

in which most job seekers are something more than competent.

Another problem is that the feedback we get, especially when it comes

to social matters, is likely to be highly unrepresentative. Consider the

finding that 94% of university professors believe they’re better-thanaverage

at their jobs (Gilovich 1991, 77). One reason for this may be that we

typically get personal feedback from students who think we were terrific

teachers (or at least who say we were terrific teachers). Seldom will students

go out of their way to make contact with professors they thought were

really mediocre (if for no other reason than, where would they begin?).

The problem of unrepresentative feedback can be made vivid with an

example that is likely familiar to everybody. Think about someone who

employs mildly (or outright) annoying interpersonal strategies, for example,

dominating conversations or name dropping. How likely are you to

tell this person that these behaviors are annoying? Some blunt folk might

always do so. But most of us, probably as a result of some combination of

politeness, pusillanimity, and prudence, let it slide. Of course, we recognize

that this behavior is annoying (or worse), and we might judge the person to

be annoying (or worse). But given the feedback he has received, he might

well go forth into the world confident that he has once again been socially

deft, charming, and deeply impressive. (We are inclined to suggest you

perform a public service. Supply accurate feedback. Call a bore a bore, a

jerk a jerk, a blowhard a blowhard. Just don’t do it to us.)

Even when the feedback we get is representative and shows that our

predictions are mistaken, we will often interpret such feedback in a way

that supports our preconceptions. For example, Gilovich (1983) asked people

who gambled on football games to tape-record their thoughts about

the outcomes of their bets. One might expect the gamblers to remember

their wins and repress their losses. In fact, just the opposite occurred:

[T]hey spent more time discussing their losses than their wins. Furthermore,

the kind of comments made about wins and losses were quite different. The

bettors tended to make ‘‘undoing’’ comments about their losses—comments

to the effect that the outcome would have been different if not for some

anomalous or ‘‘fluke’’ element. . . . In contrast, they tended to make ‘‘bolstering’’

comments about their wins—comments indicating that the outcome

either should have been as it was, or should have been even more

extreme in the same direction. . . . By carefully scrutinizing and explaining away their losses, while accepting their successes at face value, gamblers do

indeed rewrite their personal histories of success and failure. Losses are often

counted, not as losses, but as ‘‘near wins.’’ (Gilovich 1991, 55)

One interesting feature of this common interpretative strategy is that the

subject cannot be accused of ignoring negative evidence. In fact, the subject

is attending more to the negative evidence than to the positive evidence.

It’s just that he interprets the positive evidence as positive, and the

negative evidence as bad luck.

Even if we don’t start off making complex social judgments in a reliable

fashion, we can at least hope to improve our judgments by receiving and

acting on accurate feedback. If we can determine that a depressingly large

number of our past judgments were mistaken (or even that a well-defined

class of past predictions was mistaken), perhaps we can modify our reasoning

strategies and so judge more accurately. (The fact that a person

might have made such modifications might lead him to discount the

pessimism we seem to be insisting upon here.) Unfortunately, there are a

number of quite natural phenomena that keep us from getting accurate

feedback on our past judgments and behaviors.

For many irrevocable decisions we make, the feedback we receive on

our judgments is almost inevitably incomplete. Consider the grizzled

philosopher who has played a major role in hiring a number of junior

colleagues and who takes the interviews very seriously. Given the nature of

the job market in philosophy, it’s quite likely that his junior colleagues are,

by and large, a pretty impressive lot. Given this feedback, he is likely to

think quite highly of his ability to identify in interviews good young

philosophers. The problem here is that the grizzled philosopher doesn’t

know whether his predictions would have turned out better or worse

without the interviews. (And even if he did, it’s unlikely he would have a

large enough sample size to draw a reasonable conclusion.) Simply put,

most gatekeepers don’t have control groups to test the effectiveness of

their reasoning strategies. After all, the set of junior colleagues who would

have been hired without interviews (the control group) might have been

even more terrific than his actual set of junior colleagues. The problem is

The Amazing Success of Statistical Prediction Rules 41

not just that most gatekeepers don’t have control groups—that is often a

practical inevitability. The problem is that they don’t recognize that this is

a serious problem. Most gatekeepers should probably have much more

diffidence concerning their powers of prediction—especially in a job market

in which most job seekers are something more than competent.

Another problem is that the feedback we get, especially when it comes

to social matters, is likely to be highly unrepresentative. Consider the

finding that 94% of university professors believe they’re better-thanaverage

at their jobs (Gilovich 1991, 77). One reason for this may be that we

typically get personal feedback from students who think we were terrific

teachers (or at least who say we were terrific teachers). Seldom will students

go out of their way to make contact with professors they thought were

really mediocre (if for no other reason than, where would they begin?).

The problem of unrepresentative feedback can be made vivid with an

example that is likely familiar to everybody. Think about someone who

employs mildly (or outright) annoying interpersonal strategies, for example,

dominating conversations or name dropping. How likely are you to

tell this person that these behaviors are annoying? Some blunt folk might

always do so. But most of us, probably as a result of some combination of

politeness, pusillanimity, and prudence, let it slide. Of course, we recognize

that this behavior is annoying (or worse), and we might judge the person to

be annoying (or worse). But given the feedback he has received, he might

well go forth into the world confident that he has once again been socially

deft, charming, and deeply impressive. (We are inclined to suggest you

perform a public service. Supply accurate feedback. Call a bore a bore, a

jerk a jerk, a blowhard a blowhard. Just don’t do it to us.)

Even when the feedback we get is representative and shows that our

predictions are mistaken, we will often interpret such feedback in a way

that supports our preconceptions. For example, Gilovich (1983) asked people

who gambled on football games to tape-record their thoughts about

the outcomes of their bets. One might expect the gamblers to remember

their wins and repress their losses. In fact, just the opposite occurred:

[T]hey spent more time discussing their losses than their wins. Furthermore,

the kind of comments made about wins and losses were quite different. The

bettors tended to make ‘‘undoing’’ comments about their losses—comments

to the effect that the outcome would have been different if not for some

anomalous or ‘‘fluke’’ element. . . . In contrast, they tended to make ‘‘bolstering’’

comments about their wins—comments indicating that the outcome

either should have been as it was, or should have been even more

extreme in the same direction. . . . By carefully scrutinizing and explaining away their losses, while accepting their successes at face value, gamblers do

indeed rewrite their personal histories of success and failure. Losses are often

counted, not as losses, but as ‘‘near wins.’’ (Gilovich 1991, 55)

One interesting feature of this common interpretative strategy is that the

subject cannot be accused of ignoring negative evidence. In fact, the subject

is attending more to the negative evidence than to the positive evidence.

It’s just that he interprets the positive evidence as positive, and the

negative evidence as bad luck.