1.3. Random linear models
К оглавлению1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 1617 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33
34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50
51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67
68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84
85 86 87 88 89 90 91 92 93 94
Dawes and Corrigan (1974) took five bootstrapping experiments and for
each one constructed a random linear model. Random linear models do
not pretend to assign optimum weights to variables. Instead, random weights
are assigned—with one important caveat: All the cues are defined so they
are positively correlated with the target property. They found that the random
linear models were as reliable as the proper models and more reliable
than human experts. Recall we said that there was an SPR finding that was
denied by a well-known philosopher of psychology. This is it. This philosopher
is not alone. Dawes has described one dominant reaction to the
success of random linear models: ‘‘[M]any people didn’t believe them—
until they tested out random . . . models on their own data sets’’ (Dawes
1988, 209, n. 17).
The resistance to this finding is understandable (though, as we shall
later argue, misguided). It is very natural to suppose that people who make
predictions are in some sense ‘‘calculating’’ a suboptimal formula. (Of
course, the idea isn’t that the person explicitly calculates a complex formula
in order to make a prediction; rather, the idea is that there will be an
improper formula that simulates the person’s weighing of the various lines
of evidence in making some prediction.) Since we can’t calculate in our
heads the optimum weights to attach to the relevant cues, it’s understandable
that proper models outperform humans. This picture of humans
‘‘calculating’’ suboptimal formulas, of implicitly using improper models,
The Amazing Success of Statistical Prediction Rules 29
also fits with the optimistic explanation of the bootstrapping effect. A
bootstrapping model approximates the suboptimal formula a person
uses—but the bootstrapping model doesn’t fall victim to performance
errors to which humans are prone. So far, so good. But how are we to
understand random linear models outperforming expert humans? After
all, if experts are calculating some sort of suboptimal formula, how could
they be defeated by a formula that uses weights that are both suboptimal
and random? Surely we must do better than linear models that assign just
any old weights at all. But alas, we do not. Without a plausible explanation
for this apparent anomaly, our first reaction (and perhaps even our wellconsidered
reaction) may be to refuse to believe this could be true.
Dawes and Corrigan (1974) took five bootstrapping experiments and for
each one constructed a random linear model. Random linear models do
not pretend to assign optimum weights to variables. Instead, random weights
are assigned—with one important caveat: All the cues are defined so they
are positively correlated with the target property. They found that the random
linear models were as reliable as the proper models and more reliable
than human experts. Recall we said that there was an SPR finding that was
denied by a well-known philosopher of psychology. This is it. This philosopher
is not alone. Dawes has described one dominant reaction to the
success of random linear models: ‘‘[M]any people didn’t believe them—
until they tested out random . . . models on their own data sets’’ (Dawes
1988, 209, n. 17).
The resistance to this finding is understandable (though, as we shall
later argue, misguided). It is very natural to suppose that people who make
predictions are in some sense ‘‘calculating’’ a suboptimal formula. (Of
course, the idea isn’t that the person explicitly calculates a complex formula
in order to make a prediction; rather, the idea is that there will be an
improper formula that simulates the person’s weighing of the various lines
of evidence in making some prediction.) Since we can’t calculate in our
heads the optimum weights to attach to the relevant cues, it’s understandable
that proper models outperform humans. This picture of humans
‘‘calculating’’ suboptimal formulas, of implicitly using improper models,
The Amazing Success of Statistical Prediction Rules 29
also fits with the optimistic explanation of the bootstrapping effect. A
bootstrapping model approximates the suboptimal formula a person
uses—but the bootstrapping model doesn’t fall victim to performance
errors to which humans are prone. So far, so good. But how are we to
understand random linear models outperforming expert humans? After
all, if experts are calculating some sort of suboptimal formula, how could
they be defeated by a formula that uses weights that are both suboptimal
and random? Surely we must do better than linear models that assign just
any old weights at all. But alas, we do not. Without a plausible explanation
for this apparent anomaly, our first reaction (and perhaps even our wellconsidered
reaction) may be to refuse to believe this could be true.