3. Cohen’s conceptual reject-the-norm argument

К оглавлению1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 
17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 
34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 
51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 
68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 
85 86 87 88 89 90 91 92 93 94 

Perhaps the best known response from a philosopher to the HB program

is L. J. Cohen’s article ‘‘Can human irrationality be experimentally demonstrated?’’

(1981). Cohen’s thesis is that it is impossible for empirical

evidence to show that normal, adult humans are irrational. Although

Cohen’s article is more than two decades old, it still reflects the views of at

least some analytic philosophers. For example, Pollock and Cruz briefly

discuss the HB literature and state that their reading ‘‘is pretty much the

same as the assessment of the irrationality literature offered by Jonathan

Cohen’’ (1999, 130 n. 110).

Cohen begins by asking two questions: What would a normative

theory that describes how we ought to reason look like? And what would a descriptive theory that describes our reasoning competence look like?

Cohen’s strategy is to argue that these theories would be identical: A theory

of our reasoning competence just is a theory of how we ought to reason.

Cohen puts the point bluntly. ‘‘[O]rdinary human reasoning—by which I

mean the reasoning of adults who have not been systematically educated

in any branch of logic or probability theory—cannot be held to be faultily

programmed: it sets its own standards’’ (1981, 317). Cohen does not suggest

that normal human adults never reason poorly. But his explanation of

reasoning failures depends essentially on a performance-competence distinction.

A normal adult’s reasoning flaws are performance errors; her

reasoning competence is flawless—and necessarily so.

Cohen’s performance-competence distinction is familiar from linguistics.

The idea in linguistics is that we possess a cognitive system that

stores a rich body of grammatical information. A central goal of linguistics

is to infer linguistic rules from intuitions that cognitive system produces.

So linguists build theories of grammar. Such a theory for English would

accurately capture the grammatical competence of speakers of English. In

order to find out about this competence, we investigate the grammatical

judgments of English speakers (e.g., judgments about whether particular

sentences are grammatical). In order to make such judgments, speakers

must employ not only their grammatical competence, but also various

ancillary cognitive systems, such as memory, attention, and perception. As

a result, some judgments of a speaker might fail to reflect her underlying

grammatical competence because of the failure of one of these ancillary

systems. These are performance errors. Because of performance errors, the

linguist is likely to be faced with a situation in which she cannot construct

a theory because her data—speakers’ judgments—are messy or inconsistent.

As a result, the linguist must construct an idealized grammar for the

language. For Cohen, much of this story can be applied directly to the

theory of reasoning competence: We possess a cognitive system that represents

our reasoning competence. A descriptive theory of reasoning

competence will describe the rules of reasoning that make up our reasoning

competence. The data for such a theory will consist of subjects’

intuitions about particular reasoning problems. For Cohen, ‘‘an intuition

that p is . . . an immediate and untutored inclination, without evidence or

inference, to judge that p’’ (1981, 318). Such intuitions might not reflect a

subject’s reasoning competence, however, because of the potential for

performance errors. As a result, subjects’ intuitions might well be messy or

inconsistent. So the theorist must construct an idealized theory of reasoning

competence.

What about a normative theory of how we ought to reason? Here again,

we start with subjects’ intuitions about particular reasoning problems:

In order to discover what criteria of probability are appropriate for the

evaluation of lay reasoning we have to investigate what judgments of

probability are intuitively acceptable to lay adults and what rational constraints

these judgments are supposed to place on one another. (1981, 319)

Because of the possibility of performance errors, some amount of idealization

will be necessary in constructing a normative theory. According to

Cohen, we achieve an idealized normative theory through a process of

narrow reflective equilibrium: a coherent reconstruction of the subject’s

reasoning principles. Nelson Goodman describes this process as follows:

[R]ules and particular inferences alike are justified by being brought into

agreement with each other. A rule is amended if it yields an inference we are

unwilling to accept; an inference is rejected if it violates a rule we are unwilling

to amend. The process of justification is the delicate one of making

mutual adjustments between rules and accepted inferences; and in the

agreement achieved lies the only justification needed for either. (1965, 67)

An implication of Cohen’s view is that ‘‘Nothing can count as an

error of reasoning among our fellow adults unless even the author of the

error would, under ideal conditions, agree that it is an error’’ (1981, 322).

So it is not possible to empirically demonstrate that people are incompetent

reasoners, and the reason is that it is impossible for people to be

incompetent reasoners (i.e., it is impossible for people’s reasoning competence

to be defective). When we make mistakes, it is due to contingent

barriers like pesky distractions or memory failures that hamper our (in

principle) flawless execution.

Cohen believes that the primary data for our normative theory are subjects’

intuitions—their immediate, untutored inclinations—about reasoning

problems. What justifies this premise? Cohen seems to suggest that this is the

only possibility unless one invokes the standards of a Higher Power.

[I]f you claim no special revelation in matters of logic or probability, you

will have to be content there too to accept the inherent rationality of your

fellow adults. (1981, 321)

One may be tempted to ask: ‘‘how do we know that any intuition of the

relevant kind is veridical?’’ . . . The best that normative theorists can hope for

in this field (and also what they need to achieve), if they do not claim any

special revelation, is that the contents of all relevant intuitions—suitably

sifted or qualified, if necessary—can be made to corroborate one another by being exhibited as the consequences of a consistent and relatively simple set

of rules or axioms that also sanctions other intuitively acceptable, but

previously unremarked, patterns of reasoning. (1981, 322)

Keep in mind, however, that for Cohen, intuitions are immediate and

untutored inclinations to judge, derived without evidence or inference.

There are surely many other possibilities, even if one is inclined to construct

a normative theory on the basis of the judgments of reasoners. For

example, one might construct a normative theory on the basis of people’s

well-considered judgments (as opposed to their immediate judgments).

Cohen, however, rejects such possibilities by insisting that ‘‘[t]he judgments

of everyday reasoning must be evaluated in their own terms and by

their own standards’’ (1981, 320).

What does Cohen say about base rate neglect? He argues that, at best,

this is an example of subjects being ignorant of a mathematical truth.

However, he rejects even this possibility by offering a number of arguments

to the effect that ‘‘it is doubtful whether the subjects have made any

kind of mathematical error at all’’ (1981, 328). Here is one of Cohen’s

arguments:

You are suffering from a disease that, according to your manifest symptoms,

is either A or B. For a variety of demographic reasons disease A happens to

be nineteen times as common as B. The two diseases are equally fatal if

untreated, but it is dangerous to combine the respectively appropriate

treatments. Your physician orders a certain test which, through the operation

of a fairly well understood causal process, always gives a unique diagnosis

in such cases, and this diagnosis has been tried out on equal

numbers of A- and B-patients and is known to be correct on 80% of those

occasions. The tests report that you are suffering from disease B.

Let’s pause here to consider how someone might reason who was not

neglecting base rates: For every hundred patients who have either A or B in

this population, on average five will have B and 95 will have A. Of those

who test positive for B, four will have B (80% of 5) and 19 will have A

(20% of 95). So for the 23 who test positive for B, 4 actually have B. Now

back to Cohen:

Should you nevertheless opt for the treatment appropriate to A, on the

supposition . . . that the probability of your suffering from A is 19/23? Or

should you opt for the treatment appropriate to B, on the supposition

. . . that the probability of your suffering from B is 4/5? It is the former

option that would be the irrational one for you, qua patient, not the

latter. . . . Indeed, on the other view, which is the one espoused in the literature, it would be a waste of time and money even to carry out the tests,

since whatever their results, the base rates would still compel a more than

4/5 probability in favour of disease A. So the literature under criticism is

propagating an analysis that could increase the number of deaths from a

rare disease of this kind. (1981, 329)

This is a stunning line of argument. (As is Cohen’s defense of the gambler’s

fallacy [1981, 327–28].) While Cohen is right that the literature he is

criticizing defends a position that would lead to a greater number of

deaths from (rare) disease B, he recognizes that his own position would

lead to a greater number of total deaths—indeed, four times as many—

from both diseases A and B. Out of 100 people, Cohen’s strategy can be

expected to lead to 20 deaths (1 person dies from B, 19 die from A), while

the other strategy can be expected to lead to 5 deaths (all of those with

disease B). Cohen grants that ‘‘[t]he administrator who wants to secure a

high rate of diagnostic success for his hospital at minimal cost would be

right to seek to maximize just that probability, and therefore to dispense

altogether with the tests.’’ Note also that Cohen admits that in order not

to violate Bayes’ Rule, the subject must ignore the given base rate and

‘‘suppose equal predispositions’’ (1981, 329).

Cohen suggests that base rate neglect is superior to a Bayesian reasoning

strategy because, although base rate neglect will lead to more

overall deaths, it will lead to fewer deaths from a rare disease. This is a real

head-scratcher. Philosophers and their loved ones, after all, get sick too. If

you have the same symptoms as 200 other people and there are two

treatments, one with a survival rate of 95% and the other with a survival

rate of 80%, Cohen’s argument implies that the rational person will

choose the treatment with the lower survival rate. If that’s what rationality

dictates, we don’t want it. Cohen’s optimism is the result of an a priori

attachment that exacts a heavy price so that we may save cognitive face.

But then any conceptual reject-the-norm argument is by its very nature

deeply antithetical to the scientific spirit that animates not just Ameliorative

Psychology but all inquiry about the natural world.

Like with Gigerenzer’s conceptual reject-the-norm argument, Cohen

is defending a rather eccentric normative category. Cohen’s conception of

what it is to be ‘‘rational’’ is distinctly Protagorean—‘‘man is the measure

of all things.’’ But subjects who neglect the base rate might well be ‘‘rational’’

in a Protagorean sense and yet reason extraordinarily badly. In fact, in

chapter 9, we will argue that most subjects who neglect base rates are

reasoning badly. To see this intuitively, consider someone whose behavior makes him a very low risk for HIV, who takes a very reliable HIV test, and

who tests positive. What is an AIDS counselor supposed to tell this patient?

Should the counselor ignore the fact that the subject is low risk and

advise him that there is a 99% chance he has HIV? Or should the counselor

take that fact into account and tell the subject that the chances are

more like 50-50? How to reason about this situation has literally life-anddeath

consequences: ‘‘Former Senator Lawton Chiles of Florida reported

at an AIDS conference in 1987 that of 22 blood donors in Florida who

were notified that they tested HIV-positive with the ELISA test, seven

committed suicide. In the same medical text that reported this tragedy, the

reader is informed that ‘even if the results of both AIDS tests, the ELISA

and WB (Western blot), are positive, the chances are only 50-50 that the

individual is infected’ (Stine, 1996, 333, 338)’’ (Gigerenzer, Hoffrage, and

Ebert 1998). Now consider the AIDS counselor who ignores the base rate

and tells his clients that they have a 99% chance of having HIV, when in

fact only about 50% of his clients who test positive have HIV. Ignoring the

base rate in this situation may well be ‘‘rational’’ in Cohen’s sense. But it

would be lousy reasoning—the sort of reasoning that would quite properly

haunt one the rest of one’s days.

Cohen is surely right to suggest that there is some sort of a distinction

between our performance and the cognitive capacities that make it possible.

But it is not the distinction seen in linguistics. It is the distinction

familiar from any activity that requires skill and dedication. People who

have the competence to sing Weber’s desperate aria ‘‘Wo berg ich mich’’

might on occasion perform it badly, but even at our best, most of us really

can’t perform it at all. In this case, no one is inclined to suppose that

everyone’s capacities are similar or that everyone’s contingent collection of

capacities just is the measure of excellence. Some people sing better than

others. Further, everyone recognizes that there are differences in people’s

native mathematical and logical abilities. And both of these are crucial to

reasoning competence. So surely it is not too much of a stretch to suppose

that some people reason better than others. But Cohen denies that different

normal adults might have different reasoning competences; he

insists that any apparently different intuitions about particular reasoning

problems must always be resolvable.

No doubt two different people, or the same people on two different occasions,

may sometimes have apparently conflicting intuitions. But such an apparent

conflict always demands resolution. The people involved might come to

recognize some tacit misunderstanding about the terms of the problem, so that there is no real conflict; or they might repudiate a previously robust

intuition, perhaps as a result of becoming aware that an otherwise preferred

solution has unacceptable implications; or they might conclude that different

idiolects or conceptions of deducibility are at issue. (1981, 319)

This is an empirical speculation, and we will now turn to some fascinating

evidence that suggests that it is false.

Perhaps the best known response from a philosopher to the HB program

is L. J. Cohen’s article ‘‘Can human irrationality be experimentally demonstrated?’’

(1981). Cohen’s thesis is that it is impossible for empirical

evidence to show that normal, adult humans are irrational. Although

Cohen’s article is more than two decades old, it still reflects the views of at

least some analytic philosophers. For example, Pollock and Cruz briefly

discuss the HB literature and state that their reading ‘‘is pretty much the

same as the assessment of the irrationality literature offered by Jonathan

Cohen’’ (1999, 130 n. 110).

Cohen begins by asking two questions: What would a normative

theory that describes how we ought to reason look like? And what would a descriptive theory that describes our reasoning competence look like?

Cohen’s strategy is to argue that these theories would be identical: A theory

of our reasoning competence just is a theory of how we ought to reason.

Cohen puts the point bluntly. ‘‘[O]rdinary human reasoning—by which I

mean the reasoning of adults who have not been systematically educated

in any branch of logic or probability theory—cannot be held to be faultily

programmed: it sets its own standards’’ (1981, 317). Cohen does not suggest

that normal human adults never reason poorly. But his explanation of

reasoning failures depends essentially on a performance-competence distinction.

A normal adult’s reasoning flaws are performance errors; her

reasoning competence is flawless—and necessarily so.

Cohen’s performance-competence distinction is familiar from linguistics.

The idea in linguistics is that we possess a cognitive system that

stores a rich body of grammatical information. A central goal of linguistics

is to infer linguistic rules from intuitions that cognitive system produces.

So linguists build theories of grammar. Such a theory for English would

accurately capture the grammatical competence of speakers of English. In

order to find out about this competence, we investigate the grammatical

judgments of English speakers (e.g., judgments about whether particular

sentences are grammatical). In order to make such judgments, speakers

must employ not only their grammatical competence, but also various

ancillary cognitive systems, such as memory, attention, and perception. As

a result, some judgments of a speaker might fail to reflect her underlying

grammatical competence because of the failure of one of these ancillary

systems. These are performance errors. Because of performance errors, the

linguist is likely to be faced with a situation in which she cannot construct

a theory because her data—speakers’ judgments—are messy or inconsistent.

As a result, the linguist must construct an idealized grammar for the

language. For Cohen, much of this story can be applied directly to the

theory of reasoning competence: We possess a cognitive system that represents

our reasoning competence. A descriptive theory of reasoning

competence will describe the rules of reasoning that make up our reasoning

competence. The data for such a theory will consist of subjects’

intuitions about particular reasoning problems. For Cohen, ‘‘an intuition

that p is . . . an immediate and untutored inclination, without evidence or

inference, to judge that p’’ (1981, 318). Such intuitions might not reflect a

subject’s reasoning competence, however, because of the potential for

performance errors. As a result, subjects’ intuitions might well be messy or

inconsistent. So the theorist must construct an idealized theory of reasoning

competence.

What about a normative theory of how we ought to reason? Here again,

we start with subjects’ intuitions about particular reasoning problems:

In order to discover what criteria of probability are appropriate for the

evaluation of lay reasoning we have to investigate what judgments of

probability are intuitively acceptable to lay adults and what rational constraints

these judgments are supposed to place on one another. (1981, 319)

Because of the possibility of performance errors, some amount of idealization

will be necessary in constructing a normative theory. According to

Cohen, we achieve an idealized normative theory through a process of

narrow reflective equilibrium: a coherent reconstruction of the subject’s

reasoning principles. Nelson Goodman describes this process as follows:

[R]ules and particular inferences alike are justified by being brought into

agreement with each other. A rule is amended if it yields an inference we are

unwilling to accept; an inference is rejected if it violates a rule we are unwilling

to amend. The process of justification is the delicate one of making

mutual adjustments between rules and accepted inferences; and in the

agreement achieved lies the only justification needed for either. (1965, 67)

An implication of Cohen’s view is that ‘‘Nothing can count as an

error of reasoning among our fellow adults unless even the author of the

error would, under ideal conditions, agree that it is an error’’ (1981, 322).

So it is not possible to empirically demonstrate that people are incompetent

reasoners, and the reason is that it is impossible for people to be

incompetent reasoners (i.e., it is impossible for people’s reasoning competence

to be defective). When we make mistakes, it is due to contingent

barriers like pesky distractions or memory failures that hamper our (in

principle) flawless execution.

Cohen believes that the primary data for our normative theory are subjects’

intuitions—their immediate, untutored inclinations—about reasoning

problems. What justifies this premise? Cohen seems to suggest that this is the

only possibility unless one invokes the standards of a Higher Power.

[I]f you claim no special revelation in matters of logic or probability, you

will have to be content there too to accept the inherent rationality of your

fellow adults. (1981, 321)

One may be tempted to ask: ‘‘how do we know that any intuition of the

relevant kind is veridical?’’ . . . The best that normative theorists can hope for

in this field (and also what they need to achieve), if they do not claim any

special revelation, is that the contents of all relevant intuitions—suitably

sifted or qualified, if necessary—can be made to corroborate one another by being exhibited as the consequences of a consistent and relatively simple set

of rules or axioms that also sanctions other intuitively acceptable, but

previously unremarked, patterns of reasoning. (1981, 322)

Keep in mind, however, that for Cohen, intuitions are immediate and

untutored inclinations to judge, derived without evidence or inference.

There are surely many other possibilities, even if one is inclined to construct

a normative theory on the basis of the judgments of reasoners. For

example, one might construct a normative theory on the basis of people’s

well-considered judgments (as opposed to their immediate judgments).

Cohen, however, rejects such possibilities by insisting that ‘‘[t]he judgments

of everyday reasoning must be evaluated in their own terms and by

their own standards’’ (1981, 320).

What does Cohen say about base rate neglect? He argues that, at best,

this is an example of subjects being ignorant of a mathematical truth.

However, he rejects even this possibility by offering a number of arguments

to the effect that ‘‘it is doubtful whether the subjects have made any

kind of mathematical error at all’’ (1981, 328). Here is one of Cohen’s

arguments:

You are suffering from a disease that, according to your manifest symptoms,

is either A or B. For a variety of demographic reasons disease A happens to

be nineteen times as common as B. The two diseases are equally fatal if

untreated, but it is dangerous to combine the respectively appropriate

treatments. Your physician orders a certain test which, through the operation

of a fairly well understood causal process, always gives a unique diagnosis

in such cases, and this diagnosis has been tried out on equal

numbers of A- and B-patients and is known to be correct on 80% of those

occasions. The tests report that you are suffering from disease B.

Let’s pause here to consider how someone might reason who was not

neglecting base rates: For every hundred patients who have either A or B in

this population, on average five will have B and 95 will have A. Of those

who test positive for B, four will have B (80% of 5) and 19 will have A

(20% of 95). So for the 23 who test positive for B, 4 actually have B. Now

back to Cohen:

Should you nevertheless opt for the treatment appropriate to A, on the

supposition . . . that the probability of your suffering from A is 19/23? Or

should you opt for the treatment appropriate to B, on the supposition

. . . that the probability of your suffering from B is 4/5? It is the former

option that would be the irrational one for you, qua patient, not the

latter. . . . Indeed, on the other view, which is the one espoused in the literature, it would be a waste of time and money even to carry out the tests,

since whatever their results, the base rates would still compel a more than

4/5 probability in favour of disease A. So the literature under criticism is

propagating an analysis that could increase the number of deaths from a

rare disease of this kind. (1981, 329)

This is a stunning line of argument. (As is Cohen’s defense of the gambler’s

fallacy [1981, 327–28].) While Cohen is right that the literature he is

criticizing defends a position that would lead to a greater number of

deaths from (rare) disease B, he recognizes that his own position would

lead to a greater number of total deaths—indeed, four times as many—

from both diseases A and B. Out of 100 people, Cohen’s strategy can be

expected to lead to 20 deaths (1 person dies from B, 19 die from A), while

the other strategy can be expected to lead to 5 deaths (all of those with

disease B). Cohen grants that ‘‘[t]he administrator who wants to secure a

high rate of diagnostic success for his hospital at minimal cost would be

right to seek to maximize just that probability, and therefore to dispense

altogether with the tests.’’ Note also that Cohen admits that in order not

to violate Bayes’ Rule, the subject must ignore the given base rate and

‘‘suppose equal predispositions’’ (1981, 329).

Cohen suggests that base rate neglect is superior to a Bayesian reasoning

strategy because, although base rate neglect will lead to more

overall deaths, it will lead to fewer deaths from a rare disease. This is a real

head-scratcher. Philosophers and their loved ones, after all, get sick too. If

you have the same symptoms as 200 other people and there are two

treatments, one with a survival rate of 95% and the other with a survival

rate of 80%, Cohen’s argument implies that the rational person will

choose the treatment with the lower survival rate. If that’s what rationality

dictates, we don’t want it. Cohen’s optimism is the result of an a priori

attachment that exacts a heavy price so that we may save cognitive face.

But then any conceptual reject-the-norm argument is by its very nature

deeply antithetical to the scientific spirit that animates not just Ameliorative

Psychology but all inquiry about the natural world.

Like with Gigerenzer’s conceptual reject-the-norm argument, Cohen

is defending a rather eccentric normative category. Cohen’s conception of

what it is to be ‘‘rational’’ is distinctly Protagorean—‘‘man is the measure

of all things.’’ But subjects who neglect the base rate might well be ‘‘rational’’

in a Protagorean sense and yet reason extraordinarily badly. In fact, in

chapter 9, we will argue that most subjects who neglect base rates are

reasoning badly. To see this intuitively, consider someone whose behavior makes him a very low risk for HIV, who takes a very reliable HIV test, and

who tests positive. What is an AIDS counselor supposed to tell this patient?

Should the counselor ignore the fact that the subject is low risk and

advise him that there is a 99% chance he has HIV? Or should the counselor

take that fact into account and tell the subject that the chances are

more like 50-50? How to reason about this situation has literally life-anddeath

consequences: ‘‘Former Senator Lawton Chiles of Florida reported

at an AIDS conference in 1987 that of 22 blood donors in Florida who

were notified that they tested HIV-positive with the ELISA test, seven

committed suicide. In the same medical text that reported this tragedy, the

reader is informed that ‘even if the results of both AIDS tests, the ELISA

and WB (Western blot), are positive, the chances are only 50-50 that the

individual is infected’ (Stine, 1996, 333, 338)’’ (Gigerenzer, Hoffrage, and

Ebert 1998). Now consider the AIDS counselor who ignores the base rate

and tells his clients that they have a 99% chance of having HIV, when in

fact only about 50% of his clients who test positive have HIV. Ignoring the

base rate in this situation may well be ‘‘rational’’ in Cohen’s sense. But it

would be lousy reasoning—the sort of reasoning that would quite properly

haunt one the rest of one’s days.

Cohen is surely right to suggest that there is some sort of a distinction

between our performance and the cognitive capacities that make it possible.

But it is not the distinction seen in linguistics. It is the distinction

familiar from any activity that requires skill and dedication. People who

have the competence to sing Weber’s desperate aria ‘‘Wo berg ich mich’’

might on occasion perform it badly, but even at our best, most of us really

can’t perform it at all. In this case, no one is inclined to suppose that

everyone’s capacities are similar or that everyone’s contingent collection of

capacities just is the measure of excellence. Some people sing better than

others. Further, everyone recognizes that there are differences in people’s

native mathematical and logical abilities. And both of these are crucial to

reasoning competence. So surely it is not too much of a stretch to suppose

that some people reason better than others. But Cohen denies that different

normal adults might have different reasoning competences; he

insists that any apparently different intuitions about particular reasoning

problems must always be resolvable.

No doubt two different people, or the same people on two different occasions,

may sometimes have apparently conflicting intuitions. But such an apparent

conflict always demands resolution. The people involved might come to

recognize some tacit misunderstanding about the terms of the problem, so that there is no real conflict; or they might repudiate a previously robust

intuition, perhaps as a result of becoming aware that an otherwise preferred

solution has unacceptable implications; or they might conclude that different

idiolects or conceptions of deducibility are at issue. (1981, 319)

This is an empirical speculation, and we will now turn to some fascinating

evidence that suggests that it is false.