02.03.2013 Views

Thinking and Deciding

Thinking and Deciding

Thinking and Deciding

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

EVALUATING PROBABILITY JUDGMENTS 119<br />

say there is a 50% probability, <strong>and</strong> so forth. If my judgments are perfectly calibrated,<br />

in the long run, these kinds of proportions will match exactly.<br />

Note that my judgments can be coherent without being calibrated. For example,<br />

I can say that the probability of heads is .90 <strong>and</strong> the probability of tails is .10. These<br />

two are consistent with each other, but not with the facts. If my judgments are perfectly<br />

calibrated, however, they must also be coherent, for whenever I say that the<br />

probability of rain is 75%, I must also say that the probability of no rain is 25%.<br />

Good calibration helps in making decisions. If the weather forecast is so poorly<br />

calibrated that the probability of rain is really 1.00 when the forecast says .25, I will<br />

get wet, if I regard .25 as low enough so as not to require an umbrella.<br />

Calibration is a criterion for evaluation of probability judgments, not a method<br />

for making them. Calibration serves as a criterion in hindsight, after we know what<br />

happened. In principle, calibration could be assessed for judgments based on frequencies<br />

or on the logical view, but this would be superfluous. If the assumptions<br />

going into the judgments were correct, calibration would have to be perfect, <strong>and</strong> we<br />

would know what the probability judgments were without asking a judge.<br />

It may seem that calibration is related to the frequency view of probability. In<br />

a way it is. Calibration certainly captures the idea that we want our probability<br />

judgments to correspond to something about the real world, which is perhaps the<br />

major idea behind the frequency view. By the personal view, however, a judgment<br />

can be justifiable even before the frequency data are in. For the frequency view,<br />

the judgment is meaningless unless the data are in. In addition, a personalist would<br />

argue that calibration, though always desirable, is often impossible to assess. When<br />

someone predicts the probability of a nuclear-power plant explosion, there are simply<br />

not enough data to assess calibration. 3<br />

Scoring rules<br />

Coherence does not take into account what actually happens. A judge can make<br />

completely coherent judgments <strong>and</strong> still get all of the probabilities exactly backward.<br />

(When such a judge said .80, the observed probability was .20; when the judge said<br />

1.00, the event never happened.) Moreover, the idea of calibration does not tell us<br />

how to measure degree of miscalibration. A set of probability judgments is either<br />

calibrated or not; however, degrees of error surely matter.<br />

More important, calibration ignores the information provided by a judgment. A<br />

weather forecaster’s predictions would be perfectly calibrated if they stated that the<br />

probability of rain was .30 every day, provided that it does actually rain on 30% of<br />

the days. Such a forecast would also be coherent if the forecaster added that the<br />

probability of no rain was .70. These forecasts would be useless, however, because<br />

they do not distinguish days when it rains from days when it does not rain.<br />

3 It is an important empirical question whether we can assess a judge’s calibration in general,sothatwe<br />

could draw conclusions about the judge’s calibration in one area from calibration in another. The answer<br />

to this question, however, has nothing to do with the question of whether judgments are in principle<br />

justifiable in the absence of frequency data.

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!