22.03.2013 Views

et al.

et al.

et al.

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

758<br />

BOOKS ET AL.<br />

STATISTICS<br />

Mr. Bayes Goes to Washington<br />

Sam Wang 1 and Benjamin C. Campbell 2<br />

One day before the 2012 U.S. presidenti<strong>al</strong><br />

election, former Reagan speechwriter<br />

Peggy Noonan wrote that<br />

“nobody knows anything” about who would<br />

win, asserting that Republican candidate Mitt<br />

Romney’s supporters had the greater passion<br />

and enthusiasm ( 1). From a similarly datafree<br />

remove, columnist George Will predicted<br />

a Romney elector<strong>al</strong> landslide. MSNBC’s Joe<br />

Scarborough said “it could go either way …<br />

anybody that thinks that this race is anything<br />

but a tossup right now … should be kept away<br />

from typewriters, computers, laptops, and<br />

microphones, because they’re jokes.” ( 2)<br />

In the end, these pundits were the ones<br />

whose opinions proved dispensable. They<br />

were unable to d<strong>et</strong>ect a plain fact: based on<br />

public opinion polls with collectively excellent<br />

track records, President Obama had an<br />

advantage of 3 to 4 percentage points for<br />

nearly the entire campaign season. However,<br />

the world of politic<strong>al</strong> punditry measures<br />

success not by accuracy but by readership<br />

and viewership. And so it came to pass that<br />

legions of commentators expressed tot<strong>al</strong> confi<br />

dence—and were wrong.<br />

Beating the pundits has been possible<br />

since at least 2004, when one of us was<br />

among the fi rst to statistic<strong>al</strong>ly aggregate polls<br />

( 3). In 2008, Nate Silver emerged as a poster<br />

child for aggregation, armed<br />

with a degree in economics, a<br />

love of numbers, and a profession<strong>al</strong><br />

track record in an<strong>al</strong>yzing<br />

baseb<strong>al</strong>l performance and<br />

financi<strong>al</strong> data. He enlivened<br />

a mostly suspenseless presidenti<strong>al</strong><br />

race, providing timely<br />

quantitative an<strong>al</strong>ysis and color<br />

commentary on his website<br />

FiveThirtyEight, which<br />

became highly popular and<br />

was snapped up by the New<br />

York Times ( 4). His fame rose<br />

further in 2012, when he and other aggregators<br />

and modelers used hardnosed an<strong>al</strong>ysis<br />

( 3– 6) to silence skeptics.<br />

Now Silver has written The Sign<strong>al</strong> and the<br />

Noise, a book that addresses predictions not<br />

1 Department of Molecular Biology and Neuroscience Institute,<br />

Princ<strong>et</strong>on University, Princ<strong>et</strong>on, NJ 08544, USA.<br />

E-mail: sswang@princ<strong>et</strong>on.edu 2 Laboratory of Biologic<strong>al</strong><br />

Modeling, Rockefeller University, 1230 York Avenue, New<br />

York, NY 10065, USA. E-mail: bcampbell@rockefeller.edu<br />

The Sign<strong>al</strong> and the Noise<br />

Why So Many Predictions<br />

Fail—But Some Don’t /<br />

The Art and Science of<br />

Prediction<br />

by Nate Silver<br />

Penguin, New York, 2012.<br />

542 pp. $27.95, C$29.50.<br />

ISBN 9781594204111.<br />

Allen Lane, London. £25.<br />

ISBN 9781846147524.<br />

97.5 90 80 80 90 97.5<br />

Probability Obama Probability Romney<br />

V<strong>al</strong>idated by the outcome. The Princ<strong>et</strong>on Election<br />

Consortium’s fi n<strong>al</strong> elector<strong>al</strong> college predictions for<br />

November 2012. (States are sized according to their<br />

share of elector<strong>al</strong> votes.)<br />

just in politics but in <strong>al</strong>l aspects of modern<br />

life, with the eye of a hobbyist and a sense<br />

of fun. Freed from the word limits of blog<br />

essays, the book is a meandering, nerd’s-eye<br />

view of what principles, if any, are common<br />

to good forecasting in daily life, leisure activity,<br />

and science.<br />

We use predictions to guide our future<br />

actions, from planning weekend outings to<br />

taking care of our he<strong>al</strong>th, but most people<br />

have no idea how scientifi c predictions are<br />

made. This book is for them. Silver introduces<br />

some of the concepts behind data modeling,<br />

including probability, Bayes-<br />

ian inference, and uncertainty.<br />

He takes lengthy looks at topics<br />

ranging from fl u epidemics<br />

to the 1996 chess-playing triumphs<br />

of Deep Blue.<br />

A reappearing theme in<br />

The Sign<strong>al</strong> and the Noise<br />

is Bayesian reasoning, an<br />

approach that has swept the<br />

sciences. Probability had been<br />

convention<strong>al</strong>ly interpr<strong>et</strong>ed as<br />

meaning the true likelihood of<br />

an event—for instance, how<br />

often the tot<strong>al</strong> of two rolled dice will add up<br />

to seven. Such a “frequentist” point of view<br />

has in many cases given way to an approach<br />

pioneered by Reverend Thomas Bayes in the<br />

18th century, which emphasizes that probability<br />

can only be interpr<strong>et</strong>ed in terms of the<br />

hypotheses that preceded the measurement.<br />

Although Silver asserts that Bayesian<br />

politic<strong>al</strong> forecasting has more in common<br />

with poker than with hard sciences such as<br />

15 FEBRUARY 2013 VOL 339 SCIENCE www.sciencemag.org<br />

Published by AAAS<br />

physics and biology, these topics <strong>al</strong>l use the<br />

same mathematic<strong>al</strong> toolkit. Large-sc<strong>al</strong>e physics<br />

collaborations depend on sensitive models<br />

to predict the probabilistic decay rates of particles,<br />

looking for outliers that might represent<br />

sign<strong>al</strong>s in the noise and hence discoveries. In<br />

our fi eld, many neuroscientists have begun to<br />

view the brain as a prediction machine ( 7). We<br />

perceive the world around us by making inferences<br />

from noisy and incompl<strong>et</strong>e data. To do<br />

so, the brain must form a model of its environment—a<br />

s<strong>et</strong> of “priors” learned over a lif<strong>et</strong>ime<br />

that is used to interpr<strong>et</strong> incoming data.<br />

This Bayesian machine continu<strong>al</strong>ly updates<br />

its priors to correspond to its environment.<br />

Through this process, our brains spend many<br />

years honing appropriate priors for the complex<br />

tasks that we perform effortlessly.<br />

Silver gives a well-known equation for<br />

how to take into account the Bayesian prior<br />

but doesn’t show where it comes from. Readers<br />

wanting a deeper explanation of Bayes’s<br />

rule might consult another source such as<br />

B<strong>et</strong>terExplained.com ( 8), which teaches the<br />

subject by using e-mail spam fi ltering as an<br />

example. Silver’s chosen anecdotes include<br />

the classic example of mammogram interpr<strong>et</strong>ation—but<br />

<strong>al</strong>so how to interpr<strong>et</strong> that unfamiliar<br />

underwear that just showed up in your<br />

partner’s dresser drawer.<br />

At times Silver writes as if the cure for bad<br />

modeling can be reduced to “more Bayes.”<br />

Such a prescription does not do justice to the<br />

historic controversies surrounding interpr<strong>et</strong>ations<br />

of probability. A beginner might come<br />

away from this book believing that an earlier<br />

generation of frequentists were simply ignorant.<br />

In a cartoonish account, Silver lobs a<br />

broadside at a monument<strong>al</strong> fi gure in statistics,<br />

Ron<strong>al</strong>d A. Fisher, who late in life argued<br />

against the idea that smoking causes cancer—<br />

and who coined “Bayesian” as a derogatory<br />

term. Silver suggests that Fisher’s aversion to<br />

Bayes caused him to err. In fact, the re<strong>al</strong> problem<br />

was that Fisher was a smoker ( 9). Fisher’s<br />

prior beliefs prevented him from accepting<br />

epidemiologic<strong>al</strong> and biologic<strong>al</strong> evidence, an<br />

erroneous prior if ever there was one.<br />

Our biggest criticism of the book is that<br />

<strong>al</strong>though statistics and Bayesian inference are<br />

powerful ideas, they are not a cure-<strong>al</strong>l. In his<br />

enthusiasm for the good Reverend, Silver has<br />

stuffed a fair bit into the same Procrustean<br />

bed. Silver uses the old fox-hedgehog an<strong>al</strong>ogy,<br />

saying that foxes (including himself) use<br />

many ideas, whereas hedgehogs focus on one<br />

subject only. But here he is a hedgehog with<br />

one big idea: statistics.<br />

However, Bayesian reasoning works only<br />

if the prior is adapted for the task. According<br />

to Silver, many of today’s “h<strong>al</strong>f-baked<br />

CREDIT: COURTESY SAM WANG/PRINCETON ELECTION CONSORTIUM<br />

on February 14, 2013<br />

www.sciencemag.org<br />

Downloaded from

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!