11.07.2015 Views

statisticalrethinkin..

statisticalrethinkin..

statisticalrethinkin..

SHOW MORE
SHOW LESS
  • No tags were found...

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

78 3. SAMPLING THE IMAGINARYDoes this mean that the model is bad? at depends. e model will always be wrongin some sense, be mis-specified. But whether or not the mis-specification should lead us totry other models will depend upon our specific interests. In this case, if tosses do tend toswitch from W to L and L to W, then each toss will provide less information about the truecoverage of water on the globe. In the long run, even the wrong model we’ve used throughoutthe chapter will converge on the correct proportion. But it will do so more slowly than theposterior distribution may lead us to believe.Rethinking: What does more extreme mean? A common way of measuring deviation of observationfrom model is to count up the tail area that includes the observed data and any more extreme data.Ordinary p-values are an example of such a tail-area probability. When comparing observations todistributions of simulated predictions, as in FIGURE 3.6 and FIGURE 3.7, we might wonder how farout in the tail the observed data must be before we conclude that the model is a poor one. Becausestatistical contexts vary so much, it’s impossible to give a universally useful answer.But more importantly, there are usually very many ways to view data and define “extreme.” Ordinaryp-values view the data in just the way to model expects it, and so provide a very weak form ofmodel checking. For example, the far-right plot in FIGURE 3.6 evaluates model fit in the best way forthe model. Alternative ways of defining “extreme” may provide a more serious challenge to a model.e different definitions of extreme in FIGURE 3.7 can more easily embarrass it.Model fitting remains on objective procedure—everyone and every golem conducts Bayesianupdating in a way that doesn’t depend upon personal preferences. But model checking is inherentlysubjective, and this actually allows it to be quite powerful, since subjective knowledge of an empiricaldomain provides expertise. Expertise in turn allows for imaginative checks of model performance.Since golems have terrible imaginations, we need the freedom to engage our imaginations. In thisway, the objective and subjective work together. E. T. Jaynes (1922–1998) said all of this much moresuccinctly:It would be very nice to have a formal apparatus that gives us some “optimal” wayof recognizing unusual phenomena and inventing new classes of hypotheses thatare most likely to contain the true one; but this remains an art for the creativehuman mind. 523.4. Summaryis chapter introduced the basic procedures for manipulating posterior distributions.Our fundamental tool is samples of parameter values drawn from the posterior distribution.Samples transform problems in integral calculus into problems in data summary. ese samplescan be used to produce intervals, point estimates, as well as posterior predictive checks,as well as other kinds of simulations.Posterior predictive checks combine uncertainty about parameters, as described by theposterior distribution, with uncertainty about outcomes, as described by the assumed likelihoodfunction. ese checks are useful for verifying that your soware worked correctly.ey are also useful for prospecting for ways in which your models is inadequate.Once models become more complex, posterior predictive simulations will be used fora broader range of applications. Even understanding a model oen requires simulating impliedobservations. We’ll keep working with samples from the posterior, to make these tasksas easy and customizable as possible.

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!