11.07.2015 Views

2DkcTXceO

2DkcTXceO

2DkcTXceO

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

N. Cressie 431and conquer” strategy that emphasizes where scientists can put effort intounderstanding the sources of uncertainty and into designing scientific studiesthat control (and perhaps minimize some of) the entropy components.The process Y and the parameters θ are unknown, but the data Z areknown. (Nevertheless, the observed Z is still thought of as one of many possiblethat could have been observed, with a distribution [Z].) At the beginning of allstatistical inference is a step that declares what to condition on, and I proposethat EI follow the path of conditioning on what is known, namely Z. Thenthe conditional probability distribution of all the unknowns given Z is[Y,θ|Z] =[Z, Y, θ]/[Z] =[Z|Y,θ] × [Y |θ] × [θ]/[Z], (38.5)where the first equality is known as Bayes’ Theorem (Bayes, 1763); (38.5) iscalled the posterior distribution, and we call (38.1)–(38.3) a Bayesian hierarchicalmodel (BHM). Notice that [Z] on the right-hand side of (38.5) is anormalizing term that ensures that the posterior distribution integrates (orsums) to 1.There is an asymmetry associated with the role of Y and θ, since(38.2)very clearly emphasises that [Y |θ] is where the “science” resides. It is equallytrue that [Y,θ] =[θ|Y ] × [Y ]. However, probability models for [θ|Y ] and [Y ]do not follow naturally from the way that uncertainties are manifested. Theasymmetry emphasizes that Y is often the first priority for inference. As a consequence,we define the predictive distribution, [Y |Z], which can be obtainedfrom (38.5) by marginalization:[Y |Z] =∫ [Z|Y,θ] × [Y |θ] × [θ]dθ/[Z]. (38.6)Then inference on Y is obtained from (38.6). While (38.5) and (38.6) areconceptually straightforward, in EI we may be trying to evaluate them inglobal spatial or spatio-temporal settings where Z might be on the order ofGb or Tb, and Y might be of a similar order. Thus, HM requires innovativeconditional-probability modeling in (38.1)–(38.3), followed by innovativestatistical computing in (38.5) and (38.6). Leading cases involve spatial data(Cressie, 1993; Banerjee et al., 2004) and spatio-temporal data (Cressie andWikle, 2011). Examples of dynamical spatio-temporal HM are given in Chapter9 of Cressie and Wikle (2011), and we also connect the literature in dataassimilation, ensemble forecasting, blind-source separation, and so forth to theHM paradigm.38.3 Decision-making in the presence of uncertaintyLet Ŷ (Z) be one of many decisions about Y based on Z. Some decisions arebetter than others, which can be quantified through a (non-negative) loss function,L{Y,Ŷ (Z)}. TheBayesexpectedlossisE{L(Y,Ŷ )}, and we minimize

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!