11.07.2015 Views

2DkcTXceO

2DkcTXceO

2DkcTXceO

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

430 Environmental informaticsHastie, Tibshirani, and Friedman (Hastie et al., 2009) shows the fecundity ofestablishing and developing such links. EI is a young discipline, and I wouldlike to see it develop in this modern and powerful way, with uncertainty quantificationthrough Statistics at its core.In what follows, I shall develop a framework that is fundamentally aboutenvironmental data and the processes that produced them. I shall be particularlyconcerned with big, incomplete, noisy datasets generated by processesthat may be some combination of non-linear, multi-scale, non-Gaussian, multivariate,and spatio-temporal. I shall account for all the known uncertaintiescoherently using hierarchical statistical modeling, or HM (Berliner, 1996),which is based on a series of conditional-probability models. Finally, throughloss functions that assign penalties as a function of how far away an estimateis from its estimand, I shall use a decision-theoretic framework (Berger, 1985)to give environmental policy-makers a way to make rational decisions in thepresence of uncertainty, based on competing risks (i.e., probabilities).38.2 Hierarchical statistical modelingThe building blocks of HM are the data model, the (scientific) process model,and the parameter model. If Z represents the data, Y represents the process,and θ represents the parameters (e.g., measurement-error variance andreaction-diffusion coefficients), then the data model isthe process model isand the parameter model is[Z|Y,θ], (38.1)[Y |θ], (38.2)[θ], (38.3)where [A|B,C] is generic notation for the conditional-probability distributionof the random quantity A given B and C.Astatisticalapproachrepresentstheuncertaintiescoherentlythroughthejoint-probability distribution, [Z, Y, θ]. Using the building blocks (38.1)–(38.3),we can write[Z, Y, θ] =[Z|Y,θ] × [Y |θ] × [θ]. (38.4)The definition of entropy of a random quantity A is E(ln[A]) By re-writing(38.4) asE(ln[Z, Y, θ]) = E(ln[Z|Y,θ]) + E(ln[Y |θ]) + E(ln[θ]),we can see that the joint entropy can be partitioned into data-model entropy,process-model entropy, and parameter-model entropy. This results in a “divide

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!