06.06.2013 Views

Theory of Statistics - George Mason University

Theory of Statistics - George Mason University

Theory of Statistics - George Mason University

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

332 4 Bayesian Inference<br />

3. Identify the joint distribution; if densities exist, it is<br />

fX,Θ(x, θ) = fX|θ(x)fΘ(θ). (4.16)<br />

4. Determine the marginal distribution <strong>of</strong> the observable; if densities exist,<br />

it is<br />

<br />

fX(x) = fX,Θ(x, θ)dθ. (4.17)<br />

Θ<br />

This marginal distribution is also called the prior predictive distribution.<br />

In slightly different notation, we can also write it as<br />

<br />

fX(x) = fX|θ(x)fΘ(θ)dθ.<br />

Θ<br />

5. Determine the posterior conditional distribution <strong>of</strong> the parameter given<br />

the observable random variable. If densities exist, it is<br />

fΘ|x(θ) = fX,Θ(x, θ)/fX(x). (4.18)<br />

This is the PDF <strong>of</strong> the distribution QH in equation (4.2), which is <strong>of</strong>ten<br />

just called the “posterior”. The posterior conditional distribution is then<br />

the basis for whatever decisions are to be made.<br />

6. Assess the posterior conditional distribution in the light <strong>of</strong> prior beliefs.<br />

This is called a sensitivity analysis. Repeat the steps above as appropriate.<br />

These first steps in a Bayesian analysis involve identifying the components<br />

in the equation<br />

fX|θ(x)fΘ(θ) = fX,Θ(x, θ) = fΘ|x(θ)fX(x). (4.19)<br />

Although we have written the PDFs above in terms <strong>of</strong> single random variables<br />

(any <strong>of</strong> which <strong>of</strong> course could be vectors), in applications we assume we<br />

have multiple observations on X. In place <strong>of</strong> fX|θ(x|θ) in (4.14), for example,<br />

we would have the joint density <strong>of</strong> the iid random variables X1, . . ., Xn, or<br />

fX|θ(xi|θ). The other PDFs would be similarly modified.<br />

Given a posterior based on the random sample X1, . . ., Xn, we can form<br />

a posterior predictive distribution for Xn+1, . . ., Xn+k:<br />

fXn+1,...,Xn+k|x1,...,xn (xn+1, . . ., xn+k) =<br />

<br />

Θ fX,Θ(xn+i, θ)dFΘ|x1,...,xn (θ).<br />

(4.20)<br />

Rather than determining the densities in equations (4.14) through (4.18)<br />

it is generally sufficient to determine kernel functions. That is, we write the<br />

densities as<br />

fD(z) ∝ g(z). (4.21)<br />

<strong>Theory</strong> <strong>of</strong> <strong>Statistics</strong> c○2000–2013 James E. Gentle

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!