06.06.2013 Views

Theory of Statistics - George Mason University

Theory of Statistics - George Mason University

Theory of Statistics - George Mason University

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

4.2 Bayesian Analysis 327<br />

even if they don’t think <strong>of</strong> Θ as a random variable. In the Bayesian approach<br />

to testing and determining confidence sets, we do think <strong>of</strong> the parameter as a<br />

random variable and so we can make statements about the probability <strong>of</strong> the<br />

parameter taking on certain values.<br />

If the parameter is a random variable, especially if it is a continuous random<br />

variable, point estimation <strong>of</strong> the parameter or a test <strong>of</strong> an hypothesis that<br />

a parameter takes a specific value when the parameter is modeled as a continuous<br />

random variable does not make much sense. The idea <strong>of</strong> a point estimator<br />

that formally minimizes the Bayes risk, however, remains viable. Going beyond<br />

point estimation, the Bayesian paradigm provides a solid theoretical<br />

infrastructure for other aspects <strong>of</strong> statistical inference, such as confidence intervals<br />

and tests <strong>of</strong> hypotheses. The parameter random variable is different in<br />

a fundamental way from the other random variable in the estimation problem:<br />

the parameter random variable is not observable; the other random variable<br />

is — that is, we can observe and record realizations <strong>of</strong> this random variable<br />

<strong>of</strong> interest, and those observations constitute the sample, which is the basis<br />

for the statistical inference.<br />

4.2 Bayesian Analysis<br />

The starting point in ordinary Bayesian inference is the conditional distribution<br />

<strong>of</strong> the observable random variable. (In a frequentist approach, this is just<br />

the distribution — not the “conditional” distribution.)<br />

The prior density represents a probability distribution <strong>of</strong> the parameter<br />

assumed a priori, that is, without the information provided by a random<br />

sample. Bayesian inference depends on the conditional distribution <strong>of</strong> the<br />

parameter, given data from the random variable <strong>of</strong> interest.<br />

4.2.1 Theoretical Underpinnings<br />

The relationships among the conditional, marginal, and joint distributions<br />

can be stated formally in the “Bayes formula”. The simple relationship <strong>of</strong><br />

probabilities <strong>of</strong> events as in equations (1.227) and (1.228) allows us to express<br />

a conditional probability in terms <strong>of</strong> the two marginal probabilities and the<br />

conditional probability with the conditioning reversed;<br />

Pr(A|B) = Pr(B|A)Pr(A)<br />

. (4.11)<br />

Pr(B)<br />

This relationship expresses the basic approach in Bayesian statistical analysis.<br />

Instead <strong>of</strong> probabilities <strong>of</strong> discrete events, however, we wish to utilize<br />

relationships among probability densities.<br />

We consider the random variable X with range X ⊆ IR d and Θ with range<br />

Θ ⊆ IR k . We consider the product space X × Θ together with the product<br />

<strong>Theory</strong> <strong>of</strong> <strong>Statistics</strong> c○2000–2013 James E. Gentle

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!