01.06.2013 Views

Statistical Methods in Medical Research 4ed

Statistical Methods in Medical Research 4ed

Statistical Methods in Medical Research 4ed

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

distributed. The methods developed for samples from a normal distribution<br />

therefore provide a reliable approximation for samples from non-normal distributions,<br />

unless the departure from normality is severe or the sample size is very<br />

small. The same is true <strong>in</strong> Bayesian <strong>in</strong>ference, and we shall concentrate here on<br />

methods appropriate for samples from normal distributions.<br />

Figure 4.1 describes the likelihood function for a s<strong>in</strong>gle observation x from a<br />

normal distribution with unit variance, N(m, 1). It is a function of m which takes<br />

the shape of a normal curve with mean x and unit variance. This result can<br />

immediately be extended to give the likelihood from a sample mean. Suppose<br />

that x is the mean of a sample of size n from a normal distribution N(m, s2 ).<br />

From §4.2, we know that x is distributed as N(m, s2 =n), and the likelihood<br />

function is therefore a normal curve N(x, s2 =n).<br />

Suppose now that m follows a normal prior distribution N(m0, s2 0 ). Then,<br />

application of Bayes' theorem shows that the posterior distribution of m is<br />

N x ‡ m0s2 =ns2 0<br />

1 ‡ s2 =ns2 s<br />

,<br />

0<br />

2 =n<br />

1 ‡ s2 =ns2 : …6:1†<br />

0<br />

The mean of this distribution can be written <strong>in</strong> the form<br />

x n<br />

s2 ‡ m0 n 1<br />

‡<br />

s2 s<br />

1<br />

2 0<br />

6.2 Bayesian <strong>in</strong>ference for a mean 169<br />

s 2 0<br />

, …6:2†<br />

which is a weighted mean of the observed mean x and the prior mean m0, the<br />

weights be<strong>in</strong>g <strong>in</strong>versely proportional to the two variances of these quantities (the<br />

sampl<strong>in</strong>g variance of x, s2 =n, and the prior variance of m, s2 0 ). Thus, the observed<br />

data and the prior <strong>in</strong>formation contribute to the posterior mean <strong>in</strong> proportion to<br />

their precision. The fact that the posterior estimate of m is shifted from the<br />

sample mean x, <strong>in</strong> the direction of the prior mean m0, is an example of the<br />

phenomenon known as shr<strong>in</strong>kage, to be discussed further <strong>in</strong> §6.4.<br />

The variance of the posterior distribution (6.1) may be written <strong>in</strong> the form<br />

…s2 =n†s2 0<br />

…s2 =n†‡s2 , …6:3†<br />

0<br />

which is less than either of the two separate variances, s2 =n and s2 0 . In this sense,<br />

precision has been ga<strong>in</strong>ed by comb<strong>in</strong><strong>in</strong>g the <strong>in</strong>formation from the data and the<br />

prior <strong>in</strong>formation.<br />

These results illustrate various po<strong>in</strong>ts made <strong>in</strong> §6.1. First, the family chosen for<br />

the prior distributions, the normal, constitutes the conjugate family for the normal<br />

likelihood. When the prior is chosen from a conjugate family, the posterior<br />

distribution is always another member of the same family, but with parameters<br />

altered by the <strong>in</strong>corporation of the likelihood. Although this is mathematically<br />

very convenient, it does not follow that the prior should necessarily be chosen

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!