01.06.2013 Views

Statistical Methods in Medical Research 4ed

Statistical Methods in Medical Research 4ed

Statistical Methods in Medical Research 4ed

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

566 Further Bayesian methods<br />

probabilities can be found by apply<strong>in</strong>g the usual methods, i.e. the jo<strong>in</strong>t probability<br />

of the data and model M1 is<br />

p…y, M1† ˆ „ p…yju1, M1†p…u1jM1†p…M1† du1: …16:18†<br />

There is a similar expression for p…y, M2†, so Bayes' theorem (§3.3) immediately<br />

gives p…M1jy† ˆp…M1, y†=‰p…M1, y†‡p…M2, y†Š:<br />

The usual way to express these quantities is through the ratio of the posterior<br />

odds to the prior odds of model M1, a quantity known as the Bayes factor, BF<br />

(§3.3); that is,<br />

BF ˆ p…M1jy†=p…M2jy†<br />

p…M1†=p…M2†<br />

This k<strong>in</strong>d of analysis can readily be extended to consider m compet<strong>in</strong>g models,<br />

M1, ..., Mm. Bayes factors can then be used to make pairwise comparisons<br />

between models. Alternatively, <strong>in</strong>terest may be focused on the models with the<br />

largest posterior probabilities.<br />

There are some difficulties with the Bayes factor for compar<strong>in</strong>g two models.<br />

The first concerns the practicalities of comput<strong>in</strong>g BF via (16.18), as this <strong>in</strong>volves<br />

the specification of prior distributions for the parameters <strong>in</strong> each of the models.<br />

If the priors have to be established us<strong>in</strong>g an elicitation procedure, then hav<strong>in</strong>g to<br />

accommodate several possible models can make the elicitation procedure much<br />

more cumbersome and complicated. An approach which avoids the need to<br />

specify prior distributions was developed by Schwarz (1978), who showed that<br />

for samples of size n, with n large,<br />

2 log BF 2 log LR …k2 k1† log n, …16:19†<br />

where ki is the number of parameters <strong>in</strong> model Mi and LR is the usual (frequentist)<br />

likelihood ratio between the models, i.e. LR ˆ p…yj ^ u1†=p…yj ^ u2†, where ^ ui is<br />

the maximum likelihood estimator for model Mi. The quantity <strong>in</strong> (16.19) is<br />

sometimes referred to as the Bayesian <strong>in</strong>formation criterion, (BIC).<br />

It might be asked to what extent it is possible to approximate a quantity such<br />

as the Bayes factor, which depends on prior distributions, us<strong>in</strong>g a quantity which<br />

does not depend on any priors at all. However, the approximation <strong>in</strong> (16.19) is<br />

valid only for large n, and for large n (and priors that are not too concentrated)<br />

the likelihood plays a more important role <strong>in</strong> determ<strong>in</strong><strong>in</strong>g quantities such as<br />

(16.18) than does the prior.<br />

It might have been thought that an alternative to avoid<strong>in</strong>g the specification of<br />

the priors was to use non-<strong>in</strong>formative priors. This would be possible if the non<strong>in</strong>formative<br />

priors are proper but is not possible for improper priors. This is<br />

because (16.18) does not def<strong>in</strong>e a proper distribution if the prior for the parameters<br />

of the model is not proper. This occurs for exactly the same reason as was

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!