23.10.2012 Views

View PDF Version - RePub - Erasmus Universiteit Rotterdam

View PDF Version - RePub - Erasmus Universiteit Rotterdam

View PDF Version - RePub - Erasmus Universiteit Rotterdam

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

Estimation<br />

Discriminant analysis using a MLMM with a normal mixture 147<br />

The MLMM (1) can be regarded as a standard LMM (2) with X and Z matrices<br />

supplemented by zeros. However, maximum-likelihood based estimation routines<br />

that ignore the specific sparse structure of X and Z matrices are inefficient and encounter<br />

numerical problems even with normally distributed random effects (K =1<br />

in expression (3)). For example, Fieuws et al. 9 used a pairwise fitting approach of<br />

Fieuws and Verbeke17 to avoid numerical problems. Another route would be to use<br />

methods for sparse matrices, see, e.g., R package Matrix (Bates and Maechler18 ).<br />

Here we adopt the Bayesian approach with Markov chain Monte Carlo (MCMC)<br />

estimation. The MCMC approach proved to be a useful machinery for problems<br />

involving hierarchically specified models (like linear mixed models) and models involving<br />

mixture distributions. Our prior distributions for the model parameters are<br />

weakly informative such that the posterior summary statistics correspond closely to<br />

maximum-likelihood estimates.<br />

Prior distributions<br />

To specify the model from a Bayesian point of view, prior distributions have to be<br />

assigned to model parameters. The vector θ of model parameters is supplemented<br />

by latent quantities (values of random effects b =(b ′ 1,...,b ′ N) ′ ), variance hyperpa-<br />

rameters γ b =(γb,1,...,γb,q) ′ , γ ε =(γε,1,...,γε,R) ′ (see below), and further, by<br />

other parameters pertaining to the hierarchical structure of the model which simplify<br />

the subsequent computations (component allocations u =(u1,...,uN) ′ , see below)<br />

in the spirit of the Bayesian data augmentation approach of Tanner and Wong19 .<br />

This leads to the vector of parameters<br />

ψ =(θ ′ , b ′ , γ ′ b, γ ′ ε, u ′ ) ′ , (6)<br />

for which the joint prior distribution will be specified hierarchically.<br />

It is well known (see, e.g., Diebolt and Robert20 ) that, due to the problem of an<br />

unbounded likelihood mentioned above, mixture models do not allow for improper<br />

priors. Nevertheless, several proper, however weakly informative prior distributions<br />

for mixture problems have been suggested in the literature leading to the proper pos-<br />

terior distribution (see Diebolt and Robert20 , Roeder and Wasserman21 , Richardson<br />

and Green22 ). In this paper, we exploit the approach of Richardson and Green22 adapted to our needs and to hierarchical linear models (see, e.g., Gelman et al. 23 ,<br />

Chapter 15). Let p be a generic symbol for a distribution. The joint prior distribution

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!