View PDF Version - RePub - Erasmus Universiteit Rotterdam
View PDF Version - RePub - Erasmus Universiteit Rotterdam
View PDF Version - RePub - Erasmus Universiteit Rotterdam
Create successful ePaper yourself
Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.
Estimation<br />
Discriminant analysis using a MLMM with a normal mixture 147<br />
The MLMM (1) can be regarded as a standard LMM (2) with X and Z matrices<br />
supplemented by zeros. However, maximum-likelihood based estimation routines<br />
that ignore the specific sparse structure of X and Z matrices are inefficient and encounter<br />
numerical problems even with normally distributed random effects (K =1<br />
in expression (3)). For example, Fieuws et al. 9 used a pairwise fitting approach of<br />
Fieuws and Verbeke17 to avoid numerical problems. Another route would be to use<br />
methods for sparse matrices, see, e.g., R package Matrix (Bates and Maechler18 ).<br />
Here we adopt the Bayesian approach with Markov chain Monte Carlo (MCMC)<br />
estimation. The MCMC approach proved to be a useful machinery for problems<br />
involving hierarchically specified models (like linear mixed models) and models involving<br />
mixture distributions. Our prior distributions for the model parameters are<br />
weakly informative such that the posterior summary statistics correspond closely to<br />
maximum-likelihood estimates.<br />
Prior distributions<br />
To specify the model from a Bayesian point of view, prior distributions have to be<br />
assigned to model parameters. The vector θ of model parameters is supplemented<br />
by latent quantities (values of random effects b =(b ′ 1,...,b ′ N) ′ ), variance hyperpa-<br />
rameters γ b =(γb,1,...,γb,q) ′ , γ ε =(γε,1,...,γε,R) ′ (see below), and further, by<br />
other parameters pertaining to the hierarchical structure of the model which simplify<br />
the subsequent computations (component allocations u =(u1,...,uN) ′ , see below)<br />
in the spirit of the Bayesian data augmentation approach of Tanner and Wong19 .<br />
This leads to the vector of parameters<br />
ψ =(θ ′ , b ′ , γ ′ b, γ ′ ε, u ′ ) ′ , (6)<br />
for which the joint prior distribution will be specified hierarchically.<br />
It is well known (see, e.g., Diebolt and Robert20 ) that, due to the problem of an<br />
unbounded likelihood mentioned above, mixture models do not allow for improper<br />
priors. Nevertheless, several proper, however weakly informative prior distributions<br />
for mixture problems have been suggested in the literature leading to the proper pos-<br />
terior distribution (see Diebolt and Robert20 , Roeder and Wasserman21 , Richardson<br />
and Green22 ). In this paper, we exploit the approach of Richardson and Green22 adapted to our needs and to hierarchical linear models (see, e.g., Gelman et al. 23 ,<br />
Chapter 15). Let p be a generic symbol for a distribution. The joint prior distribution