Annexes 184166 C. Proust, H. Jacqmin-GaddaSpiessens and Verbeke [3] recently proposed a freeSAS-macro (HETNLMIXED) using the EM algorithmand the NLMIXED procedure for the optimization inthe M-step. This SAS-macro is an extension of theSAS-macro HETMIXED which was developed earlierfor estimating heterogeneous linear mixed modelsusing the MIXED procedure [6]. To our know<strong>le</strong>dge,HETNLMIXED and its first version HETMIXED are theonly free availab<strong>le</strong> programs developed for estimatingheterogeneous mixed models. The first versionHETMIXED was proved to be very slow and limitedto small samp<strong>le</strong>s due to very large matrices handlingand prohibitive computation; it will not beexpanded in this work. HETNLMIXED was developedto reduce these computational prob<strong>le</strong>ms and to allowestimation of both linear and generalized linearmodels. However, in the linear case, this SAS-macrohas the drawback of computing numerically an integralacross the random effects whi<strong>le</strong> it has a closedform, and thus the macro is limited to a small numberof random effects. We have also observed convergenceprob<strong>le</strong>ms when using the macro with largesamp<strong>le</strong>s except for very simp<strong>le</strong> models.Moreover, the EM algorithm, which is used inthese macros, has some general drawbacks. In particular,it does not have any good convergence criteria;the convergence is only built on a lack ofprogression of the likelihood or the parameter estimates[7]. Furthermore, the convergence is slow[8] and the EM algorithm does not provide directestimates of the variance of the parameters. In theparticular case of an heterogeneous mixed model,the M-step also requires the estimation of an homogeneousmixed model which is computationallyexpensive.Therefore, the first aim of this paper is to proposea program for estimating more general heterogeneouslinear mixed models suitab<strong>le</strong> for largesamp<strong>le</strong>s. The proposed program HETMIXLIN is writtenin Fortran90 and uses a direct maximization ofthe likelihood via a Marquardt optimization algorithm.The second objective of this paper is to illustratethe use of heterogeneous linear mixed modelthrough a study of the different patterns of evolutionin cognitive ageing.2. Computational methods and theory2.1. The heterogeneous linear mixed modelLet Y i = (Y i1 ,...,Y ini ) be the response vector forthe n i measurements of the subject i with i =1,...,N. The linear mixed model [1] for the responsevector Y i is defined as:Y i = X iˇ + Z i u i + i (1)X i is a n i ×p design matrix for the p-vector offixed effects ˇ, and Z i is an i ×q design matrix associatedto the q-vector of random effects u i whichrepresents the subject specific regression coefficients.The errors i are assumed to be normallydistributed with mean zero and covariance matrix 2 I ni , and are assumed to be independent from thevector of random effects u i .In an homogeneous mixed model [1], u i is normallydistributed with meanand covariance matrixD, i.e.u i ∼N(, D) (2)In the heterogeneous mixed model [2–4], u i is assumedto follow a mixture of G multivariate Gaussianswith different means ( g ) g=1,G and a commoncovariance matrix D, i.e.u i ∼G∑ g N( g , D) (3)g=1Each component g of the mixture has a probabilityg and the ( g ) g=1,G verify the following conditions:0 ≤ g ≤ 1∀g = 1,G andG∑ g = 1 (4)g=1In this work, we propose a slightly more generalformulation of the model described in (1) in whichthe effect of some covariates may depend on thecomponents of mixture and some of the randomeffects may have a common mean whatever thecomponent of mixture. Thus, the X i design matrix issplit in X 1i associated with the vector ˇ of fixed effectswhich are common to all the components andX 2i associated with the vectors ı g of fixed effectswhich are specific to the components. The Z i designmatrix is also split in Z 1i associated with the vectorv i of random effects following a sing<strong>le</strong> Gaussiandistribution and Z 2i associated with the vector u iof random effects following a mixture of Gaussiandistributions. The model is then written as:Y i = X 1iˇ +G∑ g X 2i ı g + Z 1i v i + Z 2i u i + i (5)g=1where v i ∼N(0,D v ) and u i ∼ ∑ Gg=1 g N( g ,D u );given the component ( g, ) the conditional (( ) distributionof the vector is N , D with)vi0u i ( ) gDv DD = vu.D uv D u
Annexes 185Estimation of linear mixed models with a mixture of distribution for the random effects 1672.2. LikelihoodFollowing the previous works [3,4], we definew ig the unobserved variab<strong>le</strong> indicating if thesubject i belongs to the component g. We haveP(w ig = 1) = g . The density for the vector y i canthen be written as:f i (y i ) =G∑ g f(y i |w ig = 1) (6)g=1Givenw ig , y i follows a linear mixed model, andthe densityf(y i |w ig = 1) denoted by ig is the multivariateGaussian density with meanE ig and covariancematrix V i given by:E ig =E(Y i |w ig = 1) = X 1iˇ + X 2i ı g + Z 2i g andV i = var(Y i |w ig = 1) = Z i DZ ′ i +2 I ni (7)Let now be the vector of the m parametersof the model. contains with ′ =(ˇ′,(ı g ) ′ g=1,G , ( g) ′ g=1,G , Vec(D)′ , 2 ) and the vectorof the G − 1 first component probabilities( g ) g=1,G−1 . Note that G is entirely determined by as 1 − ∑ G−1g=1 g. Vec(D) represents the vector ofthe upper triangular e<strong>le</strong>ments of D. The estimatesof are obtained as the vector ˆ that maximizesthe observed log-likelihood:⎛⎞N∑ N∑ G∑L(Y; ) = ln(f i (y i )) = ln ⎝ g ig (y i ) ⎠=i=1N∑i=1i=1g=1− n i2 ln(2) − 1 2 ln(|V i|)⎛G∑+ ln ⎝g=1 g e − 1 2 (Y i−E ig ) ′ V −1i2.3. Estimation procedure(Y i −E ig )⎞⎠ (8)We propose to maximize directly the observedlog-likelihood (8) using a modified Marquardt optimizationalgorithm [9], a Newton—Raphson-likealgorithm [10]. The diagonal of the Hessian atiteration k, H (k) , is inflated to obtain a positivedefinite matrix as: H ∗(k) = (H ∗(k)ij) with H ∗(k)ii=H (k)ii+[(1 −)|H (k)ii|+tr(H (k) )] and H ∗(k)ij=H (k)ijif i ≠j. Initial values for and are = 0.01 and = 0.01. They are reduced when H ∗ is positivedefinite and increased if not. The estimates (k) arethen updated to (k+1) using the current modifiedHessian H ∗(k) and the current gradient of theparametersg( (k) ) according to the formula: (k+1) = (k) −˛H ∗(k)−1 g( (k) ) (9)where if necessary, ˛ which equals 1 by defaultis modified to ensure that the log-likelihood isimproved at each iteration.To ensure that the covariance matrix D is positive,we maximize the log-likelihood on the nonzeroe<strong>le</strong>ments of U, the Cho<strong>le</strong>sky factor of D (i.e.U ′ U = D) [7]. Furthermore, to deal with the constraintson (4), we use the transformed parameters( g ) g=1,G−1 with :( )g g = ln(10) GStandard errors of the e<strong>le</strong>ments of D and( g ) g=1,G−1 are computed by the -method [11],whi<strong>le</strong> standard errors of the other parameters aredirectly computed using the inverse of the observedHessian matrix.The convergence is reached when the threefollowing convergence criteria are satisfied:∑ mj=1( (k)j− (k−1)j) 2 ≤ a , |L (k) −L (k−1) |≤ b andg( (k) ) ′ H (k)−1 g( (k) ) ≤ d . The default values are a = 10 −5 , b = 10 −5 and d = 10 −8 .As the log-likelihood of a mixture model mayhave several maxima [8], we use a grid of initial valuesto find the global maximum. The multimodalityof the log-likelihood in mixture models has been oftendiscussed and some authors proposed differentstrategies to choose the set of initial values [12].However, none of them seems to be optimal in ageneral way. We have observed, in our experience,that the results were mainly sensitive to initial valuesof ( g ) g=1,G−1 and ( g ) g=1,G and <strong>le</strong>ss sensitiveto the other parameters (Vec(U), ˇ and) for whichestimates of the homogeneous mixed models weregood initial values.A mixture model is estimated with a fixed numberof components G, otherwise the number of parametersin the model is unknown. To choose theright number of components, one has to estimatemodels with different values for G and se<strong>le</strong>ct thebest model according to a test or a criterion. Someworks favor a bootstrap approach to approximatethe asymptotic distribution of the likelihood ratiotest between models with different number ofcomponents [13] but this approach is very heavyin particular for mixture models with random effects.Criteria such as Akaike’s Information Criterion(AIC) [14] or Bayesian Information Criterion(BIC) [15] are often preferred. We use these se<strong>le</strong>ctioncriteria to se<strong>le</strong>ct the optimal number ofcomponents.
- Page 1 and 2:
Université Victor Segalen Bordeaux
- Page 3 and 4:
3RemerciementsA Monsieur Jean-Louis
- Page 5 and 6:
5Un immense merci à tous ceux que
- Page 7 and 8:
7A Delphine et ses mille et une his
- Page 9 and 10:
TABLE DES MATIÈRES 92.3.2 Estimati
- Page 11 and 12:
TABLE DES MATIÈRES 116.3.1 Adéqua
- Page 13 and 14:
Introduction 13ans atteintes d’un
- Page 15 and 16:
Introduction 15Mais, pour l’insta
- Page 17 and 18:
Introduction 171.2 Problèmes méth
- Page 19 and 20:
Introduction 191.2.3 Association en
- Page 21 and 22:
Introduction 21QUID suggère que le
- Page 23 and 24:
Chapitre 2Etat des connaissancesCe
- Page 25 and 26:
Etat des connaissances 25Plusieurs
- Page 27 and 28:
Etat des connaissances 27Prise en c
- Page 29 and 30:
Etat des connaissances 29Extensions
- Page 31 and 32:
Etat des connaissances 31données s
- Page 33 and 34:
Etat des connaissances 33variables
- Page 35 and 36:
Etat des connaissances 35(2000) ne
- Page 37 and 38:
Etat des connaissances 372.3 Modél
- Page 39 and 40:
Etat des connaissances 39chapitre 3
- Page 41 and 42:
Etat des connaissances 41de mélang
- Page 43 and 44:
Etat des connaissances 43Hawkins et
- Page 45 and 46:
Etat des connaissances 45tique de d
- Page 47 and 48:
Etat des connaissances 47L’estima
- Page 49 and 50:
Etat des connaissances 49normales s
- Page 51 and 52:
Etat des connaissances 512.4 Modél
- Page 53 and 54:
Etat des connaissances 53en tant qu
- Page 55 and 56:
Etat des connaissances 55interactio
- Page 57 and 58:
Etat des connaissances 57décrite p
- Page 59 and 60:
Etat des connaissances 59entre l’
- Page 61 and 62:
Etat des connaissances 612.4.3 Cas
- Page 63 and 64:
Etat des connaissances 63l’inform
- Page 65 and 66:
Modèle nonlinéaire à processus l
- Page 67 and 68:
Modèle nonlinéaire à processus l
- Page 69 and 70:
Modèle nonlinéaire à processus l
- Page 71 and 72:
Modèle nonlinéaire à processus l
- Page 73 and 74:
Modèle nonlinéaire à processus l
- Page 75 and 76:
Modèle nonlinéaire à processus l
- Page 77 and 78:
Modèle nonlinéaire à processus l
- Page 79 and 80:
Modèle nonlinéaire à processus l
- Page 81 and 82:
Modèle nonlinéaire à processus l
- Page 83 and 84:
Modèle nonlinéaire à processus l
- Page 85 and 86:
Modèle nonlinéaire à processus l
- Page 87 and 88:
Modèle nonlinéaire à processus l
- Page 89 and 90:
Modèle nonlinéaire à processus l
- Page 91 and 92:
Modèle nonlinéaire à processus l
- Page 93 and 94:
Chapitre 4Modèle nonlinéaire à c
- Page 95 and 96:
Modèle nonlinéaire à classes lat
- Page 97 and 98:
Modèle nonlinéaire à classes lat
- Page 99 and 100:
Modèle nonlinéaire à classes lat
- Page 101 and 102:
Modèle nonlinéaire à classes lat
- Page 103 and 104:
Modèle nonlinéaire à classes lat
- Page 105 and 106:
Modèle nonlinéaire à classes lat
- Page 107 and 108:
Modèle nonlinéaire à classes lat
- Page 109 and 110:
Modèle nonlinéaire à classes lat
- Page 111 and 112:
Modèle nonlinéaire à classes lat
- Page 113 and 114:
Modèle nonlinéaire à classes lat
- Page 115 and 116:
Modèle nonlinéaire à classes lat
- Page 117 and 118:
Modèle nonlinéaire à classes lat
- Page 119 and 120:
Modèle nonlinéaire à classes lat
- Page 121 and 122:
Modèle nonlinéaire à classes lat
- Page 123 and 124:
Modèle nonlinéaire à classes lat
- Page 125 and 126:
Modèle nonlinéaire à classes lat
- Page 127 and 128:
Modèle nonlinéaire à classes lat
- Page 129 and 130:
Modèle nonlinéaire à classes lat
- Page 131 and 132:
Modèle nonlinéaire à classes lat
- Page 133 and 134: Modèle nonlinéaire à classes lat
- Page 135 and 136: Modèle nonlinéaire à classes lat
- Page 137 and 138: Modèle nonlinéaire à classes lat
- Page 139 and 140: Modèle nonlinéaire à classes lat
- Page 141 and 142: Modèle nonlinéaire à classes lat
- Page 143 and 144: Modèle nonlinéaire à classes lat
- Page 145 and 146: Modèle nonlinéaire à classes lat
- Page 147 and 148: Modèle nonlinéaire à classes lat
- Page 149 and 150: Chapitre 6Discussion et perspective
- Page 151 and 152: Discussion et perspectives 151la pr
- Page 153 and 154: Discussion et perspectives 153Varia
- Page 155 and 156: Discussion et perspectives 1556.2 M
- Page 157 and 158: Discussion et perspectives 157à cl
- Page 160 and 161: Discussion et perspectives 160effet
- Page 162 and 163: Discussion et perspectives 162Dans
- Page 164 and 165: Chapitre 7BibliographieAmieva, H.,
- Page 166 and 167: Bibliographie 1661, S19-25.Brown, E
- Page 168 and 169: Bibliographie 168Sons, New-York.Fol
- Page 170 and 171: Bibliographie 170Hogan, J. W. et La
- Page 172 and 173: Bibliographie 172Lin, H., McCulloch
- Page 174 and 175: Bibliographie 174Park, CA.Muthén,
- Page 176 and 177: Bibliographie 176Schlattmann, P. (2
- Page 178 and 179: Bibliographie 178in the random-effe
- Page 180 and 181: Chapitre 8Annexes8.1 Liste des publ
- Page 182 and 183: Annexes 182Pau (France)Communicatio
- Page 186 and 187: Annexes 186168 C. Proust, H. Jacqmi
- Page 188 and 189: Annexes 188170 C. Proust, H. Jacqmi
- Page 190 and 191: Annexes 190172 C. Proust, H. Jacqmi
- Page 192 and 193: AnnexesAmerican Journal of Epidemio
- Page 194 and 195: Annexes 194Psychometric Tests’ Se
- Page 196 and 197: Annexes 196Psychometric Tests’ Se
- Page 198 and 199: Annexes 198Psychometric Tests’ Se
- Page 200 and 201: Annexes 200Abstract :When investiga
- Page 202 and 203: Annexes 202educated subjects. This
- Page 204 and 205: Annexes 204(iii) the recognition fo
- Page 206 and 207: Annexes 206Explanatory variablesIn
- Page 208 and 209: Annexes 208IST15 (median=28, IQR=24
- Page 210 and 211: Annexes 210and on the mean evolutio
- Page 212 and 213: Annexes 212These findings should be
- Page 214 and 215: Annexes 214Appendix : model specifi
- Page 216 and 217: Annexes 216References[1] Amieva H,
- Page 218 and 219: Annexes 218[19] Letenneur L, Commen
- Page 220 and 221: Annexes 220Figure 1 : (A) Predicted
- Page 222 and 223: Annexes 222Table 1: demographic and
- Page 224: Annexes 224MMSE 0.0037 * 0.0013,0.0