29.07.2014 Views

mixed - Stata

mixed - Stata

mixed - Stata

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

<strong>mixed</strong> — Multilevel <strong>mixed</strong>-effects linear regression 45<br />

Matrices<br />

e(b)<br />

e(N g)<br />

e(g min)<br />

e(g avg)<br />

e(g max)<br />

e(tmap)<br />

e(V)<br />

e(V modelbased)<br />

Functions<br />

e(sample)<br />

coefficient vector<br />

group counts<br />

group-size minimums<br />

group-size averages<br />

group-size maximums<br />

ID mapping for unstructured residual errors<br />

variance–covariance matrix of the estimator<br />

model-based variance<br />

marks estimation sample<br />

Methods and formulas<br />

As given by (1), in the absence of weights we have the linear <strong>mixed</strong> model<br />

y = Xβ + Zu + ɛ<br />

where y is the n × 1 vector of responses, X is an n × p design/covariate matrix for the fixed effects<br />

β, and Z is the n × q design/covariate matrix for the random effects u. The n × 1 vector of errors<br />

ɛ is for now assumed to be multivariate normal with mean 0 and variance matrix σɛ 2 I n . We also<br />

assume that u has variance–covariance matrix G and that u is orthogonal to ɛ so that<br />

[ ] [ ]<br />

u G 0<br />

Var =<br />

ɛ 0 σɛ 2 I n<br />

Considering the combined error term Zu + ɛ, we see that y is multivariate normal with mean Xβ<br />

and n × n variance–covariance matrix<br />

V = ZGZ ′ + σ 2 ɛ I n<br />

Defining θ as the vector of unique elements of G results in the log likelihood<br />

L(β, θ, σ 2 ɛ ) = − 1 2<br />

{<br />

n log(2π) + log |V| + (y − Xβ) ′ V −1 (y − Xβ) } (9)<br />

which is maximized as a function of β, θ, and σɛ 2 . As explained in chapter 6 of Searle, Casella,<br />

and McCulloch (1992), considering instead the likelihood of a set of linear contrasts Ky that do not<br />

depend on β results in the restricted log likelihood<br />

L R (β, θ, σ 2 ɛ ) = L(β, θ, σ 2 ɛ ) − 1 2 log ∣ ∣ X ′ V −1 X ∣ ∣ (10)<br />

Given the high dimension of V, however, the log-likelihood and restricted log-likelihood criteria are<br />

not usually computed by brute-force application of the above expressions. Instead, you can simplify<br />

the problem by subdividing the data into independent clusters (and subclusters if possible) and using<br />

matrix decomposition methods on the smaller matrices that result from treating each cluster one at a<br />

time.<br />

Consider the two-level model described previously in (2),<br />

y j = X j β + Z j u j + ɛ j<br />

for j = 1, . . . , M clusters with cluster j containing n j observations, with Var(u j ) = Σ, a q × q<br />

matrix.

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!