mixed - Stata
mixed - Stata
mixed - Stata
Create successful ePaper yourself
Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.
<strong>mixed</strong> — Multilevel <strong>mixed</strong>-effects linear regression 45<br />
Matrices<br />
e(b)<br />
e(N g)<br />
e(g min)<br />
e(g avg)<br />
e(g max)<br />
e(tmap)<br />
e(V)<br />
e(V modelbased)<br />
Functions<br />
e(sample)<br />
coefficient vector<br />
group counts<br />
group-size minimums<br />
group-size averages<br />
group-size maximums<br />
ID mapping for unstructured residual errors<br />
variance–covariance matrix of the estimator<br />
model-based variance<br />
marks estimation sample<br />
Methods and formulas<br />
As given by (1), in the absence of weights we have the linear <strong>mixed</strong> model<br />
y = Xβ + Zu + ɛ<br />
where y is the n × 1 vector of responses, X is an n × p design/covariate matrix for the fixed effects<br />
β, and Z is the n × q design/covariate matrix for the random effects u. The n × 1 vector of errors<br />
ɛ is for now assumed to be multivariate normal with mean 0 and variance matrix σɛ 2 I n . We also<br />
assume that u has variance–covariance matrix G and that u is orthogonal to ɛ so that<br />
[ ] [ ]<br />
u G 0<br />
Var =<br />
ɛ 0 σɛ 2 I n<br />
Considering the combined error term Zu + ɛ, we see that y is multivariate normal with mean Xβ<br />
and n × n variance–covariance matrix<br />
V = ZGZ ′ + σ 2 ɛ I n<br />
Defining θ as the vector of unique elements of G results in the log likelihood<br />
L(β, θ, σ 2 ɛ ) = − 1 2<br />
{<br />
n log(2π) + log |V| + (y − Xβ) ′ V −1 (y − Xβ) } (9)<br />
which is maximized as a function of β, θ, and σɛ 2 . As explained in chapter 6 of Searle, Casella,<br />
and McCulloch (1992), considering instead the likelihood of a set of linear contrasts Ky that do not<br />
depend on β results in the restricted log likelihood<br />
L R (β, θ, σ 2 ɛ ) = L(β, θ, σ 2 ɛ ) − 1 2 log ∣ ∣ X ′ V −1 X ∣ ∣ (10)<br />
Given the high dimension of V, however, the log-likelihood and restricted log-likelihood criteria are<br />
not usually computed by brute-force application of the above expressions. Instead, you can simplify<br />
the problem by subdividing the data into independent clusters (and subclusters if possible) and using<br />
matrix decomposition methods on the smaller matrices that result from treating each cluster one at a<br />
time.<br />
Consider the two-level model described previously in (2),<br />
y j = X j β + Z j u j + ɛ j<br />
for j = 1, . . . , M clusters with cluster j containing n j observations, with Var(u j ) = Σ, a q × q<br />
matrix.