13.11.2012 Views

Introduction to Categorical Data Analysis

Introduction to Categorical Data Analysis

Introduction to Categorical Data Analysis

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

298 RANDOM EFFECTS: GENERALIZED LINEAR MIXED MODELS<br />

treatments. By contrast, random effects apply <strong>to</strong> a sample. For a study with repeated<br />

measurement of subjects, for example, a cluster is a set of observations for a particular<br />

subject, and the model contains a random effect term for each subject. The random<br />

effects refer <strong>to</strong> a sample of clusters from all the possible clusters.<br />

10.1.1 The Generalized Linear Mixed Model<br />

Generalized linear models (GLMs) extend ordinary regression by allowing nonnormal<br />

responses and a link function of the mean (recall Chapter 3). The generalized linear<br />

mixed model, denoted by GLMM, is a further extension that permits random effects<br />

as well as fixed effects in the linear predic<strong>to</strong>r. Denote the random effect for cluster i<br />

by ui. We begin with the most common case, in which ui is an intercept term in the<br />

model.<br />

Let yit denote observation t in cluster i. Let xit be the value of the explana<strong>to</strong>ry<br />

variable for that observation. (The model extends in an obvious way for multiple<br />

predic<strong>to</strong>rs.) Conditional on ui, a GLMM resembles an ordinary GLM. Let μit =<br />

E(Yit|ui), the mean of the response variable for a given value of the random effect.<br />

With the link function g(·), the GLMM has the form<br />

g(μit) = ui + βxit, i = 1,...,n, t = 1,...,T<br />

A GLMM with random effect as an intercept term is called a random intercept model.<br />

In practice, the random effect ui is unknown, like the usual intercept parameter. It is<br />

treated as a random variable and is assumed <strong>to</strong> have a normal N(α, σ ) distribution,<br />

with unknown parameters. The variance σ 2 is referred <strong>to</strong> as a variance component.<br />

When xit = 0 in this model, the expected value of the linear predic<strong>to</strong>r is α, the<br />

mean of the probability distribution of ui. An equivalent model enters α explicitly in<br />

the linear predic<strong>to</strong>r,<br />

g(μit) = ui + α + βxit<br />

(10.1)<br />

compensating by taking ui <strong>to</strong> have an expected value of 0. The ML estimate of α<br />

is identical either way. However, it would be redundant <strong>to</strong> have both the α term in<br />

the model and permit an unspecified mean for the distribution of ui. We will specify<br />

models this second way, taking ui <strong>to</strong> have a N(0,σ) distribution. The separate α<br />

parameter in the model then is the value of the linear predic<strong>to</strong>r when xit = 0 and ui<br />

takes value at its mean of 0.<br />

Why not treat the cluster-specific {ui} terms as fixed effects (parameters)? Usually<br />

a study has a large number of clusters, and so the model would then contain a<br />

large number of parameters. Treating {ui} as random effects, we have only a single<br />

additional parameter (σ ) in the model, describing their dispersion.<br />

Section 10.5 outlines the model-fitting process for GLMMs. As in ordinary models<br />

for a univariate response, for given predic<strong>to</strong>r values ML fitting treats the observations<br />

as independent. For the GLMM, this independence is assumed conditional on the

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!