13.11.2012 Views

Introduction to Categorical Data Analysis

Introduction to Categorical Data Analysis

Introduction to Categorical Data Analysis

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

140 BUILDING AND APPLYING LOGISTIC REGRESSION MODELS<br />

With either approach, for categorical predic<strong>to</strong>rs with more than two categories,<br />

the process should consider the entire variable at any stage rather than just individual<br />

indica<strong>to</strong>r variables. Otherwise, the result depends on how you choose the baseline<br />

category for the indica<strong>to</strong>r variables. Add or drop the entire variable rather than just<br />

one of its indica<strong>to</strong>rs.<br />

Variable selection methods need not yield a meaningful model. Use them with<br />

caution! When you evaluate many terms, one or two that are not truly important may<br />

look impressive merely due <strong>to</strong> chance.<br />

In any case, statistical significance should not be the sole criterion for whether <strong>to</strong><br />

include a term in a model. It is sensible <strong>to</strong> include a variable that is important for<br />

the purposes of the study and report its estimated effect even if it is not statistically<br />

significant. Keeping it in the model may help reduce bias in estimating effects of other<br />

predic<strong>to</strong>rs and may make it possible <strong>to</strong> compare results with other studies where the<br />

effect is significant (perhaps because of a larger sample size). Likewise, with a very<br />

large sample size sometimes a term might be statistically significant but not practically<br />

significant. You might then exclude it from the model because the simpler model is<br />

easier <strong>to</strong> interpret – for example, when the term is a complex interaction.<br />

5.1.4 Example: Backward Elimination for Horseshoe Crabs<br />

When one model is a special case of another, we can test the null hypothesis that the<br />

simpler model is adequate against the alternative hypothesis that the more complex<br />

model fits better. According <strong>to</strong> the alternative, at least one of the extra parameters<br />

in the more complex model is nonzero. Recall that the deviance of a GLM is the<br />

likelihood-ratio statistic for comparing the model <strong>to</strong> the saturated model, which has<br />

a separate parameter for each observation (Section 3.4.3). As Section 3.4.4 showed,<br />

the likelihood-ratio test statistic −2(L0 − L1) for comparing the models is the difference<br />

between the deviances for the models. This test statistic has an approximate<br />

chi-squared null distribution.<br />

Table 5.2 summarizes results of fitting and comparing several logistic regression<br />

models. To select a model, we use a modified backward elimination procedure. We<br />

start with a complex model, check whether the interaction terms are needed, and then<br />

successively take out terms.<br />

We begin with model (1) in Table 5.2, symbolized by C ∗ S + C ∗ W + S ∗ W.It<br />

contains all the two-fac<strong>to</strong>r interactions and main effects. We test all the interactions<br />

simultaneously by comparing it <strong>to</strong> model (2) containing only the main effects. The<br />

likelihood-ratio statistic equals the difference in deviances, which is 186.6 − 173.7 =<br />

12.9, with df = 166 − 155 = 11. This does not suggest that the interactions terms<br />

are needed (P = 0.30). If they were, we could check individual interactions <strong>to</strong> see<br />

whether they could be eliminated (see Problem 5.3).<br />

The next stage considers dropping a term from the main effects model. Table 5.2<br />

shows little consequence from removing spine condition S (model 3c). Both remaining<br />

variables (C and W ) then have nonnegligible effects. For instance, removing C<br />

increases the deviance (comparing models 4b and 3c) by 7.0 on df = 3(P = 0.07).<br />

The analysis in Section 4.4.3 revealed a noticeable difference between dark crabs

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!