13.11.2012 Views

Introduction to Categorical Data Analysis

Introduction to Categorical Data Analysis

Introduction to Categorical Data Analysis

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

158 BUILDING AND APPLYING LOGISTIC REGRESSION MODELS<br />

relative <strong>to</strong> the sample size, conditional ML estimates of parameters work better than<br />

ordinary ML estima<strong>to</strong>rs.<br />

Exact inference for a parameter uses the conditional likelihood function that eliminates<br />

all the other parameters. Since that conditional likelihood does not involve<br />

unknown parameters, probabilities such as P -values use exact distributions rather<br />

than approximations. When the sample size is small, conditional likelihood-based<br />

exact inference in logistic regression is more reliable than the ordinary large-sample<br />

inferences.<br />

5.4.2 Small-Sample Tests for Contingency Tables<br />

Consider first logistic regression with a single explana<strong>to</strong>ry variable,<br />

logit[π(x)]=α + βx<br />

When x takes only two values, the model applies <strong>to</strong> 2 × 2 tables of counts {nij }<br />

for which the two columns are the levels of Y . The usual sampling model treats the<br />

responses on Y in the two rows as independent binomial variates. The row <strong>to</strong>tals,<br />

which are the numbers of trials for those binomial variates, are naturally fixed.<br />

For this model, the hypothesis of independence is H0: β = 0. The unknown parameter<br />

α refers <strong>to</strong> the relative number of outcomes of y = 1 and y = 0, which are<br />

the column <strong>to</strong>tals. Software eliminates α from the likelihood by conditioning also<br />

on the column <strong>to</strong>tals, which are the information in the data about α. Fixing both<br />

sets of marginal <strong>to</strong>tals yields a hypergeometric distribution for n11, for which the<br />

probabilities do not depend on unknown parameters. The resulting exact test of H0:<br />

β = 0 is the same as Fisher’s exact test (Section 2.6.1).<br />

Next, suppose the model also has a second explana<strong>to</strong>ry fac<strong>to</strong>r, Z, with K levels.<br />

If Z is nominal-scale, a relevant model is<br />

logit(π) = α + βx + β Z k<br />

Section 4.3.4 presented this model for 2 × 2 × K contingency tables. The test of<br />

H0: β = 0 refers <strong>to</strong> the effect of X on Y , controlling for Z. The exact test eliminates<br />

the other parameters by conditioning on the marginal <strong>to</strong>tals in each partial table. This<br />

gives an exact test of conditional independence between X and Y , controlling for Z.<br />

For 2 × 2 × K tables {nij k}, conditional on the marginal <strong>to</strong>tals in each partial table,<br />

the Cochran–Mantel–Haenszel test of conditional independence (Section 4.3.4) is a<br />

large-sample approximate method that compares �<br />

k n11k <strong>to</strong> its null expected value.<br />

Exact tests use �<br />

k n11k in the way they use n11 in Fisher’s exact test for 2 × 2<br />

tables. Hypergeometric distributions in each partial table determine the probability<br />

distribution of �<br />

k n11k. The P -value for Ha: β>0 equals the right-tail probability<br />

that �<br />

k n11k is at least as large as observed, for the fixed marginal <strong>to</strong>tals. Two-sided<br />

alternatives can use a two-tail probability of those outcomes that are no more likely<br />

than the observed one.

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!