13.11.2012 Views

Introduction to Categorical Data Analysis

Introduction to Categorical Data Analysis

Introduction to Categorical Data Analysis

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

152 BUILDING AND APPLYING LOGISTIC REGRESSION MODELS<br />

Figure 5.2. Observed proportion (x) and estimated probability of heart disease (curve) for linear logit<br />

model.<br />

patterns <strong>to</strong> what might be chance variation from a model. Also, these deletion diagnostics<br />

all relate <strong>to</strong> removing an entire binomial sample at a blood pressure level instead<br />

of removing a single subject’s binary observation. Such subject-level deletions have<br />

little effect for this model.<br />

Another useful graphical display for showing lack of fit compares observed and<br />

fitted proportions by plotting them against each other, or by plotting both of them<br />

against explana<strong>to</strong>ry variables. For the linear logit model, Figure 5.2 plots both the<br />

observed proportions and the estimated probabilities of heart disease against blood<br />

pressure. The fit seems decent.<br />

5.3 EFFECTS OF SPARSE DATA<br />

The log likelihood function for logistic regression models has a concave (bowl) shape.<br />

Because of this, the algorithm for finding the ML estimates (Section 3.5.1) usually<br />

converges quickly <strong>to</strong> the correct values. However, certain data patterns present difficulties,<br />

with the ML estimates being infinite or not existing. For quantitative or<br />

categorical predic<strong>to</strong>rs, this relates <strong>to</strong> observing only successes or only failures over<br />

certain ranges of predic<strong>to</strong>r values.<br />

5.3.1 Infinite Effect Estimate: Quantitative Predic<strong>to</strong>r<br />

Consider first the case of a single quantitative predic<strong>to</strong>r. The ML estimate for its effect<br />

is infinite when the predic<strong>to</strong>r values having y = 0 are completely below or completely<br />

above those having y = 1, as Figure 5.3 illustrates. In it, y = 0atx = 10, 20, 30,<br />

40, and y = 1atx = 60, 70, 80, 90. An ideal (perfect) fit has ˆπ = 0 for x ≤ 40 and<br />

ˆπ = 1 for x ≥ 60. One can get a sequence of logistic curves that gets closer and closer<br />

<strong>to</strong> this ideal by letting ˆβ increase without limit, with ˆα =−50 ˆβ. (This ˆα value yields

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!