13.11.2012 Views

Introduction to Categorical Data Analysis

Introduction to Categorical Data Analysis

Introduction to Categorical Data Analysis

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

10.2 EXAMPLES OF RANDOM EFFECTS MODELS FOR BINARY DATA 303<br />

the logits of the probabilities vary according <strong>to</strong> a normal distribution, the fitting process<br />

“borrows from the whole,” using data from all the areas <strong>to</strong> estimate the probability in<br />

any given one. The estimate for a given area is then a weighted average of the sample<br />

proportion for that area alone and the overall proportion for all the areas.<br />

Software provides ML estimates ˆα and ˆσ and predicted values {ûi} for the random<br />

effects. The predicted value ûi depends on all the data, not only the data for area i.<br />

The estimate of the probability πi in area i is then<br />

ˆπi = exp(ûi +ˆα)/[1 + exp(ûi +ˆα)]<br />

A benefit of using data from all the areas instead of only area i <strong>to</strong> estimate πi is that<br />

the estima<strong>to</strong>r ˆπi tends <strong>to</strong> be closer than the sample proportion pi <strong>to</strong> πi. The {ˆπi}<br />

result from shrinking the sample proportions <strong>to</strong>ward the overall sample proportion.<br />

The amount of shrinkage increases as ˆσ decreases. If ˆσ = 0, then {ˆπi} are identical.<br />

In fact, they then equal the overall sample proportion after pooling all n samples.<br />

When truly all πi are equal, ˆπi is a much better estima<strong>to</strong>r of that common value than<br />

the sample proportion from sample i alone.<br />

Foragiven ˆσ >0, the {ˆπi} give more weight <strong>to</strong> the sample proportions as {Ti}<br />

grows. As each sample has more data, we put more trust in the separate sample<br />

proportions.<br />

The simple random effects model (10.4), which is natural for small-area estimation,<br />

can be useful for any application that estimates a large number of binomial parameters<br />

when the sample sizes are small. The following example illustrates this.<br />

10.2.2 Example: Estimating Basketball Free Throw Success<br />

In basketball, the person who plays center is usually the tallest player on the team.<br />

Often, centers shoot well from near the basket but not so well from greater distances.<br />

Table 10.2 shows results of free throws (a standardized shot taken from a distance<br />

of 15 feet from the basket) for the 15 <strong>to</strong>p-scoring centers in the National Basketball<br />

Association after one week of the 2005–2006 season.<br />

Let πi denote the probability that player i makes a free throw (i = 1,...,15). For<br />

Ti observations of player i, we treat the number of successes yi as binomial with<br />

index Ti and parameter πi. Table 10.2 shows {Ti} and {pi = yi/Ti}.<br />

For the ML fit of model (10.4), ˆα = 0.908 and ˆσ = 0.422. For a player<br />

with random effect ui = 0, the estimated probability of making a free throw is<br />

exp(0.908)/[1 + exp(0.908)] =0.71. We predict that 95% of the logits fall within<br />

0.908 ± 1.96(0.422), which is (0.08, 1.73). This interval corresponds <strong>to</strong> probabilities<br />

in the range (0.52, 0.85).<br />

The predicted random effect values (obtained using PROC NLMIXED in SAS)<br />

yield probability estimates {ˆπi}, also shown in Table 10.2. Since {Ti} are small and<br />

since ˆσ is relatively small, these estimates shrink the sample proportions substantially<br />

<strong>to</strong>ward the overall sample proportion of free throws made, which was 101/143 =<br />

0.706. The {ˆπi} vary only between 0.61 and 0.76, whereas the sample proportions

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!