17.01.2015 Views

LibraryPirate

LibraryPirate

LibraryPirate

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

566 CHAPTER 11 REGRESSION ANALYSIS: SOME ADDITIONAL TECHNIQUES<br />

is motivated by a desire to describe, understand, and make use of the relationship<br />

between independent variables and a dependent (or outcome) variable that is discrete.<br />

Particularly plentiful are circumstances in which the outcome variable is dichotomous.<br />

A dichotomous variable, we recall, is a variable that can assume only one of two<br />

mutually exclusive values. These values are usually coded Y = 1 for a success and<br />

Y = 0 for a nonsuccess, or failure. Dichotomous variables include those whose two<br />

possible values are such categories as died–did not die; cured–not cured; disease<br />

occurred–disease did not occur; and smoker–nonsmoker. The health sciences professional<br />

who either engages in research or needs to understand the results of research<br />

conducted by others will find it advantageous to have, at least, a basic understanding<br />

of logistic regression, the type of regression analysis that is usually employed when<br />

the dependent variable is dichotomous. The purpose of the present discussion is to<br />

provide the reader with this level of understanding. We shall limit our presentation to<br />

the case in which there is only one independent variable that may be either continuous<br />

or dichotomous.<br />

The Logistic Regression Model Recall that in Chapter 9 we referred to<br />

regression analysis involving only two variables as simple linear regression analysis. The<br />

simple linear regression model was expressed by the equation<br />

y = b 0 + b 1 x +P<br />

(11.4.1)<br />

in which y is an arbitrary observed value of the continuous dependent variable. When<br />

the observed value of Y is m yƒx , the mean of a subpopulation of Y values for a given value<br />

of X, the quantity P, the difference between the observed Y and the regression line (see<br />

Figure 9.2.1) is zero, and we may write Equation 11.4.1 as<br />

m y ƒ x = b 0 + b 1 x<br />

(11.4.2)<br />

which may also be written as<br />

E 1yƒx2 = b 0 + b 1 x<br />

(11.4.3)<br />

Generally the right-hand side of Equations 11.4.1 through 11.4.3 may assume any value<br />

between minus infinity and plus infinity.<br />

Even though only two variables are involved, the simple linear regression model<br />

is not appropriate when Y is a dichotomous variable because the expected value (or mean)<br />

of Y is the probability that Y = 1 and, therefore, is limited to the range 0 through 1,<br />

inclusive. Equations 11.4.1 through 11.4.3, then, are incompatible with the reality of the<br />

situation.<br />

If we let p = P1Y = 12, then the ratio p>11 - p2 can take on values between 0 and<br />

plus infinity. Furthermore, the natural logarithm (ln) of p>11 - p2 can take on values

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!