10.04.2013 Views

SPSS Categories® 11.0

SPSS Categories® 11.0

SPSS Categories® 11.0

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

8 Chapter 1<br />

Categorical Regression<br />

The use of Categorical Regression is most appropriate when the goal of your analysis is<br />

to predict a dependent (response) variable from a set of independent (predictor) variables.<br />

As with all optimal scaling procedures, scale values are assigned to each category<br />

of every variable such that these values are optimal with respect to the regression. The<br />

solution of a categorical regression maximizes the squared correlation between the<br />

transformed response and the weighted combination of transformed predictors.<br />

Relation to other Categories procedures. Categorical regression with optimal scaling is<br />

comparable to optimal scaling canonical correlation analysis with two sets, one of which<br />

contains only the dependent variable. In the latter technique, similarity of sets is derived<br />

by comparing each set to an unknown variable that lies somewhere between all of the<br />

sets. In categorical regression, similarity of the transformed response and the linear<br />

combination of transformed predictors is assessed directly.<br />

Relation to standard techniques. In standard linear regression, categorical variables can<br />

either be recoded as indicator variables or can be treated in the same fashion as interval<br />

level variables. In the first approach, the model contains a separate intercept and slope<br />

for each combination of the levels of the categorical variables. This results in a large<br />

number of parameters to interpret. In the second approach, only one parameter is estimated<br />

for each variable. However, the arbitrary nature of the category codings makes<br />

generalizations impossible.<br />

If some of the variables are not continuous, alternative analyses are available. If the<br />

response is continuous and the predictors are categorical, analysis of variance is often<br />

employed. If the response is categorical and the predictors are continuous, logistic<br />

regression or discriminant analysis may be appropriate. If the response and the predictors<br />

are both categorical, loglinear models are often used.<br />

Regression with optimal scaling offers three scaling levels for each variable. Combinations<br />

of these levels can account for a wide range of nonlinear relationships for which<br />

any single “standard” method is ill-suited. Consequently, optimal scaling offers greater<br />

flexibility than the standard approaches with minimal added complexity.<br />

In addition, nonlinear transformations of the predictors usually reduce the dependencies<br />

among the predictors. If you compare the eigenvalues of the correlation matrix for<br />

the predictors with the eigenvalues of the correlation matrix for the optimally scaled predictors,<br />

the latter set will usually be less variable than the former. In other words, in<br />

categorical regression, optimal scaling makes the larger eigenvalues of the predictor<br />

correlation matrix smaller and the smaller eigenvalues larger.

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!