14.03.2014 Views

Basic Analysis and Graphing - SAS

Basic Analysis and Graphing - SAS

Basic Analysis and Graphing - SAS

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

104 Performing Bivariate <strong>Analysis</strong> Chapter 4<br />

Fit Line <strong>and</strong> Fit Polynomial<br />

Table 4.6 Description of the <strong>Analysis</strong> of Variance Report (Continued)<br />

Sum of Squares<br />

Mean Square<br />

F Ratio<br />

Prob > F<br />

The sum of squares (SS for short) for each source of variation:<br />

• In this example, the total (C. Total) sum of squared distances of each response<br />

from the sample mean is 57,258.157, as shown in Figure 4.12. That is the sum<br />

of squares for the base model (or simple mean model) used for comparison with<br />

all other models.<br />

• For the linear regression, the sum of squared distances from each point to the<br />

line of fit reduces from 12,012.733. This is the residual or unexplained (Error)<br />

SS after fitting the model. The residual SS for a second degree polynomial fit is<br />

6,906.997, accounting for slightly more variation than the linear fit. That is, the<br />

model accounts for more variation because the model SS are higher for the<br />

second degree polynomial than the linear fit. The C. total SS less the Error SS<br />

gives the sum of squares attributed to the model.<br />

The sum of squares divided by its associated degrees of freedom. The F-ratio for a<br />

statistical test is the ratio of the following mean squares:<br />

• The Model mean square for the linear fit is 45,265.424. This value estimates the<br />

error variance, but only under the hypothesis that the model parameters are zero.<br />

• The Error mean square is 245.2. This value estimates the error variance.<br />

The model mean square divided by the error mean square. The underlying<br />

hypothesis of the fit is that all the regression parameters (except the intercept) are<br />

zero. If this hypothesis is true, then both the mean square for error <strong>and</strong> the mean<br />

square for model estimate the error variance, <strong>and</strong> their ratio has an F-distribution. If<br />

a parameter is a significant model effect, the F-ratio is usually higher than expected<br />

by chance alone.<br />

The observed significance probability (p-value) of obtaining a greater F-value by<br />

chance alone if the specified model fits no better than the overall response mean.<br />

Observed significance probabilities of 0.05 or less are often considered evidence of a<br />

regression effect.<br />

Parameter Estimates Report<br />

The terms in the Parameter Estimates report for a linear fit are the intercept <strong>and</strong> the single x variable.<br />

For a polynomial fit of order k, there is an estimate for the model intercept <strong>and</strong> a parameter estimate for<br />

each of the k powers of the X variable.

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!