04.01.2013 Views

Springer - Read

Springer - Read

Springer - Read

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

Figure 9-4<br />

The first 132 values of the<br />

data set AIRPASS.TSM<br />

and predictors of the last<br />

12 values obtained by<br />

direct application of<br />

the ARAR algorithm.<br />

9.4 Choosing a Forecasting Algorithm 329<br />

observations in the series WINE.TSM (Figure 1.1) and compare the average mean<br />

squared errors of the forecasts of Y131,...,Y142, we find (Problem 9.2) that an MA(12)<br />

model fitted to the mean corrected differenced series {Yt − Yt−12} does better than<br />

seasonal Holt–Winters (with period 12), which in turn does better than ARAR and<br />

(not surprisingly) dramatically better than nonseasonal Holt–Winters. An interesting<br />

empirical comparison of these and other methods applied to a variety of economic<br />

time series is contained in Makridakis et al. (1998).<br />

The versions of the Holt–Winters algorithms we have discussed in Sections 9.2<br />

and 9.3 are referred to as “additive,” since the seasonal and trend components enter the<br />

forecasting function in an additive manner. “Multiplicative” versions of the algorithms<br />

can also be constructed to deal directly with processes of the form<br />

Yt mtstZt, (9.4.1)<br />

where mt, st, and Zt are trend, seasonal, and noise factors, respectively (see, e.g.,<br />

Makridakis et al., 1983). An alternative approach (provided that Yt > 0 for all t)isto<br />

apply the linear Holt–Winters algorithms to {ln Yt} (as in the case of WINE.TSM in<br />

the preceding paragraph). Because of the rather general memory shortening permitted<br />

by the ARAR algorithm, it gives reasonable results when applied directly to series<br />

of the form (9.4.1), even without preliminary transformations. In particular, if we<br />

consider the first 132 observations in the series AIRPASS.TSM and apply the ARAR<br />

algorithm to predict the last 12 values in the series, we obtain (Problem 9.4) an<br />

observed root mean squared error of 18.21. On the other hand if we use the same<br />

data take logarithms, difference at lag 12, subtract the mean and then fit an AR(13)<br />

model by maximum likelihood using ITSM and use it to predict the last 12 values, we<br />

(thousands)<br />

0.1 0.2 0.3 0.4 0.5 0.6<br />

1949 1951 1953 1955 1957 1959 1961

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!