Sunteți pe pagina 1din 15

See discussions, stats, and author profiles for this publication at: https://www.researchgate.

net/publication/247087596

Out-of sample tests of forecasting accuracy: a tutorial and review

Article  in  International Journal of Forecasting · January 2000

CITATIONS READS

11 3,531

1 author:

Len Tashman
International Institute of Forecasters
23 PUBLICATIONS   364 CITATIONS   

SEE PROFILE

All content following this page was uploaded by Len Tashman on 25 May 2016.

The user has requested enhancement of the downloaded file.


International Journal of Forecasting 16 (2000) 437–450
www.elsevier.com / locate / ijforecast

Out-of-sample tests of forecasting accuracy: an analysis and review

Leonard J. Tashman*
School of Business Administration, University of Vermont, Burlington, Vermont 05405, USA

Abstract

In evaluations of forecasting accuracy, including forecasting competitions, researchers have paid attention to the selection
of time series and to the appropriateness of forecast-error measures. However, they have not formally analyzed choices in the
implementation of out-of-sample tests, making it difficult to replicate and compare forecasting accuracy studies. In this
paper, I (1) explain the structure of out-of-sample tests, (2) provide guidelines for implementing these tests, and (3) evaluate
the adequacy of out-of-sample tests in forecasting software. The issues examined include series-splitting rules, fixed versus
rolling origins, updating versus recalibration of model coefficients, fixed versus rolling windows, single versus multiple test
periods, diversification through multiple time series, and design characteristics of forecasting competitions. For individual
time series, the efficiency and reliability of out-of-sample tests can be improved by employing rolling-origin evaluations,
recalibrating coefficients, and using multiple test periods. The results of forecasting competitions would be more
generalizable if based upon precisely described groups of time series, in which the series are homogeneous within group and
heterogeneous between groups. Few forecasting software programs adequately implement out-of-sample evaluations,
especially general statistical packages and spreadsheet add-ins.  2000 International Institute of Forecasters. Published by
Elsevier Science B.V. All rights reserved.

Keywords: Out-of-sample; Fit period; Test period; Fixed origin; Rolling origin; Updating; Recalibration; Rolling window; Sliding
simulation; Forecasting competitions

1. Introduction to an individual time series: Issues addressed


are rules for splitting the series between fit and
In this paper, I discuss the implementation of test periods, updating versus recalibrating model
out-of-sample tests of forecasting accuracy. coefficients, single versus multiple test periods
Section 2 summarizes the rationale for out-of- and the use of rolling windows. Section 5
sample testing. Section 3 compares fixed-origin considers the role of out-of-sample testing in
and rolling-origin procedures. Section 4 ex- method selection. Section 6 describes the exten-
amines the application of out-of-sample testing sion of out-of-sample testing from an individual
time series to multiple time series and forecast-
ing competitions. Section 7 evaluates the
*Tel.: 11-802-425-3805; fax: 11-802-425-3806. adequacy of out-of-sample tests in forecasting
E-mail address: lentashman@compuserve.com (L.J. software. Section 8 contains my conclusions and
Tashman). recommendations.

0169-2070 / 00 / $ – see front matter  2000 International Institute of Forecasters. Published by Elsevier Science B.V. All rights reserved.
PII: S0169-2070( 00 )00065-0
438 L. J. Tashman / International Journal of Forecasting 16 (2000) 437 – 450

2. In-sample versus out-of-sample methods selected by best in-sample fit may not
evaluation best predict post-sample data. Bartolomei and
Sweet (1989) and Pant and Starbuck (1990)
Forecasters generally agree that forecasting provide particularly convincing evidence on this
methods should be assessed for accuracy using point.
out-of-sample tests rather than goodness of fit to One way to ascertain post-sample forecasting
past data (in-sample tests). ‘The performance of performance is to wait and see in real time. The
a model on data outside that used in its con- M2-competition (Makridakis et al., 1993) did
struction remains the touchstone for its utility in exactly this. In one phase, forecasts (for 1–15
all applications; (Fildes and Makridakis, 1995, months ahead) made in September 1987 were
p. 293). evaluated at the conclusion of 1988.
The argument has two related aspects. First, Real time assessment has practical limitations
for a given forecasting method, in-sample errors for forecasting practitioners, since a long wait
are likely to understate forecasting errors. Meth- may be necessary before a reliable picture of a
od selection and estimation are designed to forecasting track record will materialize. As a
calibrate a forecasting procedure to the histori- result, tests based on holdout samples have
cal data. But the nuances of past history are become commonplace. The fit period is used to
unlikely to persist into the future, and the identify and estimate a model (or method) while
nuances of the future may not have revealed the test period is reserved to assess the model’s
themselves in the past. forecasting accuracy.
Overfitting and structural changes may fur- If the forecaster withholds all data about
ther aggravate the divergence between in-sam- events occurring after the end of the fit period,
ple and post-sample performance. The M- the forecast-accuracy evaluation is structurally
competition (Makridakis et al., 1982) and many identical to the real-world-forecasting environ-
subsequent empirical studies show that forecast- ment, in which we stand in the present and
ing errors generally exceed in-sample errors, forecast the future. However, ‘peeking’ at the
even at reasonably short horizons. As well, held-out data while selecting the forecasting
prediction intervals built on in-sample standard method pollutes the evaluation environment.
errors are likely to be too narrow (Chatfield,
1993, p.131).
Moreover, common extrapolative forecasting 3. Fixed-origin versus rolling-origin
methods, such as exponential smoothing, are procedures
based on updating procedures, in which one
makes each forecast as if one were standing in An out-of-sample evaluation of forecasting
the immediately prior period. For updating accuracy begins with the division of the histori-
methods, the traditional measurement of good- cal data series into a fit period and a test period.
ness-of-fit is based on one step-ahead errors — The final time in the fit period (T ) — the point
errors made in estimating the next time period from which the forecasts are generated — is the
from the current time period. However, research forecasting origin. The number of time periods
shows (e.g., Schnaars, 1986, Exhibit 2, p.76) between the origin and the time being forecast
that errors in forecasting into the more distant is the lead time or the forecasting horizon. The
future will be larger than those made in fore- longest lead time is the N step-ahead forecast.
casting one step ahead. Equivalently, N denotes the length of the test
The second aspect to the argument is that period.
L. J. Tashman / International Journal of Forecasting 16 (2000) 437 – 450 439

3.1. Fixed-origin evaluations (1985, p. 343) provides a schematic illustration


of the rolling-origin procedure.
In performing an out-of-sample test, we can When N54, for example, the fixed-origin
use either a single forecasting origin or multiple evaluation results in four forecasts, all from
forecasting origins. The former can be called a origin T. The rolling-origin evaluation also
fixed-origin evaluation: Standing at origin (T ), generates four forecasts from this origin, but
we generate forecasts for time periods T 1 1, then supplies an additional three forecasts from
T 1 2, . . . T 1 N. By subtracting each of these origin T 11, two from origin T 12, and one
forecasts from the known data values of the test from origin T 13, for a total of 10 forecasts.
period, we determine the forecast errors. We can The total number of forecasts grows from 4 to
average the errors in various ways to obtain 10. In general, the rolling-origin procedure
summary statistics. provides N(N 1 1) / 2 forecasts, against N from
Applied to a single time series, the fixed- the fixed-origin. With eight time periods form-
origin evaluation has several shortcomings. ing the test set, for example, the rolling-origin
Because it yields only one forecast (and hence, evaluation supplies 36 forecasts, a multiple of
only one forecast error) for each lead time, it 4.5 times N.
requires a fairly long test period to produce a
forecasting track record. Second, forecasts gen- 3.3. Analysis of forecasting errors by lead
erated from a single origin are susceptible to time
corruption by occurrences unique to that origin.
Third, in the usual software implementation of a In contrast to the fixed-origin evaluation, the
fixed-origin evaluation, summary error measures rolling out-of-sample evaluation produces multi-
are computed by averaging forecasting errors ple forecasts for every lead time but the longest,
across lead times. The resulting summary statis- N. As a result, it permits us to assess the
´
tic is a melange of near-term and far-term forecasting accuracy of an individual time series
forecast errors. at each lead time. Moreover, the errors for a
We can partly overcome the three problems given lead time form a coherent empirical
by successively updating the forecasting origin. distribution, one we can profitably analyze for
We can also mitigate the problems by using further distributional information, such as out-
multiple time series. Still, even within a single- liers. Makridakis and Winkler (1989) describe
series context, the fixed-origin evaluation can such analysis.
play a useful role: it is the only way we can
assess the post-sample accuracy of forecasts,
such as judgmental forecasts, when we do not 4. Issues in implementing out-of-sample
know or can not replicate the underlying fore- evaluations
casting methodology.
In designing an out-of-sample test for an
3.2. Rolling-origin evaluations individual time series, the most fundamental
choice is how to split the series between fit and
In a rolling-origin evaluation, we successive- test periods. This decision determines the
ly update the forecasting origin and produce amount of data that will be available to identify
forecasts from each new origin. One of the first and fit a forecasting model and the number of
explicit descriptions of the procedure was Arm- forecasts generated for the out-of-sample
strong and Grohman’s (1972). Armstrong evaluation of the model’s performance.
440 L. J. Tashman / International Journal of Forecasting 16 (2000) 437 – 450

4.1. Series splitting rules Recalibration is the preferred procedure. Up-


dating without recalibrating imposes an arbit-
In deciding upon the appropriate number of rary handicap on the forecasting method. Re-
periods N to withhold from the time series, we calibration, moreover, desensitizes error mea-
can be guided by several considerations, the sures to events unique to the original fit period.
most important of which is the longest-term However, recalibration is more computationally
forecast required. Denote this maximal length intensive than simply updating, and only two of
forecast by H. Manifestly, N must be at least as 15 forecasting software packages examined by
large as H. Tashman and Hoover (2001) recalibrate as they
However, we may wish to increase the length update the forecasting origin.
of the test period to insure a certain minimum When it is a (causal) regression model under
number of forecasts M at lead time H. We evaluation, failure to recalibrate transforms a
would then set the length of the test period to rolling-origin evaluation into a fixed-origin
equal H 1 M 2 1 forecasts. If this minimum is evaluation at one step ahead and into meaning-
M 5 3, we should design a rolling-origin less figures at longer horizons. Without re-
evaluation with a test period of length H 1 2. calibration, the addition of a new data point
For example, if the longest-term forecast re- changes neither the inputs to nor the coefficients
quired is a five-year ahead forecast (H 5 5), we of the forecasting equation.
would specify a test period of seven years, thus For extrapolative methods, research is lacking
insuring that the assessment of accuracy in on the extent to which recalibration of the
forecasting five years ahead is based on a smoothing weights across the test period in-
minimum of three forecasts. We would need a fluences the reported absolute and relative ac-
much larger number of forecasts than this to curacy of forecasting methods. Fildes, Hibon,
examine a distribution of forecast errors, rather Makridakis and Meade (1998) provide evidence
than simply measures of average error. that recalibrating weights in fitting an exponen-
Short time series impose restrictions on the tial smoothing method improves the out-of-sam-
length of the test period, since truncating the ple accuracy of the method. However, they did
data could leave too few observations to fit the not recalibrate the smoothing weights within the
model. In this circumstance, we might profit test period of a rolling-origin evaluation.
from the efficiency of the rolling-origin pro- Similarly, no one has examined the empirical
cedure and still be able to examine one-step significance of recalibration in the context of
ahead forecast errors without greatly truncating out-of-sample evaluations of regression models.
the period of fit. If the model contains dynamic terms, such as a
lagged dependent variable or a lagged error,
4.2. Updating versus recalibrating each forecast will adjust as the origin is succes-
sively updated. Unless the sample size is small,
In the rolling-origin evaluation, each update these effects may be more substantial than the
of the forecasting origin leads to a revision of changes that arise from recalibrating the regres-
the forecasting equation.The successive revi- sion coefficients.
sions to the forecasting equation may arise
simply from the addition of a data point to the 4.3. Multiple test periods
fit period, or may arise as well from recalibra-
tion (reoptimization) of the smoothing weights Fildes (1992, p.82) observed that replacing a
as the new data point comes in. fixed-origin design with a rolling-origin design
L. J. Tashman / International Journal of Forecasting 16 (2000) 437 – 450 441

removes ‘the possibility that the arbitrary choice prudent. Perhaps individual test-period MAPEs
of time origin might unduly affect the [forecast- should be averaged. The average MAPE for
ing accuracy] results’ Distinguishing sensitivity Method A at four-steps head is 4.5 percent,
to outliers in the test period from sensitivity to which is the most broad-based indication of this
the phase of the business cycle, however, is method’s expected accuracy in forecasting four
useful. The test period marks a single calendar months into the future.
interval. Especially for monthly and quarterly Fildes et al. (1998) used multiple test periods,
data, therefore, it is likely to reflect a single which they called multiple origins, to compare
phase of the business cycle or single period of the accuracy of five designated extrapolative
business activity. To attain cyclical diversity in methods on a batch of monthly telecommunica-
analyzing an individual time series, we should tions time series. While they found that one
use multiple test periods. method was uniformly most accurate (across
Pack (1990) illustrated the virtues of multiple lead time and for every test period), the relative
test periods using a retail sales series of 95 accuracy of three of the other methods was not
consecutive months. For each of three forecast- consistent across test periods.
ing methods, he designated three distinct test Schnaars (1986) examined the cyclical sen-
periods, and performed a rolling-origin evalua- sitivity of forecast error measures by sorting all
tion for each test period. Table 1 is a portion of one year-ahead forecast errors by calendar year
his Exhibit 5 (p. 217). (1978–1984). He then compared forecast errors
The MAPEs are sensitive to the choice of test for (a) years in which cyclical turning points
period. For lead time 4, for example, forecasting occurred and (b) years in which the overall
method A earned a MAPE of 3.1 percent over direction of the economy did not change. For
test period 61–71; however, the same measure almost all of the methods included, he found
applied to test period 73–83 yielded a MAPE of that one-year-ahead forecasting accuracy was
5.8 percent, nearly twice as high. At lead time 1 poorer during the years of cyclical turning
in test period 85–95, the three methods appear points.
about equally accurate (MAPEs of 3.1%, 3.3% Using multiple test periods may be particu-
and 3.4%), while, in test period 73–83, method larly beneficial when we are limited by software
B looks significantly worse (at both lead times) to fixed-origin evaluations. However, the pro-
than the others. cedure requires a long time series.
Diversifying into multiple test periods seems
4.4. Rolling windows
Table 1
How the MAPE varies by lead time and test period in In a rolling-origin evaluation, each update of
comparing three methods
the forecasting origin adds one new observation
Lead time Method Test periods Average to the fit period. Alternatively, in some studies,
61–71 73–83 85–95 researchers have maintained a fit period (or
sample or window) of constant length. They do
1 A 3.0 4.1 3.1 3.4
B 3.2 5.0 3.3 3.8
this by pruning the oldest observation at each
C 2.3 2.7 3.4 2.8 update, much as we would in taking a moving
average. The procedure is called a fixed-size,
4 A 3.1 4.6 5.8 4.5 rolling window (Swanson and White, 1997) or
B 5.3 7.4 6.0 6.2 fixed-size rolling sample (Callen, Kwan, Yip
C 3.5 3.9 7.0 4.8
and Yuan, 1996).
442 L. J. Tashman / International Journal of Forecasting 16 (2000) 437 – 450

Why prune the fit period at each update of the for out-of-sample analysis.) Fildes (1989) also
forecasting origin? One reason is to ‘clean out used the procedure — under the name rolling
old data’ in an attempt to update model co- horizon — to compare the efficacy of various
efficients. Doing so may be unnecessary in method-selection rules.
common time-series methods, however, because The sliding simulation requires a three-way
the weighting systems in these methods mitigate division of the time series. N observations
the influence of data from the distant past. withheld from the time series serve as a test set.
Swanson and White (1997) discussed the The remaining period of fit is subdivided be-
usefulness of rolling windows in econometric tween the first T observations, which represent
modeling, particularly in determining how the in-sample fit period and the remaining P
econometric models evolve over time to fixed observations, T 11 to T 1 P, which constitute
specifications. the post-sample fit period.
For out-of-sample testing, the principal pur- For each method under consideration, the
pose of a rolling window is to level the playing sliding simulation entails a pair of rolling out-
field in a multiperiod comparison of forecasting of-sample evaluations. In the first, we optimize
accuracy. We might analyze whether a particular the smoothing weights to the post-sample fit
method’s performance deteriorates between an period, and select a best method for each lead
earlier and later test period. The comparison time. The second is performed on the test set,
would be confounded if the second fit period with the traditional purpose of evaluating the
were longer than the first. accuracy of the forecasts made with this meth-
Swanson and White (1997) further pruned od.
their rolling windows to generate the same In the same spirit, Weiss and Anderson (1984,
frequency of forecasts at each horizon of the p.485) proposed that, for cumulative forecasts, a
test period. They wished to ensure equality model be calibrated to minimize a cumulative
between the number of one step-ahead forecasts post-sample error measure.
and the number of four step-ahead forecasts. Makridakis (1990) applied variants of the
That procedure, however, results in a different sliding simulation to a subsample of 111 time
calendar fit period for each forecast horizon: the series used in the M-competition (Makridakis et
fit period for a four- step-ahead forecast will al., 1982). For each of three exponential
begin and end three periods earlier than the fit smoothing methods, post-sample forecasting
period underlying the one step-ahead forecasts. accuracy improved when he calibrated smooth-
As a result of the calendar shift, the evidence on ing weights to minimize a post-sample error
how forecasting accuracy of any method de- measure instead of calibrating weights in-sam-
teriorates as the forecasting horizon increases ple, as is traditional.
may be confounded. Results reported in the M2-Competition
(Makridakis et al., 1993) were not so positive
for the sliding simulation process. There, the
5. ‘Sliding simulations’ method chosen as best — from among simple,
damped, and linear-trend smoothing — did not
Makridakis (1990) extended the rolling-origin systematically outperform any individual
design to serve as a process for method selec- smoothing method (Exhibit 3, p.9). In fact, two
tion and estimation. He called this process a of the three smoothing methods performed more
sliding simulation. (He did not intend the term poorly when calibrated post-sample, the linear
simulation to mean a resampling or Monte trend being the exception.
Carlo process; he used it rather as a synonym Fildes (1989) used the sliding simulation to
L. J. Tashman / International Journal of Forecasting 16 (2000) 437 – 450 443

compare individual-selection and aggregate- competition (Makridakis et al., 1982) included


selection rules. When following an individual- 1001 time series, a compendium of annual,
selection rule, we identify a best method for quarterly, and monthly as well as firm, industry,
each time series in a batch. When following macroeconomic, and demographic data. ‘‘Al-
aggregate-selection rule, we apply to every though the [M-competition] sample is not ran-
series in the batch the method that works best in dom, efforts were made to select series covering
the aggregate. a wide spectrum of possibilities. This included
Fildes considered two extrapolative methods, different sources of statistical data and different
both involving damping of trends and smooth- starting / ending dates.’’ (p.113).
ing of outliers. He calibrated each method to a In contrast, selectivity was the principal ob-
post-sample fit period and chose the better of jective for Schnaars (1986). Schnaars wished to
the two methods based on post-sample fit. He ‘‘discover how well extrapolations are able to
concluded that the extra effort needed in in- perform on a specific type of data series —
dividual rather than aggregate selection was not annual unit sales by industry — rather than a
worth the small potential gain in accuracy for wide assortment of potentially disparate series.’’
forecasting one month ahead, the most impor- (p.72). Selectivity was also an objective for the
tant horizon when forecasting for inventory M2-competition (Makridakis et al., 1993). Of
control. At longer lead times, individual selec- its 29 time series, 23 were monthly firm-level
tion has more potential to improve accuracy. series, chosen to compare the accuracy of
designated methods in forecasting for budgeting
and capital investment.
6. Multiple time series: forecasting The diversity objective for the M-competition
competitions returns with the M3-competition (Makridakis
and Hibon, 2000), in which the database is
For a single time series, desirable characteris- enlarged from 1001 to 3003 time series. Again,
tics of an out-of-sample test are adequacy, the authors chose time series to represent data
enough forecasts at each lead time, and diversi- of different periodicities (yearly, quarterly,
ty, desensitizing forecast error measures to monthly, and other) and types (micro, industry,
special events and specific phases of business. macro, finance, demographic, and other). The
To achieve these goals with an individual time selection process was essentially downloading a
series, we must use rolling origins and multiple convenience sample of data from the Internet.
test periods. The emphasis in a forecasting competition
Alternatively, we can attain adequacy and affects both the selection of time series and the
diversity by using multiple time series. To implementation of the out-of-sample tests. With
promote adequacy, we need to select component the emphasis on diversity, the authors of the
series that are homogeneous in some relevant M-competition and the M3-competition amas-
characteristic. For diversity, we should collect sed a large collection of heterogeneous time
time series that are heterogeneous in both nature series, but relied on fixed-origin evaluations and
and calendar time, thus establishing a broad- a single test period per series to obtain post-
based track record for a forecasting method. sample error measures. In emphasizing selectivi-
Diversity was the primary motivation in the ty, Schnaars and the authors of the M2-competi-
early forecasting competitions. Newbold and tion employed a relatively small number of
Granger (1974) amassed 106 economic series, a homogeneous series and used rolling-origin
mixture of monthly and quarterly as well as of evaluations (Schnaars) and multiple test periods
micro-level and macro-level data. The M- (M2-competition) for diversity.
444 L. J. Tashman / International Journal of Forecasting 16 (2000) 437 – 450

The reliance on fixed-origin rather than roll- cycles. Moreover, the attributes are interdepen-
ing-origin evaluations in the three M-competi- dent in many ways: Seasonality is likely to be
tions was probably also essential for keeping the most pronounced in quarterly and monthly data,
forecasting process manageable. In these volatility greatest in micro level series, and
studies, participants provided forecasts to the trends strongest in macroeconomic data.
researchers, who had withheld the test period A perfectly stratified random sample, hence,
data. To implement a rolling-origin evaluation, is not a realistic possibility. Nevertheless, the
the participants would have had to be shown the competitions can be faulted for a lack of
test period data, so that they could successively formality in the collection of data. Series were
update the forecasting origins. In contrast, collected and retrospectively classified by attri-
Schnaars (1986) produced his own forecasts. bute. For this reason alone, tabulations based on
In principle, a synthesis of the diversity and ‘all series’ are suspect.
selectivity strategies is to be recommended.
Ideally, a forecasting competition would begin 6.1. Pooled data structure
with precisely described groups of time series,
The use of multiple time series, as in a
in which the series are homogeneous within forecasting competition, creates a pooled data
group but heterogeneous between groups. Ran- structure: S time series, s 5 1 to S, and up to
domized selection could then be used to obtain T 1 N time periods per series. Individual time
a sample of series from each group. series need not be of equal length nor need they
Armstrong et al. (1998, p. 360) observed that cover the same calendar period. Hence, the
within-group homogeneity abets method selec- periods of fit can vary in both length and
tion by helping the forecaster to determine calendar interval.
which methods are best suited to the specific The length of the test period, however, is
characteristics of the data. Within-group homo- normally fixed for all time series of a given
geneity can also be of value for forecasting periodicity. For example, Schnaars (1986) with-
product hierarchies. At the same time, the held the last five years from all the historical
forecaster needs heterogeneity among groups to series. In the three M-competitions, the test
draw general inferences about the relative fore- period was specified to be six years, eight
casting accuracy of different methods. quarters and 18 months for annual, quarterly
In practice, it is difficult to implement a and monthly data respectively.
random-sampling design. Time series are multi- Fixing the length of the test period is partly a
attributed: periodicity and type were the two matter of statistical convenience: it simplifies
explicit attributes in the forecasting competi- the calculation and presentation of forecast-error
tions. However, type is really a catchall de- averages. Still, considerable obfuscation can
scriptor, comprising level of aggregation (item, result if the forecast error measures are tabu-
product, brand, company, industry, economy), lated for an aggregate of series of different
domain (financial, marketing, operations), geog- periodicities. For the M-competition results, the
raphic area (country, region) and data charac- ‘all data’ tables combined monthly, quarterly
teristics (seasonal versus nonseasonal, stable and annual series. Thus, a one step-ahead error
versus volatile, trended versus untrended). figure blended the one-month-ahead, one-quar-
Another dimension of importance is calendar ter-ahead and one-year-ahead forecast errors.
time interval: Series differ in starting date, The M2-competition and M3-competition have
ending date, and length, and span different avoided this confusion by separately reporting
stages of economic cycles and product life results for series of different periodicities.
L. J. Tashman / International Journal of Forecasting 16 (2000) 437 – 450 445

6.2. Pooled averages cent error (MdAPE) may be preferable to


using the MAPE. MdAPE is the principal
To calculate forecast error statistics in a error statistic used in Vokurka, Flores and
multiseries data set, we can average errors Pearce (1996). Still another alternative to the
across time series, o s ; across lead time, o n ; or MAPE is the symmetric MAPE (Armstrong,
both, o sn . Precisely how the averaging is done 1985, p. 348), which makes underforecasts
can be important. and overforecasts of the same percent equal.
This statistic is being featured in the M3-
6.2.1. Choice of error statistic for averaging competition.
over series 3. Use relative error measures when it is neces-
Much has been written about the choice of sary to average over time series that differ in
forecast-error statistics. A good overview is volatility. Collopy and Armstrong proposed a
provided in a series of articles and commen- ratio of an absolute error from a designated
taries in the International Journal of Forecast- method to the analogous absolute error from
¨ method, which they call the relative
a naıve
ing (Armstrong and Collopy, 1992; Fildes,
absolute error, RAE (Collopy and Arm-
1992; Ahlburg et al., 1992).
strong, 1992, p.71). They showed that the
There are two arithmetic issues. One concerns
RAE is not only scale independent but also
the choice of error measure: Should we be
serves to standardize the component series
averaging squared errors, percent errors or
for degree of change and, hence, degree of
relative errors? The second deals with the
forecasting difficulty. Tashman and Kruk
appropriate statistical operator: should we use a
(1996) used relative error measures to com-
median, an arithmetic mean or geometric mean? pare the accuracy of a forecasting method
The lessons from the research are at least between distinct groups of time series. Rec-
threefold: When averaging over series o s , we ommended operators for averaging relative
should: errors are the median (MdRAE) and geomet-
ric mean (GMRAE). Fildes (1992, p.84)
1. Avoid scale dependent error measures, such
endorses a variant (calculable only in a
as root mean squared error RMSE or mean
rolling-origin evaluation) called the relative
absolute deviation MAD. With these, if you
geometric root mean square.
rescale the measurement (for example, from
one currency to another or from millions of By using a single summation o s , we obtain
units to thousands of units), you alter the an average error for an individual method at a
numerical value of the error measure. More- specific horizon. In reporting the M-competition
over, a subset of the time series with large results, the authors refer an average of absolute
numerical values may dominate the error percent errors (APEs) as an average MAPE
measures, and that subset would change with (Makridakis et al., 1982, Table 2). For an
the scaling. individual lead time, however, it may be called
2. Use percent error measures instead, such as simply a MAPE, without the preceding average,
the absolute percent error (APE), because since we are averaging a single APE per time
(for data with a natural zero) they are scale series.
independent. However, the distribution of
percent errors can be badly skewed, especial- 6.2.2. Cumulating over lead times
ly if the series contains values close to zero. For cumulative lead time error measures,
In this case using the median absolute per- such as 1–4 quarters or 1–12 months ahead, we
446 L. J. Tashman / International Journal of Forecasting 16 (2000) 437 – 450

can use a double summation o sn , summing relative accuracy (ranking) of different forecast-
individual APEs over both the series and the ing methods changed appreciably as the fore-
lead times. Doing so gives equal weight to casting origin varied. Such instability, they
errors at short and long lead times. Alternative- concluded, should discourage forecasters from
ly, we can start with each individual lead time using a single forecasting origin.
MAPE and then take an average or weighted Whether their concern extends to the fore-
average across lead times, o n MAPE. The latter casting competitions is uncertain. Their time
properly requires a modifier such as average series were of equal length and had identical
MAPE. starting and ending dates. The series in the
The route taken for calculating cumulative M-competition and in the M3-competition have
lead time error measures can make a difference. considerable diversity in length and calendar
Using the o n approach maintains the distinctive- dates.
ness of the individual lead times and thus Calendar diversity plays the same role in
permits flexibility in assigning weights to reflect multiseries evaluations that multiple test periods
the relative importance of the individual play in individual-series evaluations: Both miti-
horizons. Moreover, in a rolling-origin evalua- gate the sensitivity of forecast error measures to
tion, the alternative o sn approach would assign the phase of the business cycle.
greater weight for the first lead time, successive-
ly smaller weights for each longer lead. If equal 6.4. Method selection rules
weighting of each lead time is desired, the o n
MAPE calculation is preferred. In the forecasting competitions, every fore-
Sensitivity to outliers can be mitigated in casting method was applied to every time series,
both approaches. With the doubly summed whether or not the method was appropriate for
measure, we can calculate a median absolute the series. For example, Holt’s exponential
percent error MdAPE or we can employ the smoothing method was applied to nontrended
median MAPE, as do Tashman and Kruk (1996, series, and simple exponential smoothing was
Table 7). applied to trended series. Tashman and Kruk
For measuring forecast accuracy over a (1996, p. 5) call this unselective application and
cumulative lead time, Collopy and Armstrong argue that, by fusing appropriate and inappro-
propose the cumulative RAE (Collopy and priate cases, unselective application tends to
Armstrong, 1992, p. 75–76). denigrate a method’s expected performance.
The alternative is to first screen out those series
6.3. Stability of error measures across for which a method is judged inappropriate.
forecasting origins Effective screening, however, requires a reliable
method-selection rule.
Pooling time series and cross-sectional data Fildes (1989) articulated the distinction be-
can create analytical and interpretational dif- tween (a) knowledge of a method’s forecasting
ficulties. Normally, as a precondition of pooling, accuracy after a test and (b) the ability to select
we perform tests to see if the parameters of a best method in advance. ‘Forecasting competi-
cross-sectional models are stable over time. tions, such as the M-competition, only offer the
Fildes et al. (1998) used a data set of 263 forecaster information on the relative accuracy
telecommunications series to examine the of (methods) A and B, ex post; these show
stability of error measures across forecasting which of the two turned out to be better; but
origins. Their results, similar to those reported they do not demonstrate how to pick a winner’
earlier from Pack (1990), indicate that the (1989, p. 1057).
L. J. Tashman / International Journal of Forecasting 16 (2000) 437 – 450 447

Effective method selection, ex ante, requires product-class aggregate, thus effectively impos-
effective method-selection rules. Among the ing the structure of the product-class series on
forecasting competitions, the M3-competition the individual components. Doing so is ap-
(Makridakis and Hibon, 2000) is the first to pealing when individual item series are short
examine automatic forecasting systems, many of and irregular.
which incorporate method-selection rules. Al- Testing product hierarchy methodologies
though the M3-competition summary tables do should be a high priority for future research.
not include a direct comparison of the category
of automatic forecasting systems against the
aggregate of single-method procedures, auto- 7. Out-of-sample evaluations in forecasting
matic systems were found to be among the software
methods that give best results for many types of
time series. In a review of 13 business-forecasting pro-
This result is more promising than prior grams with automatic forecasting features, Tas-
research would have suggested. Gardner and hman and Leach (1991) reported that only six
McKenzie (1988) offered selection rules for programs included post-sample tests of forecast-
choosing among exponential smoothing proce- ing accuracy. Of these, moreover, all but two
dures. Tashman and Kruk (1996) compared the were limited to fixed-origin evaluations on a
Gardner–McKenzie protocol with two other single series. In the two packages that offered
protocols for method selection. They found that rolling-origin evaluations, the implementation
(1) none of method-selection protocols effec- was based on a single series in a single test
tively identified an appropriate smoothing pro- period and model coefficients that were held
cedure for time series that lacked strong trends, fixed rather than recalibrated through the test
(2) the protocols frequently disagreed as to what period. While the authors warned forecasting
constituted an appropriate method, and (3) even practitioners to evaluate those methods the
when they agreed on an appropriate method, software selected automatically, the forecasting
following their advice did not ensure improved software of the early 1990s did not facilitate this
forecasting accuracy (1996, p. 252). process.
Has out-of-sample testing in forecasting soft-
6.5. Product hierarchies ware been upgraded during the past decade? Of
the 13 programs Tashman and Leach investi-
While the authors of the forecasting competi- gated, 10 have ceased to exist. In the remaining
tions have classified time series by periodicity three, Autobox, Forecast Pro and SmartFore-
and level of aggregation, they have not incorpo- casts, the developers have enhanced their post-
rated hierarchical data structures. New tech- sample testing options All three now offer
niques for demand forecasting have emerged in rolling out-of-sample evaluations and a variety
the past decade that link forecasts for one item of forecast error measures.
(stock keeping unit) to the product class to During the 1990s, the forecasting software
which the item belongs. For example, Bunn and market has seen many new entrants. Tashman
Vassilopoulis (1993) showed how the seasonal and Hoover (2001) examined 15 forecasting
pattern in the product class aggregate could be software programs, of which 9 had their roots in
applied effectively to forecast the seasonality in the 1990s. They divided the forecasting pack-
individual items. Several forecasting programs ages into four categories: spreadsheet add-ins,
permit automatic adjustment of forecasts for forecasting modules of general statistical pro-
individual items to reconcile them with the grams, neural-network programs, and dedicated
448 L. J. Tashman / International Journal of Forecasting 16 (2000) 437 – 450

business-forecasting programs. The last cate- market segment have been published. Develop-
gory included the three aforementioned pack- ers of demand planning packages have focused
ages plus Time Series Expert and tsMetrix. on the technology of managing forecasting
Tashman and Hoover (2001, Table 4) re- databases and automating forecasting methods.
ported that only one of the three spreadsheet This focus has come at the expense of trans-
add-ins and one of the four general statistical parency regarding how forecasts are made and
programs effectively distinguished within-sam- what forecast errors to expect. Useful out-of-
ple from out-of-sample forecasting accuracy. In sample tests are seldom included in this type of
contrast, two of the three neural-network pack- program.
ages and three of the five dedicated business- Forecast Pro, SmartForecasts and Autobox,
forecasting programs made this distinction ef- which can serve as forecasting engines in a
fectively. demand planning package, are major excep-
In my further analysis of the 12 non-neural tions. These programs enable users to view
network programs (software references are at average forecast errors made on an entire batch
the end of the paper), I found that none of the of time series. The programs perform rolling-
four general statistical programs and none of the origin evaluations on individual time series,
three spreadsheet add-ins offered a rolling out- sorts the forecasting errors by lead time and
of-sample evaluation. In addition, most of these then report averages of the forecast errors across
include a limited set of error measures: their time series.
developers essentially ignore the recent litera-
ture on forecast error measurement.
Within the category of dedicated business- 8. Summary
forecasting software, tsMetrix comes closest to
providing the opportunity for systematic out-of- For an individual time series, out-of-sample
sample tests on individual series. Once the user testing of forecasting accuracy is facilitated by
selects a test period, the program will perform a use of rolling-origin evaluations. The rolling-
rolling-origin evaluation, recalibrating the co- origin procedure permits more efficient series-
efficients of the forecasting equations at each splitting rules, allows for distinct error distribu-
update of the origin. This option is available for tions by lead time, and desensitizes the error
smoothing, ARIMA, and regression methods. measures to special events at any single origin.
Users can define multiple test periods; however, Applying the procedure across multiple test
the program does not integrate error measures periods is desirable to mitigate the sensitivity of
across test periods. error measures to single phases of the business
The post-sample procedure in Autobox mat- cycle. In an implementation of a rolling-origin
ches that in tsMetrix, although it is available evaluation, recalibration of the parameters of a
only for ARIMA modeling. The Forecast Pro forecasting equation can be important in general
procedure is also similar, except that it does not and is essential in the context of a regression
recalibrate coefficients with each update of the model.
forecasting origin. Forecasting software does not always nurture
A major growth segment of the forecasting the proper implementation of post-sample tests.
software market has been demand planning Many programs permit only fixed-origin evalua-
packages, which incorporate automatic batch tions and report few error measures. Those that
forecasting for large product hierarchies. Un- offer rolling-origin evaluations often restrict
fortunately, few reviews and evaluations of this them to certain methods, usually extrapolative.
L. J. Tashman / International Journal of Forecasting 16 (2000) 437 – 450 449

Few demand planning packages incorporate Chatfield, C. (1993). Calculating interval forecasts. Jour-
useful out-of-sample evaluations. nal of Business and Economic Statistics 11, 121–135.
Collopy, F., & Armstrong, J. S. (1992). Rule-based
Forecasting competitions would be more
forecasting. Management Science 38, 1394–1414.
generalizable if based upon precisely described Fildes, R. (1989). Evaluation of aggregate versus indi-
groups of time series, in which the series were vidual forecast method selection rules. Management
homogeneous within group and heterogeneous Science 35, 1056–1065.
between groups. Even a large collection of time Fildes, R. (1992). The evaluation of extrapolative forecast-
series does not automatically ensure diversity of ing methods. International Journal of Forecasting 8,
forecasting situations, especially if calendar 81–98.
dates are more or less coterminous. Measures Fildes, R., Hibon, M., Makridakis, S., & Meade, N.
(1998). Generalising about univariate forecasting meth-
based on a single cross-section can be unstable ods: further empirical evidence. International Journal of
over time. Error statistics that are calculated by Forecasting 14, 339–358.
applying every method to every time series may Fildes, R., & Makridakis, S. (1995). The impact of
give misleading results. Evaluating methods empirical accuracy studies on time series analysis and
used in forecasting product hierarchies remain forecasting. International Statistical Review 63, 289–
an important avenue for further research. 308.
Gardner, Jr. E. S., & McKenzie, E. (1988). Model
identification in exponential smoothing. Journal of the
Operational Research Society 3, 863–867.
References
Makridakis, S. (1990). Sliding simulation: a new approach
to time series forecasting. Management Science 36,
Ahlburg, D. A., Chatfield, C., Taylor, S. J., Thompson, P.
505–512.
A., Winkler, R. L., Murphy, A. H., Collopy, F., &
Makridakis, S., Anderson, A., Carbone, R., Fildes, R.,
Fildes, R. (1992). A commentary on error measures.
Hibon, M., Lewandowski, R., Newton, J., Parzen, P., &
International Journal of Forecasting 8, 99–111.
Winkler, R. (1982). The accuracy of extrapolation (time
Armstrong, J. S. (1985). Long-range forecasting, Wiley-
series) methods: results of a forecasting competition.
Interscience, New York.
Journal of Forecasting 1, 111–153.
Armstrong, J. S., & Collopy, F. (1992). Error measures for
generalising about forecasting methods: empirical com- Makridakis, S., Chatfield, C., Hibon, M., Lawrence, M.,
parisons. International Journal of Forecasting 8, 69– Mills, T., Ord, J. K., & Simmons, L. F. (1993). The M2
80. competition: a real life judgmentally-based forecasting
Armstrong, J. S., & Grohman, M. C. (1972). A compara- study. International Journal of Forecasting 9, 5–29.
tive study of methods for long-range market forecasting. Makridakis, S., & Hibon, M. (2000). The M3-competition:
Management Science 19, 211–221. results, conclusions and implications. International
Armstrong, J. S., Koehler, A. B., Fildes, R., Hibon, M., Journal of Forecasting 16, 451–476.
Makridakis, S., & Meade, N. (1998). Commentaries on Makridakis, S., & Winkler, R. L. (1989). Sampling
‘Generalizing about univariate forecasting methods: distribution of post-sample forecasting errors. Applied
further empirical evidence’. International Journal of Statistics 38, 331–342.
Forecasting 14, 359–366. Newbold, P., & Granger, C. W. J. (1974). Experience with
Bartolomei, S. M., & Sweet, A. L. (1989). A note on a forecasting univariate time series and the combination
comparison of exponential smoothing methods for of forecasts. Journal of the Royal Statistical Society (A)
forecasting seasonal series. International Journal of 137, 131–165.
Forecasting 5, 111–116. Pack, D. J. (1990). In defense of ARIMA modeling.
Bunn, D. W., & Vassilopoulis, A. I. (1993). Using group International Journal of Forecasting 6, 211–218.
seasonal indices in multi-item short-term forecasting. Pant, P. N., & Starbuck, W. H. (1990). Innocents in the
International Journal of Forecasting 9, 517–526. forest: forecasting and research methods. Journal of
Callen, J. L., Kwan, C. C. Y., Yip, P. C. Y., & Yuan, Y. Management 16, 433–460.
(1996). Neural network forecasting of quarterly ac- Schnaars, S. P. (1986). A comparison of extrapolation
counting earnings. International Journal of Forecasting procedures on yearly sales forecasts. International Jour-
12, 475–482. nal of Forecasting 2, 71–85.
450 L. J. Tashman / International Journal of Forecasting 16 (2000) 437 – 450

Swanson, N. R., & White, H. (1997). Forecasting econ- CB Predictor: forecasting software for Microsoft Excel,
omic time series using flexible versus fixed specification Version 1 (1999). Decisioneering, Inc., 1515 Arapahoe
and linear versus nonlinear econometric models. Inter- Street, Suite 1330, Denver, CO 80202
national Journal of Forecasting 13, 439–461. Forecast Pro, Version 4 (1999) and Forecast Pro Un-
Tashman, L.J., and Hoover, J.H. (2001). Diffusion of limited (1999). Business Forecast Systems, Inc., 68
forecasting principles: an assessment of forecasting Leonard Street, Belmont, MA. 02178
software programs. In J. Scott Armstrong, Principles of SAS /ETS, Version 7 (1997–99). SAS Institute, Inc., SAS
forecasting: a handbook for researchers and practition- Campus Drive, Cary, NC 27513-2414
ers. Norwell, MA: Kluwer Academic Publishers (in Insight.xla: business analysis software for Microsoft Excel,
press). Version 1 (1998). Sam Savage, Duxbury Press.
Tashman, L. J., & Kruk, J. M. (1996). The use of protocols Minitab, Release 11 (1997). Minitab, Inc., 3081 Enterprise
to select exponential smoothing methods: a reconsidera- Drive, State College, PA 16801-3008
tion of forecasting competitions. International Journal SmartForecasts for Windows, Version 5 (1999). Smart
of Forecasting 12, 235–253. Software, Inc., 4 Hill Road, Belmont, MA 02178
Tashman, L. J., & Leach, M. L. (1991). Automatic Soritec for Windows, Version 1 (1998). Full Information
forecasting software: a survey and evaluation. Interna- Software, Inc., 6417 Loisdale Road, Suite 200, Spring-
tional Journal of Forecasting 7, 209–230. field, VA, 2215-1811
Vokurka, R. J., Flores, B. E., & Pearce, S. (1996). SPSS Trends, Version 8 for Windows (1998). SPSS, Inc.,
Automatic feature identification and graphical support in 444 North Michigan Avenue, Chicago, IL 60611
rule-based forecasting: a comparison. International Time Series Expert, Version 2.31 (1998). Statistical Insti-
Journal of Forecasting 12, 495–512. tute of the Free University of Brussels (Contact person:
Weiss, A. A., & Anderson, A. P. (1984). Estimating time Professor Guy Melard, gmelard@ulb.ac.be)
series models using relevant forecast evaluation criteria. tsMetrix, Version 2 (1997). RER, Inc.,12520 High Bluff
Journal of the Royal Statistical Society (A) 147, 484– Drive, Suite 220, San Diego, CA 92130
487.
Biography: Len TASHMAN is on the faculty of the
School of Business Administration of the University of
Software References Vermont. He has contributed articles to several forecasting
journals and has published many evaluations of forecasting
Autobox for Windows, Version 5 (1999). AFS Inc., PO Box software.
563, Hatboro, PA 19040

View publication stats

S-ar putea să vă placă și