Business Forecasting John E. Hanke Dean Wichern Ninth Edition
Business Forecasting John E. Hanke Dean Wichern Ninth Edition
Table of Contents
1. Introduction to Forecasting
John E. Hanke/Dean Wichern 1
2. Exploring Data Patterns and an Introduction to Forecasting Techniques
John E. Hanke/Dean Wichern 15
3. Moving Averages and Smoothing Methods
John E. Hanke/Dean Wichern 61
4. Time Series and Their Components
John E. Hanke/Dean Wichern 119
5. Simple Linear Regression
John E. Hanke/Dean Wichern 175
6. Multiple Regression Analysis
John E. Hanke/Dean Wichern 235
7. Regression with Time Series Data
John E. Hanke/Dean Wichern 295
8. The Box-Jenkins (ARIMA) Methodology
John E. Hanke/Dean Wichern 355
9. Judgmental Forecasting and Forecast Adjustments
John E. Hanke/Dean Wichern 437
10. Managing the Forecasting Process
John E. Hanke/Dean Wichern 459
Appendix: Tables
John E. Hanke/Dean Wichern 477
Appendix: Data Sets and Databases
John E. Hanke/Dean Wichern 487
Index 501
I
EXPLORING DATA PATTERNS
AND AN INTRODUCTION TO
FORECASTING TECHNIQUES
One of the most time-consuming and difficult parts of forecasting is the collection of
valid and reliable data. Data processing personnel are fond of using the expression
“garbage in, garbage out” (GIGO). This expression also applies to forecasting. A fore-
cast can be no more accurate than the data on which it is based. The most sophisticated
forecasting model will fail if it is applied to unreliable data.
Modern computer power and capacity have led to an accumulation of an incred-
ible amount of data on almost all subjects. The difficult task facing most forecasters
is how to find relevant data that will help solve their specific decision-making
problems.
Four criteria can be applied when determining whether data will be useful:
1. Data should be reliable and accurate. Proper care must be taken that data are col-
lected from a reliable source with proper attention given to accuracy.
2. Data should be relevant. The data must be representative of the circumstances for
which they are being used.
3. Data should be consistent. When definitions concerning data collection change,
adjustments need to be made to retain consistency in historical patterns. This
can be a problem, for example, when government agencies change the mix or
“market basket” used in determining a cost-of-living index. Years ago personal
computers were not part of the mix of products being purchased by consumers;
now they are.
4. Data should be timely. Data collected, summarized, and published on a timely
basis will be of greatest value to the forecaster. There can be too little data (not
enough history on which to base future outcomes) or too much data (data from
irrelevant historical periods far in the past).
Generally, two types of data are of interest to the forecaster. The first is data col-
lected at a single point in time, be it an hour, a day, a week, a month, or a quarter. The
second is observations of data made over time. When all observations are from the
same time period, we call them cross-sectional data. The objective is to examine such
data and then to extrapolate or extend the revealed relationships to the larger popu-
lation. Drawing a random sample of personnel files to study the circumstances of the
employees of a company is one example. Gathering data on the age and current
maintenance cost of nine Spokane Transit Authority buses is another. A scatter dia-
gram such as Figure 1 helps us visualize the relationship and suggests age might be
used to help in forecasting the annual maintenance budget.
From Chapter 3 of Business Forecasting, Ninth Edition. John E. Hanke, Dean W. Wichern.
Copyright © 2009 by Pearson Education, Inc. All rights reserved.
Exploring Data Patterns and an Introduction to Forecasting Techniques
Any variable that consists of data that are collected, recorded, or observed over
successive increments of time is called a time series. Monthly U.S. beer production is an
example of a time series.
A time series consists of data that are collected, recorded, or observed over suc-
cessive increments of time.
25
Cyclical Peak
20
Cost
15
Cyclical Valley
Trend Line
10
0 10 20
Year
Many macroeconomic variables, like the U.S. gross national product (GNP),
employment, and industrial production exhibit trendlike behavior. Figure 10 contains
another example of a time series with a prevailing trend. This figure shows the growth
of operating revenue for Sears from 1955 to 2004.
The trend is the long-term component that represents the growth or decline in
the time series over an extended period of time.
When observations exhibit rises and falls that are not of a fixed period, a cyclical
pattern exists. The cyclical component is the wavelike fluctuation around the trend that
is usually affected by general economic conditions. A cyclical component, if it exists,
typically completes one cycle over several years. Cyclical fluctuations are often influ-
enced by changes in economic expansions and contractions, commonly referred to as
the business cycle. Figure 2 also shows a time series with a cyclical component. The
cyclical peak at year 9 illustrates an economic expansion and the cyclical valley at year
12 an economic contraction.
1,100
1,000
Kilowatts 900
800
700
600
500
Water Power residential customers is highest in the first quarter (winter months) of
each year. Figure 14 shows that the quarterly sales for Coastal Marine are typically low
in the first quarter of each year. Seasonal variation may reflect weather conditions,
school schedules, holidays, or length of calendar months.
The seasonal component is a pattern of change that repeats itself year after year.
When a variable is measured over time, observations in different time periods are fre-
quently related or correlated. This correlation is measured using the autocorrelation
coefficient.
Data patterns, including components such as trend and seasonality, can be studied
using autocorrelations. The patterns can be identified by examining the autocorrela-
tion coefficients at different time lags of a variable.
The concept of autocorrelation is illustrated by the data presented in Table 1. Note
that variables Yt - 1 and Yt - 2 are actually the Y values that have been lagged by one and
two time periods, respectively. The values for March, which are shown on the row for
time period 3, are March sales, Yt = 125; February sales, Yt - 1 = 130; and January sales,
Yt - 2 = 123.
Exploring Data Patterns and an Introduction to Forecasting Techniques
1 January 123
2 February 130 123
3 March 125 130 123
4 April 138 125 130
5 May 145 138 125
6 June 142 145 138
7 July 141 142 145
8 August 146 141 142
9 September 147 146 141
10 October 157 147 146
11 November 150 157 147
12 December 160 150 157
Equation 1 is the formula for computing the lag k autocorrelation coefficient (rk)
between observations, Yt and Yt - k that are k periods apart.
a 1Yt - Y 21Yt - k - Y 2
n
t=k+1
rk = k = 0, 1, 2, Á (1)
a 1Yt - Y 2
n
2
t=1
where
rk = the autocorrelation coefficient for a lag of k periods
Y = the mean of the values of the series
Yt = the observation in time period t
Yt - k = the observation k time periods earlier or at time period t - k
Example 1
Harry Vernon has collected data on the number of VCRs sold last year by Vernon’s Music
Store. The data are presented in Table 1. Table 2 shows the computations that lead to the cal-
culation of the lag 1 autocorrelation coefficient. Figure 4 contains a scatter diagram of the
pairs of observations (Yt, Yt - 1). It is clear from the scatter diagram that the lag 1 correlation
will be positive.
The lag 1 autocorrelation coefficient (r1), or the autocorrelation between Yt and Yt-1 ,
is computed using the totals from Table 2 and Equation 1. Thus,
a 1Yt-1 - Y 21Yt - Y2
n
t=1+1 843
r1 = = = .572
a 1Yt - Y2
n
2 1,474
t=1
As suggested by the plot in Figure 4, positive lag 1 autocorrelation exists in this time
series. The correlation between Yt and Yt-1 or the autocorrelation for time lag 1, is .572. This
means that the successive monthly sales of VCRs are somewhat correlated with each other.
This information may give Harry valuable insights about his time series, may help him pre-
pare to use an advanced forecasting method, and may warn him about using regression
analysis with his data.
Exploring Data Patterns and an Introduction to Forecasting Techniques
a 1Yt - Y21Yt - 2 - Y2
n
t=2+1 682
r2 = = = .463
a 1Yt - Y2
n
2 1,474
t=1
Yt
Yt!1
It appears that moderate autocorrelation exists in this time series lagged two time peri-
ods. The correlation between Yt and Yt - 2, or the autocorrelation for time lag 2, is .463. Notice
that the autocorrelation coefficient at time lag 2 (.463) is less than the autocorrelation coef-
ficient at time lag 1 (.572). Generally, as the number of time lags (k) increases, the magni-
tudes of the autocorrelation coefficients decrease.
Figure 5 shows a plot of the autocorrelations versus time lags for the Harry Vernon
data used in Example 1. The horizontal scale on the bottom of the graph shows each time
lag of interest: 1, 2, 3, and so on.The vertical scale on the left shows the possible range of an
autocorrelation coefficient, -1 to 1. The horizontal line in the middle of the graph repre-
sents autocorrelations of zero.The vertical line that extends upward above time lag 1 shows
an autocorrelation coefficient of .57, or r1 = .57. The vertical line that extends upward
above time lag 2 shows an autocorrelation coefficient of .46, or r2 = .46. The dotted lines
and the T (test) and LBQ (Ljung-Box Q) statistics displayed in the Session window will be
discussed in Examples 2 and 3. Patterns in a correlogram are used to analyze key features
of the data, a concept demonstrated in the next section. The Minitab computer package
(see the Minitab Applications section at the end of the chapter for specific instructions) can
be used to compute autocorrelations and develop correlograms.
With a display such as that in Figure 5, the data patterns, including trend and sea-
sonality, can be studied. Autocorrelation coefficients for different time lags for a vari-
able can be used to answer the following questions about a time series:
1. Are the data random?
2. Do the data have a trend (are they nonstationary)?
3. Are the data stationary?
4. Are the data seasonal?
If a series is random, the autocorrelations between Yt and Yt - k for any time lag k
are close to zero. The successive values of a time series are not related to each other.
If a series has a trend, successive observations are highly correlated, and the auto-
correlation coefficients typically are significantly different from zero for the first sev-
eral time lags and then gradually drop toward zero as the number of lags increases.
The autocorrelation coefficient for time lag 1 will often be very large (close to 1). The
autocorrelation coefficient for time lag 2 will also be large. However, it will not be as
large as for time lag 1.
If a series has a seasonal pattern, a significant autocorrelation coefficient will occur
at the seasonal time lag or multiples of the seasonal lag. The seasonal lag is 4 for quar-
terly data and 12 for monthly data.
How does an analyst determine whether an autocorrelation coefficient is signifi-
cantly different from zero for the data of Table 1? Quenouille (1949) and others have
demonstrated that the autocorrelation coefficients of random data have a sampling dis-
tribution that can be approximated by a normal curve with a mean of zero and an
approximate standard deviation of 1> 1n. Knowing this, the analyst can compare the
sample autocorrelation coefficients with this theoretical sampling distribution and deter-
mine whether, for given time lags, they come from a population whose mean is zero.
Actually, some software packages use a slightly different formula, as shown in
Equation 2, to compute the standard deviations (or standard errors) of the autocorre-
lation coefficients. This formula assumes that any autocorrelation before time lag k is
different from zero and any autocorrelation at time lags greater than or equal to k is
zero. For an autocorrelation at time lag 1, the standard error 1> 1n is used.
1 + 2 a r 2i
k-1
i=1
SE1rk2 = (2)
T n
where
SE1rk2 = the standard error 1estimated standard deviation2
of the autocorrelation at time lag k
ri = the autocorrelation at time lag i
k = the time lag
n = the number of observations in the time series
This computation will be demonstrated in Example 2. If the series is truly random,
almost all of the sample autocorrelation coefficients should lie within a range specified
by zero, plus or minus a certain number of standard errors. At a specified confidence
level, a series can be considered random if each of the calculated autocorrelation coef-
ficients is within the interval about 0 given by 0 ± t × SE(rk), where the multiplier t is an
appropriate percentage point of a t distribution.
Exploring Data Patterns and an Introduction to Forecasting Techniques
Q = n1n + 22 a
m
r 2k
(3)
k=1 n - k
where
n = the number of observations in the time series
k = the time lag
m = the number of time lags to be tested
rk = the sample autocorrelation function of the residuals lagged k time periods
Yt = c + "t (4)
Are the data in Table 1 consistent with this model? This issue will be explored in
Examples 2 and 3.
Example 2
A hypothesis test is developed to determine whether a particular autocorrelation coefficient
is significantly different from zero for the correlogram shown in Figure 5. The null and alter-
native hypotheses for testing the significance of the lag 1 population autocorrelation coeffi-
cient are
H0 : r1 = 0
H1 : r1 Z 0
r1 - r1 r1 - 0 r1
t = = = (5)
SE1r12 SE1r12 SE1r12
Exploring Data Patterns and an Introduction to Forecasting Techniques
The critical values #2.2 are the upper and lower .025 points of a t distribution with 11
degrees of freedom. The standard error of r1 is SE1r12 = 11>12 = 1.083 = .289, and the
value of the test statistic becomes
r1 .572
t = = = 1.98
SE1r12 .289
Using the decision rule above, H0 : r1 = 0 cannot be rejected, since -2.2 < 1.98 < 2.2.
Notice the value of our test statistic, t = 1.98, is the same as the quantity in the Lag 1 row
under the heading T in the Minitab output in Figure 5. The T values in the Minitab out-
put are simply the values of the test statistic for testing for zero autocorrelation at the
various lags.
To test for zero autocorrelation at time lag 2, we consider
H0 : r2 = 0
H1 : r2 Z 0
r2 - r2 r2 - 0 r2
t = = =
SE1r22 SE122 SE1r22
Using Equation 2,
1 + 2 a r 2i 1 + 2 a r 2i
k-1 2-1
and
.463
t = = 1.25
.371
This result agrees with the T value for Lag 2 in the Minitab output in Figure 5.
Using the decision rule above, H0 : r1 = 0 cannot be rejected at the .05 level, since
-2.2 $ 1.25 $ 2.2. An alternative way to check for significant autocorrelation is to construct,
say, 95% confidence limits centered at 0. These limits for time lags 1 and 2 are as follows:
Autocorrelation significantly different from 0 is indicated whenever a value for rk falls out-
side the corresponding confidence limits. The 95% confidence limits are shown in Figure 5 by
the dashed lines in the graphical display of the autocorrelation function.
Example 3
Minitab was used to generate the time series of 40 pseudo-random three-digit numbers
shown in Table 3. Figure 6 shows a time series graph of these data. Because these data are
random (independent of one another and all from the same population), autocorrelations for
Exploring Data Patterns and an Introduction to Forecasting Techniques
t Yt t Yt t Yt t Yt
all time lags should theoretically be equal to zero. Of course, the 40 values in Table 3 are only
one set of a large number of possible samples of size 40. Each sample will produce different
autocorrelations. Most of these samples will produce sample autocorrelation coefficients that
are close to zero. However, it is possible that a sample will produce an autocorrelation coef-
ficient that is significantly different from zero just by chance.
Next, the autocorrelation function shown in Figure 7 is constructed using Minitab. Note
that the two dashed lines show the 95% confidence limits. Ten time lags are examined, and all
the individual autocorrelation coefficients lie within these limits. There is no reason to doubt
that each of the autocorrelations for the first 10 lags is zero. However, even though the individ-
ual sample autocorrelations are not significantly different from zero, are the magnitudes of
the first 10 rk’s as a group larger than one would expect under the hypothesis of no autocorre-
lation at any lag? This question is answered by the Ljung-Box Q (LBQ in Minitab) statistic.
If there is no autocorrelation at any lag, the Q statistic has a chi-square distribution
with, in this case, df = 10. Consequently, a large value for Q (in the tail of the chi-square dis-
tribution) is evidence against the null hypothesis. From Figure 7, the value of Q (LBQ) for
10 time lags is 7.75. From Table 4 in Appendix: Tables, the upper .05 point of a chi-square
distribution with 10 degrees of freedom is 18.31. Since 7.75 < 18.31, the null hypothesis can-
not be rejected at the 5% significance level. These data are uncorrelated at any time lag, a
result consistent with the model in Equation 4.
Yt
1955 3,307 1967 7,296 1979 17,514 1991 57,242 2003 23,253
1956 3,556 1968 8,178 1980 25,195 1992 52,345 2004 19,701
1957 3,601 1969 8,844 1981 27,357 1993 50,838
1958 3,721 1970 9,251 1982 30,020 1994 54,559
1959 4,036 1971 10,006 1983 35,883 1995 34,925
1960 4,134 1972 10,991 1984 38,828 1996 38,236
1961 4,268 1973 12,306 1985 40,715 1997 41,296
1962 4,578 1974 13,101 1986 44,282 1998 41,322
1963 5,093 1975 13,639 1987 48,440 1999 41,071
1964 5,716 1976 14,950 1988 50,251 2000 40,937
1965 6,357 1977 17,224 1989 53,794 2001 36,151
1966 6,769 1978 17,946 1990 55,972 2002 30,762
four time lags are significantly different from zero (.96, .92, .87, and .81) and that the values
then gradually drop to zero. As a final check, Maggie looks at the Q statistic for 10 time lags.
The LBQ is 300.56, which is greater than the chi-square value 18.3 (the upper .05 point of a
chi-square distribution with 10 degrees of freedom). This result indicates the autocorrela-
tions for the first 10 lags as a group are significantly different from zero. She decides that the
data are highly autocorrelated and exhibit trendlike behavior.
Maggie suspects that the series can be differenced to remove the trend and to create a
stationary series. She differences the data (see the Minitab Applications section at the end
of the chapter), and the results are shown in Figure 12. The differenced series shows no evi-
dence of a trend, and the autocorrelation function, shown in Figure 13, appears to support
this conclusion. Examining Figure 13, Maggie notes that the autocorrelation coefficient at
time lag 3, .32, is significantly different from zero (tested at the .05 significance level). The
autocorrelations at lags other than lag 3 are small, and the LBQ statistic for 10 lags is also
relatively small, so there is little evidence to suggest the differenced data are autocorrelated.
Yet Maggie wonders whether there is some pattern in these data that can be modeled by
one of the more advanced forecasting techniques.
Exploring Data Patterns and an Introduction to Forecasting Techniques
0 ; 1.96 11>52
0 ; .272
Then Perkin computes the autocorrelation coefficients shown in Figure 15. He notes
that the autocorrelation coefficients at time lags 1 and 4 are significantly different from zero
(r1 = .39 7 .272 and r4 = .74 7 .333). He concludes that Coastal Marine sales are seasonal
on a quarterly basis.
patterns can be recognized, then techniques that are capable of effectively extrapolating
these patterns can be selected.
Forecasting Techniques for Stationary Data
A stationary series was defined earlier as one whose mean value is not changing over
time. Such situations arise when the demand patterns influencing the series are rela-
tively stable. It is important to recognize that stationary data do not necessarily vary
randomly about a mean level. Stationary series can be autocorrelated, but the nature
of the association is such that the data do not wander away from the mean for any
extended period of time. In its simplest form, forecasting a stationary series involves
using the available history of the series to estimate its mean value, which then becomes
the forecast for future periods. More-sophisticated techniques allow the first few fore-
casts to be somewhat different from the estimated mean but then revert to the mean
for additional future periods. Forecasts can be updated as new information becomes
available. Updating is useful when initial estimates are unreliable or when the stability
of the average is in question. In the latter case, updating provides some degree of
responsiveness to a potential change in the underlying level of the series.
Stationary forecasting techniques are used in the following circumstances:
• The forces generating a series have stabilized, and the environment in which the
series exists is relatively unchanging. Examples are the number of breakdowns per
week on an assembly line having a uniform production rate, the unit sales of a
product or service in the maturation stage of its life cycle, and the number of sales
resulting from a constant level of effort.
• A very simple model is needed because of a lack of data or for ease of explanation or
implementation. An example is when a business or organization is new and very
few historical data are available.
• Stability may be obtained by making simple corrections for factors such as popula-
tion growth or inflation. Examples are changing income to per capita income and
changing dollar sales to constant dollar amounts.
• The series may be transformed into a stable one. Examples are transforming a
series by taking logarithms, square roots, or differences.
• The series is a set of forecast errors from a forecasting technique that is considered
adequate. (See Example 7.)
Techniques that should be considered when forecasting stationary series include
naive methods, simple averaging methods, moving averages, and autoregressive mov-
ing average (ARMA) models (Box-Jenkins methods).
• Market acceptance increases. An example is the growth period in the life cycle of a
new product.
Techniques that should be considered when forecasting trending series include
moving averages, Holt’s linear exponential smoothing, simple regression, growth
curves, exponential models, and autoregressive integrated moving average (ARIMA)
models (Box-Jenkins methods).
these techniques become less applicable. For instance, moving averages, exponential
smoothing, and ARIMA models are poor predictors of economic turning points,
whereas econometric models are more useful. Regression models are appropriate for
the short, intermediate, and long terms. Means, moving averages, classical decomposi-
tion, and trend projections are quantitative techniques that are appropriate for the
short and intermediate time horizons. The more complex Box-Jenkins and economet-
ric techniques are also appropriate for short- and intermediate-term forecasts.
Qualitative methods are frequently used for longer time horizons.
The applicability of forecasting techniques is generally something a forecaster
bases on experience. Managers frequently need forecasts in a relatively short time.
Exponential smoothing, trend projection, regression models, and classical decomposi-
tion methods have an advantage in this situation. (See Table 6.)
Computer costs are no longer a significant part of technique selection. Desktop
computers (microprocessors) and forecasting software packages are becoming com-
monplace for many organizations. Due to these developments, other criteria will likely
overshadow computer cost considerations.
Ultimately, a forecast will be presented to management for approval and use in the
planning process. Therefore, ease of understanding and interpreting the results is an
important consideration. Regression models, trend projections, classical decomposi-
tion, and exponential smoothing techniques all rate highly on this criterion.
It is important to point out that the information displayed in Table 6 should be
used as a guide for the selection of a forecasting technique. It is good practice to try
Naive ST, T, S S TS 1
Simple averages ST S TS 30
Moving averages ST S TS 4–20
Exponential smoothing ST S TS 2
Linear exponential smoothing T S TS 3
Quadratic exponential smoothing T S TS 4
Seasonal exponential smoothing S S TS 2×s
Adaptive filtering S S TS 5×s
Simple regression T I C 10
Multiple regression C, S I C 10 × V
Classical decomposition S S TS 5×s
Exponential trend models T I, L TS 10
S-curve fitting T I, L TS 10
Gompertz models T I, L TS 10
Growth curves T I, L TS 10
Census X-12 S S TS 6×s
Box-Jenkins ST, T, C, S S TS 24 3×s
Leading indicators C S C 24
Econometric models C S C 30
Time series multiple regression T, S I, L C 6×s
Pattern of the data: ST, stationary; T, trending; S, seasonal; C, cyclical
Time horizon: S, short term (less than three months); I, intermediate term; L, long term
Type of model: TS, time series; C, causal
Seasonal: s, length of seasonality
Variable: V, number of variables
Exploring Data Patterns and an Introduction to Forecasting Techniques
more than one forecasting method for a particular problem, holding out some recent
data, and then to compute forecasts of these holdout observations using the different
methods. The performance of the methods for these holdout test cases can be
determined using one or more of the accuracy measures defined in Equations 7
through 11, discussed below. Assuming an adequate fit to the data, the most accurate
method (the one with the smallest forecast error) is a reasonable choice for the “best”
method. It may not be the best method in the next situation.
Empirical Evaluation of Forecasting Methods
Empirical research has found that the forecast accuracy of simple methods is often as
good as that of complex or statistically sophisticated techniques (see Fildes et al., 1997;
Makridakis et al., 1993; and Makridakis and Hibon, 2000). Results of the M3–IJF
Competition, where different experts using their favorite forecasting methodology
each generated forecasts for 3,003 different time series, tended to support this finding
(Makridakis and Hibon, 2000). It would seem that the more statistically complex a
technique is, the better it should predict time series patterns. Unfortunately, estab-
lished time series patterns can and do change in the future. Thus, having a model that
best represents the historical data (the thing complex methods do well) does not neces-
sarily guarantee more accuracy in future predictions. Of course, the ability of the fore-
caster also plays an important role in the development of a good forecast.
The M3–IJF Competition was held in 1997. The forecasts produced by the various
forecasting techniques were compared across the sample of 3,003 time series, with the
accuracy assessed using a range of measures on a holdout set. The aim of the 1997
study was to check the four major conclusions of the original M-Competition on a
larger data set (see Makridakis et al., 1982). Makridakis and Hibon (2000) summarized
the latest competition as follows:
1. As discussed previously, statistically sophisticated or complex methods do not nec-
essarily produce more accurate forecasts than simpler methods.
2. Various accuracy measures produce consistent results when used to evaluate dif-
ferent forecasting methods.
3. The combination of three exponential smoothing methods outperforms, on aver-
age, the individual methods being combined and does well in comparison with
other methods.
4. The performance of the various forecasting methods depends on the length of the
forecasting horizon and the kind of data (yearly, quarterly, monthly) analyzed.
Some methods perform more accurately for short horizons, whereas others are
more appropriate for longer ones. Some methods work better with yearly data, and
others are more appropriate for quarterly and monthly data.
As part of the final selection, each technique must be evaluated in terms of its reli-
ability and applicability to the problem at hand, its cost effectiveness and accuracy
compared with competing techniques, and its acceptance by management. Table 6 sum-
marizes forecasting techniques appropriate for particular data patterns. As we have
pointed out, this table represents a place to start—that is, methods to consider for data
with certain characteristics. Ultimately, any chosen method should be continuously
monitored to be sure it is adequately doing the job for which it was intended.
A residual is the difference between an actual observed value and its forecast
value.
Equation 6 is used to compute the error or residual for each forecast period.
et = Yt - YNt (6)
where
et = the forecast error in time period t
Yt = the actual value in time period t
YNt = the forecast value for time period t
One method for evaluating a forecasting technique uses the sum of the absolute
errors. The mean absolute deviation (MAD) measures forecast accuracy by averaging
the magnitudes of the forecast errors (the absolute values of the errors). The MAD is
in the same units as the original series and provides an average size of the “miss”
regardless of direction. Equation 7 shows how the MAD is computed.
n ta
1 n
MAD = ƒ Yt - YNt ƒ (7)
=1
The mean squared error (MSE) is another method for evaluating a forecasting
technique. Each error or residual is squared; these are then summed and divided by the
number of observations. This approach penalizes large forecasting errors, since the
errors are squared. This is important because a technique that produces moderate
errors may well be preferable to one that usually has small errors but occasionally
yields extremely large ones. The MSE is given by Equation 8.
n ta
1 n
MSE = 1Yt - YNt 22 (8)
=1
The square root of the MSE, or the root mean squared error (RMSE), is also used
to evaluate forecasting methods. The RMSE, like the MSE, penalizes large errors but
Exploring Data Patterns and an Introduction to Forecasting Techniques
has the same units as the series being forecast so its magnitude is more easily inter-
preted. The RMSE is displayed below.
A n ta
1 n
RMSE = 1Yt - YNt22 (9)
=1
n ta
1 n ƒ Yt - YNt ƒ
MAPE = (10)
=1 ƒ Yt ƒ
Notice that MAPE cannot be calculated if any of the Yt are zero.
Sometimes it is necessary to determine whether a forecasting method is biased (con-
sistently forecasting low or high). The mean percentage error (MPE) is used in these
cases. It is computed by finding the error in each period, dividing this by the actual value
for that period, and then averaging these percentage errors. The result is typically multi-
plied by 100 and expressed as a percentage. If the forecasting approach is unbiased, the
MPE will produce a number that is close to zero. If the result is a large negative percent-
age, the forecasting method is consistently overestimating. If the result is a large positive
percentage, the forecasting method is consistently underestimating. The MPE is given by
n ta
1 n 1Yt - YNt 2
MPE = (11)
=1 Yt
Example 6
Table 7 shows the data for the daily number of customers requiring repair work, Yt, and a
forecast of these data, YNt, for Gary’s Chevron station. The forecasting technique used the
number of customers serviced in the previous period as the forecast for the current period.
The following computations were employed to evaluate this model using the MAD, MSE,
RMSE, MAPE, and MPE.
n ta
1 n 34
MAD = ƒ Yt - YNt ƒ = = 4.3
=1 8
Exploring Data Patterns and an Introduction to Forecasting Techniques
n ta
1 n 188
MSE = 1Yt - YNt22 = = 23.5
=1 8
n ta
1 n ƒ Yt - YNt ƒ .556
MAPE = = = .0695 16.95%2
=1 Yt 8
n ta
1 n 1Yt - YNt 2 .162
MPE = = = .0203 12.03%2
=1 Yt 8
1 58 — — — — — —
2 54 58 -4 4 16 .074 -.074
3 60 54 6 6 36 .100 .100
4 55 60 -5 5 25 .091 -.091
5 62 55 7 7 49 .113 .113
6 62 62 0 0 0 .000 .000
7 65 62 3 3 9 .046 .046
8 63 65 -2 2 4 .032 -.032
9 70 63 7 7 49 .100 .100
Totals 12 34 188 .556 .162
The MAD indicates that each forecast deviated by an average of 4.3 customers. The
MSE of 23.5 (or the RMSE of 4.8) and the MAPE of 6.95% would be compared to the MSE
(RMSE) and the MAPE for any other method used to forecast these data. Finally, the small
MPE of 2.03% indicates that the technique is not biased: Since the value is close to zero, the
technique does not consistently over- or underestimate the number of customers serviced
daily.
Before forecasting with a selected technique, the adequacy of the choice should be
evaluated. The forecaster should answer the following questions:
• Are the autocorrelation coefficients of the residuals indicative of a random series?
This question can be answered by examining the autocorrelation function for the
residuals, such as the one to be demonstrated in Example 7.
• Are the residuals approximately normally distributed? This question can be
answered by analyzing a histogram of the residuals or a normal probability plot.
• Do all parameter estimates have significant t ratios? Applications of t ratios are
demonstrated in Example 2.
• Is the technique simple to use, and can planners and policy makers understand it?
The basic requirement that the residual pattern be random is verified by examining
the autocorrelation coefficients of the residuals. There should be no significant autocor-
relation coefficients. Example 2 illustrated how the autocorrelation coefficients can be
used to determine whether a series is random. The Ljung-Box Q statistic is also used to
Exploring Data Patterns and an Introduction to Forecasting Techniques
test that the autocorrelations for all lags up to a given lag K equal zero. Example 7 illus-
trates this procedure with the residuals from two fitted models.
Example 7
Maggie Trymane, the analyst for Sears, has been asked to forecast sales for 2005. The data
are shown in Table 4 for 1955 to 2004. First, Maggie tries to forecast the data using a five-
month moving average. The residuals, the differences between the actual values and the
predicted values, are computed and stored. The autocorrelation coefficients for these resid-
uals are shown in Figure 16. An examination of these autocorrelation coefficients indicates
that two are significantly different from zero, r1 = .77 and r2 = .58. Significant autocorre-
lation coefficients indicate some association or pattern in the residuals. Furthermore, the
autocorrelation function itself has a pattern of smoothly declining coefficients. Examining
the first 10 autocorrelations as a group, the Q statistic for 10 lags is 73.90, much greater
than the upper .05 value of a chi-square variable with 10 degrees of freedom, 18.3. The
hypothesis that the first 10 autocorrelations are consistent with those for a random series is
clearly rejected at the 5% level. Since one of the basic requirements for a good forecasting
technique is that it give a residual or error series that is essentially random, Maggie judges
the five-month moving average technique to be inadequate.
Maggie now tries Holt’s linear exponential smoothing. The autocorrelation function
for the residual series generated by this technique is shown in Figure 17. An examination
of these autocorrelation coefficients indicates that none is significantly different from
zero (at the 5% level). The Q statistic for 10 time lags is also examined. The LBQ value
of 7.40 in the Minitab output is less than the upper .05 value of a chi-square variable
with eight degrees of freedom, 15.5. (In this case, the degrees of freedom are equal to the
number of lags to be tested minus the number of parameters in the linear exponential
smoothing model that have been fitted to the data.) As a group, the first 10 residual
autocorrelations are not unlike those for a completely random series. Maggie decides to
consider Holt’s linear exponential smoothing technique as a possible model to forecast
2005 operating revenue for Sears.
APPLICATION TO MANAGEMENT
The concepts in this chapter provide a basis for selecting a proper forecasting tech-
nique in a given situation. It is important to note that in many practical situations,
more than one forecasting method or model may produce acceptable and nearly
indistinguishable forecasts. In fact, it is good practice to try several reasonable fore-
casting techniques. Often judgment, based on ease of use, cost, external environmen-
tal conditions, and so forth, must be used to select a particular set of forecasts from,
say, two sets of nearly indistinguishable values.
The following are a few examples of situations constantly arising in the busi-
ness world where a sound forecasting technique would help the decision-making
process:
• A soft-drink company wants to project the demand for its major product over the
next two years, by month.
• A major telecommunications company wants to forecast the quarterly dividend
payments of its chief rival for the next three years.
• A university needs to forecast student credit hours by quarter for the next four
years in order to develop budget projections for the state legislature.
• A public accounting firm needs monthly forecasts of dollar billings so it can plan
for additional accounting positions and begin recruiting.
• The quality control manager of a factory that makes aluminum ingots needs a
weekly forecast of production defects for top management of the company.
• A banker wants to see the projected monthly revenue of a small bicycle manufac-
turer that is seeking a large loan to triple its output capacity.
• A federal government agency needs annual projections of average miles per gallon
of American-made cars over the next 10 years in order to make regulatory
recommendations.
• A personnel manager needs a monthly forecast of absent days for the company
workforce in order to plan overtime expenditures.
• A savings and loan company needs a forecast of delinquent loans over the next
two years in an attempt to avoid bankruptcy.
• A company that makes computer chips needs an industry forecast for the number
of personal computers sold over the next 5 years in order to plan its research and
development budget.
• An Internet company needs forecasts of requests for service over the next six
months in order to develop staffing plans for its call centers.
Exploring Data Patterns and an Introduction to Forecasting Techniques
Glossary
Autocorrelation. Autocorrelation is the correla- Seasonal component. The seasonal component is a
tion between a variable lagged one or more peri- pattern of change that repeats itself year after
ods and itself. year.
Correlogram or autocorrelation function. The cor- Stationary series. A stationary series is one whose
relogram (autocorrelation function) is a graph of basic statistical properties, such as the mean and
the autocorrelations for various lags of a time series. variance, remain constant over time.
Cross-sectional. Cross-sectional data are observa- Time series. A time series consists of data that are
tions collected at a single point in time. collected, recorded, or observed over successive
Cyclical component. The cyclical component is the increments of time.
wavelike fluctuation around the trend. Trend. The trend is the long-term component that
Residual. A residual is the difference between an represents the growth or decline in the time series
actual observed value and its forecast value. over an extended period of time.
Key Formulas
a 1Yt - Y21Yt - k - Y 2
n
t=k+1
rk = (1)
a 1Yt - Y 2
n
2
t=1
1 + 2 a r 2i
k-1
i=1
SE1rk2 = (2)
T n
Ljung-Box Q statistic
Q = n1n + 22 a
m
r k2
(3)
k = 1n - k
Random model
Yt = c + "t (4)
r1
t = (5)
SE1r12
et = Yt - YNt (6)
Exploring Data Patterns and an Introduction to Forecasting Techniques
n ta
1 n
MAD = ƒ Yt - YNt ƒ (7)
=1
n ta
1 n
MSE = 1Yt - YNt 22 (8)
=1
A n ta
1 n
RMSE = 1Yt - YN t22 (9)
=1
n ta
1 n ƒ Yt - YNt ƒ
MAPE = (10)
=1 ƒ Yt ƒ
Mean percentage error
n ta
1 n 1Yt - YNt 2
MPE = (11)
=1 Yt
Problems
TABLE P-13
16. Which of the following statements is true concerning the accuracy measures used
to evaluate forecasts?
a. The MAPE takes into consideration the magnitude of the values being forecast.
b. The MSE and RMSE penalize large errors.
c. The MPE is used to determine whether a model is systematically predicting too
high or too low.
d. The advantage of the MAD method is that it relates the size of the error to the
actual observation.
17. Allie White, the chief loan officer for the Dominion Bank, would like to ana-
lyze the bank’s loan portfolio for the years 2001 to 2006. The data are shown in
Table P-17.
a. Compute the autocorrelations for time lags 1 and 2. Test to determine whether
these autocorrelation coefficients are significantly different from zero at the
.05 significance level.
b. Use a computer program to plot the data and compute the autocorrelations for
the first six time lags. Is this time series stationary?
18. This question refers to Problem 17. Compute the first differences of the quarterly
loan data for Dominion Bank.
a. Compute the autocorrelation coefficient for time lag 1 using the differenced
data.
b. Use a computer program to plot the differenced data and compute the autocor-
relations for the differenced data for the first six time lags. Is this time series
stationary?
19. Analyze the autocorrelation coefficients for the series shown in Figures 18 through
21. Briefly describe each series.
20. An analyst would like to determine whether there is a pattern to earnings per
share for the Price Company, which operated a wholesale/retail cash-and-carry
business in several states under the name Price Club. The data are shown in Table
P-20. Describe any patterns that exist in these data.
a. Find the forecast value of the quarterly earnings per share for Price Club for
each quarter by using the naive approach (i.e., the forecast for first-quarter 1987
is the value for fourth-quarter 1986, .32).
b. Evaluate the naive forecast using MAD.
Exploring Data Patterns and an Introduction to Forecasting Techniques
1.0
0.8
0.6
Autocorrelation
0.4
0.2
0.0
−0.2
−0.4
−0.6
−0.8
−1.0
5 10 15
1.0
0.8
0.6
Autocorrelation
0.4
0.2
0.0
−0.2
−0.4
−0.6
−0.8
−1.0
2 7 12 17
1.0
0.8
0.6
Autocorrelation
0.4
0.2
0.0
−0.2
−0.4
−0.6
−0.8
−1.0
5 15 25
1.0
0.8
0.6
Autocorrelation 0.4
0.2
0.0
−0.2
−0.4
−0.6
−0.8
−1.0
5 10 15
Source: The Value Line Investment Survey (New York: Value Line,
1994), p. 1646.
21. Table P-21 contains the weekly sales for a food item for 52 consecutive weeks.
a. Plot the sales data as a time series.
b. Do you think this series is stationary or nonstationary?
c. Using Minitab or a similar program, compute the autocorrelations of the sales
series for the first 10 time lags. Is the behavior of the autocorrelations consistent
with your choice in part b? Explain.
Exploring Data Patterns and an Introduction to Forecasting Techniques
b. Using the forecasts in part a, calculate the MAD, RMSE, and MAPE.
c. Given the results in part b and the nature of the patterns in the income series,
do you think this naive forecasting method is viable? Can you think of another
naive method that might be better?
CASES
In 1958, the Murphy brothers established a furniture many federal publications. After looking through a
store in downtown Dallas. Over the years, they were recent copy of the Survey of Current Business, she
quite successful and extended their retail coverage found the history on monthly sales for all retail
throughout the West and Midwest. By 1996, their stores in the United States and decided to use this
chain of furniture stores had become well estab- variable as a substitute for her variable of interest,
lished in 36 states. Murphy Brothers sales dollars. She reasoned that,
Julie Murphy, the daughter of one of the if she could establish accurate forecasts for national
founders, had recently joined the firm. Her sales, she could relate these forecasts to Murphy’s
father and uncle were sophisticated in many ways own sales and come up with the forecasts she
but not in the area of quantitative skills. In particu- wanted.
lar, they both felt that they could not accu- Table 8 shows the data that Julie collected, and
rately forecast the future sales of Murphy Brothers Figure 22 shows a data plot provided by Julie’s com-
using modern computer techniques. For this rea- puter program. Julie began her analysis by using the
son, they appealed to Julie for help as part of her computer to develop a plot of the autocorrelation
new job. coefficients.
Julie first considered using Murphy sales dollars After examining the autocorrelation function
as her variable but found that several years of his- produced in Figure 23, it was obvious to Julie that
tory were missing. She asked her father, Glen, about her data contain a trend. The early autocorrelation
this, and he told her that at the time he “didn’t think coefficients are very large, and they drop toward
it was that important.” Julie explained the impor- zero very slowly with time. To make the series
tance of past data to Glen, and he indicated that he stationary so that various forecasting methods could
would save future data. be considered, Julie decided to first difference her
Julie decided that Murphy sales were probably data to see if the trend could be removed. The auto-
closely related to national sales figures and decided correlation function for the first differenced data is
to search for an appropriate variable in one of the shown in Figure 24.
QUESTIONS
1. What should Julie conclude about the retail 3. What forecasting techniques should Julie try?
sales series? 4. How will Julie know which technique works
2. Has Julie made good progress toward finding a best?
forecasting technique?
TABLE 8 Monthly Sales ($ billions) for All Retail Stores, 1983–1995
1983 1984 1985 1986 1987 1988 1989 1990 1991 1992 1993 1994 1995
Jan. 81.3 93.1 98.8 105.6 106.4 113.6 122.5 132.6 130.9 142.1 148.4 154.6 167.0
Feb. 78.9 93.7 95.6 99.7 105.8 115.0 118.9 127.3 128.6 143.1 145.0 155.8 164.0
Mar. 93.8 103.3 110.2 114.2 120.4 131.6 141.3 148.3 149.3 154.7 164.6 184.2 192.1
Apr. 93.8 103.9 113.1 115.7 125.4 130.9 139.8 145.0 148.5 159.1 170.3 181.8 187.5
May 97.8 111.8 120.3 125.4 129.1 136.0 150.3 154.1 159.8 165.8 176.1 187.2 201.4
Jun. 100.6 112.3 115.0 120.4 129.0 137.5 149.0 153.5 153.9 164.6 175.7 190.1 202.6
Jul. 99.4 106.9 115.5 120.7 129.3 134.1 144.6 148.9 154.6 166.0 177.7 185.8 194.9
Aug. 100.1 111.2 121.1 124.1 131.5 138.7 153.0 157.4 159.9 166.3 177.1 193.8 204.2
Sep. 97.9 104.0 113.8 124.4 124.5 131.9 144.1 145.6 146.7 160.6 171.1 185.9 192.8
Oct. 100.7 109.6 115.8 123.8 128.3 133.8 142.3 151.5 152.1 168.7 176.4 189.7 194.0
Nov. 103.9 113.5 118.1 121.4 126.9 140.2 148.8 156.1 155.6 167.2 180.9 194.7 202.4
Dec. 125.8 132.3 138.6 152.1 157.2 171.0 176.5 179.7 181.0 204.1 218.3 233.3 238.0
Source: Based on Survey of Current Business, various years.
Jan. Feb. Mar. Apr. May Jun. Jul. Aug. Sep. Oct. Nov. Dec.
1992 4,906 5,068 4,710 4,792 4,638 4,670 4,574 4,477 4,571 4,370 4,293 3,911
1993 5,389 5,507 4,727 5,030 4,926 4,847 4,814 4,744 4,844 4,769 4,483 4,120
1994 5,270 5,835 5,147 5,354 5,290 5,271 5,328 5,164 5,372 5,313 4,924 4,552
1995 6,283 6,051 5,298 5,659 5,343 5,461 5,568 5,287 5,555 5,501 5,201 4,826
Glen Murphy was not happy to be chastised by his Table 9. He was surprised to find out that Julie did not
daughter. He decided to conduct an intensive search of share his enthusiasm. She knew that acquiring actual
Murphy Brothers’ records. Upon implementing an sales data for the past 4 years was a positive occur-
investigation, he was excited to discover sales data for rence. Julie’s problem was that she was not quite sure
the past four years, 1992 through 1995, as shown in what to do with the newly acquired data.
QUESTIONS
1. What should Julie conclude about Murphy 3. Which data should Julie use to develop a fore-
Brothers’ sales data? casting model?
2. How does the pattern of the actual sales data
compare to the pattern of the retail sales data
presented in Case 1A?
Exploring Data Patterns and an Introduction to Forecasting Techniques
John Mosby, owner of several Mr. Tux rental stores, is Finally, John uses another computer program to
beginning to forecast his most important business vari- calculate the percentage of the variance in the origi-
able, monthly dollar sales. One of his employees, nal data explained by the trend, seasonal, and ran-
Virginia Perot, has gathered the sales data. John decides dom components.
to use all 96 months of data he has collected. He runs The program calculates the percentage of the
the data on Minitab and obtains the autocorrelation variance in the original data explained by the factors
function shown in Figure 25. Since all the autocorrela- in the analysis:
tion coefficients are positive and they are trailing off
very slowly, John concludes that his data have a trend. FACTOR % EXPLAINED
Next, John asks the program to compute the
Data 100
first differences of the data. Figure 26 shows the
]
Trend 6
autocorrelation function for the differenced data. Seasonal 45
The autocorrelation coefficients for time lags 12 and Random 49
24, r12 = .68 and r24 = .42, respectively, are both sig-
nificantly different from zero.
QUESTIONS
1. Summarize the results of John’s analysis in one 3. How would you explain the line “Random 49%.”?
paragraph that a manager, not a forecaster, can 4. Consider the significant autocorrelations, r12 and
understand. r24, for the differenced data. Would you conclude
2. Describe the trend and seasonal effects that that the sales first differenced have a seasonal
appear to be present in the sales data for component? If so, what are the implications for
Mr. Tux. forecasting, say, the monthly changes in sales?
Marv Harnishfeger, executive director, was con- various data exploration techniques, agreed to ana-
cerned about the size and scheduling of staff for lyze the problem. She asked Marv to provide
the remainder of 1993. He explained the problem monthly data for the number of new clients seen.
to Dorothy Mercer, recently elected president of Marv provided the monthly data shown in Table 10
the executive committee. Dorothy thought about for the number of new clients seen by CCC for the
the problem and concluded that Consumer Credit period January 1985 through March 1993. Dorothy
Counseling (CCC) needed to analyze the number then analyzed these data using a time series plot
of new clients it acquired each month. Dorothy, and autocorrelation analysis.
who worked for a local utility and was familiar with
QUESTIONS
1. Explain how Dorothy used autocorrelation 3. What type of forecasting technique did Dorothy
analysis to explore the data pattern for the num- recommend for this data set?
ber of new clients seen by CCC.
2. What did she conclude after she completed this
analysis?
The president of Alomega, Julie Ruth, had collected which of the predictor variables was best for this
data from her company’s operations. She found sev- purpose.
eral months of sales data along with several possible Julie investigated the relationships between
predictor variables. While her analysis team was sales and the possible predictor variables. She now
working with the data in an attempt to forecast realizes that this step was premature because she
monthly sales, she became impatient and wondered doesn’t even know the data pattern of her sales. (See
Table 11.)
QUESTION
1. What did Julie conclude about the data pattern
of Alomega Sales?
GAMESA Company is the largest producer of Surtido is a mixture of different cookies in one
cookies, snacks, and pastry in Mexico and Latin presentation. Jame knows that this product is com-
America. monly served at meetings, parties, and reunions. It is
Jame Luna, an analyst for SINTEC, which is an also popular during the Christmas holidays. So Jame
important supply-chain consulting firm based in is pretty sure there will be a seasonal component in
Mexico, is working with GAMESA on an optimal the time series of monthly sales, but he is not sure if
vehicle routing and docking system for all the distri- there is a trend in sales. He decides to plot the sales
bution centers in Mexico. Forecasts of the demand series and use autocorrelation analysis to help him
for GAMESA products will help to produce a plan determine if these data have a trend and a seasonal
for the truck float and warehouse requirements component.
associated with the optimal vehicle routing and Karin, one of the members of Jame’s team, sug-
docking system. gests that forecasts of future monthly sales might be
As a starting point, Jame decides to focus on generated by simply using the average sales for each
one of the main products of the Cookies Division of month. Jame decides to test this suggestion by “hold-
GAMESA. He collects data on monthly aggregated ing out” the monthly sales for 2003 as test cases. Then
demand (in kilograms) for Surtido cookies from the average of the January sales for 2000–2002 will
January 2000 through May 2003. The results are be used to forecast January sales for 2003 and so
shown in Table 12. forth.
QUESTIONS
1. What should Jame conclude about the data 2. What did Jame learn about the forecast accu-
pattern of Surtido cookie sales? racy of Karin’s suggestion?
Minitab Applications
The problem. In Example 4, Maggie Trymane, an analyst for Sears, wants to forecast
sales for 2005. She needs to determine the pattern for the sales data for the years from
1955 to 2004.
Minitab Solution
1. Enter the Sears data shown in Table 4 into column C1. Since the data are already
stored in a file called Tab4.MTW, click on
File>Open Worksheet
and double-click the Minitab Ch3 file. Click on Tab4.MTW and Open the file. The
Sears data will appear in C1.
2. To construct an autocorrelation function, click on the following menus, as shown in
Figure 27:
Stat>Time Series>Autocorrelation
The Differences option is above the Autocorrelation option shown in Figure 27.
5. The Differences dialog box shown in Figure 29 appears.
a. Double-click on the variable Revenue and it will appear to the right of Series.
b. Tab to Store differences in: and enter C2. The differenced data will now appear
in the worksheet in column C2.
Exploring Data Patterns and an Introduction to Forecasting Techniques
Excel Applications
The problem. Harry Vernon wants to use Excel to compute the autocorrelation coeffi-
cients and a correlogram for the data presented in Table 1.
Excel Solution
1. Create a new file by clicking on the following menus:
File>New
2. Position the mouse at A1. Notice that, whenever you position the mouse on a cell,
it is highlighted. Type the heading VERNON’S MUSIC STORE. Position the
mouse at A2 and type NUMBER OF VCRS SOLD.
3. Position the mouse at A4 and type Month. Press <enter> and the A5 cell is high-
lighted. Now enter each month, starting with January in A5 and ending with
December in A16.
4. Position the mouse at B4 and type Y. Enter the data from Table 1, beginning in cell
B5. Position the mouse at C4 and type Z.
5. Highlight cells B4:C16 and click on the following menus:
Insert>Name>Create
In the Create Names dialog box, click on the Top row check box, and click on OK.
This step creates the name Y for the range B5:B16 and the name Z for the range
C5:C16.
6. Highlight C5 and enter the formula
=(B5-AVERAGE(Y))/STDEV(Y)
Copy C5 to the rest of the column by highlighting it and then clicking on the Fill
handle in the lower right corner and dragging it down to cell C16. With cells
C5:C16 still highlighted, click the Decrease Decimal button (shown in Figure 30
toward the upper middle) until three decimal places are displayed. The Decrease
Decimal button is on the Formatting taskbar. This taskbar can be displayed by
right-clicking File and then left-clicking Formatting.
7. Enter the labels LAG and ACF in cells E4 and F4. In order to examine the first six
time lags, enter the digits 1 through 6 in cells E5:E10.
8. Highlight F5 and enter the formula
=SUMPRODUCT(OFFSET(Z,E5,0,12–E5),OFFSET(Z,0,0,12-E5))/11
Highlight F5, click the Fill handle in the lower right corner, and drag it down to cell
F10. With cells F5:F10 still highlighted, click the Decrease Decimal button until
three decimal places are displayed. The results are shown in Figure 30.
9. To develop the autocorrelation function, highlight cells F5:F10. Click on the
ChartWizard tool (shown in Figure 31 toward the upper middle).
10. The ChartWizard—Step 1 to 4 dialog box appears. In step 1, select a chart type by
clicking on Column and then on Next. In the dialog box for step 2, click on the Series
dialog box. In the blank next to Name, type Corr.. Click Next and the step 3 dialog
box appears. Under Chart title, delete Corr.. Under Category (X) axis, type Time
Lags. Now click on the Data Table dialog box and on the box next to Show data
table. Click on Next to obtain the step 4 dialog box and then on Finish to produce the
Exploring Data Patterns and an Introduction to Forecasting Techniques
autocorrelation function shown in Figure 31. Click on one of the corners of the chart,
and move it outward in order to enlarge the autocorrelation function.
11. In order to save the data, click on
File>Save As
In the Save As dialog box, type Tab1 in the space to the right of File name. Click on
Save and the file will be saved as Tab1.xls.
MOVING AVERAGES AND
SMOOTHING METHODS
This chapter will describe three simple approaches to forecasting a time series: naive,
averaging, and smoothing methods. Naive methods are used to develop simple models
that assume that very recent data provide the best predictors of the future. Averaging
methods generate forecasts based on an average of past observations. Smoothing
methods produce forecasts by averaging past values of a series with a decreasing
(exponential) series of weights.
Figure 1 outlines the forecasting methods discussed in this chapter. Visualize your-
self on a time scale. You are at point t in Figure 1 and can look backward over past
observations of the variable of interest 1Yt 2 or forward into the future. After you
select a forecasting technique, you adjust it to the known data and obtain forecast val-
ues 1YNt2. Once these forecast values are available, you compare them to the known
observations and calculate the forecast error 1e t2.
A good strategy for evaluating forecasting methods involves the following steps:
From Chapter 4 of Business Forecasting, Ninth Edition. John E. Hanke, Dean W. Wichern.
Copyright © 2009 by Pearson Education, Inc. All rights reserved.
Moving Averages and Smoothing Methods
NAIVE MODELS
Often young businesses face the dilemma of trying to forecast with very small data sets.
This situation creates a real problem, since many forecasting techniques require large
amounts of data. Naive forecasts are one possible solution, since they are based solely
on the most recent information available.
Naive forecasts assume that recent periods are the best predictors of the future.
The simplest model is
YNt + 1 = Yt (1)
where YNt + 1 is the forecast made at time t (the forecast origin) for time t + 1.
The naive forecast for each period is the immediately preceding observation. One
hundred percent of the weight is given to the current value of the series. The naive
forecast is sometimes called the “no change” forecast. In short-term weather forecast-
ing, the “no change” forecast occurs often. Tomorrow’s weather will be much like
today’s weather.
Since the naive forecast (Equation 1) discards all other observations, this scheme
tracks changes very rapidly. The problem with this approach is that random fluctua-
tions are tracked as faithfully as other fundamental changes.
Example 1
Figure 2 shows the quarterly sales of saws from 2000 to 2006 for the Acme Tool Company.
The naive technique is used to forecast sales for the next quarter to be the same as for the
previous quarter. Table 1 shows the data from 2000 to 2006. If the data from 2000 to 2005 are
used as the initialization part and the data from 2006 as the test part, the forecast for the
first quarter of 2006 is
YN24 + 1 = Y24
YN25 = 650
FIGURE 2 Time Series Plot for Sales of Saws for Acme Tool
Company, 2000–2006, for Example 1
Moving Averages and Smoothing Methods
2000 1 1 500
2 2 350
3 3 250
4 4 400
2001 1 5 450
2 6 350
3 7 200
4 8 300
2002 1 9 350
2 10 200
3 11 150
4 12 400
2003 1 13 550
2 14 350
3 15 250
4 16 550
2004 1 17 550
2 18 400
3 19 350
4 20 600
2005 1 21 750
2 22 500
3 23 400
4 24 650
2006 1 25 850
2 26 600
3 27 450
4 28 700
The forecasting error is determined using the following equation. The error for period 25 is
In a similar fashion, the forecast for period 26 is 850, with an error of -250. Figure 2 shows
that these data have an upward trend and that there appears to be a seasonal pattern (the
first and fourth quarters are relatively high), so a decision is made to modify the naive model.
Examination of the data in Example 1 leads us to conclude that the values are
increasing over time. When data values increase over time, they are said to be
nonstationary in level or to have a trend. If Equation 1 is used, the projections will be
consistently low. However, the technique can be adjusted to take trend into considera-
tion by adding the difference between this period and the last period. The forecast
equation is
Equation 2 takes into account the amount of change that occurred between quarters.
Moving Averages and Smoothing Methods
Example 1 (cont.)
Using Equation 2, the forecast for the first quarter of 2006 is
For some purposes, the rate of change might be more appropriate than the
absolute amount of change. If this is the case, it is reasonable to generate forecasts
according to
Yt
YNt + 1 = Yt (3)
Yt - 1
Visual inspection of the data in Table 1 indicates that seasonal variation seems to
exist. Sales in the first and fourth quarters are typically larger than those in the second
and third quarters. If the seasonal pattern is strong, then an appropriate forecast equa-
tion for quarterly data might be
YNt + 1 = Yt - 3 (4)
Equation 4 says that in next quarter the variable will take on the same value that it
did in the corresponding quarter one year ago.
The major weakness of this approach is that it ignores everything that has occurred
since last year and also any trend. There are several ways of introducing more recent
information. For example, the analyst can combine seasonal and trend estimates and
forecast the next quarter using
(Yt - Yt - 1) + Á + 1Yt - 3 - Yt - 42 Yt - Yt - 4
YNt!1 = Yt - 3 + = Yt - 3 + (5)
4 4
where the Yt - 3 term forecasts the seasonal pattern and the remaining term averages
the amount of change for the past four quarters and provides an estimate of the trend.
The naive forecasting models in Equations 4 and 5 are given for quarterly data.
Adjustments can be made for data collected over different time periods. For monthly
data, for example, the seasonal period is 12, not 4, and the forecast for the next period
(month) given by Equation 4 is Y N t + 1 = Yt - 11.
It is apparent that the number and complexity of possible naive models are limited
only by the ingenuity of the analyst, but use of these techniques should be guided by
sound judgment.
Naive methods are also used as the basis for making comparisons against which
the performance of more sophisticated methods is judged.
Example 1 (cont.)
The forecasts for the first quarter of 2006 using Equations 3, 4, and 5 are
Y24 Y24
YN24 + 1 = Y24 = Y24
Y24 - 1 Y23
650 (Equation 3)
YN25 = 650 = 1,056
400
Moving Averages and Smoothing Methods
Simple Averages
Historical data can be smoothed in many ways. The objective is to use past data to
develop a forecasting model for future periods. In this section, the method of simple
averages is considered. As with the naive methods, a decision is made to use the first t
data points as the initialization part and the remaining data points as the test part.
Next, Equation 6 is used to average (compute the mean of) the initialization part of the
data and to forecast the next period.
YNt + 1 = a Yi
1 t
(6)
t i=1
When a new observation becomes available, the forecast for the next period, YNt + 2, is the
average or the mean computed using Equation 6 and this new observation.
When forecasting a large number of series simultaneously (e.g., for inventory man-
agement), data storage may be an issue. Equation 7 solves this potential problem. Only
the most recent forecast and the most recent observation need be stored as time moves
forward.
tYNt + 1 + Yt + 1
YNt + 2 = (7)
t + 1
The method of simple averages is an appropriate technique when the forces gener-
ating the series to be forecast have stabilized and the environment in which the series
exists is generally unchanging. Examples of this type of series are the quantity of sales
resulting from a consistent level of salesperson effort; the quantity of sales of a product
in the mature stage of its life cycle; and the number of appointments per week
requested of a dentist, doctor, or lawyer whose patient or client base is fairly stable.
Moving Averages and Smoothing Methods
A simple average uses the mean of all relevant historical observations as the
forecast for the next period.
Example 2
The Spokane Transit Authority operates a fleet of vans used to transport both persons with
disabilities and the elderly. A record of the gasoline purchased for this fleet of vans is shown
in Table 2. The actual amount of gasoline consumed by a van on a given day is determined
by the random nature of the calls and the destinations. Examination of the gasoline pur-
chases plotted in Figure 3 shows the data are very stable. Since the data seem stationary, the
method of simple averages is used for weeks 1 to 28 to forecast gasoline purchases for
weeks 29 and 30. The forecast for week 29 is
28 ia
1 28
YN28 + 1 = Yi
=1
7,874
YN29 = = 281.2
28
The forecast error is
The forecast for week 30 includes one more data point (302) added to the initialization
period. The forecast using Equation 7 is
Using the method of simple averages, the forecast of gallons of gasoline purchased for
week 31 is
30 ia
1 30 8,461
YN30 + 1 = Yi = = 282
=1 30
Moving Averages
The method of simple averages uses the mean of all the data to forecast. What if the
analyst is more concerned with recent observations? A constant number of data points
can be specified at the outset and a mean computed for the most recent observations.
The term moving average is used to describe this approach. As each new observation
becomes available, a new mean is computed by adding the newest value and dropping
the oldest. This moving average is then used to forecast the next period. Equation 8
gives the simple moving average forecast. A moving average of order k, MA(k), is
computed by
Yt + Yt - 1 + # # # + Yt - k + 1
YNt + 1 = (8)
k
where
YNt + 1 = the forecast value for the next period
Yt = the actual value at period t
k = the number of terms in the moving average
The moving average for time period t is the arithmetic mean of the k most recent
observations. In a moving average, equal weights are assigned to each observation.
Each new data point is included in the average as it becomes available, and the earliest
data point is discarded. The rate of response to changes in the underlying data pattern
depends on the number of periods, k, included in the moving average.
Note that the moving average technique deals only with the latest k periods of
known data; the number of data points in each average does not change as time
advances. The moving average model does not handle trend or seasonality very well,
although it does better than the simple average method.
Moving Averages and Smoothing Methods
The analyst must choose the number of periods, k, in a moving average. A moving
average of order 1, MA(1), would take the current observation, Yt, and use it to forecast
Y for the next period. This is simply the naive forecasting approach of Equation 1.
Example 3
Table 3 demonstrates the moving average forecasting technique with the Spokane Transit
Authority data, using a five-week moving average.The moving average forecast for week 29 is
t Gallons Ynt et
1 275 — —
2 291 — —
3 307 — —
4 281 — —
5 295 — —
6 268 289.8 -21.8
7 252 288.4 -36.4
8 279 280.6 -1.6
9 264 275.0 -11.0
10 288 271.6 16.4
11 302 270.2 31.8
12 287 277.0 10.0
13 290 284.0 6.0
14 311 286.2 24.8
15 277 295.6 -18.6
16 245 293.4 -48.4
17 282 282.0 0.0
18 277 281.0 -4.0
19 298 278.4 19.6
20 303 275.8 27.2
21 310 281.0 29.0
22 299 294.0 5.0
23 285 297.4 -12.4
24 250 299.0 -49.0
25 260 289.4 -29.4
26 245 280.8 -35.8
27 271 267.8 3.2
28 282 262.2 19.8
29 302 261.6 40.4
30 285 272.0 13.0
Moving Averages and Smoothing Methods
When the actual value for week 29 is known, the forecast error is calculated:
Minitab can be used to compute a five-week moving average (see the Minitab
Applications section at the end of the chapter for instructions). Figure 4 shows the five-
week moving average plotted against the actual data; the MAPE, MAD, and MSD; and the
basic Minitab instructions. Note that Minitab calls the mean squared error MSD (mean
squared deviation).
Figure 5 shows the autocorrelation function for the residuals from the five-week mov-
ing average method. Error limits for the individual autocorrelations centered at zero and
the Ljung-Box Q statistic (with six degrees of freedom, since no model parameters are esti-
mated) indicate that significant residual autocorrelation exists. That is, the residuals are not
random. The association contained in the residuals at certain time lags can be used to
improve the forecasting model.
The analyst must use judgment when determining how many days, weeks, months,
or quarters on which to base the moving average. The smaller the number, the larger
the weight given to recent periods. Conversely, the larger the number, the smaller the
weight given to more recent periods. A small number is most desirable when there are
sudden shifts in the level of the series. A small number places heavy weight on recent
history, which enables the forecasts to catch up more rapidly to the current level.
A large number is desirable when there are wide, infrequent fluctuations in the series.
Moving averages are frequently used with quarterly or monthly data to help smooth
the components within a time series. For quarterly data, a four-quarter moving average,
MA(4), yields an average of the four quarters, and for monthly data, a 12-month moving
average, MA(12), eliminates or averages out the seasonal effects. The larger the order of
the moving average, the greater the smoothing effect.
In Example 3, the moving average technique was used with stationary data. In
Example 4, we show what happens when the moving average method is used with
trending data. The double moving average technique, which is designed to handle
trending data, is introduced next.
1 654 — — —
2 658 — — —
3 665 1,977 — —
4 672 1,995 659 13
5 673 2,010 665 8
6 671 2,016 670 1
7 693 2,037 672 21
8 694 2,058 679 15
9 701 2,088 686 15
10 703 2,098 696 7
11 702 2,106 699 3
12 710 2,115 702 8
13 712 2,124 705 7
14 711 2,133 708 3
15 728 2,151 711 17
16 — — 717 —
730 Rentals
720
Moving
Average
710
700
Rentals
690 Double
Moving
680 Average
670
660
650
5 10 15
Week
YNt+p = at + bt p (12)
where
k = the number of periods in the moving average
p = the number of periods ahead to be forecast
Example 4
The Movie Video Store operates several videotape rental outlets in Denver, Colorado. The
company is growing and needs to expand its inventory to accommodate the increasing
demand for its services. The president of the company assigns Jill Ottenbreit to forecast
rentals for the next month. Rental data for the last 15 weeks are available and are presented
in Table 5. At first, Jill attempts to develop a forecast using a three-week moving average.
The MSE for this model is 133 (see Table 4). Because the data are obviously trending, she
finds that her forecasts are consistently underestimating actual rentals. For this reason, she
decides to try a double moving average. The results are presented in Table 5. To understand
the forecast for week 16, the computations are presented next. Equation 8 is used to com-
pute the three-week moving average (column 3).
Y15 + Y15-1 + Y15-3+1
M15 = YN15+1 =
3
728 + 711 + 712
M15 = YN16 = = 717
3
TABLE 5 Double Moving Average Forecast for the Movie Video Store for
Example 4
1 654 — — — — — —
2 658 — — — — — —
3 665 659 — — — — —
4 672 665 — — — — —
5 673 670 665 675 5 — —
6 671 672 669 675 3 680 -9
7 693 679 674 684 5 678 15
8 694 686 679 693 7 689 5
9 701 696 687 705 9 700 1
10 703 699 694 704 5 714 -11
11 702 702 699 705 3 709 -7
12 710 705 702 708 3 708 2
13 712 708 705 711 3 711 1
14 711 711 708 714 3 714 -3
15 728 717 712 722 5 717 11
16 — — — — — 727 —
Then Equation 9 is used to compute the double moving average (column 4).
¿
M15 + M15-1 + M15-3+1
M 15 =
3
¿ 717 + 711 + 708
M 15 = = 712
3
Equation 10 is used to compute the difference between the two moving averages
(column 5).
¿
a15 = 2M15 - M 15 = 217172 - 712 = 722
2 ¿ 2
b15 = 1M - M 15 2 = 1717 - 7122 = 5
3 - 1 15 2
Equation 12 is used to make the forecast one period into the future (column 7).
Note that the MSE has been reduced from 133 to 63.7.
Whereas the method of moving averages takes into account only the most recent
observations, simple exponential smoothing provides an exponentially weighted mov-
ing average of all previously observed values. The model is often appropriate for data
with no predictable upward or downward trend. The aim is to estimate the current
level. This level estimate is then used as the forecast of future values.
Exponential smoothing continually revises an estimate in the light of more-recent
experiences. This method is based on averaging (smoothing) past values of a series in
an exponentially decreasing manner. The most recent observation receives the largest
weight, a (where 0 6 a 6 1); the next most recent observation receives less weight,
a11 - a2; the observation two time periods in the past receives even less weight,
a11 - a22; and so forth.
In one representation of exponential smoothing, the new forecast (for time t + 1)
may be thought of as a weighted sum of the new observation (at time t) and the old
forecast (for time t). The weight a (0 6 a 6 1) is given to the newly observed value,
and the weight 11 - a2 is given to the old forecast. Thus,
where
YNt+1 = the new smoothed value or the forecast value for the next period
a = the smoothing constant 10 6 a 6 12
Yt = the new observation or the actual value of the series in period t
YNt = the old smoothed value or the forecast for period t
Equation 13 can be written as
In this form, the new forecast (YN t + 1) is the old forecast (YN t) adjusted by a times the
error Yt - YN t in the old forecast.
In Equation 13, the smoothing constant, a, serves as the weighting factor. The
value of a determines the extent to which the current observation influences the fore-
cast of the next observation. When a is close to 1, the new forecast will be essentially
the current observation. (Equivalently, the new forecast will be the old forecast plus a
substantial adjustment for any error that occurred in the preceding forecast.)
Conversely, when a is close to zero, the new forecast will be very similar to the old fore-
cast, and the current observation will have very little impact.
Finally, Equation 13 implies, for time t, that YNt = #YNt - 1 + (1 - #)Yt - 1 , and substi-
tution for YNt in Equation 13 gives
That is, YN t + 1 is an exponentially smoothed value. The speed at which past observations
lose their impact depends on the value of a, as demonstrated in Table 6.
Equations 13 and 14 are equivalent, but Equation 13 is typically used to calculate
the forecast YN t + 1 because it requires less data storage and is easily implemented.
The value assigned to a is the key to the analysis. If it is desired that predictions be
stable and random variations be smoothed, a small value of a is required. If a rapid
response to a real change in the pattern of observations is desired, a larger value of a is
appropriate. One method of estimating α is an iterative procedure that minimizes the
mean squared error (MSE ). Forecasts are computed for, say, α equal to .1, .2, . . . , .9,
and the sum of the squared forecast errors is computed for each.
Moving Averages and Smoothing Methods
# = .1 # = .6
Period Calculations Weight Calculations Weight
t .100 .600
t-1 .9 * .1 .090 .4 * .6 .240
t-2 .9 * .9 * .1 .081 .4 * .4 * .6 .096
t-3 .9 * .9 * .9 * .1 .073 .4 * .4 * .4 * .6 .038
t-4 .9 * .9 * .9 * .9 * .1 .066 .4 * .4 * .4 * .4 * .6 .015
All others .590 .011
The value of α producing the smallest error is chosen for use in generating future
forecasts.
To start the algorithm for Equation 13, an initial value for the old smoothed series
must be set. One approach is to set the first smoothed value equal to the first observa-
tion. Example 5 will illustrate this approach. Another method is to use the average of
the first five or six observations for the initial smoothed value.
Example 5
The exponential smoothing technique is demonstrated in Table 7 and Figure 7 for Acme
Tool Company for the years 2000 to 2006, using smoothing constants of .1 and .6. The data
for the first quarter of 2006 will be used as test data to help determine the best value of α
(among the two considered). The exponentially smoothed series is computed by initially set-
ting YN1 equal to 500. If earlier data are available, it might be possible to use them to develop
a smoothed series up to 2000 and then use this experience as the initial value for the
smoothed series. The computations leading to the forecast for periods 3 and 4 are demon-
strated next.
1. Using Equation 13, at time period 2, the forecast for period 3 with # = .1 is
From Table 7, when the smoothing constant is .1, the forecast for the first quarter of 2006
is 469, with a squared error of 145,161. When the smoothing constant is .6, the forecast for
the first quarter of 2006 is 576, with a squared error of 75,076. On the basis of this limited
evidence, exponential smoothing with a = .6 performs better than exponential smoothing
with a = .1.
In Figure 7, note how stable the smoothed values are for the .1 smoothing con-
stant. On the basis of minimizing the mean squared error, MSE (MSE is called MSD
on the Minitab output), over the first 24 quarters, the .6 smoothing constant is better.
Moving Averages and Smoothing Methods
Note: The numbers in parentheses refer to the explanations given in the text in Example 5.
If the mean absolute percentage errors (MAPEs) are compared, the .6 smoothing
constant is also better. To summarize:
However, the MSE and MAPE are both large, and on the basis of these summary
statistics, it is apparent that exponential smoothing does not represent these data well.
As we shall see, a smoothing method that allows for seasonality does a better job of
predicting the Acme Tool Company saw sales.
A factor, other than the choice of α, that affects the values of subsequent forecasts
is the choice of the initial value, YN1 for the smoothed series. In Table 7 (see Example 5),
YN1 = Y1 was used as the initial smoothed value. This choice tends to give Y1 too much
weight in later forecasts. Fortunately, the influence of the initial forecast diminishes
greatly as t increases.
Moving Averages and Smoothing Methods
YN1 = a Yt
1 k
k t=1
Often k is chosen to be a relatively small number. For example, the default approach in
Minitab is to set k = 6.
Moving Averages and Smoothing Methods
Example 6
The computation of the initial value as an average for the Acme Tool Company data
presented in Example 5 is shown next. If k is chosen to equal 6, then the initial value is
The MSE and MAPE for each α when an initial smoothed value of 383.3 is used are
shown next.
a = .1 MSE = 21,091 MAPE = 32.1%
a = .6 MSE = 22,152 MAPE = 36.7%
The initial value, YN1 = 383.3, led to a decrease in the MSE and MAPE for a = .1 but did
not have much effect when a = .6. Now the best model, based on the MSE and MAPE sum-
mary measures, appears to be one that uses a = .1 instead of .6.
Figure 8 shows results for Example 5 when the data are run on Minitab (see the
Minitab Applications section at the end of the chapter for instructions). The smoothing con-
stant of a = .266 was automatically selected by minimizing the MSE. The MSE is reduced
to 19,447, the MAPE equals 32.2%, and although not shown, the MPE equals –6.4%. The
forecast for the first quarter of 2006 is 534.
Figure 9 shows the autocorrelation function for the residuals of the exponential
smoothing method using an alpha of .266. When the Ljung-Box test is conducted for six
time lags, the large value of LBQ (33.86) shows that the first six residual autocorrelations as
a group are larger than would be expected if the residuals were random. In particular, the
significantly large residual autocorrelations at lags 2 and 4 indicate that the seasonal varia-
tion in the data is not accounted for by simple exponential smoothing.
A tracking signal involves computing a measure of forecast errors over time and
setting limits so that, when the cumulative error goes outside those limits, the
forecaster is alerted.
For example, a tracking signal might be used to determine when the size of the
smoothing constant α should be changed. Since a large number of items are usually
being forecast, common practice is to continue with the same value of α for many peri-
ods before attempting to determine if a revision is necessary. Unfortunately, the sim-
plicity of using an established exponential smoothing model is a strong motivator for
not making a change. But at some point, it may be necessary to update α or abandon
exponential smoothing altogether. When the model produces forecasts containing a
great deal of error, a change is appropriate.
A tracking system is a method for monitoring the need for change. Such a system
contains a range of permissible deviations of the forecast from actual values. As long as
forecasts generated by exponential smoothing fall within this range, no change in α is
necessary. However, if a forecast falls outside the range, the system signals a need to
update α.
For instance, if things are going well, the forecasting technique should over-
and underestimate equally often. A tracking signal based on this rationale can be
developed.
Let U equal the number of underestimates out of the last n forecasts. In other
words, U is the number of errors out of the last k that are positive. If the process is in
control, the expected value of U is k/2, but sampling variability is involved, so values
close to k/2 would not be unusual. On the other hand, values that are not close to k/2
would indicate that the technique is producing biased forecasts.
Example 7
Suppose that Acme Tool Company has decided to use the exponential smoothing technique
with α equal to .1, as shown in Example 5. If the process is in control and the analyst decides
to monitor the last 10 error values, U has an expected value of 5. Actually,
Moving Averages and Smoothing Methods
For this example, the permissible absolute error is 279. If for any future forecast the magni-
tude of the error is greater than 279, there is reason to believe that the optimal smoothing
constant α should be updated or a different forecasting method considered.
The preceding discussion on tracking signals also applies to the smoothing meth-
ods yet to be discussed in the rest of the chapter.
Simple exponential smoothing works well when the data vary about an infre-
quently changing level. Whenever a sustained trend exists, exponential smoothing will
lag behind the actual values over time. Holt’s linear exponential smoothing technique,
which is designed to handle data with a well-defined trend, addresses this problem and
is introduced next.
When a trend in the time series is anticipated, an estimate of the current slope, as
well as the current level, is required. Holt’s technique smoothes the level and slope
directly by using different smoothing constants for each. These smoothing constants
provide estimates of level and slope that adapt over time as new observations become
available. One of the advantages of Holt’s technique is that it provides a great deal of
flexibility in selecting the rates at which the level and trend are tracked.
The three equations used in Holt’s method are
1. The exponentially smoothed series, or current level estimate
Lt = #Yt + 11 - #21Lt-1 + Tt-12 (15)
2. The trend estimate
Tt = $1Lt - Lt-12 + 11 - $)Tt-1 (16)
3. The forecast for p periods into the future
YNt+p = Lt + pTt (17)
where
Lt = the new smoothed value (estimate of current level)
a = 1 the smoothing constant for the level 10 6 a 6 12
Yt = the new observation or actual value of the series in period t
b = the smoothing constant for the trend estimate 10 6 b 6 12
Tt = the trend estimate
p = the periods to be forecast into the future
YNt+p " the forecast for p periods into the future
Example 9
In Example 6, simple exponential smoothing did not produce successful forecasts of Acme
Tool Company saw sales. Because Figure 8 suggests that there might be a trend in these data,
Holt’s linear exponential smoothing is used to develop forecasts. To begin the computations
shown in Table 8, two estimated initial values are needed, namely, the initial level and the ini-
tial trend value. The estimate of the level is set equal to the first observation. The trend is esti-
mated to equal zero. The technique is demonstrated in Table 8 for a = .3 and b = .1.
The value for α used here is close to the optimal α (a = .266) for simple exponential
smoothing in Example 6. α is used to smooth the data to eliminate randomness and estimate
level. The smoothing constant b is like α, except that it is used to smooth the trend in the
data. Both smoothing constants are used to average past values and thus to remove ran-
domness. The computations leading to the forecast for period 3 are shown next.
1. Update the exponentially smoothed series or level:
YN t+p = Lt + pTt
Year t Yt Lt Tt YN t+p et
MSE = 20,515.5
On the basis of minimizing the MSE over the period 2000 to 2006, Holt’s linear
smoothing (with a = .3 and b = .1) does not reproduce the data any better than sim-
ple exponential smoothing that used a smoothing constant of .266. A comparison of the
MAPEs shows them to be about the same. When the forecasts for the actual sales for
Moving Averages and Smoothing Methods
the first quarter of 2006 are compared, again Holt smoothing and simple exponential
smoothing are comparable. To summarize:
a = .266 MSE = 19,447 MAPE = 32.2%
a = .3, b = .1 MSE = 20,516 MAPE = 35.4%
Figure 10 shows the results when Holt’s method using a = .3 and b = .1 is run on
Minitab.2 The autocorrelation function for the residuals from Holt’s linear exponential
smoothing is given in Figure 11. The autocorrelation coefficients at time lags 2 and 4
appear to be significant. Also, when the Ljung-Box Q statistic is computed for six time
lags, the large value of LBQ (36.33) shows that the residuals contain extensive autocor-
relation; they are not random. The large residual autocorrelations at lags 2 and 4
suggest a seasonal component may be present in Acme Tool Company data.
The results in Examples 6 and 9 (see Figures 8 and 10) are not much different
because the smoothing constant α is about the same in both cases and the smoothing
constant b in Example 9 is small. (For b = 0, Holt’s linear smoothing reduces to simple
exponential smoothing.)
2In the Minitab program, the trend parameter gamma (g) is identical to our beta (b).
Moving Averages and Smoothing Methods
and an estimate of the seasonal index given by Yt >Lt is added to 11 - g2 times the pre-
vious seasonal component, St-s. This procedure is equivalent to smoothing current and
previous values of Yt >Lt. Yt is divided by the current level estimate, Lt, to create an
index (ratio) that can be used in a multiplicative fashion to adjust a forecast to account
for seasonal peaks and valleys.
The four equations used in Winters’ (multiplicative) smoothing are
1. The exponentially smoothed series, or level estimate:
Yt
Lt = a + 11 - a21Lt-1 + Tt-12 (18)
St-s
2. The trend estimate:
Tt = b1Lt - Lt-12 + 11 - b2Tt-1 (19)
3. The seasonality estimate:
Yt
St = % + 11 - %2St-s (20)
Lt
4. The forecast for p periods into the future:
YNt+p = 1Lt + pTt2St-s+p (21)
where
Lt = the new smoothed value or current level estimate
a = the smoothing constant for the level
Yt = the new observation or actual value in period t
b = the smoothing constant for the trend estimate
Tt = the trend estimate
g = the smoothing constant for the seasonality estimate
St = the seasonal estimate
Moving Averages and Smoothing Methods
Yt
Lt = a + 11 - a21Lt - 1 + Tt - 12
St - s
Y24
L24 = .4 + 11 - .421L24 - 1 + T24 - 12
S24 - 4
650
L24 = .4 + 11 - .421501.286 + 9.1482
1.39628
L24 = .41465.522 + .61510.4342 = 492.469
3In the Minitab program, the trend parameter gamma (g) is identical to our beta (b) and the seasonal param-
eter delta (δ) is identical to our gamma (g) in Equations 19 and 20, respectively.
Moving Averages and Smoothing Methods
Year t Yt Lt Tt St YN t+p et
For the parameter values considered, Winters’ technique is better than both of the
previous smoothing procedures in terms of minimizing the MSE. When the forecasts for
the actual sales for the first quarter of 2006 are compared, Winters’ technique also
appears to do a better job. Figure 13 shows the autocorrelation function for the Winters’
exponential smoothing residuals. None of the residual autocorrelation coefficients
appears to be significantly larger than zero. When the Ljung-Box Q statistic is calculated
for six time lags, the small value of LBQ (5.01) shows that the residual series is random.
Winters’ exponential smoothing method seems to provide adequate forecasts for the
Acme Tool Company data.
Winters’ method provides an easy way to account for seasonality when data have a
seasonal pattern. An alternative method consists of first deseasonalizing or seasonally
adjusting the data. Deseasonalizing is a process that removes the effects of seasonality
from the raw data. The forecasting model is applied to the deseasonalized data, and if
required, the seasonal component is reinserted to provide accurate forecasts.
Exponential smoothing is a popular technique for short-run forecasting. Its major
advantages are its low cost and simplicity. When forecasts are needed for inventory sys-
tems containing thousands of items, smoothing methods are often the only acceptable
approach.
Simple moving averages and exponential smoothing base forecasts on weighted
averages of past measurements. The rationale is that past values contain information
about what will occur in the future. Since past values include random fluctuations as
well as information concerning the underlying pattern of a variable, an attempt is made
Moving Averages and Smoothing Methods
to smooth these values. Smoothing assumes that extreme fluctuations represent ran-
domness in a series of historical observations.
Moving averages are the means of a certain number, k, of values of a variable. The
most recent average is then the forecast for the next period. This approach assigns an
equal weight to each past value involved in the average. However, a convincing argu-
ment can be made for using all the data but emphasizing the most recent values.
Exponential smoothing methods are attractive because they generate forecasts by
using all the observations and assigning weights that decline exponentially as the
observations get older.
APPLICATION TO MANAGEMENT
Forecasts are one of the most important inputs managers have to aid them in the
decision-making process. Virtually every important operating decision depends to
some extent on a forecast. The production department has to schedule employment
needs and raw material orders for the next month or two; the finance department must
determine the best investment opportunities; marketing must forecast the demand for
a new product. The list of forecasting applications is quite lengthy.
Executives are keenly aware of the importance of forecasting. Indeed, a great deal
of time is spent studying trends in economic and political affairs and how events might
affect demand for products and/or services. One issue of interest is the importance
executives place on quantitative forecasting methods compared to their own opinions.
This issue is especially sensitive when events that have a significant impact on demand
are involved. One problem with quantitative forecasting methods is that they depend
on historical data. For this reason, they are probably least effective in determining the
dramatic change that often results in sharply higher or lower demand.
The averaging and smoothing forecasting methods discussed in this chapter are
useful because of their relative simplicity. These simple methods tend to be less costly,
easier to implement, and easier to understand than complex methods. Frequently, the
cost and potential difficulties associated with constructing more sophisticated models
outweigh any gains in their accuracy. For these reasons, small businesses find simple
Moving Averages and Smoothing Methods
Glossary
Exponential smoothing. Exponential smoothing is Simple average. A simple average uses the mean
a procedure for continually revising a forecast in of all relevant historical observations as the fore-
the light of more-recent experience. cast for the next period.
Moving average. A moving average of order k is Tracking signal. A tracking signal involves com-
the mean value of k consecutive observations. The puting a measure of forecast errors over time and
most recent moving average value provides a fore- setting limits so that, when the cumulative error
cast for the next period. goes outside those limits, the forecaster is alerted.
Key Formulas
Naive model
YNt + 1 = Yt (1)
Naive trend model
Yt
YNt + 1 = Yt (3)
Yt - 1
Naive seasonal model for quarterly data
YNt + 1 = Yt - 3 (4)
Naive trend and seasonal model for quarterly data
Yt - Yt - 4
YNt + 1 = Yt - 3 + (5)
4
Simple average model
YNt + 1 = a Yi
1 t
(6)
t i=1
Moving Averages and Smoothing Methods
tYNt + 1 + Yt + 1
YNt + 2 = (7)
t + 1
Moving average for k time periods
Yt + Yt - 1 + Á + Yt - k + 1
YNt + 1 = (8)
k
Double moving average
Mt + Mt-1 + Mt-2 + Á + Mt-k + 1
M ¿t = (9)
k
at = 2Mt - M t¿ (10)
2
bt = 1M - M ¿t2 (11)
k - 1 t
YNt + p = at + bt p (12)
Problems
1. Which forecasting technique continually revises an estimate in the light of more-
recent experiences?
2. Which forecasting technique uses the value for the current period as the forecast
for the next period?
3. Which forecasting technique assigns equal weight to each observation?
4. Which forecasting technique(s) should be tried if the data are trending?
5. Which forecasting technique(s) should be tried if the data are seasonal?
6. Apex Mutual Fund invests primarily in technology stocks. The price of the fund at
the end of each month for the 12 months of 2006 is shown in Table P-6.
a. Find the forecast value of the mutual fund for each month by using a naive
model (see Equation 1). The value for December 2005 was 19.00.
b. Evaluate this forecasting method using the MAD.
c. Evaluate this forecasting method using the MSE.
d. Evaluate this forecasting method using the MAPE.
e. Evaluate this forecasting method using the MPE.
f. Using a naive model, forecast the mutual fund price for January 2007.
g. Write a memo summarizing your findings.
7. Refer to Problem 6. Use a three-month moving average to forecast the mutual
fund price for January 2007. Is this forecast better than the forecast made using the
naive model? Explain.
8. Given the series Yt in Table P-8:
a. What is the forecast for period 9, using a five-period moving average?
b. If simple exponential smoothing with a smoothing constant of .4 is used, what is
the forecast for time period 4?
c. In part b, what is the forecast error for time period 3?
9. The yield on a general obligation bond for the city of Davenport fluctuates with
the market. The monthly quotations for 2006 are given in Table P-9.
TABLE P-6
January 19.39
February 18.96
March 18.20
April 17.89
May 18.43
June 19.98
July 19.51
August 20.63
September 19.78
October 21.25
November 21.18
December 22.14
Moving Averages and Smoothing Methods
TABLE P-8
1 200 200 —
2 210 — —
3 215 — —
4 216 — —
5 219 — —
6 220 — —
7 225 — —
8 226 — —
TABLE P-9
Month Yield
January 9.29
February 9.99
March 10.16
April 10.25
May 10.61
June 11.07
July 11.52
August 11.09
September 10.80
October 10.50
November 10.86
December 9.97
a. Find the forecast value of the yield for the obligation bonds for each month,
starting with April, using a three-month moving average.
b. Find the forecast value of the yield for the obligation bonds for each month,
starting with June, using a five-month moving average.
c. Evaluate these forecasting methods using the MAD.
d. Evaluate these forecasting methods using the MSE.
e. Evaluate these forecasting methods using the MAPE.
f. Evaluate these forecasting methods using the MPE.
g. Forecast the yield for January 2007 using the better technique.
h. Write a memo summarizing your findings.
10. This question refers to Problem 9. Use exponential smoothing with a smoothing
constant of .2 and an initial value of 9.29 to forecast the yield for January 2007. Is
this forecast better than the forecast made using the better moving average
model? Explain.
11. The Hughes Supply Company uses an inventory management method to deter-
mine the monthly demands for various products. The demand values for the last 12
months of each product have been recorded and are available for future forecast-
ing. The demand values for the 12 months of 2006 for one electrical fixture are pre-
sented in Table P-11.
Moving Averages and Smoothing Methods
TABLE P-11
Month Demand
January 205
February 251
March 304
April 284
May 352
June 300
July 241
August 284
September 312
October 289
November 385
December 256
Quarter
Year 1 2 3 4
Source: The Value Line Investment Survey (New York: Value Line,
1990, 1993, 1996).
Moving Averages and Smoothing Methods
Quarter
Year 1 2 3 4
Source: The Value Line Investment Survey (New York: Value Line,
1990, 1993, 1996, 1999).
13. Southdown, Inc., the nation’s third largest cement producer, is pushing ahead
with a waste fuel burning program. The cost for Southdown will total about
$37 million. For this reason, it is extremely important for the company to have an
accurate forecast of revenues for the first quarter of 2000. The data are presented
in Table P-13.
a. Use exponential smoothing with a smoothing constant of .4 and an initial value
of 77.4 to forecast the quarterly revenues for the first quarter of 2000.
b. Now use a smoothing constant of .6 and an initial value of 77.4 to forecast the
quarterly revenues for the first quarter of 2000.
c. Which smoothing constant provides the better forecast?
d. Refer to part c. Examine the residual autocorrelations. Are you happy with sim-
ple exponential smoothing for this example? Explain.
14. The Triton Energy Corporation explores for and produces oil and gas. Company
president Gail Freeman wants to have her company’s analyst forecast the com-
pany’s sales per share for 2000. This will be an important forecast, since Triton’s
restructuring plans have hit a snag. The data are presented in Table P-14.
Determine the best forecasting method and forecast sales per share for 2000.
15. The Consolidated Edison Company sells electricity (82% of revenues), gas (13%),
and steam (5%) in New York City and Westchester County. Bart Thomas, company
forecaster, is assigned the task of forecasting the company’s quarterly revenues for
the rest of 2002 and all of 2003. He collects the data shown in Table P-15.
Determine the best forecasting technique and forecast quarterly revenue for
the rest of 2002 and all of 2003.
16. A job-shop manufacturer that specializes in replacement parts has no forecasting
system in place and manufactures products based on last month’s sales. Twenty-
four months of sales data are available and are given in Table P-16.
Moving Averages and Smoothing Methods
Source: The Value Line Investment Survey (New York: Value Line,
1990, 1993, 1996, 1999)
Source: The Value Line Investment Survey (New York: Value Line, 1990, 1993,
1996, 1999, 2001).
a. Plot the sales data as a time series. Are the data seasonal?
Hint: For monthly data, the seasonal period is s = 12. Is there a pattern (e.g.,
summer sales relatively low, fall sales relatively high) that tends to repeat
itself every 12 months?
Moving Averages and Smoothing Methods
TABLE P-16
b. Use a naive model to generate monthly sales forecasts (e.g., the February
2005 forecast is given by the January 2005 value, and so forth). Compute the
MAPE.
c. Use simple exponential smoothing with a smoothing constant of .5 and an initial
smoothed value of 430 to generate sales forecasts for each month. Compute the
MAPE.
d. Do you think either of the models in parts b and c is likely to generate accurate
forecasts for future monthly sales? Explain.
e. Use Minitab and Winters’ multiplicative smoothing method with smoothing con-
stants a = b = g = .5 to generate a forecast for January 2007. Save the residuals.
f. Refer to part e. Compare the MAPE for Winters’ method from the computer
printout with the MAPEs in parts b and c. Which of the three forecasting proce-
dures do you prefer?
g. Refer to part e. Compute the autocorrelations (for six lags) for the residuals
from Winters’ multiplicative procedure. Do the residual autocorrelations sug-
gest that Winters’ procedure works well for these data? Explain.
17. Consider the gasoline purchases for the Spokane Transit Authority given in Table
2. In Example 3, a five-week moving average is used to smooth the data and gener-
ate forecasts.
a. Use Minitab to smooth the Spokane Transit Authority data using a four-week
moving average. Which moving average length, four weeks or five weeks,
appears to represent the data better? Explain.
b. Use Minitab to smooth the Spokane Transit Authority data using simple expo-
nential smoothing. Compare your results with those in part a. Which procedure,
four-week moving average or simple exponential smoothing, do you prefer for
these data? Explain.
18. Table P-18 contains the number of severe earthquakes (those with a Richter scale
magnitude of 7 and above) per year for the years 1900–1999.
a. Use Minitab to smooth the earthquake data with moving averages of orders
k = 5, 10, and 15. Describe the nature of the smoothing as the order of the mov-
ing average increases. Do you think there might be a cycle in these data? If so,
provide an estimate of the length (in years) of the cycle.
Moving Averages and Smoothing Methods
b. Use Minitab to smooth the earthquake data using simple exponential smooth-
ing. Store the residuals and generate a forecast for the number of severe earth-
quakes in the year 2000. Compute the residual autocorrelations. Does simple
exponential smoothing provide a reasonable fit to these data? Explain.
c. Is there a seasonal component in the earthquake data? Why or why not?
19. This problem was intentionally excluded from this text.
20. Table P-20 contains quarterly sales ($MM) of The Gap for fiscal years 1980–2004.
Plot The Gap sales data as a time series and examine its properties. The objec-
tive is to generate forecasts of sales for the four quarters of 2005. Select an appro-
priate smoothing method for forecasting and justify your choice.
Moving Averages and Smoothing Methods
TABLE P-20 Quarterly Sales for The Gap, Fiscal Years 1980–2004
Year Q1 Q2 Q3 Q4
Source: Based on The Value Line Investment Survey (New York: Value Line, various years),
and 10K filings with the Securities and Exchange Commission.
CASES
4This case was contributed by William P. Darrow of Towson State University, Towson, Maryland.
Moving Averages and Smoothing Methods
energy systems, they were able to put together a than the first. Many of the second-year customers
solar system for heating domestic hot water. The sys- are neighbors of people who had purchased the sys-
tem consists of a 100-gallon fiberglass storage tank, tem in the first year. Apparently, after seeing the
two 36-foot solar panels, electronic controls, PVC system operate successfully for a year, others were
pipe, and miscellaneous fittings. willing to try the solar concept. Sales occur through-
The payback period on the system is 10 years. out the year. Demand for the system is greatest in
Although this situation does not present an attrac- late summer and early fall, when homeowners typi-
tive investment opportunity from a financial point cally make plans to winterize their homes for the
of view, there is sufficient interest in the novelty of upcoming heating season.
the concept to provide a moderate level of sales. With the anticipated growth in the business,
The Johnsons clear about $75 on the $2,000 price of the Johnsons felt that they needed a sales forecast
an installed system, after costs and expenses. to manage effectively in the coming year. It usually
Material and equipment costs account for 75% of takes 60 to 90 days to receive storage tanks after
the installed system cost. An advantage that helps placing the order. The solar panels are available
to offset the low profit margin is the fact that the off the shelf most of the year. However, in the late
product is not profitable enough to generate any summer and throughout the fall, the lead time can
significant competition from heating contractors. be as much as 90 to 100 days. Although there is
The Johnsons operate the business out of their limited competition, lost sales are nevertheless a
home. They have an office in the basement, and real possibility if the potential customer is asked to
their one-car garage is used exclusively to store the wait several months for installation. Perhaps more
system components and materials. As a result, over- important is the need to make accurate sales pro-
head is at a minimum. The Johnsons enjoy a modest jections to take advantage of quantity discount
supplemental income from the company’s opera- buying. These factors, when combined with the
tion. The business also provides a number of tax high cost of system components and the limited
advantages. storage space available in the garage, make it nec-
Bob and Mary are pleased with the growth of essary to develop a reliable forecast. The sales his-
the business. Although sales vary from month to tory for the company’s first two years is given in
month, overall the second year was much better Table 10.
TABLE 10
January 5 17 July 23 44
February 6 14 August 26 41
March 10 20 September 21 33
April 13 23 October 15 23
May 18 30 November 12 26
June 15 38 December 14 17
ASSIGNMENT
1. Identify the model Bob and Mary should use as 2. Forecast sales for 2007.
the basis for their business planning in 2007, and
discuss why you selected this model.
Moving Averages and Smoothing Methods
John Mosby, owner of several Mr. Tux rental stores, the program finds the optimum α value to be
is beginning to forecast his most important business .867. Again he records the appropriate error
variable, monthly dollar sales. One of his employees, measurements:
Virginia Perot, gathered the sales data. John now
MAD = 46,562
wants to create a forecast based on these sales data
using moving average and exponential smoothing MAPE = 44.0%
techniques. John asks the program to use Holt’s linear exponential
John used Minitab to determine that these data smoothing on his data. This program uses the expo-
are both trending and seasonal. He has been told nential smoothing method but can account for a trend
that simple moving averages and exponential in the data as well. John has the program use a
smoothing techniques will not work with these data smoothing constant of .4 for both α and b. The two
but decides to find out for himself. summary error measurements for Holt’s method are
He begins by trying a three-month moving
MAD = 63,579
average. The program calculates several summary
forecast error measurements. These values summa- MAPE = 59.0%
rize the errors found in predicting actual historical John is surprised to find larger measurement
data values using a three-month moving average. errors for this technique. He decides that the seasonal
John decides to record two of these error measure- aspect of the data is the problem. Winters’ multiplica-
ments: tive exponential smoothing is the next method John
tries. This method can account for seasonal factors as
MAD = 54,373
well as trend. John uses smoothing constants of
MAPE = 47.0% a = .2, b = .2, and g = .2. Error measurements are
The MAD (mean absolute deviation) is the average MAD = 25,825
absolute error made in forecasting past values. Each
MAPE = 22.0%
forecast using the three-month moving average
method is off by an average of 54,373. The MAPE When John sits down and begins studying the results
(mean absolute percentage error) shows the error as of his analysis, he is disappointed. The Winters’
a percentage of the actual value to be forecast. The method is a big improvement; however, the MAPE
average error using the three-month moving aver- is still 22%. He had hoped that one of the methods
age technique is 47%, or almost half as large as the he used would result in accurate forecasts of past
value to be forecast. periods; he could then use this method to forecast
Next, John tries simple exponential smoothing. the sales levels for coming months over the next
The program asks him to input the smoothing year. But the average absolute errors (MADs) and
constant (α) to be used or to ask that the optimum percentage errors (MAPEs) for these methods lead
α value be calculated. John does the latter, and him to look for another way of forecasting.
QUESTIONS
1. Summarize the forecast error level for the best methods. What should John do, for example, to
method John has found using Minitab. determine the adequacy of the Winters’ fore-
2. John used the Minitab default values for α, b, casting technique?
and g. John thinks there are other choices for 4. Although not calculated directly in Minitab, the
these parameters that would lead to smaller MPE (mean percentage error) measures fore-
error measurements. Do you agree? cast bias. What is the ideal value for the MPE?
3. Although disappointed with his initial results, What is the implication of a negative sign on the
this may be the best he can do with smoothing MPE?
Moving Averages and Smoothing Methods
The executive director, Marv Harnishfeger, con- seen by CCC for the period from January 1985
cluded that the most important variable that through March 1993. Dorothy then used autocorre-
Consumer Credit Counseling (CCC) needed to lation analysis to explore the data pattern. Use the
forecast was the number of new clients that would results of this investigation to complete the follow-
be seen in the rest of 1993. Marv provided Dorothy ing tasks.
Mercer monthly data for the number of new clients
ASSIGNMENT
1. Develop a naive model to forecast the number 4. Evaluate these forecasting methods using the
of new clients seen by CCC for the rest of 1993. forecast error summary measures.
2. Develop a moving average model to forecast 5. Choose the best model and forecast new clients
the number of new clients seen by CCC for the for the rest of 1993.
rest of 1993. 6. Determine the adequacy of the forecasting
3. Develop an exponential smoothing procedure model you have chosen.
to forecast the number of new clients seen by
CCC for the rest of 1993.
Julie Murphy knows that most important operat- investment opportunities and must forecast the
ing decisions depend, to some extent, on a fore- demand for a new furniture line.
cast. For Murphy Brothers Furniture, sales fore- Julie Murphy used monthly sales for all retail
casts impact adding new furniture lines or stores from 1983 through 1995 to develop a pattern
dropping old ones; planning purchases; setting for Murphy Brothers Furniture sales. Glen Murphy
sales quotas; and making personnel, advertising, discovered actual sales data for the past four years,
and financial decisions. Specifically, Julie is aware 1992 through 1995. Julie was not excited about her
of several current forecasting needs. She knows father’s discovery because she was not sure which
that the production department has to schedule set of data to use to develop a forecast for 1996. She
employees and determine raw material orders for determined that sales for all retail stores had some-
the next month or two. She also knows that her what the same pattern as actual Murphy Brothers’
dad, Glen Murphy, needs to determine the best sales data.
QUESTIONS
1. Do any of the forecasting models studied in this 3. Which data set and forecasting model should
chapter work with the national sales data? Julie use to forecast sales for 1996?
2. Do any of the forecasting models studied in this
chapter work with the actual Murphy Brothers’
sales data?
Moving Averages and Smoothing Methods
is apparent and can be modeled using Holt’s two- Downtown Radiology was analyzed from August
parameter linear exponential smoothing. Smooth-ing 1982 to May 1984. Figure 17 shows the data pattern.
constants of a = .5 and b = .1 are used, and the The data were not seasonal and had no trend or
forecast for period 36, June 1984, is 227. cyclical pattern. For this reason, simple exponential
smoothing was chosen as the appropriate forecast-
Nuclear Medicine Procedures ing method. A smoothing factor of a = .5 was found
The number of nuclear medicine procedures to provide the best model. The forecast for period
performed by the two mobile units owned by 23, June 1984, is 48 nuclear medicine procedures.
Moving Averages and Smoothing Methods
QUESTION
1. Downtown Radiology’s accountant projected Downtown Radiology’s management must
that revenue would be considerably higher make a decision concerning the accuracy of
than that provided by Professional Marketing Professional Marketing Associates’ projections.
Associates. Since ownership interest will be You are asked to analyze the report. What rec-
made available in some type of public offering, ommendations would you make?
Pat Niebuhr and his team are responsible for develop- of the total orders and contacts per order (CPO) sup-
ing a global staffing plan for the contact centers of a plied by the finance department and ultimately
large web retailer. Pat needs to take a monthly forecast forecast the number of customer contacts (phone,
Moving Averages and Smoothing Methods
email, and so forth) arriving at the retailer’s contact Pat thinks it might be a good idea to use the
centers weekly. The contact centers are open 24 hours historical data to generate forecasts of orders and
7 days a week and must be appropriately staffed to contacts per order directly. He is interested in
maintain a high service level. The retailer recognizes determining whether these forecasts are more accu-
that excellent customer service will likely result in rate than the forecasts for these quantities that the
repeat visits and purchases. finance department derives from revenue projec-
The key equation for Pat and his team is tions. As a start, Pat and his team are interested in
the data patterns of the monthly historical orders
Contacts = Orders * CPO
and contacts per order and decide to plot these time
Historical data provide the percentage of contacts series and analyze the autocorrelations. The data
for each day of the week. For example, historically are given in Table 12 and plotted in Figures 19 and
9.10% of the weekly contacts occur on Sundays, 20. The autocorrelations are shown in Figures 21
17.25% of the weekly contacts occur on Mondays, and 22.
and so forth. Keeping in mind the number of Pat is intrigued with the time series plots and
Sundays, Mondays, and so on in a given month, the autocorrelation functions and feels a smoothing
monthly forecasts of contacts can be converted to procedure might be the right tool for fitting the time
weekly forecasts of contacts. It is the weekly fore- series for orders and contacts per order and for gen-
casts that are used for staff planning purposes. erating forecasts.
QUESTIONS
1. What did Pat and his team learn about the data 2. Fit an appropriate smoothing procedure to the
patterns for orders and contacts per order from the orders time series and generate forecasts for
time series plots and autocorrelation functions? the next four months. Justify your choice.
Moving Averages and Smoothing Methods
3. Fit an appropriate smoothing procedure to the contacts directly instead of multiplying forecasts
contacts per order time series and generate fore- of orders and contacts per order to get a fore-
casts for the next four months. Justify your choice. cast. Does this seem reasonable? Why?
4. Use the results for Questions 2 and 3 to generate 6. Many orders consist of more than one item
forecasts of contacts for the next four months. (unit). Would it be better to focus on number of
5. Pat has access to a spreadsheet with historical units and contacts per unit to get a forecast of
actual contacts. He is considering forecasting contacts? Discuss.
Moving Averages and Smoothing Methods
Mary Beasley is responsible for keeping track of the If so, how many new doctors should be hired and/or
number of billable visits to the Medical Oncology be reassigned from other areas?
group at Southwest Medical Center. Her anecdotal To provide some insight into the nature of the
evidence suggests that the number of visits has been demand for service, Mary opens her Excel spreadsheet
increasing and some parts of the year seem to be and examines the total number of billable visits on a
busier than others. Some doctors are beginning to monthly basis for the last several fiscal years. The data
complain about the work load and suggest they are listed in Table 13.
don’t always have enough time to interact with indi- A time series plot of Mary’s data is shown in
vidual patients. Will additional medical staff be Figure 23. As expected, the time series shows
required to handle the apparently increasing demand? an upward trend, but Mary is not sure if there is a
Year Sept. Oct. Nov. Dec. Jan. Feb. Mar. Apr. May Jun. Jul. Aug.
FY1994–95 725 789 893 823 917 811 1,048 970 1,082 1,028 1,098 1,062
FY1995–96 899 1,022 895 828 1,011 868 991 970 934 784 1,028 956
FY1996–97 916 988 921 865 998 963 992 1,118 1,041 1,057 1,200 1,062
FY1997–98 1,061 1,049 829 1,029 1,120 1,084 1,307 1,458 1,295 1,412 1,553 1,480
FY1998–99 1,554 1,472 1,326 1,231 1,251 1,092 1,429 1,399 1,341 1,409 1,367 1,483
FY1999–00 1,492 1,650 1,454 1,373 1,466 1,477 1,466 1,182 1,208 1,132 1,094 1,061
FY2000–01 1,018 1,233 1,112 1,107 1,305 1,181 1,391 1,324 1,259 1,236 1,227 1,294
FY2001–02 1,083 1,404 1,329 1,107 1,313 1,156 1,184 1,404 1,310 1,200 1,396 1,373
FY2002–03 1,259 1,295 1,100 1,097 1,357 1,256 1,350 1,318 1,271 1,439 1,441 1,352
FY2003–04 1,339 1,351 1,197 1,333 1,339 1,307 — — — — — —
seasonal component in the total visits series. She and so forth. Mary knows from a course in her
decides to investigate this issue by constructing the Executive MBA program that Winters’ smoothing
autocorrelation function. If a seasonal component procedure might be a good way to generate fore-
exists with monthly data, Mary expects to see fairly casts of future visits if trend and seasonal compo-
large autocorrelations at the seasonal lags: 12, 24, nents are present.
QUESTIONS
1. What did Mary’s autocorrelation analysis 3. Given the results in Question 2, do you think it
show? is likely another forecasting method would gen-
2. Fit an appropriate smoothing procedure to erate “better” forecasts? Discuss.
Mary’s data, examine the residual autocor- 4. Do you think additional medical staff might be
relations, and generate forecasts for the remain- needed to handle future demand? Write a brief
der of FY2003–04. Do these forecasts seem report summarizing Mary’s data analysis and
reasonable? the implications for additional staff.
Jame Luna investigated the data pattern of monthly steps in selecting a forecasting method are plotting
Surtido cookie sales. In that case, Karin, one of the the sales time series and conducting an autocorrela-
members of Jame’s team, suggested forecasts of tion analysis. He knows you can often learn a lot by
future sales for a given month might be generated simply examining a plot of your time series.
by simply using the historical average sales for that Moreover, the autocorrelations tend to reinforce the
month. After learning something about smoothing pattern observed from the plot. Jame is ready to
methods however, Jame thinks a smoothing proce- begin with the goal of generating forecasts of
dure might be a better way to construct forecasts of monthly cookie sales for the remaining months
future sales. Jame recognizes that important first of 2003.
QUESTIONS
1. What pattern(s) did Jame observe from a time and produce forecasts for the remaining months
series plot of Surtido cookie sales? of 2003.
2. Are the autocorrelations consistent with the pat- 4. Use Karin’s historical monthly average sugges-
tern(s) Jame observed in the time series plot? tion to construct forecasts for the remaining
3. Select and justify an appropriate smoothing months of 2003. Which forecasts, yours or
procedure for forecasting future cookie sales Karin’s, do you prefer? Why?
Minitab Applications
The problem. In Example 3, the Spokane Transit Authority data need to be forecast
using a five-week moving average.
Minitab Solution
1. Enter the Spokane Transit Authority data shown in Table 2 into column C1. Click
on the following menus:
Stat>Time Series>Moving Average
Moving Averages and Smoothing Methods
Excel Applications
The problem. In Example 5, the Acme Tool Company data were forecast using single
exponential smoothing with a smoothing constant equal to .6.
Excel Solution
1. Open the file containing the data presented in Table 1 by clicking on the following
menus:
File>Open
References
Aaker, D. A., and R. Jacobson. “The Sophistication Work Force. Englewood Cliffs, N.J.: Prentice-Hall,
of ‘Naive’ Modeling.” International Journal of 1960.
Forecasting 3 (314) (1987): 449–452. Koehler, A. B., R. D. Snyder, and D. K. Ord.
Bowerman, B. L., R. T. O’Connell, and A. B. Koehler. “Forecasting Models and Prediction Intervals for
Forecasting, Time Series and Regression, 4th ed. the Multiplicative Holt-Winters Method.”
Belmont, CA: Thomson Brooks/Cole, 2005. International Journal of Forecasting 17 (2001):
Dalrymple, D. J., and B. E. King. “Selecting 269–286.
Parameters for Short-Term Forecasting Ledolter, J., and B. Abraham. “Some Comments on
Techniques.” Decision Sciences 12 (1981): the Initialization of Exponential Smoothing.”
661–669. Journal of Forecasting 3 (1) (1984): 79–84.
Gardner, E. S., Jr. “Exponential Smoothing: The State Makridakis, S., S. C. Wheelwright, and R. Hyndman.
of the Art.” Journal of Forecasting 4 (1985): 1–28. Forecasting Methods and Applications. New York:
Gardner, E. S., Jr., and D. G. Dannenbring. Wiley, 1998.
“Forecasting with Exponential Smoothing: Some McKenzie, E. “An Analysis of General Exponential
Guidelines for Model Selection.” Decision Smoothing.” Operations Research 24 (1976):
Sciences 11 (1980): 370–383. 131–140.
Holt, C. C. “Forecasting Seasonals and Trends by Newbold, P., and T. Bos. Introductory Business and
Exponentially Weighted Moving Averages.” Economic Forecasting, 2nd ed. Cincinnati, Ohio:
International Journal of Forecasting 20 (2004): South-Western, 1994.
5–10. Winters, P. R. “Forecasting Sales by Exponentially
Holt, C. C., F. Modigliani, J. F. Muth, and H. A. Weighted Moving Averages.” Management
Simon. Planning Production Inventories and Science 6 (1960): 324–342.
TIME SERIES AND THEIR
COMPONENTS
Observations of a variable Y that become available over time are called time series data, or
simply a time series. These observations are often recorded at fixed time intervals. For
example, Y might represent sales, and the associated time series could be a sequence of
annual sales figures. Other examples of time series include quarterly earnings, monthly
inventory levels, and weekly exchange rates. In general, time series do not behave like ran-
dom samples and require special methods for their analysis. Observations of a time series
are typically related to one another (autocorrelated). This dependence produces patterns
of variability that can be used to forecast future values and assist in the management of
business operations. Consider these situations.
American Airlines (AA) compares current reservations with forecasts based on
projections of historical patterns. Whether current reservations are lagging behind or
exceeding the projections, AA adjusts the proportion of discounted seats accordingly.
The adjustments are made for each flight segment in the AA system.
A Canadian importer of cut flowers buys from growers in the United States,
Mexico, Central America, and South America. However, because these sources pur-
chase their growing stock and chemicals from the United States, all the selling prices
are quoted in U.S. dollars at the time of the sale. An invoice is not paid immediately,
and since the Canadian–U.S. exchange rate fluctuates, the cost to the importer in
Canadian dollars is not known at the time of purchase. If the exchange rate does not
change before the invoice is paid, there is no monetary risk to the importer. If the index
rises, the importer loses money for each U.S. dollar of purchase. If the index drops, the
importer gains. The importer uses forecasts of the weekly Canadian-dollar-to-U.S.-dollar
exchange rate to manage the inventory of cut flowers.
Although time series are often generated internally and are unique to the organi-
zation, many time series of interest in business can be obtained from external sources.
Publications such as Statistical Abstract of the United States, Survey of Current
Business, Monthly Labor Review, and Federal Reserve Bulletin contain time series of
all types. These and other publications provide time series data on prices, production,
sales, employment, unemployment, hours worked, fuel used, energy produced, earn-
ings, and so on, reported on a monthly, quarterly, or annual basis. Today, extensive col-
lections of time series are available on the World Wide Web at sites maintained by U.S.
government agencies, statistical organizations, universities, and individuals.
It is important that managers understand the past and use historical data and
sound judgment to make intelligent plans to meet the demands of the future. Properly
constructed time series forecasts help eliminate some of the uncertainty associated
with the future and can assist management in determining alternative strategies.
From Chapter 5 of Business Forecasting, Ninth Edition. John E. Hanke, Dean W. Wichern.
Copyright © 2009 by Pearson Education, Inc. All rights reserved.
Time Series and Their Components
DECOMPOSITION
One approach to the analysis of time series data involves an attempt to identify the
component factors that influence each of the values in a series. This identification pro-
cedure is called decomposition. Each component is identified separately. Projections of
each of the components can then be combined to produce forecasts of future values of
the time series. Decomposition methods are used for both short-run and long-run fore-
casting. They are also used to simply display the underlying growth or decline of a
series or to adjust the series by eliminating one or more of the components.
Analyzing a time series by decomposing it into its component parts has a long
history. Recently, however, decomposition methods of forecasting have lost some of
their luster. Projecting the individual components into the future and recombining
these projections to form a forecast of the underlying series often does not work very
well in practice. The difficulty lies in getting accurate forecasts of the components. The
development of more-flexible model-based forecasting procedures has made decom-
position primarily a tool for understanding a time series rather than a forecasting
method in its own right.
To understand decomposition, we start with the four components of a time series.
These are the trend component, the cyclical component, the seasonal component, and
the irregular or random component.
1. Trend. The trend is the component that represents the underlying growth (or
decline) in a time series. The trend may be produced, for example, by consistent
population change, inflation, technological change, and productivity increases. The
trend component is denoted by T.
2. Cyclical. The cyclical component is a series of wavelike fluctuations or cycles of
more than one year’s duration. Changing economic conditions generally produce
cycles. C denotes the cyclical component.
In practice, cycles are often difficult to identify and are frequently regarded as
part of the trend. In this case, the underlying general growth (or decline) compo-
nent is called the trend-cycle and denoted by T. We use the notation for the trend
because the cyclical component often cannot be separated from the trend.
3. Seasonal. Seasonal fluctuations are typically found in quarterly, monthly, or weekly
data. Seasonal variation refers to a more or less stable pattern of change that
appears annually and repeats itself year after year. Seasonal patterns occur
Time Series and Their Components
To study the components of a time series, the analyst must consider how the
components relate to the original series. This task is accomplished by specifying a model
(mathematical relationship) that expresses the time series variable Y in terms of the
components T, C, S, and I. A model that treats the time series values as a sum of the
components is called an additive components model. A model that treats the time series
values as the product of the components is called a multiplicative components model.
Both models are sometimes referred to as unobserved components models, since, in
practice, although we observe the values of the time series, the values of the components
are not observed. The approach to time series analysis described in this chapter involves
an attempt, given the observed series, to estimate the values of the components. These
estimates can then be used to forecast or to display the series unencumbered by the sea-
sonal fluctuations. The latter process is called seasonal adjustment.
It is difficult to deal with the cyclical component of a time series. To the extent that
cycles can be determined from historical data, both their lengths (measured in years)
and their magnitudes (differences between highs and lows) are far from constant. This
lack of a consistent wavelike pattern makes distinguishing cycles from smoothly evolv-
ing trends difficult. Consequently, to keep things relatively simple, we will assume any
cycle in the data is part of the trend. Initially, then, we consider only three components,
T, S, and I. A brief discussion of one way to handle cyclical fluctuations in the decom-
position approach to time series analysis is available in the Cyclical and Irregular
Variations section of this chapter.
The two simplest models relating the observed value (Yt) of a time series to the
trend (Yt ), seasonal (St), and irregular (It) components are the additive components
model
Yt = Tt + St + It (1)
Yt = Tt * St * It (2)
The additive components model works best when the time series being analyzed
has roughly the same variability throughout the length of the series. That is, all the val-
ues of the series fall essentially within a band of constant width centered on the trend.
The multiplicative components model works best when the variability of the time
series increases with the level.1 That is, the values of the series spread out as the trend
increases, and the set of observations has the appearance of a megaphone or funnel. A
time series with constant variability and a time series with variability increasing with
1,000
900
Milk Production
800
700
600
50 100 150
Month
900
800
700
600
Monthly Sales
500
400
300
200
100
0
10 20 30 40 50 60 70
Month
level are shown in Figure 1. Both of these monthly series have an increasing trend and
a clearly defined seasonal pattern.2
Trend
Trends are long-term movements in a time series that can sometimes be described by a
straight line or a smooth curve. Examples of the basic forces producing or affecting the
trend of a series are population change, price change, technological change, productiv-
ity increases, and product life cycles.
2Variants of the decomposition models (see Equations 1 and 2) exist that contain both multiplicative and
additive terms. For example, some software packages do “multiplicative” decomposition using the model
Y = T * S + I.
Time Series and Their Components
A population increase may cause retail sales of a community to rise each year for
several years. Moreover, the sales in current dollars may have been pushed upward
during the same period because of general increases in the prices of retail goods—even
though the physical volume of goods sold did not change.
Technological change may cause a time series to move upward or downward. The
development of high-speed computer chips, enhanced memory devices, and improved
display panels, accompanied by improvements in telecommunications technology, has
resulted in dramatic increases in the use of personal computers and cellular
telephones. Of course, the same technological developments have led to a downward
trend in the production of mechanical calculators and rotary telephones.
Productivity increases—which, in turn, may be due to technological change—give
an upward slope to many time series. Any measure of total output, such as manufactur-
ers’ sales, is affected by changes in productivity.
For business and economic time series, it is best to view the trend (or trend-cycle)
as smoothly changing over time. Rarely can we realistically assume that the trend can
be represented by some simple function such as a straight line over the whole period
for which the time series is observed. However, it is often convenient to fit a trend
curve to a time series for two reasons: (1) It provides some indication of the general
direction of the observed series, and (2) it can be removed from the original series to
get a clearer picture of the seasonality.
If the trend appears to be roughly linear—that is, if it increases or decreases like a
straight line—then it is represented by the equation
TNt = b0 + b1t (3)
Here, TNt is the predicted value for the trend at time t. The symbol t represents time, the
independent variable, and ordinarily assumes integer values 1, 2, 3, . . . corresponding
to consecutive time periods. The slope coefficient, b1, is the average increase or
decrease in T for each one-period increase in time.
Time trend equations, including the straight-line trend, can be fit to the data using
the method of least squares. Recall that the method of least squares selects the values
of the coefficients in the trend equation (b0 and b1 in the straight-line case) so that the
estimated trend values (TNt) are close to the actual values (Yt ) as measured by the sum
of squared errors criterion
FIGURE 3 Trend Line for the Car Registrations Time Series for Example 1
Trend
Sales, Units
curve are quite different depending on the stage of the cycle. A curve, other than a
straight line, is needed to model the trend over a new product’s life cycle.
A simple function that allows for curvature is the quadratic trend
TNt = b0 + b1t + b2t2 (5)
As an illustration, Figure 5 shows a quadratic trend curve fit to the passenger car regis-
trations data of Example 1 using the SSE criterion. The quadratic trend can be pro-
jected beyond the data for, say, two additional years, 1993 and 1994. We will consider
the implications of this projection in the next section, Forecasting Trend.
Based on the MAPE, MAD, and MSD accuracy measures, a quadratic trend appears
to be a better representation of the general direction of the car registrations series than
the linear trend in Figure 3. Which trend model is appropriate? Before considering this
issue, we will introduce a few additional trend curves that have proved useful.
When a time series starts slowly and then appears to be increasing at an increasing
rate (see Figure 4) such that the percentage difference from observation to observation
is constant, an exponential trend can be fitted. The exponential trend is given by
The coefficient b1 is related to the growth rate. If the exponential trend is fit to annual
data, the annual growth rate is estimated to be 1001b1 - 12%.
Figure 6 indicates the number of mutual fund salespeople employed by a particu-
lar company for several consecutive years. The increase in the number of salespeople is
not constant. It appears as if increasingly larger numbers of people are being added in
the later years.
An exponential trend curve fit to the salespeople data has the equation
TNt = 10.01611.3132t
implying an annual growth rate of about 31%. Consequently, if the model estimates
51 salespeople for 2005, the increase for 2006 would be 16 151 * .312 for an estimated
total of 67. This can be compared to the actual value of 68 salespeople.
A linear trend fit to the salespeople data would indicate a constant average
increase of about nine salespeople per year. This trend overestimates the actual
increase in the earlier years and underestimates the increase in the last year. It does not
model the apparent trend in the data as well as the exponential curve does.
It is clear that extrapolating an exponential trend with a 31% growth rate will quickly
result in some very big numbers. This is a potential problem with an exponential trend
model. What happens when the economy cools off and stock prices begin to retreat? The
demand for mutual fund salespeople will decrease, and the number of salespeople could
even decline. The trend forecast by the exponential curve will be much too high.
Growth curves of the Gompertz or logistic type represent the tendency of many
industries and product lines to grow at a declining rate as they mature. If the plotted
data reflect a situation in which sales begin low, then increase as the product catches on,
and finally ease off as saturation is reached, the Gompertz curve or Pearl–Reed logistic
model might be appropriate. Figure 7 shows a comparison of the general shapes of
(a) the Gompertz curve and (b) the Pearl–Reed logistic model. Note that the logistic
curve is very similar to the Gompertz curve but has a slightly gentler slope. Figure 7
shows how the Y intercepts and maximum values for these curves are related to some of
the coefficients in their functional forms. The formulas for these trend curves are com-
plex and not within the scope of this text. Many statistical software packages, including
Minitab, allow one to fit several of the trend models discussed in this section.
Although there are some objective criteria for selecting an appropriate trend, in
general, the correct choice is a matter of judgment and therefore requires experience
and common sense on the part of the analyst. As we will discuss in the next section, the
line or curve that best fits a set of data points might not make sense when projected as
the trend of the future.
Forecasting Trend
Suppose we are presently at time t = n (end of series) and we want to use a trend model
to forecast the value of Y, p steps ahead. The time period at which we make the forecast,
n in this case, is called the forecast origin. The value p is called the lead time. For the lin-
ear trend model, we can produce a forecast by evaluating TNn + p = b0 + b11n + p2.
Using the trend line fitted to the car registration data in Example 1, a forecast of the
trend for 1993 (t = 34) made in 1992 (t = n = 33) would be the p = 1 step ahead forecast
^ ^
Tt Tt
b0 1
b0
1
b 0b 1 b0 + b1
t t
0 0
(a) Gompertz Trend Curve (b) Logistic (Pearl-Reed) Trend Curve
Figure 5 shows the fitted quadratic trend curve for the car registration data. Using
the equation shown in the figure, we can calculate forecasts of the trend for 1993 and
1994 by setting t = 33 + 1 = 34 and t = 33 + 2 = 35. The reader can verify that
TN33 + 1 = 8.688 and TN33 + 2 = 8.468. These numbers were plotted in Figure 5 as extrapola-
tions of the quadratic trend curve.
Recalling that car registrations are measured in millions, the two forecasts of trend
produced from the quadratic curve are quite different from the forecasts produced by
the linear trend equation. Moreover, they are headed in the opposite direction. If we
were to extrapolate the linear and quadratic trends for additional time periods, their
differences would be magnified.
The car registration example illustrates why great care must be exercised in using
fitted trend curves for the purpose of forecasting future trends. Two equations, both of
which may reasonably represent the observed time series, can give very different
results when projected over future time periods. These differences can be substantial
for large lead times (long-run forecasting).
Trend curve models are based on the following assumptions:
1. The correct trend curve has been selected.
2. The curve that fits the past is indicative of the future.
These assumptions suggest that judgment and expertise play a substantial role in the
selection and use of a trend curve. To use a trend curve for forecasting, we must be able
to argue that the correct trend has been selected and that, in all likelihood, the future
will be like the past.
There are objective criteria for selecting a trend curve. However, although these
and other criteria help to determine an appropriate model, they do not replace good
judgment.
Seasonality
A seasonal pattern is one that repeats itself year after year. For annual data, seasonal-
ity is not an issue because there is no chance to model a within-year pattern with data
recorded once per year. Time series consisting of weekly, monthly, or quarterly obser-
vations, however, often exhibit seasonality.
The analysis of the seasonal component of a time series has immediate short-term
implications and is of greatest importance to mid- and lower-level management.
Marketing plans, for example, have to take into consideration expected seasonal pat-
terns in consumer purchases.
Several methods for measuring seasonal variation have been developed. The basic
idea in all of these methods is to first estimate and remove the trend from the original
series and then smooth out the irregular component. Keeping in mind our decomposi-
tion models, this leaves data containing only seasonal variation. The seasonal values
are then collected and summarized to produce a number (generally an index number)
for each observed interval of the year (week, month, quarter, and so on).
Thus, the identification of the seasonal component in a time series differs from
trend analysis in at least two ways:
1. The trend is determined directly from the original data, but the seasonal compo-
nent is determined indirectly after eliminating the other components from the
data, so that only the seasonality remains.
2. The trend is represented by one best-fitting curve, or equation, but a separate sea-
sonal value has to be computed for each observed interval (week, month, quarter)
of the year and is often in the form of an index number.
Time Series and Their Components
With monthly data, for example, a seasonal index of 1.0 for a particular month means
the expected value for that month is 1/12 the total for the year. An index of 1.25 for a dif-
ferent month implies the observation for that month is expected to be 25% more than
1/12 of the annual total.A monthly index of 0.80 indicates that the expected level of activ-
ity that month is 20% less than 1/12 of the total for the year, and so on.The index numbers
indicate the expected ups and downs in levels of activity over the course of a year after the
effects due to the trend (or trend-cycle) and irregular components have been removed.
To highlight seasonality, we must first estimate and remove the trend. The trend
can be estimated with one of the trend curves we discussed previously, or it can be esti-
mated using a moving average.
Assuming a multiplicative decomposition model, the ratio to moving average is a
popular method for measuring seasonal variation. In this method, the trend is esti-
mated using a centered moving average. We illustrate the ratio-to-moving-average
method using the monthly sales of the Cavanaugh Company, shown in Figure 1
(Bottom) in the next example.
Example 2
To illustrate the ratio-to-moving-average method, we use two years of the monthly sales of
the Cavanaugh Company.3 Table 2 gives the monthly sales from January 2004 to December
2005 to illustrate the beginning of the computations. The first step for monthly data is to
compute a 12-month moving average (for quarterly data, a four-month moving average
would be computed). Because all of the months of the year are included in the calculation of
this moving average, effects due to the seasonal component are removed, and the moving
average itself contains only the trend and the irregular components.
The steps (identified in Table 2) for computing seasonal indexes by the ratio-to-moving-
average method follow:
Step 1. Starting at the beginning of the series, compute the 12-month moving total, and
place the total for January 2004 through December 2004 between June and July
2004.
Step 2. Compute a two-year moving total so that the subsequent averages are centered on
July rather than between months.
Step 3. Since the two-year total contains the data for 24 months (January 2004 once,
February 2004 to December 2004 twice, and January 2005 once), this total is cen-
tered on July 2004.
Step 4. Divide the two-year moving total by 24 in order to obtain the 12-month centered
moving average.
Step 5. The seasonal index for July is calculated by dividing the actual value for July by the
12-month centered moving average.4
3The units have been omitted and the dates and name have been changed to protect the identity of the company.
4This is the ratio-to-moving-average operation that gives the procedure its name.
Time Series and Their Components
12-Month
12-Month Two-Year Centered
Moving Moving Moving Seasonal
Period Sales Total Total Average Index
2004
January
4
518
February 404
March 300
April 210
May 196
June 186 1 4,869 2
4
July 247 4,964 9,833} 3 409.7} 4 0.60} 5
August 343 4,952 9,916 413.2 0.83
September 464 4,925 9,877 411.5 1.13
October 680 5,037 9,962 415.1 1.64
November 711 5,030 10,067 419.5 1.69
December 610 10,131 422.1 1.45
2005
January 613 5,101 10,279 428.3 1.43
February 392 5,178 10,417 434.0 0.90
March 273 5,239 10,691 445.5 0.61
April 322 5,452 11,082 461.8 0.70
May 189 5,630 11,444 476.8 0.40
June 257 5,814 11,682 486.8 0.53
July 324 5,868
August 404
September 677
October 858
November 895
December 664
Repeat steps 1 to 5 beginning with the second month of the series, August 2004, and so on.
The process ends when a full 12-month moving total can no longer be calculated.
Because there are several estimates (corresponding to different years) of the seasonal
index for each month, they must be summarized to produce a single number. The median,
rather than the mean, is used as the summary measure. Using the median eliminates the
influence of data for a month in a particular year that are unusually large or small. A sum-
mary of the seasonal ratios, along with the median value for each month, is contained in
Table 3.
The monthly seasonal indexes for each year must sum to 12, so the median for each
month must be adjusted to get the final set of seasonal indexes.5 Since this multiplier should
be greater than 1 if the total of the median ratios before adjustment is less than 12 and
smaller than 1 if the total is greater than 12, the multiplier is defined as
12
Multiplier =
Actual total
5The monthly indexes must sum to 12 so that the expected annual total equals the actual annual total.
Time Series and Their Components
Adjusted
Seasonal
Index (Median
Month 2000 2001 2002 2003 2004 2005 2006 Median ! 1.0044)
January — 1.208 1.202 1.272 1.411 1.431 — 1.272 1.278
February — 0.700 0.559 0.938 1.089 0.903 — 0.903 0.907
March — 0.524 0.564 0.785 0.800 0.613 — 0.613 0.616
April — 0.444 0.433 0.480 0.552 0.697 — 0.480 0.482
May — 0.424 0.365 0.488 0.503 0.396 — 0.424 0.426
June — 0.490 0.459 0.461 0.465 0.528 0.465 0.467
July 0.639 0.904 0.598 0.681 0.603 0.662 0.651 0.654
August 1.115 0.913 0.889 0.799 0.830 0.830 0.860 0.864
September 1.371 1.560 1.346 1.272 1.128 1.395 1.359 1.365
October 1.792 1.863 1.796 1.574 1.638 1.771 1.782 1.790
November 1.884 2.012 1.867 1.697 1.695 1.846 1.857 1.865
December 1.519 1.088 1.224 1.282 1.445 — 1.282 1.288
11.948 12.002
12
Multiplier = = 1.0044
11.948
The final column in Table 3 contains the final seasonal index for each month, determined by
making the adjustment (multiplying by 1.0044) to each of the median ratios.6 The final sea-
sonal indexes, shown in Figure 8, represent the seasonal component in a multiplicative
decomposition of the sales of the Cavanaugh Company time series. The seasonality in sales
Seasonal Indices
1.9
1.4
Index
0.9
0.4
1 2 3 4 5 6 7 8 9 10 11 12
Month
6The seasonal indexes are sometimes multiplied by 100 and expressed as percentages.
Time Series and Their Components
is evident from Figure 8. Sales for this company are periodic, with relatively low sales in the
late spring and relatively high sales in the late fall.
Our analysis of the sales series in Example 2 assumed that the seasonal pattern
remained constant from year to year. If the seasonal pattern appears to change over
time, then estimating the seasonal component with the entire data set can produce mis-
leading results. It is better, in this case, either (1) to use only recent data (from the last
few years) to estimate the seasonal component or (2) to use a time series model that
allows for evolving seasonality.
The seasonal analysis illustrated in Example 2 is appropriate for a multiplicative
decomposition model. However, the general approach outlined in steps 1 to 5 works
for an additive decomposition if, in step 5, the seasonality is estimated by subtracting
the trend from the original series rather than dividing by the trend (moving average) to
get an index. In an additive decomposition, the seasonal component is expressed in the
same units as the original series.
In addition, it is apparent from our sales example that using a centered moving
average to determine trend results in some missing values at the ends of the series. This
is particularly problematic if forecasting is the objective. To forecast future values using
a decomposition approach, alternative methods for estimating the trend must be used.
The results of a seasonal analysis can be used to (1) eliminate the seasonality in
data; (2) forecast future values; (3) evaluate current positions in, for example, sales,
inventory, and shipments; and (4) schedule production.
Yt - St = Tt + It (7a)
Yt
= Tt * It (7b)
St
Most economic series published by government agencies are seasonally adjusted
because seasonal variation is not of primary interest. Rather, it is the general pattern of
economic activity, independent of the normal seasonal fluctuations, that is of interest.
For example, new car registrations might increase by 10% from May to June, but is this
increase an indication that new car sales are completing a banner quarter? The answer
is “no” if the 10% increase is typical at this time of year largely due to seasonal factors.
In a survey concerned with the acquisition of seasonally adjusted data, Bell and
Hillmer (1984) found that a wide variety of users value seasonal adjustment. They
identified three motives for seasonal adjustment:
Bell and Hillmer concluded that “seasonal adjustment is done to simplify data so that
they may be more easily interpreted by statistically unsophisticated users without a sig-
nificant loss of information”.
Yt Tt * Ct * St * It
= = Ct * It (8)
Tt * St Tt * St
A moving average can be used to smooth out the irregularities, It, leaving the cyclical
component, Ct. To eliminate the centering problem encountered when a moving aver-
age with an even number of time periods is used, the irregularities are smoothed using a
moving average with an odd number of time periods. For monthly data, a 5-, a 7-, a
9-, or even an 11-period moving average will work. For quarterly data, an estimate of C
can be computed using a three-period moving average of the values.8
Finally, the irregular component is estimated by
Ct * It
It = (9)
Ct
The irregular component represents the variability in the time series after the other
components have been removed. It is sometimes called the residual or error. With a
multiplicative decomposition, both the cyclical and the irregular components are
expressed as indexes.
Summary Example
One reason for decomposing a time series is to isolate and examine the components of
the series. After the analyst is able to look at the trend, seasonal, cyclical, and irregular
components of a series one at a time, insights into the patterns in the original data val-
ues may be gained. Also, once the components have been isolated, they can be recom-
bined or synthesized to produce forecasts of future values of the time series.
Example 3
In Example 3.5, Perkin Kendell, the analyst for the Coastal Marine Corporation, used auto-
correlation analysis to determine that sales were seasonal on a quarterly basis. Now he uses
decomposition to understand the quarterly sales variable. Perkin uses Minitab (see the
Minitab Applications section at the end of the chapter) to produce Table 4 and Figure 9. In
order to keep the seasonal pattern current, only the last seven years (2000 to 2006) of sales
data, Y, were analyzed.
The original data are shown in the chart in the upper left of Figure 10. The trend is com-
puted using the linear model: TNt = 261.24 + .759t. Since 1 represented the first quarter of
2000, Table 4 shows the trend value equal to 262.000 for this time period, and estimated sales
(the T column) increased by .759 each quarter.
7Notice that we have added the cyclical component, C, to the multiplicative decomposition shown in
Equation 2.
8For annual data, there is no seasonal component, and the cyclical * irregular component is obtained by
simply removing the trend from the original series.
Time Series and Their Components
The chart in the upper right of Figure 10 shows the detrended data. These data are also
shown in the SCI column of Table 4. The detrended value for the first quarter of 2000 was9
Y 232.7
SCI = = = .888
T 262.000
The seasonally adjusted data are shown in the TCI column in Table 4 and in Figure 10.
The seasonally adjusted value for the first quarter of 2000 was
232.7
TCI = = 298.458
.77967
Sales in the first quarter of 2005 were 242.6. However, examination of the seasonally
adjusted column shows that the sales for this quarter were actually high when the data were
adjusted for the fact that the first quarter is typically a very weak quarter.
9To simplify the notation in this example, we omit the subscript t on the original data, Y, and each of its com-
ponents, T, S, C, and I. We also omit the multiplication sign, !, between components, since it is clear we are
considering a multiplicative decomposition.
FIGURE 9 Minitab Output for Decomposition of Coastal Marine
Quarterly Sales for Example 3
The chart on the left of Figure 11 shows the seasonal components relative to 1.0. We see that
first-quarter sales are 22% below average, second-quarter sales are about as expected,
third-quarter sales are almost 12% above average, and fourth-quarter sales are almost 9%
above normal.
The cyclical-irregular value for first quarter of 2000 was10
Y 232.7
CI = = = 1.139
TS 1262.00021.779672
In order to calculate the cyclical column, a three-period moving average was computed. The
value for the second quarter of 2000 was
1.139
1.159
1.056
3.354 3.354/3 = 1.118
Notice how smooth the C column is compared to the CI column. The reason is that using the
moving average has smoothed out the irregularities.
Finally, the I column was computed. For example, for the second quarter of 2000,
CI 1.159
I = = = 1.036
C 1.118
Examination of the I column shows that there were some large changes in the irregular
component. The irregular index dropped from 111.4% in the fourth quarter of 2002 to 87%
10Minitab calculates the cyclical * irregular component (or simply the irregular component if no cyclical
component is contemplated) by subtracting the trend * seasonal component from the original data. In sym-
bols, Minitab sets CI = Y - TS. The Minitab CI component is shown in the lower right-hand chart of Figure
10. Moreover, Minitab fits the trend line to the seasonally adjusted data. That is, the seasonal adjustment is
done before the trend is determined.
Time Series and Their Components
in the first quarter of 2003 and then increased to 106.2% in the second quarter of 2003. This
behavior results from the unusually low sales in the first quarter of 2003.
Business Indicators
The cyclical indexes can be used to answer the following questions:
1. Does the series cycle?
2. If so, how extreme is the cycle?
3. Does the series follow the general state of the economy (business cycle)?
One way to investigate cyclical patterns is through the study of business indicators.
A business indicator is a business-related time series that is used to help assess the gen-
eral state of the economy, particularly with reference to the business cycle. Many busi-
nesspeople and economists systematically follow the movements of such statistical
series to obtain economic and business information in the form of an unfolding picture
that is up to date, comprehensive, relatively objective, and capable of being read and
understood with a minimum expenditure of time.
Business indicators are business-related time series that are used to help assess
the general state of the economy.
The most important list of statistical indicators originated during the sharp busi-
ness setback from 1937 to 1938. Secretary of the Treasury Henry Morgenthau
requested the National Bureau of Economic Research (NBER) to devise a system that
would signal when the setback was nearing an end. Under the leadership of Wesley
Mitchell and Arthur F. Burns, NBER economists selected 21 series that from past per-
formance promised to be fairly reliable indicators of business revival. Since then, the
business indicator list has been revised several times. The current list consists of 21
indicators—10 classified as leading, 4 as coincident, and 7 as lagging.
1. Leading indicators. In practice, components of the leading series are studied to help
anticipate turning points in the economy. The Survey of Current Business publishes
this list each month, along with the actual values of each series for the past several
months and the most recent year. Also, a composite index of leading indicators is
computed for each month and year; the most recent monthly value is frequently
reported in the popular press to indicate the general direction of the future economy.
Examples of leading indicators are building permits and an index of stock prices.
2. Coincident indicators. The four coincident indicators provide a measure of how the
U.S. economy is currently performing. An index of these four series is computed
each month. Examples of coincident indicators are industrial production and man-
ufacturing and trade sales.
3. Lagging indicators. The lagging indicators tend to lag behind the general state of
the economy, both on the upswing and on the downswing. A composite index is
also computed for this list. Examples of lagging indicators are the prime interest
rate and the unemployment rate.
Cycles imply turning points. That is, turning points come into existence only as a
consequence of a subsequent decline or gain in the business cycle. Leading indicators
change direction ahead of turns in general business activity, coincident indicators turn at
about the same time as the general economy, and turns in the lagging indicators follow
Time Series and Their Components
those of the general economy. However, it is difficult to identify cyclical turning points
at the time they occur, since all areas of the economy do not expand at the same time
during periods of expansion or contract at the same time during periods of contraction.
Hence, several months may go by before a genuine cyclical upturn or downturn is finally
identified with any assurance.
Leading indicators are the most useful predictive tool, since they attempt to signal
economic changes in advance. Coincident and lagging indicators are of minimal inter-
est from a forecasting perspective, but they are used to assess the effectiveness of cur-
rent and past economic policy and so to help formulate future policy.
In their article entitled “Early Warning Signals for the Economy,” Geoffrey H.
Moore and Julius Shiskin (1976) have the following to say on the usefulness of business
cycle indicators:
It seems clear from the record that business cycle indicators are helpful in
judging the tone of current business and short-term prospects. But because of
their limitations, the indicators must be used together with other data and with
full awareness of the background of business and consumer confidence and
expectations, governmental policies, and international events. We also must
anticipate that the indicators will often be difficult to interpret, that interpreta-
tions will sometimes vary among analysts, and that the signals they give will
not be correctly interpreted. Indicators provide a sensitive and revealing pic-
ture of the ebb and flow of economic tides that a skillful analyst of the eco-
nomic, political, and international scene can use to improve his chances of
making a valid forecast of short-run economic trends. If the analyst is aware of
their limitations and alert to the world around him, he will find the indicators
useful guideposts for taking stock of the economy and its needs.
1. Trend. The quarterly trend equation is TNt = 261.24 + .759t. The forecast origin is the
fourth quarter of 2006, or time period t = n = 28. Sales for the first quarter of 2007
occurred in time period t = 28 + 1 = 29. This notation shows we are forecasting p = 1
Time Series and Their Components
period ahead from the end of the time series. Setting t = 29, the trend projection is
then
2. Seasonal. The seasonal index for the first quarter, .77967, is given in Figure 9.
3. Cyclical. The cyclical projection must be determined from the estimated cyclical pat-
tern (if any) and from any other information generated by indicators of the general
economy for 2007. Projecting the cyclical pattern for future time periods is fraught with
uncertainty, and as we indicated earlier, it is generally assumed, for forecasting pur-
poses, to be included in the trend. To demonstrate the completion of this example, we
set the cyclical index to 1.0.
4. Irregular. Irregular fluctuations represent random variation that can’t be explained by
the other components. For forecasting, the irregular component is set to the average
value 1.0.11
The forecast for the first quarter of 2007 is
The multiplicative decomposition fit for Coastal Marine Corporation sales, along with
the forecasts for 2007, is shown in Figure 12. We can see from the figure that the fit, con-
structed from the trend and seasonal components, represents the actual data reasonably
well. However, the fit is not good for the last two quarters of 2006, time periods 27 and 28.
The forecasts for 2007 mimic the pattern of the fit.
11For forecasts generated from an additive model, the irregular index is set to the average value 0.
Time Series and Their Components
Time series decomposition methods have a long history. In the 1920s and early 1930s,
the Federal Reserve Board and the National Bureau of Economic Research were
heavily involved in the seasonal adjustment and smoothing of economic time series.
However, before the development of computers, the decomposition calculations were
laborious, and practical application of the methods was limited. In the early 1950s,
Julius Shiskin, chief economic statistician at the U.S. Bureau of the Census, developed
a large-scale computer program to decompose time series. The first computer program
essentially approximated the hand methods that were used up to that time and was
replaced a year later by an improved program known as Method II. Over the years,
improved variants of Method II followed. The current variant of the Census Bureau
time series decomposition program is known as X-12-ARIMA. This program is avail-
able from the Census Bureau at no charge and is widely used by government agencies
and private companies.12
Census II decomposition is usually multiplicative, since most economic time series
have seasonal variation that increases with the level of the series. Also, the decomposi-
tion assumes three components: trend-cycle, seasonal, and irregular.
The Census II method iterates through a series of steps until the components are
successfully isolated. Many of the steps involve the application of weighted moving
averages to the data. This results in inevitable loss of data at the beginning and end of
the series because of the averaging. The ARIMA part of X-12-ARIMA provides the
facility to extend the original series in both directions with forecasts so that more of
the observations are adjusted using the full weighted moving averages. These forecasts
are generated from an ARIMA time series model.
The steps for each iteration of the Census II method as implemented in X-12-
ARIMA are outlined next. It may seem that the method is complicated because of the
many steps involved. However, the basic idea is quite simple—to isolate the trend-
cycle, seasonal, and irregular components one by one. The various iterations are
designed to improve the estimate of each component. Good references for additional
study are Forecasting: Methods and Applications (Makridakis, Wheelwright, and
Hyndman 1998), Forecasting: Practice and Process for Demand Management
(Levenbach and Cleary 2006), and “New Capabilities and Methods of the X-12-
ARIMA Seasonal-Adjustment Program” (Findley et al. 1998).
Step 1. An s-period moving average is applied to the original data to get a rough
estimate of the trend-cycle. (For monthly data, s = 12; for quarterly data,
s = 4; and so forth.)
Step 2. The ratios of the original data to these moving average values are calculated
as in classical multiplicative decomposition, illustrated in Example 2.
12The PC version of the X-12-ARIMA program can be downloaded from the U.S. Census Bureau website.
At the time this text was written, the web address for the download page was www.census.gov/srd/www/
x12a/x12down_pc.html.
Time Series and Their Components
Step 3. The ratios from step 2 contain both the seasonal and the irregular compo-
nents. They also include extreme values resulting from unusual events
such as strikes or wars. The ratios are divided by a rough estimate of the
seasonal component to give an estimate of the irregular component.
A large value for an irregular term indicates an extreme value in the orig-
inal data. These extreme values are identified and the ratios in step 2
adjusted accordingly. This effectively eliminates values that do not fit the
pattern of the remaining data. Missing values at the beginning and end of
the series are also replaced by estimates at this stage.
Step 4. The ratios created from the modified data (with extreme values replaced
and estimates for missing values) are smoothed using a moving average to
eliminate irregular variation. This creates a preliminary estimate of the
seasonal component.
Step 5. The original data are then divided by the preliminary seasonal component
from step 4 to get the preliminary seasonally adjusted series. This season-
ally adjusted series contains the trend-cycle and irregular components. In
symbols,
Yt Tt * St * It
= = Tt * It
St St
The X-12-ARIMA program contains many additional features that we have not
described. For example, adjustments can be made for different numbers of trading days
and for holiday effects, missing values within the series can be estimated and replaced,
the effects of outliers can be removed before decomposition, and other changes in
trend such as shifts in level or temporary ramp effects can be modeled.
APPLICATION TO MANAGEMENT
Time series analysis is a widely used statistical tool for forecasting future events that
are intertwined with the economy in some fashion. Manufacturers are extremely inter-
ested in the boom–bust cycles of our economy as well as of foreign economies so that
they can better predict demand for their products, which, in turn, impacts their inven-
tory levels, employment needs, cash flows, and almost all other business activities
within the firm.
The complexity of these problems is enormous. Take, for example, the problem of
predicting demand for oil and its by-products. In the late 1960s, the price of oil per bar-
rel was very low, and there seemed to be an insatiable worldwide demand for gas and
oil. Then came the oil price shocks of the early and mid-1970s. What would the future
demand for oil be? What about prices? Firms such as Exxon and General Motors were
obviously very interested in the answers to these questions. If oil prices continued to
escalate, would the demand for large cars diminish? What would be the demand for
electricity? By and large, analysts predicted that the demand for energy, and therefore
oil, would be very inelastic; thus, prices would continue to outstrip inflation. However,
these predictions did not take into account a major downswing in the business cycle in
the early 1980s and the greater elasticity of consumer demand for energy than pre-
dicted. By 1980, the world began to see a glut of oil on the market and radically falling
prices. At the time, it seemed hard to believe that consumers were actually benefiting
once again from gasoline price wars. At the time this edition text was written, substan-
tial unrest in the Middle East created a shortage of oil once again. The price of a barrel
of oil and the cost of a gallon of gas in the United States were at record highs.
Oil demand is affected not only by long-term cyclical events but also by seasonal and
random events, as are most other forecasts of demand for any type of product or service.
For instance, consider the service and retail industries. We have witnessed a continued
movement of employment away from manufacturing to the retail and service fields.
However, retailing (whether in-store, catalog, or web-based) is an extremely seasonal
and cyclical business and demand and inventory projections are critical to retailers, so
time series analysis will be used more widely by increasingly sophisticated retailers.
Manufacturers will have a continued need for statistical projections of future
events. Witness the explosive growth in the technology and telecommunications fields
during the 1990s and the substantial contraction of these industries in the early 2000s.
This growth and contraction resulted, to a large extent, from projections of demand that
never completely materialized. Questions that all manufacturers must address include
these: What will the future inflation rate be? How will it affect the cost-of-living adjust-
ments that may be built into a company’s labor contract? How will these adjustments
affect prices and demand? What is the projected pool of managerial skills for 2025?
What will be the effect of the government’s spending and taxing strategies?
What will the future population of young people look like? What will the ethnic mix
be? These issues affect almost all segments of our economy. Demographers are closely
watching the current fertility rate and using almost every available time series forecast-
ing technique to try to project population variables. Very minor miscalculations will
Time Series and Their Components
have major impacts on everything from the production of babies’ toys to the financial
soundness of the Social Security system. Interestingly, demographers are looking at very
long-term business cycles (20 years or more per cycle) in trying to predict what this gen-
eration’s population of women of childbearing age will do with regard to having chil-
dren. Will they have one or two children, as families in the 1960s and 1970s did, or will
they return to having two or three, as preceding generations did? These decisions will
determine the age composition of our population for the next 50 to 75 years.
Political scientists are interested in using time series analysis to study the changing
patterns of government spending on defense and social welfare programs. Obviously,
these trends have great impact on the future of whole industries.
Finally, one interesting microcosm of applications of time series analysis has shown
up in the legal field. Lawyers are making increasing use of expert witnesses to testify
about the present value of a person’s or a firm’s future income, the cost incurred from
the loss of a job due to discrimination, and the effect on a market of an illegal strike.
These questions can often be best answered through the judicious use of time series
analysis.
Satellite technology and the World Wide Web have made the accumulation and
transmission of information almost instantaneous. The proliferation of personal comput-
ers, the availability of easy-to-use statistical software programs, and increased access to
databases have brought information processing to the desktop. Business survival during
periods of major competitive change requires quick, data-driven decision making. Time
series analysis and forecasting play a major role in these decision-making processes.
Several of the series on production, sales, and other economic situations contain data
available only in dollar values. These data are affected by both the physical quantity of
goods sold and their prices. Inflation and widely varying prices over time can cause
analysis problems. For instance, an increased dollar volume may hide decreased sales
in units when prices are inflated. Thus, it is frequently necessary to know how much of
the change in dollar value represents a real change in physical quantity and how much
is due to change in price because of inflation. It is desirable in these instances to
express dollar values in terms of constant dollars.
The concept of purchasing power is important. The current purchasing power of $1
is defined as follows:
100
Current purchasing power of $1 = (10)
Consumer Price Index
Thus, if in November 2006 the consumer price index (with 2002 as 100) reaches 150, the
current purchasing power of the November 2006 consumer dollar is
100
Current purchasing power of $1 = = .67
150
The 2006 dollar purchased only two-thirds of the goods and services that could have
been purchased with a base period (2002) dollar.
To express dollar values in terms of constant dollars, Equation 11 is used.
Suppose that car sales rose from $300,000 in 2005 to $350,000 in 2006, while the
new-car price index (with 2002 as the base) rose from 135 to 155. Deflated sales for
2005 and 2006 would be
100
Deflated 2005 sales = 1$300,0002 ¢ ≤ = $222,222
135
100
Deflated 2006 sales = 1$350,0002 ¢ ≤ = $225,806
155
Note that actual dollar sales had a sizable increase of $350,000 - $300,000 = $50,000.
However, deflated sales increased by only $225,806 - $222,222 = $3,584.
The purpose of deflating dollar values is to remove the effect of price changes.
This adjustment is called price deflation or is referred to as expressing a series in con-
stant dollars.
The sales are deflated for 1999 in terms of 2002 purchasing power, so that
100
Deflated 1999 sales = 142.12 ¢ ≤ = 46.0
91.45
Table 5 shows that, although actual sales gained steadily from 1999 to 2006, physical volume
remained rather stable from 2004 to 2006. Evidently, the sales increases were due to price
markups that were generated, in turn, by the inflationary tendency of the economy.
Time Series and Their Components
Retail
Burnham Retail Furniture Appliance Price Price
Sales Price Index Index Indexa Deflated Salesb
Year ($1,000s) (2002=100) (2002=100) (2002=100) ($1,000s of 2002)
Glossary
Business indicators. Business indicators are business- Price deflation. Price deflation is the process of
related time series that are used to help assess the expressing values in a series in constant dollars.
general state of the economy.
Index numbers. Index numbers are percentages that
show changes over time.
Key Formulas
Linear trend
TNt = b0 + b1t (3)
Quadratic trend
TNt = b0 + b1t + b2t2 (5)
Exponential trend
TNt = b0bt1 (6)
Time Series and Their Components
Yt - St = Tt + It (7a)
Yt
= Tt * It (7b)
St
Yt
Ct * It = (8)
Tt * St
Ct * It
It = (9)
Ct
100
(10)
Consumer Price Index
Problems
TABLE P-7
Year Y Year Y
26,891
29,073
28,189
30,450
31,698
35,435
37,828
42,484
44,580
Time Series and Their Components
9. The expected trend value for October is $850. Assuming an October seasonal
index of 1.12 (112%) and the multiplicative model given by Equation 2, what
would be the forecast for October?
10. The following specific percentage seasonal indexes are given for the month of
December:
75.4 86.8 96.9 72.6 80.0 85.4
TABLE P-12
Adjusted
Seasonal
Month Sales ($1,000s) Index (%)
January 125 51
February 113 50
March 189 87
April 201 93
May 206 95
June 241 99
July 230 96
August 245 89
September 271 103
October 291 120
November 320 131
December 419 189
TABLE P-13
Quarter
Year 1 2 3 4
a. Would you use the trend component, the seasonal component, or both to forecast?
b. Forecast for third and fourth quarters of 1996.
c. Compare your forecasts to Value Line’s.
14. The monthly sales of the Cavanaugh Company, pictured in Figure 1 (bottom), are
given in Table P-14.
a. Perform a multiplicative decomposition of the Cavanaugh Company sales time
series, assuming trend, seasonal, and irregular components.
b. Would you use the trend component, the seasonal component, or both to forecast?
c. Provide forecasts for the rest of 2006.
15. Construct a table similar to Table P-14 with the natural logarithms of monthly
sales. For example, the value for January 2000 is ln11542 = 5.037.
a. Perform an additive decomposition of ln(sales), assuming the model
Y = T + S + I.
TABLE P-14
Month 2000 2001 2002 2003 2004 2005 2006
b. Would you use the trend component, the seasonal component, or both to forecast?
c. Provide forecasts of ln(sales) for the remaining months of 2006.
d. Take the antilogs of the forecasts calculated in part c to get forecasts of the
actual sales for the remainder of 2006.
e. Compare the forecasts in part d with those in Problem 14, part c. Which set of
forecasts do you prefer? Why?
16. Table P-16 contains the quarterly sales (in millions of dollars) for the Disney
Company from the first quarter of 1980 to the third quarter of 1995.
a. Perform a multiplicative decomposition of the time series consisting of Disney’s
quarterly sales.
b. Does there appear to be a significant trend? Discuss the nature of the seasonal
component.
c. Would you use both trend and seasonal components to forecast?
d. Forecast sales for the fourth quarter of 1995 and the four quarters of 1996.
17. The monthly gasoline demand (in thousands of barrels/day) for Yukong Oil
Company of South Korea for the period from January 1986 to September 1996 is
contained in Table P-17.
a. Plot the gasoline demand time series. Do you think an additive or a multiplica-
tive decomposition would be appropriate for this time series? Explain.
b. Perform a decomposition analysis of gasoline demand.
c. Interpret the seasonal indexes.
d. Forecast gasoline demand for the last three months of 1996.
18. Table P-18 contains data values that represent the monthly sales (in billions of dol-
lars) of all retail stores in the United States. Using the data through 1994, perform
a decomposition analysis of this series. Comment on all three components of the
series. Forecast retail sales for 1995 and compare your results with the actual val-
ues provided in the table.
TABLE P-16
Quarter
Year 1 2 3 4
TABLE P-17
Month 1986 1987 1988 1989 1990 1991 1992 1993 1994 1995 1996
January 15.5 20.4 26.9 36.0 52.1 64.4 82.3 102.7 122.2 145.8 170.0
February 17.8 20.8 29.4 39.0 53.1 68.1 83.6 102.2 121.4 144.4 176.3
March 18.1 22.2 29.9 42.2 56.5 68.5 85.5 104.7 125.6 145.2 174.2
April 20.5 24.1 32.4 44.3 58.4 72.3 91.0 108.9 129.7 148.6 176.1
May 21.3 25.5 33.3 46.6 61.7 74.1 92.1 112.2 133.6 153.7 185.3
June 19.8 25.9 34.5 46.1 61.0 77.6 95.8 109.7 137.5 157.9 182.7
July 20.5 26.1 34.8 48.5 65.5 79.9 98.3 113.5 143.0 169.7 197.0
August 22.3 27.5 39.1 52.6 71.0 86.7 102.2 120.4 149.0 184.2 216.1
September 22.9 25.8 39.0 52.2 68.1 84.4 101.5 124.6 149.9 163.2 192.2
October 21.1 29.8 36.5 50.8 67.5 81.4 98.5 116.7 139.5 155.4
November 22.0 27.4 37.5 51.9 68.8 85.1 101.1 120.6 147.7 168.9
December 22.8 29.7 39.7 55.1 68.1 81.7 102.5 124.9 154.7 178.3
TABLE P-18
TABLE P-19
Adjusted Adjusted
Month Seasonal Index Month Seasonal Index
19. The adjusted seasonal indexes presented in Table P-19 reflect the changing volume
of business of the Mt. Spokane Resort Hotel, which caters to family tourists in the
summer and skiing enthusiasts during the winter months. No sharp cyclical varia-
tions are expected during 2007.
Time Series and Their Components
TABLE P-24
Commodity
Sales Price Index
Volume ($) (2001 = 100)
a. If 600 tourists were at the resort in January 2007, what is a reasonable estimate
for February?
b. The monthly trend equation is TN = 140 + 5t where t = 0 represents January
15, 2001. What is the forecast for each month of 2007?
c. What is the average number of new tourists per month?
20. Discuss the performance of the composite index of leading indicators as a barom-
eter of business activity in recent years.
21. What is the present position of the business cycle? Is it expanding or contracting?
When will the next turning point occur?
22. What is the purpose of deflating a time series that is measured in dollars?
23. In the base period of June, the price of a selected quantity of goods was $1,289.73.
In the most recent month, the price index for these goods was 284.7. How much
would the selected goods cost if purchased in the most recent month?
24. Deflate the dollar sales volumes in Table P-24 using the commodity price index.
These indexes are for all commodities, with 2001 = 100.
25. Table P-25 contains the number (in thousands) of men 16 years of age and
older who were employed in the United States for the months from January
TABLE P-25
Year Jan. Feb. Mar. Apr. May Jun. Jul. Aug. Sep. Oct. Nov. Dec.
1993 63,344 63,621 64,023 64,482 65,350 66,412 67,001 66,861 65,808 65,961 65,779 65,545
1994 64,434 64,564 64,936 65,492 66,340 67,230 67,649 67,717 66,997 67,424 67,313 67,292
1995 65,966 66,333 66,758 67,018 67,227 68,384 68,750 68,326 67,646 67,850 67,219 67,049
1996 66,006 66,481 66,961 67,415 68,258 69,298 69,819 69,533 68,614 69,099 68,565 68,434
1997 67,640 67,981 68,573 69,105 69,968 70,619 71,157 70,890 69,890 70,215 70,328 69,849
1998 68,932 69,197 69,506 70,348 70,856 71,618 72,049 71,537 70,866 71,219 71,256 70,930
1999 69,992 70,084 70,544 70,877 71,470 72,312 72,803 72,348 71,603 71,825 71,797 71,699
2000 71,862 72,177 72,501 73,006 73,236 74,267 74,420 74,352 73,391 73,616 73,497 73,338
2001 72,408 72,505 72,725 73,155 73,313 74,007 74,579 73,714 73,483 73,228 72,690 72,547
2002 71,285 71,792 71,956 72,483 73,230 73,747 74,210 73,870 73,596 73,513 72,718 72,437
2003 71,716 72,237 72,304 72,905 73,131 73,894 74,269 74,032 73,715 73,979
Source: Based on Labor force statistics from the Current Population Survey.
Time Series and Their Components
TABLE P-27
Quarter
Year 1 2 3 4
1990 6,768 7,544 7,931 10,359
1991 9,281 10,340 10,628 13,639
1992 11,649 13,028 13,684 17,122
1993 13,920 16,237 16,827 20,361
1994 17,686 19,942 20,418 24,448
1995 20,440 22,723 22,914 27,550
1996 22,772 25,587 25,644 30,856
1997 25,409 28,386 28,777 35,386
1998 29,819 33,521 33,509 40,785
1999 35,129 28,913 40,899 51,868
2000 43,447 46,588 46,181 57,079
2001 48,565 53,187 53,185 64,735
2002 52,126 56,781 55,765 66,905
2003 57,224 63,231 63,036 75,190
2004 65,443 70,466 69,261 82,819
Source: Based on S&P Compustat North American Industrial Quarterly database.
Time Series and Their Components
CASES
The Small Engine Doctor is the name of a business the dealer had to back-order one or more parts for
developed by Thomas Brown, who is a mail carrier any given repair job. Parts ordered from the manu-
for the U.S. Postal Service. He had been a tinkerer facturer had lead times of anywhere from 30 to 120
since childhood, always taking discarded household days. As a result, Tom changed his policy and began
gadgets apart in order to understand “what made to order parts directly from the factory. He found
them tick.” As Tom grew up and became a typical that shipping and handling charges ate into his prof-
suburbanite, he acquired numerous items of lawn its, even though the part price was only 60% of retail.
and garden equipment. When Tom found out about a However, the two most important problems created
course in small engine repair offered at a local com- by the replacement parts were lost sales and storage
munity college, he jumped at the opportunity. Tom space. Tom attracted customers because of his qual-
started small engine repair by dismantling his own ity service and reasonable repair charges, which were
equipment, overhauling it, and then reassembling it. possible because of his low overhead. Unfortunately,
Soon after completing the course in engine repair, he many potential customers would go to equipment
began to repair lawn mowers, rototillers, snowblow- dealers rather than wait several months for repair.
ers, and other lawn and garden equipment for friends The most pressing problem was storage space. While
and neighbors. In the process, he acquired various a piece of equipment was waiting for spare parts, it
equipment manuals and special tools. had to be stored on the premises. It did not take long
It was not long before Tom decided to turn his for both his workshop and his one-car garage to
hobby into a part-time business. He placed an adver- overflow with equipment while he was waiting for
tisement in a suburban shopping circular under the spare parts. In the second year of operation, Tom
name of the Small Engine Doctor. Over the last two actually had to suspend advertising as a tactic to limit
years, the business has grown enough to provide a customers due to lack of storage space.
nice supplement to his regular salary. Although the Tom has considered stocking inventory for his
growth was welcomed, as the business is about to third year of operation. This practice will reduce pur-
enter its third year of operation there are a number of chasing costs by making it possible to obtain quantity
concerns.The business is operated out of Tom’s home. discounts and more favorable shipping terms. He also
The basement is partitioned into a family room, a hopes that it will provide much better turnaround
workshop, and an office. Originally, the office area time for the customers, improving both cash flow and
was used to handle the advertising, order processing, sales.The risks in this strategy are uncontrolled inven-
and bookkeeping. All engine repair was done in the tory carrying costs and part obsolescence.
workshop. Tom’s policy has been to stock only a lim- Before committing himself to stocking spare
ited number of parts, ordering replacement parts as parts, Tom wants to have a reliable forecast for busi-
they are needed. This seemed to be the only practical ness activity in the forthcoming year. He is confident
way of dealing with the large variety of parts involved enough in his knowledge of product mix to use an
in repairing engines made by the dozen or so manu- aggregate forecast of customer repair orders as a
facturers of lawn and garden equipment. basis for selectively ordering spare parts. The fore-
Spare parts have proved to be the most aggra- cast is complicated by seasonal demand patterns and
vating problem in running the business. Tom started a trend toward increasing sales.
his business by buying parts from equipment dealers. Tom plans to develop a sales forecast for the
This practice had several disadvantages. First, he had third year of operation. A sales history for the first
to pay retail for the parts. Second, most of the time two years is given in Table 6.
13This case was contributed by William P. Darrow of the Towson State University, Towson, Maryland.
Time Series and Their Components
January 5 21 July 28 46
February 8 20 August 20 32
March 10 29 September 14 27
April 18 32 October 8 13
May 26 44 November 6 11
June 35 58 December 26 52
ASSIGNMENT
1. Plot the data on a two-year time horizon from constants: 1" = .1, # = .12, 1" = .25, # = .252,
2005 through 2006. Connect the data points to and 1" = .5, # = .52. Plot the three sets of
make a time series plot. smoothed values on the time series graph.
2. Develop a trend-line equation using linear Generate forecasts through the end of the third
regression and plot the results. year for each of the trend-adjusted exponential
3. Estimate the seasonal adjustment factors for smoothing possibilities considered.
each month by dividing the average demand 5. Calculate the MAD values for the two models
for corresponding months by the average of that visually appear to give the best fits (the
the corresponding trend-line forecasts. Plot the most accurate one-step-ahead forecasts).
fitted values and forecasts for 2007 given by 6. If you had to limit your choice to one of the models
Trend * Seasonal. in Questions 2 and 4, identify the model you would
4. Smooth the time series using Holt’s linear expo- use for your business planning in 2007, and dis-
nential smoothing with three sets of smoothing cuss why you selected that model over the others.
John Mosby has been looking forward to the decom- solid forecasts on which to base their investment
position of his time series, monthly sales dollars. He decisions. John knows that his business is improving
knows that the series has a strong seasonal effect and that future prospects look bright, but investors
and would like it measured for two reasons. First, his want documentation.
banker is reluctant to allow him to make variable The monthly sales volumes for Mr. Tux for the
monthly payments on his loan. John has explained years 1999 through 2005 are entered into Minitab.
that because of the seasonality of his sales and Since 1998 was the first year of business, the sales
monthly cash flow, he would like to make extra pay- volumes were extremely low compared to the rest of
ments in some months and reduced payments, and the years. For this reason, John decides to eliminate
even no payments, in others. His banker wants to see these values from the analysis. The seasonal indexes
some verification of John’s assertion that his sales are shown in Table 7. The rest of the computer print-
have a strong seasonal effect. out is shown in Table 8.
Second, John wants to be able to forecast his John is not surprised to see the seasonal indexes
monthly sales. He needs such forecasts for planning shown in Table 7, and he is pleased to have some
purposes, especially since his business is growing. hard numbers to show his banker. After reviewing
Both bankers and venture capitalists want some these figures, the banker agrees that John will make
Time Series and Their Components
Seasonal Index
Period Index
1 0.3144
2 0.4724
3 0.8877
4 1.7787
5 1.9180
6 1.1858
7 1.0292
8 1.2870
9 0.9377
10 0.8147
11 0.6038
12 0.7706
Accuracy of Model
MAPE: 20
MAD: 21,548
MSD: 9.12E $ 08
TABLE 8 (Continued)
TABLE 8 (Continued)
double payments on his loan in April, May, June, and 2005 November 104
August and make no payments at all in January, December 105
February, November, and December. His banker 2006 January 105
asks for a copy of the seasonal indexes to show his February 106
boss and include in John’s loan file. March 107
Turning to a forecast for the first six months of April 109
2006, John begins by projecting trend values using May 110
the trend equation TNt = 12,133 + 3,033t. The trend June 111
estimate for January 2006 is
TN85 = 12,133 + 3,0331852 = 269,938
Turning to the irregular (I ) value for these
Next, John obtains the seasonal index from months, John does not foresee any unusual events
Table 7. The index for January is 31.44%. John has except for March 2006. In that month, he plans to hold
been reading The Wall Street Journal and watching an open house and reduce the rates in one of his stores
the business news talk shows on a regular basis, so that he is finishing remodeling. Because of this promo-
he already has an idea of the overall nature of the tion, to be accompanied with radio and TV advertis-
economy and its future course. He also belongs to a ing, he expects sales in that store to be 50% higher
business service club that features talks by local eco- than normal. For his overall monthly sales, he thinks
nomic experts on a regular basis. As he studies the C this effect will result in about 15% higher sales overall.
column of his computer output, showing the cyclical Using all the figures he has estimated, along
history of his series, he thinks about how he will with his computer output, John makes the forecasts
forecast this value for the first six months of 2006. for Mr. Tux sales for the first six months of 2006,
Since the forecasts of national and local experts call shown in Table 9.
for an improvement in business in 2006 and since After studying the 2006 forecasts, John is
the last C value for October 2005 has turned up disturbed to see the wide range of monthly sales
(103.4%), he decides to use the following C values projections—from $89,112 to $595,111. Although he
for his forecasts: knew his monthly volume had considerable variability,
Time Series and Their Components
Sales Forecast = T ! S ! C ! I
he is concerned about such wide fluctuations. John His real concern, however, is focused on his
has been thinking about expanding from his current worst months, January and February. He has recently
Spokane location into the Seattle area. He has been considering buying a tuxedo-shirt-making
recently discovered that there are several “dress up” machine that he saw at a trade show, thinking that he
events in Seattle that make it different from his pres- might be able to concentrate on that activity during
ent Spokane market. Formal homecoming dances, in the winter months. If he receives a positive reaction
particular, are big in Seattle but not in Spokane. from potential buyers of shirts for this period of time,
Since these take place in the fall, when his Spokane he might be willing to give it a try. As it is, the
business is slower (see the seasonal indexes for seasonal indexes on his computer output have
October and November), he sees the advantage of focused his attention on the extreme swings in his
leveling his business by entering the Seattle market. monthly sales levels.
QUESTIONS
1. Suppose John’s banker asked for two sentences Determine the Seattle seasonal indexes that
to show his boss that would justify John’s would be ideal to balance out the monthly rev-
request to make extra loan payments in some enues for Mr. Tux.
months and no payments in others. Write these 3. Disregarding Seattle, how much volume would
two sentences. John have to realize from his shirt-making
2. Assume that John will do exactly twice as much machine to make both January and February
business in Seattle as Spokane next year. “average”?
The executive director, Marv Harnishfeger, con- Dorothy gave you these data and asked you to
cluded that the most important variable that complete a time series decomposition analysis. She
Consumer Credit Counseling (CCC) needed to fore- emphasized that she wanted to understand com-
cast was the number of new clients that would be pletely the trend and seasonality components.
seen for the rest of 1993. Marv provided Dorothy Dorothy wanted to know the importance of each
Mercer monthly data for the number of new clients component. She also wanted to know if any unusual
seen by CCC for the period from January 1985 irregularities appeared in the data. Her final instruc-
through March 1993. tion required you to forecast for the rest of 1993.
Time Series and Their Components
ASSIGNMENT
Write a report that provides Dorothy with the infor-
mation she has requested.
Julie Murphy developed a naive model that com- production requirements, Julie is very anxious to
bined seasonal and trend estimates. One of the prepare short-term forecasts for the company that
major reasons why she chose this naive model was are based on the best available information concern-
its simplicity. Julie knew that her father, Glen, would ing demand.
need to understand the forecasting model used by For forecasting purposes, Julie has decided to
the company. use only the data gathered since 1996, the first full
It is now October of 2002, and a lot has changed. year Murphy Brothers manufactured its own line of
Glen Murphy has retired. Julie has completed sev- furniture (Table 10). Julie can see (Figure 13) that
eral business courses, including business forecasting, her data have both trend and seasonality. For this
at the local university. Murphy Brothers Furniture reason, she decides to use a time series decomposi-
built a factory in Dallas and began to manufacture tion approach to analyze her sales variable.
its own line of furniture in October 1995. Since Figure 13 shows that the time series she is
Monthly sales data for Murphy Brothers Furni- analyzing has roughly the same variability through-
ture from 1996 to October 2002 are shown in Table out the length of the series, Julie decides to use an
10. As indicated by the pattern of these data demon- additive components model to forecast. She runs the
strated in Figure 13, sales have grown dramatically model Yt = Tt + St + It. A summary of the results is
since 1996. Unfortunately, Figure 13 also demon- shown in Table 11. Julie checks the autocorrelation
strates that one of the problems with demand is that pattern of the residuals (see Figure 14) for random-
it is somewhat seasonal. The company’s general pol- ness. The residuals are not random, and the model
icy is to employ two shifts during the summer and does not appear to be adequate.
early fall months and then work a single shift Julie is stuck. She has tried a naive model that
through the remainder of the year. Thus, substantial combined seasonal and trend estimates, Winters’
inventories are developed in the late summer and exponential smoothing, and classical decomposition.
fall months until demand begins to pick up in Julie finally decides to adjust seasonality out of the
November and December. Because of these data so that forecasting techniques that cannot
Jan. Feb. Mar. Apr. May Jun. Jul. Aug. Sep. Oct. Nov. Dec.
1996 4,964 4,968 5,601 5,454 5,721 5,690 5,804 6,040 5,843 6,087 6,469 7,002
1997 5,416 5,393 5,907 5,768 6,107 6,016 6,131 6,499 6,249 6,472 6,946 7,615
1998 5,876 5,818 6,342 6,143 6,442 6,407 6,545 6,758 6,485 6,805 7,361 8,079
1999 6,061 6,187 6,792 6,587 6,918 6,920 7,030 7,491 7,305 7,571 8,013 8,727
2000 6,776 6,847 7,531 7,333 7,685 7,518 7,672 7,992 7,645 7,923 8,297 8,537
2001 7,005 6,855 7,420 7,183 7,554 7,475 7,687 7,922 7,426 7,736 8,483 9,329
2002 7,120 7,124 7,817 7,538 7,921 7,757 7,816 8,208 7,828
Trend-Line Equation
TNt = 5,672 + 31.4t
Seasonal Index
Period Index
1 -674.60
2 -702.56
3 -143.72
4 -366.64
5 -53.52
6 -173.27
7 -42.74
8 222.32
9 -57.95
10 145.76
11 612.30
12 1234.63
Accuracy Measures
MAPE: 1.9
MAD: 135.1
MSD: 30,965.3
Time Series and Their Components
Jan. Feb. Mar. Apr. May Jun. Jul. Aug. Sep. Oct. Nov. Dec.
1996 5,621 5,671 5,745 5,821 5,775 5,863 5,847 5,818 5,901 5,941 5,857 5,767
1997 6,091 6,096 6,051 6,135 6,161 6,189 6,174 6,277 6,307 6,326 6,334 6,380
1998 6,551 6,521 6,486 6,510 6,496 6,580 6,588 6,536 6,543 6,659 6,749 6,844
1999 6,736 6,890 6,936 6,954 6,972 7,093 7,073 7,269 7,363 7,425 7,401 7,492
2000 7,451 7,550 7,675 7,700 7,739 7,691 7,715 7,770 7,703 7,777 7,685 7,302
2001 7,680 7,558 7,564 7,550 7,608 7,648 7,730 7,700 7,484 7,590 7,871 8,094
2002 7,795 7,827 7,961 7,905 7,975 7,930 7,859 7,986 7,886
handle seasonal data can be applied. Julie deseason- adds 674.60 to the data for each January and sub-
alizes the data by adding or subtracting the seasonal tracts 1,234.63 from the data for each December.
index for the appropriate month. For example, she Table 12 shows the seasonally adjusted data.
ASSIGNMENT
1. Using the data through 2001 in Table 12, subtracting the appropriate seasonal index in
develop a model to forecast the seasonally Table 11. Are these forecasts accurate when
adjusted sales data and generate forecasts for compared with the actual values?
the first nine months of 2002. 3. Forecast sales for October 2002 using the same
2. Using the forecasts from part 1, forecast sales procedure as in part 2.
for the first nine months of 2002 by adding or
Time Series and Their Components
In 1993, AAA Washington was one of the two agency; and an insurance agency. The club provided
regional automobile clubs affiliated with the Ameri- these services through a network of offices located
can Automobile Association (AAA or Triple A) in Bellevue, Bellingham, Bremerton, Everett,
operating in Washington State. At that time, 69% of Lynnwood, Olympia, Renton, Seattle, Tacoma, the
all people belonging to automobile clubs were mem- Tri-Cities (Pasco, Richland, and Kennewick),
bers of the American Automobile Association, mak- Vancouver, Wenatchee, and Yakima, Washington.
ing it the largest automobile club in North America. Club research had consistently shown that the
AAA was a national association that serviced its emergency road service benefit was the primary rea-
individual members through a federation of approx- son that people join AAA. The importance of emer-
imately 150 regional clubs that chose to be affiliated gency road service in securing members was
with the national association. The national associa- reflected in the three types of memberships offered
tion set a certain number of minimum standards by AAA Washington: Basic, AAA Plus, and AAA
with which the affiliated clubs had to comply in Plus RV. Basic membership provided members five
order to retain their affiliation with the association. miles of towing from the point at which their vehicle
Each regional club was administered locally by its was disabled. AAA Plus provided members with 100
own board of trustees and management staff. The miles of towing from the point at which their vehicle
local trustees and managers were responsible for was disabled. AAA Plus RV provided the 100-mile
recruiting and retaining members within their towing service to members who own recreational
assigned territories and for ensuring the financial vehicles in addition to passenger cars and light
health of the regional club. Beyond compliance with trucks. Providing emergency road service was also
the minimum standards set by the AAA, each the club’s single largest operating expense. It was
regional club was free to determine what additional projected that delivering emergency road service
products and services it would offer and how it would cost $9.5 million, 37% of the club’s annual
would price these products and services. operating budget, in the next fiscal year.
AAA Washington was founded in 1904. Its ser- Michael DeCoria, a CPA and MBA graduate of
vice territory consisted of the 26 Washington coun- Eastern Washington University, had recently joined
ties west of the Columbia River. The club offered its the club’s management team as vice president of
members a variety of automobile and automobile- operations. One of the responsibilities Michael
travel-related services. Member benefits provided in assumed was the management of emergency road
cooperation with the national association included service. Early in his assessment of the emergency
emergency road services; a rating service for lodging, road service operation, Mr. DeCoria discovered that
restaurants, and automotive repair shops; tour guides emergency road service costs had increased at a rate
to AAA-approved lodging, restaurants, camping, and faster than could be justified by the rate of inflation
points of interest; and advocacy for legislation and and the growth in club membership. Michael began
public spending in the best interests of the motor- by analyzing the way the club delivered emergency
ing public. In addition to these services, AAA road service to determine if costs could be con-
Washington offered its members expanded protec- trolled more tightly in this area.
tion plans for emergency road service; financial ser- Emergency road service was delivered in one of
vices, including affinity credit cards, personal lines of four ways: the AAA Washington service fleet, con-
credit, checking and savings accounts, time deposits, tracting companies, reciprocal reimbursement, and
and no-fee American Express Travelers Cheques; direct reimbursement. AAA Washington’s fleet of
access to a fleet of mobile diagnostic vans for deter- service vehicles responded to emergency road ser-
mining the “health” of a member’s vehicle; a travel vice calls from members who became disabled in the
14This case was provided by Steve Branton, former student and MBA graduate, Eastern Washington
University.
Time Series and Their Components
downtown Seattle area. Within AAA Washington’s association. Finally, members could contact a towing
service area, but outside of downtown Seattle, com- company of their choice directly, paying for the
mercial towing companies that had contracted with towing service and then submitting a request for
AAA Washington to provide this service responded reimbursement to the club. AAA Washington
to emergency road service calls. Members arranged reimbursed the actual cost of the towing or $50,
for both of these types of emergency road service by whichever was less, directly to the member. After a
calling the club’s dispatch center. Should a member careful examination of the club’s four service deliv-
become disabled outside of AAA Washington’s ser- ery methods, Michael concluded that the club was
vice area, the member could call the local AAA- controlling the cost of service delivery as tightly as
affiliated club to receive emergency road service. was practical.
The affiliate club paid for this service and then billed Another possible source of the increasing costs
AAA Washington for reciprocal reimbursement was a rise in the use of emergency road service.
through a clearing service provided by the national Membership had been growing steadily for several
years, but the increased cost was more than what road service calls per member grew by 3.28%, from
could be attributed to simple membership growth. an average of 0.61 calls per member to 0.63 calls.
Michael then checked to see if there was a growth in Concerned that a continuation of this trend would
emergency road service use on a per-member basis. have a negative impact on the club financially,
He discovered that between fiscal year 1990 and fis- Mr. DeCoria gathered the data on emergency road
cal year 1991, the average number of emergency service call volume presented in Table 13.
ASSIGNMENT
1. Perform a time series decomposition on the road service call volume that you discovered
AAA emergency road service calls data. from your time series decomposition analysis.
2. Write a memo to Mr. DeCoria summarizing the
important insights into changes in emergency
Julie Ruth, Alomega Food Stores president, had After reviewing the results of this regression
collected data on her company’s monthly analysis, including the low r2 value (36%), she
sales along with several other variables she decided to try time series decomposition on the sin-
thought might be related to sales Julie used her gle variable, monthly sales. Figure 15 shows the plot
Minitab program to calculate a simple regression of sales data that she obtained. It looked like
equation using the best predictor for monthly sales were too widely scattered around the
sales. trend line for accurate forecasts. This impression
was confirmed when she looked at the MAPE value She also noted that the MAPE had dropped to
of 28. She interpreted this to mean that the average 12%, a definite improvement over the value
percentage error between the actual values and the obtained using the trend equation alone.
trend line was 28%, a value she considered too high. Finally, Julie had the program provide fore-
She next tried a multiplicative decomposition of the casts for the next 12 months, using the trend equa-
data. The results are shown in Figure 16. tion projections modified by the seasonal indexes.
In addition to the trend equation shown on the She thought she might use these as forecasts for
printout, Julie was interested in the seasonal her planning purposes but wondered if another
(monthly) indexes that the program calculated. She forecasting method might produce better fore-
noted that the lowest sales month was December casts. She was also concerned about what her pro-
(month 12, index = 0.49) and the highest was duction manager, Jackson Tilson, might say about
January (month 1, index = 1.74). She was aware of her forecasts, especially since he had expressed
the wide swing between December and January but concern about using the computer to make
didn’t realize how extreme it was. predictions.
QUESTION
1. What might Jackson Tilson say about her fore-
casts?
Jame Luna examined the autocorrelations for a smoothing method to generate forecasts of future
cookie sales at Surtido Cookies to determine if there sales. A member of Jame’s team has had some expe-
might be a seasonal component. He also considered rience with decomposition analysis and suggests that
Time Series and Their Components
they try a multiplicative decomposition of the throughout the year. Jame doesn’t have a lot of faith
cookie sales data. Not only will they get an indica- in a procedure that tries to estimate components
tion of the trend in sales, but also they will be able to that cannot be observed directly. However, since
look at the seasonal indexes. The latter can be Minitab is available, he agrees to give decomposi-
important pieces of information for determining tion a try and to use it to generate forecasts of sales
truck float and warehouse (inventory) requirements for the remaining months of 2003.
QUESTIONS
1. Perform a multiplicative decomposition of the 3. Compute the residual autocorrelations. Examin-
Surtido Cookie sales data, store the residuals, ing the residual autocorrelations and the fore-
and generate forecasts of sales for the remain- casts of sales for the remainder of 2003, should
ing months of 2003. Jame change his thinking about the value of
2. What did Jame learn about the trend in sales? decomposition analysis? Explain.
What did the seasonal indexes tell him?
Mary Beasley’s goal was to predict the number of evidence that some periods of the year were busier
future total billable visits to Medical Oncology at than others. (Scheduling doctors is never an easy
Southwest Medical Center. Mary learned that there task.) A colleague of Mary’s suggested she consider
was a seasonal component in her data and consid- decomposition analysis, a technique with which
ered using Winters’ smoothing method to generate Mary was not familiar. However, since she had a sta-
forecasts of future visits. Mary was not completely tistical software package available that included
satisfied with this analysis, since there appeared to decomposition, Mary was willing to give it a try. She
be some significant residual autocorrelations. also realized she had to understand decomposition
Moreover, Mary was interested in isolating the sea- well enough to sell the results to central administra-
sonal indexes because she wanted to have hard tion if necessary.
QUESTIONS
1. Write a brief memo to Mary explaining decom- 3. Interpret the trend component for Mary. What
position of a time series. did she learn from the seasonal indexes?
2. Perform a multiplicative decomposition of total 4. Compute the residual autocorrelations. Given
billable visits, save the residuals, and generate the residual autocorrelations and the forecasts,
forecasts of visits for the next 12 months, using should Mary be pleased with the decomposition
February FY2003–04 as the forecast origin. analysis of total billable visits? Explain.
Minitab Applications
The problem. In Example 1, a trend equation was developed for annual registration of
new passenger cars in the United States from 1960 to 1992.
Time Series and Their Components
Minitab Solution
1. After the new passenger car registration data are entered into column C1 of the
worksheet, click on the following menus to run the trend analysis:
Stat>Time Series>Trend Analysis
The problem. Table 1 was constructed to show the trend estimates and errors com-
puted for the new passenger car registration data.
Minitab Solution
1. Column C1 is labeled Year, C2 is labeled Y, C3 is labeled t, C4 is labeled Estimates,
and C5 is labeled Error. Clicking on the following menus creates the years:
Calc>Make Patterned Data>Simple Set of Numbers
The problem. In Examples 3 and 4, Perkin Kendell, the analyst for the Coastal Marine
Corporation, wanted to forecast quarterly sales for 2007.
Minitab Solution
1. Enter the appropriate years into column C1, quarters into C2, and the data into
C3. To run a decomposition model, click on the following menus:
Stat>Time Series>Decomposition
Time Series and Their Components
File>Print Graph
6. Label column C8 CI. Using the Calc>Calculator menu in Minitab, generate col-
umn C8 by dividing column C7 by column C4, so CI = TCI>T.
7. Label column C9 C. Generate column C9 by taking a centered moving average of
order 3 of the values in column C8. Use the menu
Excel Applications
The problem. Figure 6 shows data and a graph for mutual fund salespeople. An expo-
nential trend model is needed to fit these data.
Excel Solution
Tools>Data Analysis
5. The Data Analysis dialog box appears. Under Analysis Tools, choose Regression
and click on OK. The Regression dialog box shown in Figure 19 appears.
Time Series and Their Components
YN = 110.016211.3132t
References
Bell, W. R., and S. C. Hillmer. “Issues Involved with Program,” Journal of Business and Economic
the Seasonal Adjustment of Economic Time Statistics 16 (1998): 127–152.
Series,” Journal of Business and Economic Levenbach, H., and J. P. Cleary. Forecasting: Practice
Statistics 2 (1984): 291–320. and Process for Demand Management. Belmont,
Bowerman, B. L., R. T. O’connell, and A. B. Koehler. Calif.: Thomson Brooks/Cole, 2006.
Forecasting, Time Series and Regression, 4th ed. Makridakis, S., S. C. Wheelwright, and R. J.
Belmont, Calif.: Thomson Brooks/Cole, 2005. Hyndman. Forecasting Methods and Applications,
Diebold, F. X. Elements of Forecasting, 3rd ed. 3rd ed. New York: Wiley, 1998.
Cincinnati, Ohio: South-Western, 2004. Moore, G. H., and J. Shiskin. “Early Warning Signals
Findley, D. F., B. C. Monsell, W. R. Bell, M. C. Otto, for the Economy.” In Statistics: A Guide to
and B. Chen. “New Capabilities and Methods of Business and Economics, J. M. Tanur et al., eds.
the X-12-ARIMA Seasonal-Adjustment San Francisco: Holden-Day, 1976.