BOOK REVIEW
Principles of Forecasting
Edited by
J. Scott Armstrong
FUTURECASTS online magazine
www.futurecasts.com
Vol. 5, No. 2, 2/1/03.
Suggested best practices:
& |
"Principles of Forecasting: A Handbook for
Researchers and Practitioners" - J. Scott Armstrong, The
Wharton School, Univ. of Pennsylvania - provides an essential reference work for
any serious practitioner or student of forecasting and estimating methods. Most
impressive - given the extent to which phony forecasting is used as a propaganda
ploy - is the book's objective evaluations of reliability of the various
forecasting methods covered, and candid explanations of the degrees of
uncertainty. & |
The "principles" are "advice, guidelines, prescriptions, condition-action statements, and rules" - all supported by empirical evidence of efficacy to the extent that such evidence has been developed by scholarship. They are general recommendations that have to be fine tuned to best apply to individual circumstances.
|
By "principles," Armstrong explains, the
book broadly means "advice, guidelines, prescriptions, condition-action
statements, and rules" - all supported by empirical evidence of efficacy to
the extent that such evidence has been developed by scholarship. The book also
deals with such "principles" supported primarily by expert judgment -
and some that are admittedly speculative. However, it clearly provides the bases
for evaluation and covers the specific limited "conditions" for which
each principle is applicable. These forecasting principles are general
recommendations that have to be fine tuned to best apply to individual
circumstances.
|
While judgmental forecasting methods are clearly limited to the human skills applied, mathematical forecasting methods are likely to provide a false sense of precision and certainty. |
Uncertainty is always a factor - overconfidence a constant risk - Armstrong wisely warns. While judgmental forecasting methods are clearly limited to the human skills applied, mathematical forecasting methods are likely to provide a false sense of precision and certainty. With respect to models derived primarily from data - rather than theory: "An immense amount of research effort has so far produced little evidence that data-mining models can improve forecasting accuracy." However, when based on credible theory, econometric models can usefully "integrate judgmental and statistical sources."
All of these techniques are still evolving, and there is a pervasive need for further research and testing of both reliability and best practices. Each chapter of the handbook contains suggestions for further study. |
Judgmental Estimating and Forecasting Methods
Role Playing:
& |
"Role Playing: A Method to Forecast
Decisions" - by Armstrong - covers "a method of predicting the
decisions by people or groups engaged in conflicts." Role playing forces
participants to walk in the shoes of the various parties - and consider
interactions as various parties react to actions taken by other parties. & |
These are skills that expert judgment is weak on
because of difficulties that arise in thinking through several rounds of
interactions. Role playing is most useful when there are just two or three
parties involved - not many but more than one. It is superior to expert judgment
for predicting individual behavior. It is particularly useful in predicting
large changes - something else that expert judgment is frequently weak at. |
Even in instances where role playing missed the actual outcome of a conflict, it proved better than expert judgment or arbitrary chance in indicating the possibility of the actual outcome in a substantial majority of instances. |
War games and mock trials are typical uses.
However, role playing can also be used in commercial evaluations for such things
as changes in pricing policy or product design - and for union strike threats or
other labor actions. |
Proper technique includes "realism in casting, role instructions, situation description, and session administration" - factors the author outlines and evaluates.
Thus, it is advisable to run about 10 sessions, with half
using one description and half an alternative description - with more sessions
indicated if responses vary substantially. |
Surveys of Intentions:
& |
"Methods of Forecasting
from Intentions Data" - by Vicki G. Morwitz, Stern School of Business,
New York Univ. - covers the use of surveys of what people intend to do as a means
of predicting what they will do. Surveys are frequently used to predict the market
for commercial goods and services. They are regularly used by the vast
majority of market research clients. They can also be used "to develop
best- and worst-case scenarios. & |
The utility of such surveys varies with the type of product, the people surveyed, and the skill with which the questions are phrased. Morwitz advises:
Thus:
Studies have shown that groups of people who were asked about auto or computer purchases actually bought more of them than similar groups not asked about such intentions. However, the impact on certain subgroups varied, with both positive and negative intentions apparently reinforced by the questioning. Brand loyalty seems to be enhanced by survey questions.
|
Responses about behavior that is either socially desirable or undesirable will be impacted by bias. |
Biases are a major problem. Respondents
frequently tend to overstate or understate intentions, and the causes of such
biases are still poorly understood. There are a variety of biases - some obvious
and some subtle - that can affect results. |
Adjustments to eliminate bias are suggested.
Studies have shown that purchase probability data for durable goods - like autos
- generally substantially understate actual purchases, while data for non
durables overstate actual purchases. Historical data - from the same or
similar behavior - where available - can be used to adjust for these systemic
inaccuracies, but that will still leave some substantial margin of error.
The accuracy of forecasts involving intentions surveys is
frequently improved when used with other data - such as from role playing or
sales programs or expert opinions. |
Expert opinions:
The Delphi expert opinion survey techniques are the most powerful. |
"Improving Judgment in
Forecasting" - by Nigel Harvey, Univ. College of London - discusses
methods of reducing the impact of inconsistency and bias in forecasts based on
expert judgment. By definition - those with experience, learning and a proven
track record in the relevant field - the "experts" - are a good source
of reliable forecasts. However, there are reasons why experts fail - and there
are techniques to counter those reasons. Among these, the Delphi expert opinion
survey techniques are the most powerful. & |
Checklists help "because people can rarely bring to mind all the information relevant to a task when they need to do so." Irrelevant information can confuse judgment. The weight given individual factors can vary from the forecaster's specific judgment of their relative importance. |
Inconsistency and bias are the two primary
negative influences affecting expert opinion.
|
Requiring not just an expert opinion, but also the justification for the opinion, is a common method of reducing and/or revealing bias. Response to feedback can also counter the influence of bias, as in Delphi survey techniques. Finally, an ongoing measurement of forecast accuracy undermines biased judgment. |
Bias may arise from the self interest of the expert, or
may be inherent in the judgmental or statistical methods applied. |
Forecasting performance can change with changes in the quality of the data as well as changes in forecast method or error measure. |
Experts have a variety of forecasting methods to
employ, and should rely on the one that achieves the best results for particular
problems. Comparisons of the results of different forecasting software should be
ongoing. However, performance over a few periods is not assurance of future
performance, and performance can change with changes in the quality of the data
as well as changes in forecast method or error measure. |
"People tend to search for information that confirms rather than falsifies their hypotheses." |
Accurate records and feedback are vital for the
reduction of both inconsistency and bias. Personal memory of results becomes
unreliable over time. "The hindsight bias is likely to cause forecasters to
overestimate the quality of their forecasts." & Even with adequate records, objective evaluation is not always achieved. There may be "confirmation bias." "People tend to search for information that confirms rather than falsifies their hypotheses." Also undermining feedback in some cases are those instances where the forecast itself may be "self fulfilling" by inducing actions that increase the likelihood of the forecast. & Graphical presentation of data has been shown to facilitate evaluation better than tabular lists of numbers - but primarily for gradual trends, as in business and financial series. Extrapolation from graphical presentations - by drawing a "best fitting line" through and beyond the points on the graph - has been similarly shown to facilitate evaluation. & |
Information acquisition and processing: |
"Improving Reliability of Judgmental
Forecasts" - by Thomas R. Stewart, Univ. of Albany, State Univ. of New
York - can be achieved by adopting techniques that deal with the unreliability of
information acquisition and information processing. |
People instinctively use only a subset of available information in their forecasting and planning. Error is introduced into judgment forecasts "by the natural inconsistency of the human judgment process."
For validity - the accuracy of forecasts - the process must not only be reliably performed, but all the forecasting techniques and data must also be free of errors.
While rigorous analytical process may experience fewer errors, errors can still be introduced by such things as errors in inputs or various types of system failure - typically causing large, even catastrophic, errors.
Increasing the amount of information available may not improve the quality of the forecast. |
The reliability of judgment forecasts is
undermined by complexity - an uncertain environment - reliance on perception,
memory, or pattern recognition - and reliance on intuition instead of analysis.
Indeed, "accuracy of forecasts does not increase as information
increases," because of the increase in mental burden and complexity. People
instinctively use only a subset of available information in their forecasting
and planning. Error is introduced into judgment forecasts "by the natural
inconsistency of the human judgment process."
Stewart concludes:
To improve reliability, Stewart suggests:
|
Analyzing components of forecasting problems: |
"Decomposition for
Judgmental Forecasting and Estimation" - by Donald G. MacGregor,
Decision Research, Eugene, Or. - provides suggestions for when and how to break
down complex forecasting problems into their component parts. & |
By breaking down complex forecasting problems into
average portions or discreet segments - "decomposition" - and then
combining the results, forecasters can facilitate certain forecasting estimates.
This technique is applicable to either multiplicative or segmented forecasts -
although multiplicative decomposition can also greatly enlarge component errors.
Decomposition facilitates numerical estimating based on partial or incomplete
knowledge. It is a manner of proceeding when time and/or money constraints limit
information acquisition - or when much of the information for some segments is
simply not accessible. |
|
"Multiplicative decomposition is based on the assumption that errors in estimation are unrelated and therefore will tend to cancel each other out." |
Decomposition is useful when overall uncertainty is
high. It should not be used merely to refine fairly reliable estimates,
since risks of propagating component errors outweigh benefits.
"Multiplicative decomposition is based on the assumption that errors in
estimation are unrelated and therefore will tend to cancel each other out."
When uncertainty levels are low, individual errors may not adequately cancel
out. |
Delphi expert opinion surveys: |
"Expert Opinions in
Forecasting: The Role of the Delphi Technique" - by Gene Rowe, Norwich
Research Park, U.K., and George Wright, Strathclyde Univ. Business School, U.K. - covers the use of Delphi surveys - one of the most widely used and successful
techniques for improving forecasts and planning based on expert judgment.
"Delphi is a useful method for eliciting and aggregating expert
opinion." & |
Delphi groups should include between 5 and 20 experts with disparate knowledge of the subject - should continue through 2 or 3 rounds - with each expert providing average estimates and justification. The procedure includes anonymity, iteration, controlled feedback, and statistical aggregation. |
Delphi techniques provide a feedback mechanism for
groups of experts that substantially improves expert judgment. It is applicable
wherever expert judgment must be relied upon because statistical techniques are
not viable or practical. & Delphi surveys provide an exchange of information among anonymous experts over a number of rounds - "iterations" - permitting panelists to react to the information thus gathered in each round to perfect their forecasts. The estimates on the final round are averaged to provide a group judgment. & The authors conclude that Delphi groups should include between 5 and 20 experts with disparate knowledge of the subject - should continue through 2 or 3 rounds - with each expert providing average estimates and justification. The procedure includes anonymity, iteration, controlled feedback, and statistical aggregation.
"Domain knowledge" is "expert's knowledge about a situation" - such as managers' knowledge about a brand and its market, or of episodic events - both in the past and likely in the future. Strikes, government regulations, major product or price changes or advertising campaigns or changes in the competitive environment are typical examples of causal factors that will impact outcomes.
|
Delphi techniques may be more useful for "noncasting - for establishing current plans of action in reaction to current conditions - than for forecasting" the likelihood of change. |
The optimal size of Delphi panels is uncertain, with some
Delphi surveys including scores or hundreds of members. However, administrative
costs in terms of time and money are always considerable and increase with the
size of the panel. For quantifiable forecasts, large panels may have no
advantage over smaller panels of at least 5 experts. |
Conjoint analysis:
& |
"Forecasting with Conjoint
Analysis" - by Dick R. Wittink, Yale School of Management, and Trond
Bergestuen, American Express - covers the use of conjoint analysis - a method
"used in marketing and other fields to quantify how individuals confront
trade-offs when they choose between multidimensional alternatives." & |
In order to get realistic responses for market
surveys, it is necessary to do more than just ask what customers want. The
survey must emphasize the tradeoffs participants will have to make. For example,
a range of particular product features will be offered for increasing sums of
money. Conjoint analysis is a method for designing appropriate questions,
administering the survey, and analyzing responses to quantify various
tradeoffs. |
A well designed survey should also permit breakdowns according to types of potential customers or for other segmentation of interests. With proper segmentations, a respondent's preferences will depend on the pertinent product characteristics and there will be "no doubt about the causal direction --- the relevant variables are known and specified, and the variations within attributes and covariations between attributes are controlled."
The survey must
A distinction is drawn between surveys about changes in
existing products - for which the procedure is readily applicable and analytical
results can be repeatedly tested against actual market results - and the
introduction of totally new products about which respondents lack familiarity.
In the latter case, respondents must first be thoroughly familiarized with the
product before the survey. The effectiveness of the survey - and the reliability
of the analytical efforts - will inevitably vary widely when applied to totally
new products. |
|
Generally, it is the success stories that get published. Failures frequently go unreported. |
The validity of conjoint-based predictions is
inherently difficult to determine because of the dynamic nature of pertinent
influences. The authors caution that many other factors besides customer
response will impact sales and market share. Among other things, product
availability may be limited, advertising levels may limit customer awareness of
availability, competing offerings may appear in the market, economic conditions
may change, word of mouth influences may develop over time. The authors cover
the efforts to test validity, the problems inherent in such efforts, and the
relative strengths and weaknesses of various testing methods individually and in
combination. |
Judgmental bootstrapping:
& |
"Inferring
Experts' Rules for Forecasting" - by Armstrong - by examining the inputs
and resulting predictions, and the characteristics of the forecasting problem,
is called "judgmental bootstrapping." It is a type of limited expert
system - also called "policy capturing." Bootstrapping - or
"linear" - models resemble econometric models except that they use
expert forecasts rather than actual outcomes as the dependent variable. & |
Once developed, bootstrap models can provide reliable expert forecasts inexpensively and rapidly, and aid learning and the conduct of experiments. They are less expensive than more elaborate expert systems. They can reveal best practices. They are especially useful when large numbers of forecasts are needed - as in hiring decisions. Armstrong reports that several successful sports franchises use bootstrapping models to assist in selecting athletes.
Bootstrap models can be combined with conjoint analysis
to forecast sales for new products. They can be used to assure that decisions
that might be controversial - like university admissions - are being made fairly
and consistently. (But, today, many universities prefer political correctness in
admissions rather than any fair estimates based on rates of student success.) |
|
Although not as comprehensive or flexible as an expert, they frequently achieve better results than the unaided experts themselves because the bootstrapping models achieve consistency in application. |
Bootstrapping models are based only on data that
experts use. They apply "only to studies in which an expert's rules are
inferred by regression analysis." |
It is important to use all the variables the expert might use - something that takes careful interviewing to do since the experts might not be aware of some of the variables that influence their judgment. |
Bootstrapping models are most appropriate
"for complex situations, where judgments are unreliable, and where experts'
judgments have some validity." They should be considered for complex
problems - but not too complex for practical modeling - where reliable expert
estimates can be obtained, feedback is sufficient to judge the reliability of
the estimates, and where the alternative is to use the judgment of unskilled
individuals. Also, for situations where many different experts are needed, a
bootstrap model can be cost effective. If relevant historical
data is available, models can be tested for reliability.
|
If invalid variables are used, the model's consistency will increase unreliability instead of decreasing it.
By employing a wide range of stimulus cases, the applicability of the bootstrapping model can be broadened to cover a wide range of possibilities. Historical examples should be used. |
Thus, the model must be competently simplified to
include only the vital variables. These then must be quantified in meaningful
ways, sometimes based on experts' ratings or on evaluation of previous history.
Where good feedback is available - such as in weather forecasting and economic
market analyses - the best experts can be identified and their evaluation of
variables relied upon. If invalid variables are used, the model's consistency
will increase unreliability instead of decreasing it. Developing models based on
several different types of experts or groups of experts should help reveal
errors and biases and establish best practices.
Obviously, if and as actual results become available, the model should be recalibrated to improve results. All models fail when major unexpected changes occur - but since this is infrequent, it has not been a problem in practice. |
Combined Judgmental and Statistical Estimating and Forecasting Methods
|
Trend analysis:
& |
"Forecasting Analogous Time Series"
- by George T. Duncan, Wilpen L. Gorr and Janusz Szczypula, School of
Public Policy, Carnegie Mellon Univ. - covers the ancient and widespread use of analogies for
trend forecasts - both judgmental and statistical. For commercial purposes,
trend analysis can provide highly reliable forecasts. |
Pooled data provides additional data, extending the sample size.
Pooling methods must be checked over time to eliminate divergent time series. |
Techniques for examining analogous time series - with mechanisms to recognize
step jumps, turning points, and time trend shape changes - are discussed.
Pooling can include similar products, services or other subjects in
the same geographic areas - or the same products, services or other subjects in
different geographic areas or time periods. "To be useful in forecasting,
analogous time-series should correlate positively - - - over time." This
co-variation can enhance precision and quickly adapt to pattern changes such as
step jumps or turning points. Analogous groups may be identified by several
methods - expert judgment, correlated movements, model-based clustering, and/or
total population pooling. |
Bayesian pooling is emphasized by the authors. Various
statistical methods based on sample means or sample standard deviations or
percentages of totals allow the use of less than ideal data by removing
differences in magnitudes and variances. Standardized data is constantly
recalculated. The construction and combining of local models and group models
for Bayesian pooling is explained.
|
Trend analysis:
& |
"Extrapolation for Time Series and Cross Sectional
Data" - by Armstrong - is a widely used, reliable forecasting method
whenever past behavior is a good predictor of future behavior. It is objective,
inexpensive, quick, easily automated, and replicable. Risks arise if relevant
variables are overlooked - or if there is a sudden break in trend. & |
Developing, recognizing and using appropriate data is essential for reliable analysis, and depends on the skill - and intentions - of the analyst. |
Extrapolation is typically used for inventory and production
forecasts up to two years into the future, and for some long-term forecasts of
stable trends such as population forecasting. "Researchers and practitioners have developed many sensible
and inexpensive procedures for extrapolating reliable forecasts."
|
Data may be derived from historical records of the situation
studied, or of analogous situations - from laboratory experiments - or from
field experiments. Historical data is best for forecasting small changes but
worst for forecasting large changes, where field experiments perform best.
Drawing data from analogous situations is the worst of the three methods for
small changes and the middle choice for large changes. & |
Armstrong advises against relying on cycle forecasts unless the length and amplitude are fairly certain - as for daily use of electricity. Broader cycle forecasts for economic or sociological trends have proven very inaccurate. |
Analyst skill (and intentions) also play major roles - in
structuring the problem - segmentation of the problem - making inflation
and similar adjustments - "cleaning the data" to correct for
erroneous data - identifying and eliminating "outlier" data - applying
statistical techniques such as various averaging methods - aggregating intermittent
data across time or geographic area - adjusting for the impact of sporadic
events such as strikes or wars - and adjusting for seasonal influences by a
variety of statistical methods.
Armstrong advises that simplicity usually aids reliability when
choosing an extrapolation method. He also advises that the most recent data
should be the most heavily weighted unless "the series is subject to
substantial random measurement errors or to instabilities" which might be
magnified by such weighting. "Exponential smoothing" of moving
averages may be especially useful for short period forecasts, but its benefits
decline for longer forecasts. He also advises conservatism - especially when
dealing with volatile trend lines or periods of trend acceleration in long term
forecasts. |
Methods for estimating forecast reliability - "prediction
intervals" - are also reviewed by Armstrong. Reliability problems naturally
increase greatly with the length of the forecast period. Prediction intervals
generally err on the optimistic side and are set too narrowly. However, where
asymmetric errors are common - as in management and social sciences studies -
errors will occur on both the narrow and wide side. Domain knowledge must be
relied upon to evaluate the likelihood of substantial changes in causal forces. |
Neural networks:
& |
"Neural Networks for Time-Series Forecasting"
- by William Remus, College of Business Administration, Univ. of Hawaii, and Marcus O'Connor,
Univ. of New South Wales, Australia - perform best for (1) monthly and quarterly
forecasting, rather than annual forecasting, (2) for forecasting based on
discontinuous series, and (3) for forecasts that are several periods out on the
forecast horizon. Less complex and less expensive time-series methods should be
used for less complex and shorter term forecasting problems. & |
Many of the suggestions applicable to traditional extrapolation
models apply to neural networks. This is a relatively new kind of technique, and
experimentation and development are ongoing. |
|
Particular suggestions include:
|
Rule-based time-series extrapolation:
& |
"Rule-Based Forecasting
Using Judgment in Time-Series Extrapolation" - by Armstrong, Monica
Adya, Dep't. of Management, DePaul Univ., and Fred Callopy, The Weatherhead
School, Case Western Reserve Univ., -applies judgment
from forecasting expertise and domain knowledge to develop and apply rules for
combining extrapolations. Expectations about trends - "causal forces"
- are identified to assign weights to extrapolations. & |
Conditions where rule-based forecasting improves accuracy include: when long interval - annual or longer - data are used, good domain expertise is available, causal forces are clearly identified, causal forces conflict with trend, significant trends exist, uncertainty and instability are not too high, and/or there is a long-range forecast horizon. Domain knowledge seems to be the most important of these.
Even where these conditions don't apply and rule-based
forecasting offers little or no advantage, rules based on causal forces still
improved "the selection of forecasting methods, the structuring of time
series, and the assessment of prediction intervals."
|
In commercial forecasting, domain knowledge is managers' knowledge of episodic events - both past and likely in the future. |
In commercial forecasting, domain knowledge is managers' knowledge of episodic events - both past and likely in the future. Strikes, government regulations, major product or price changes or advertising campaigns or changes in the competitive environment are typical examples of causal factors that will impact trends. Expert knowledge is also used to employ such statistical methods as combining forecasts and dampening trends.
|
Applicable experts rules and their proper usage can be found in the growing literature about the subject. Research about applicable techniques has been derived not just from interviewing expert practitioners, but from actually observing them in action. There are now roughly 100 conditional statements derived for rule-based forecasting. Some of the most broadly applicable are:
|
|
Factors affecting trends are obtained from the cognizant domain
experts and are divided into factors promoting "growth" or
"decay" or "opposing" historical trends or tendencies - or
"regressing" towards the mean, or "supporting" the trend.
The latter is theoretical, since they are generally assumed in the identified
trend. Major events whose sum impact is "unknown" should also
nevertheless be identified. Adjustments for episodic events in the past or
expected should be made by various statistical methods. |
Expert systems:
& |
"Expert Systems for Forecasting" - by
Callopy, Adya, and Armstrong - should be used
in cases involving many forecasts - for complex problems where data on the
dependent variable are of poor quality - and where problem structure is inadequate for
statistical techniques but adequate enough for the development of applicable
rules. Examples include oil well drilling, product introductions, and choice of
medical drugs. & |
Expert systems are far less costly to use than employing experts for estimation and forecasting work. And, they assure consistency. |
Establishment and improvement of expert systems requires
considerable research involving textbooks, research papers, interviews, surveys,
and especially protocol analyses into the methods of cognizant experts. However,
once developed, they should be easy to use. They should incorporate best
practices and knowledge, and reveal applicable reasoning. They are far less
costly to use than employing experts for estimation and forecasting work. And,
they assure consistency. & |
Actual analysis of the experts in action is often needed where the experts themselves cannot explain or have no conscious knowledge of their actual practices. |
The authors discuss the major methods for detailing how experts
work. Actual analysis of the experts in action is often needed where the experts
themselves cannot explain or have no conscious knowledge of their actual
practices. However, this is very costly in time and money. Several experts
should be studied. Where econometric studies are available, they can be very
helpful. Judgmental bootstrapping and conjoint analysis can also be helpful
where applicable. |
Econometric models:
& |
"Econometric Forecasting" - by P.
Geoffrey Allen, Dep't. of Resource Economics, Univ. of Massachusetts, and Robert
Fildes,
The Management School, Univ. of Lancaster,
U.K. - has had its problems. However, results seem to have improved since the
1980s with the development of the "vector autoregression" approach -
now the most widely used type of model - which increases emphasis on dynamics. & |
Econometric forecasting methods have been beaten by less sophisticated methods "more often than they should." "The problem is that we are some way from knowing the correct strategy and principles, although we do seem to be making progress." |
Still, model building efforts can go wrong at a number of
points, so development of a well-specified model requires high levels of
professional skill. As might be expected, broad validation of econometric
methods is difficult and scarce since all models are so different.
|
"If your carefully developed and tested econometric model cannot forecast better than such a simple alternative, all your effort has been in vain." |
However, for a wide variety of microeconomic and other narrower
purposes - where the quantitative inputs are known and the causal variables
can be accurately evaluated - econometric analysis has been performing with
increasing reliability.
|
Other uses for econometric models include as assistance in strategy analysis or policy analysis.
|
Selecting Methods
Selection: |
"Selecting Forecasting Methods" - by Armstrong - provides guidance in determining the efficacy of the various methods for various problems. |
|
Many analysts will force the technique or techniques they are most familiar with on the problem at hand. Selection of an inappropriate method may lead to large errors.
Structured judgment - when competently done - has been shown to enhance reliability and validity.
The relative success of financial advisers, for instance, is seldom known or accurately evaluated. Also, past success is no guarantee of future success - especially if the relevant past was relatively stable. |
Armstrong examines six general frequently applicable considerations.
Usage is frequently unrelated to efficacy. "Forecasters use expert opinions even when ample evidence exists that other methods are more accurate, - - -." Since usage surveys frequently overlook methods such as role playing, judgmental bootstrapping, conjoint analysis, and expert systems, "market popularity is the enemy of innovation."
|
"An operational definition of simple is that an approach can be explained to a manager so clearly that he could then explain it to others." |
Armstrong provides suggestions - "principles" - from published research.
The author provides a "selection tree" graphic
for forecasting methods - with explanations for usage - much of which duplicates
material in earlier chapters - but which is usefully brought together here. |
Combining:
& |
"Combining Forecasts" - by
Armstrong - emphasizes one of the book's most important repetitive themes. Accuracy can be
increased by combining forecasts based on different methods and drawing from
different sources of information - and also by using different forecasters.
Uncertainty - about the situation or best methodology - and a need to avoid
large errors - favor the combining of forecasts. & |
"Combining can reduce errors arising from faulty assumptions, bias, or mistakes in data." Weather forecasters have found that combining longer term forecasts with short term forecasts as they arise improves results. However, combining similar forecasts risks accentuating positively correlated errors. "When inexpensive, it is sensible to combine forecasts from at least five methods."
The various familiar averaging methods and weighting
methods are discussed by the author. Unequal weighting risks the introduction of
bias, but is useful when track records are strong.
Generally, the gains from combining increase as the
forecasting horizon lengthens and also as the differences in the different
forecasts increase. Indeed, a "structured approach for combining
independent forecasts is invariably more accurate" than forecasts made by
groups of experts. |
Integrating judgment with quantitative forecasts:
& |
"Judgmental Time-Series Forecasting Using Domain Knowledge" - by
Richard Webby, Telecordia Technologies, and O'Connor and Michael
Lawrence, School of Information Systems, Univ. of New South Wales, Australia - emphasizes
a particular aspect of the combining theme - that reliability and accuracy can be enhanced
by making appropriate judgmental adjustments in quantitative
methods. Studies demonstrate that "available and valid domain knowledge
improves forecast accuracy." This is a theme repeated in all the
statistical methods chapters. & |
Reliability and accuracy can be enhanced by making appropriate judgmental adjustments in quantitative methods.
"Forecasters are typically left making judgmental adjustments to objective forecasts for a variety of hard-to-model factors," like a new promotional campaign of a client or competitor - product diffusion rates - government regulations - new technological developments - production problems - things that cause discontinuities or trend changes. |
Expert judgment may be essential in dealing with "soft" information or relatively inaccessible domain knowledge, but risks introducing biases or inefficiencies.
"Forecasters are typically left making judgmental
adjustments to objective forecasts for a variety of hard-to-model factors,"
like a new promotional campaign of a client or competitor - product diffusion
rates - government regulations - new technological developments - production
problems - things that cause discontinuities or trend changes. |
Adjustments:
& |
"Judgmental Adjustments of
Statistical Forecasts" - by Nada R. Sanders, Dep't. of Management
Sciences and Information Systems, Wright State Univ., and Larry P. Ritzman,
Operations and Strategic Management, Boston College - emphasizes the need to
adjust statistical forecasts to conform to judgmental factors in appropriate
cases. & |
"Statistical models are only as good as the data upon which they are based."
Judgmental adjustments have been shown to be useful for macroeconomic forecasts. |
Judgment must be used to assure valid data inputs. "Statistical models are only as good as the data upon which they are based." (GIGO - "garbage in, garbage out" - is the first law of mathematical reasoning.) As might be expected, studies show significant increases in forecast accuracy when domain knowledge is applied to statistical forecasts. Judgmental adjustments have been shown to be useful for macroeconomic forecasts. (No surprise here!)
|
Poor record keeping is a widespread forecasting weakness that undermines evaluation and hinders efforts at improvement of forecasting methods.
Biases include optimism, wishful thinking, lack of consistency, and manipulation for political, ideological, economic, or other self serving interests. They can be innocent or deliberate. |
Adjustment processes should be structured,
documented, and periodically checked for accuracy. The authors mention several
methods for structuring the judgmental adjustment process. If the process can be
mechanically integrated, that should be considered. The magnitude of
adjustments, the process used and the reasoning should all be recorded. Poor
record keeping is a widespread forecasting weakness that undermines evaluation
and hinders efforts at improvement of forecasting methods.
Inefficient consideration of data becomes
increasingly likely as the number of judgmental factors increases and causes
"judgment overload." However, both this problem and bias can be
countered by structured approaches. |
Achieving Best Practices
Evaluation:
& |
"Evaluating Forecasting Methods" - by Armstrong - emphasizes the testing of assumptions, data and methods - replicating outputs - and assessing outputs. Evaluation should involve the comparison of the proposed method against reasonable alternatives. The author provides extensive suggestions for best evaluation practices to ascertain and improve reliability. |
|
Uncertainty levels: Prediction intervals provide an upper and lower limit between which the future value is expected to lie with a prescribed accuracy. |
"Prediction Intervals for
Time-Series Forecasting" - by Chris Chatfield, Dep't. of Mathematical
Sciences, Univ. of Bath, U.K. - covers a method of indicating the degree of
uncertainty in a forecast. Chatfield discusses the various methods used, and
ways to overcome tendencies towards overconfidence. & Predicting "an upper and lower limit between which the future value is expected to lie with a prescribed accuracy" helps indicate the level of uncertainty - provides a basis for planning for high and low possibilities - and helps assess different forecasting methods. & |
Unfortunately, little effort is made to indicate
forecast prediction intervals - something especially lacking in econometric
forecasts. Chatfield discusses the approaches available for computing prediction
intervals. & |
Overconfidence:
& |
"Overconfidence in Judgmental
Forecasting" - by Hal R. Arkes, Dep't. of Psychology, Ohio State Univ.
- leads to a variety of forecasting weaknesses. It not only undermines
reliability, it also leads forecasters to neglect available methods that might
increase reliability - to overlook significant data - to go against the average occurrence
rate [the "base rate"] without adequate evidence - to succumb to
"groupthink" where all go along with the most confident member - to
tend to become over committed to a favorite method - to overrate the level of
expertise - or to become tied to initial impressions - among other things. & |
Arkes advises structural requirements - to consider
alternatives, especially in new situations - to require an explicit list of the
reasons why the forecast might be wrong - to employ a "devil's
advocate" to assure consideration of contrary evidence - to test specific
predictions against feedback - and to require experiments where possible. & |
Scenarios: |
"Scenarios and Acceptance of
Forecasts" - by W. Larry Gregory and Anne Duran, Dep't. of Psychology,
New Mexico State Univ. - notes the many proper uses of scenarios - which don't
include as an aid to forecasting. & |
Scenarios have a very dubious record as a forecasting method. |
Scenarios are "detailed stories about 'what happened in the future.'" They can be used properly as an aid for planning for possible outcomes - to get people to think a situation through - and to increase the likelihood that a forecaster's results will be accepted. They are proven propaganda ploys - effective methods for changing perceptions along desired lines. Studies have confirmed that "imagined scenarios could influence behavior." But they have a very dubious record as a forecasting method.
|
Monitoring forecasts:
& |
"Learning from Experience: Coping
with Hindsight Bias and Ambiguity" - by Baruch Fischhoff, Dep't of
Social and Decision Sciences, Carnegie Mellon Univ. - emphasizes that the
obtaining of accurate feedback is essential for evaluation, learning and
professional development - and for dealing with hindsight bias and ambiguity. Of
course, both hindsight bias and ambiguity may be "motivational" rather
than "cognitive" - an attempt to achieve or retain personal status.
(It might also be part of an intentional effort to further some political,
economic or ideological objective.) & |
The obtaining of accurate feedback is essential for evaluation, learning and professional development - and for dealing with hindsight bias and ambiguity. |
Hindsight bias is "the tendency to exaggerate
in hindsight what one was able to predict in foresight -- or would have been
able to predict had one been asked." It obscures past forecasts, the bases
of those forecasts, and the real achievements of the forecaster. |
Feedback should be obtained in a formal way - with accurate written records of forecasts, reasoning, results and analyses. |
Feedback should be obtained in a formal way - with
accurate written records of forecasts, reasoning, results and analyses. Detailed
records facilitate the testing of methodology and data sources, and evaluation
of whether error was avoidable or just within a normal range of variability.
Detailed written records also protect the forecaster from Monday morning
quarterbacks by setting forth all the known conditions and alternative
possibilities that existed in the absence of 20/20 hindsight. Most important of
all, detailed written records also provide information to the recipients of
forecasts to help proper understanding and use. & |
Examples of Use of Best Practices
Extrapolation methods: |
"Population
Forecasting" - by Dennis A. Ahburg, Carlson School of Management, Univ.
of Minnesota - demonstrates how forecasting methods and principles have been
applied in demographic studies. Even this most precise of forecasting fields has
significant problems. & |
Experts tend to simply extrapolate trends - and are routinely caught short when trends change. |
Disaggregation is a predominant method. Population
groupings such as age, sex and race - per time unit - within particular
geographic areas - are common characteristics broken out for separate
measurement. |
Extrapolation methods:
& |
"Forecasting the Diffusion
of Innovations: Implications for Time-Series Extrapolation" - by Niger
Meade, The Management School, Univ. of London, U.K., and Towhidul Islam, Faculty of
Management, Univ. of Northern British Columbia, Canada - discusses a variety of
diffusion models and their limitations and appropriate application. The authors
stress that "no single mathematical model is suitable for heterogeneous
diffusion processes." As is often the case, simpler models seem to
outperform more complex models. & |
Performance monitoring for short segments of the
forecast period is suggested as an appropriate way of judging how a
model will perform for the whole period. They also state that diffusion models
are inappropriate for ordinary consumption forecasts. & |
Econometric models:
& |
"Econometric Models for
Forecasting Market Share" - by Roderick J. Brodie and Peter J. Danaher,
Univ. of Auckland, New Zealand, V. Kumar, Univ. of Houston, and Peter S. H.
Leeflang,
Univ. of Groningen, The Netherlands, - discusses the limitations and appropriate usage of
econometric market share models. & |
For volatile situations where large changes in
established markets may occur, econometric models enable the analyst to
incorporate causal reasoning about the nature of competitive effects - of price,
advertising, promotion and other competitive activity that may be responsible
for the volatility. However, the authors explain that extrapolation forecasting
methods may be more appropriate in mature markets where only marginal changes
are likely - and judgmental methods like conjoint analysis and intentions
surveys may be more appropriate if historic data about the market is sparse,
as in the case of new products. |
Test market models: |
"Forecasting Trial Sales of New Consumer Packaged
Goods" - by Peter S. Fader, The Wharton School, Univ. of Pennsylvania,
and Bruce G. S. Hardie, London Business School, U.K. - discusses appropriate usage and
best practices for forecasting new packaged goods sales. & |
Since promotional expenses can be many times more
costly than development expenses - especially for inexpensive packaged goods
like shampoo - accurate forecasts of both initial sales and repeat purchase
rates are vital factors in the decision to launch. Analogy based methods drawing
on prior experience with similar goods, as well as other judgment based methods,
are inexpensive and appropriate at certain early stages of product development,
but as the need for major commitments arises, more formal pretest market models
and test market models become advisable. There have been no reported studies of
the reliability of subjective or analogy based methods, and the frequency of new
product failure speaks volumes for the need for better forecasting methods. & |
|
Test market models measure both initial and repeat
purchases during some initial test marketing period to quickly project the likely success of the product.
The applicability of models is
important since full fledged test marketing is expensive in time as well as in
money - giving competitors time to react. |
Implementation Sources
Books:
& |
"Diffusion of Forecasting Principles
through Books" - by James E. Cox, Jr., and David G. Loomis, Illinois
State Univ. - evaluates 18 forecasting books - and finds that they concentrate
on understanding and application of technique with at best limited attention to
the identity and evaluation of best practices. & |
Software:
These programs are especially poor in assessment of uncertainty and all have other significant weaknesses that make forecasting expertise essential for proper use. |
"Diffusion of Forecasting Principles
through Software" - by Leonard J. Tashman, School of Business
Administration, Univ. of Vermont, and Jim Hoover, U.S. Navy Dep't. - evaluates
four categories of commercially available statistical forecasting method
software that employ time series data - including 15 individual programs: (1) spreadsheet
add-ins, (2) forecasting modules of general statistical programs, (3) neural
network programs, and (4) dedicated business forecasting programs. & These offer many advantages, and are rated somewhat higher than the books at covering best practices. However, they are especially poor in assessment of uncertainty and all have other significant weaknesses that make forecasting expertise essential for proper use. Judgment and skill are required for such tasks as "setting objectives, structuring problems and identifying information sources," among others. & |
"[I]t has yet to be shown that multi-equation models add value to forecasting." |
Dedicated business forecasting programs were rated
as the best - and are especially strong in method selection, implementation, and
evaluation. Three of them "contain features designed to reconcile forecasts
across a product hierarchy, a task this group performs so commendably it can
serve as a role model for forecasting engines in demand-planning systems."
Software that does not employ time series data - such as
conjoint analysis software and other software designed to enhance and support
judgmental forecasting - are also not covered. Many of the time series programs
covered do offer elements that employ some of these methods. |
FUTURECASTS Methods
The "Cassandra" forecasting method:
& |
FUTURECASTS relies upon judgmental methods for its forecasts since it is
concerned predominantly with subject matter in the non scientific practical arts. It relies heavily on trend
analysis - but is always alert to the need to recognize and take into account
the new influences constantly arising. All historic trends are subject to the
rapid changes in the modern world's political, economic, military, societal, and
technological environment. & |
History opens windows on the future.
There is nothing that is completely new under the sun - but nothing ever precisely repeats itself.
Crisis is opportunity! |
However, FUTURECASTS has another - unfortunately
extremely useful and reliable - forecasting method. The "Cassandra"
forecasting method is one of the most ancient in recorded history. Armstrong's
otherwise excellent handbook totally ignores this very useful and incredibly
accurate forecasting method - which the publisher of FUTURECASTS has been using
successfully since the middle of the 1960s.
|
Please return to our Homepage and e-mail your name and comments.
Copyright © 2003 Dan Blatt