Live from the Institute of Business Forecasting conference in Orlando this week, where I'm excited to announce that my longtime SAS colleague Charlie Chase just received the 2013 Lifetime Achievement Award for his contributions to the IBF and the forecasting profession. More details and and pictures later this week... Reminder:
Author
In this quarter's installment of the SAS/Foresight Webinar Series, Steve Morlidge will discuss his promising new approach to evaluating the "Avoidability of Forecast Error." Based on his article in the Summer 2013 issue of Foresight, and examined in a four-part series on The BFD blog (Part 1, Part 2, Part
Forecast Value Added (FVA) is a metric for comparing the performance of your organization’s forecasting process to “doing nothing” and using a naïve model to generate your forecasts. The idea is, if all the resources and effort we put into forecasting are not providing forecasts that are better than using
SAS/Foresight Webinar Series The next installment of the quarterly SAS/Foresight webinar series is next Thursday, September 19, at 1:00pm EDT. Demand and Supply Integration: Achieving S&OP's Strategic Promise Join Dr. Mark Moon, Department of Marketing and Supply Chain Management at the University of Tennessee, to discover the benefits of integrating demand
Message from Mohsen Hamoudia (IIF President): For the eleventh year, the International Institute of Forecasters, in collaboration with SAS®, is proud to announce financial support for research on how to improve forecasting methods and business forecasting practice. The award for this year will be two (2) $5,000 grants. The
If you need an excuse to get out of the office and perhaps learn a thing or two this fall, here are three upcoming events: Foresight Practitioner Conference: S&OP and Collaborative Forecasting (Columbus, OH, September 25-26) From the campus of Ohio State University, Foresight's editor Len Tashman and S&OP column
The Empirical Evidence Steve Morlidge presents results from two test datasets (the first with high levels of manual intervention, the second with intermittent demand patterns), intended to challenge the robustness of the avoidability principle. The first dataset contained one year of weekly forecasts for 124 product SKUs at a fast-moving consumer
Suppose we have a perfect forecasting algorithm. This means that we know the "rule" guiding the behavior we are forecasting (i.e., we know the signal), and we have properly expressed the rule in our forecasting algorithm. As long as the rule governing the behavior doesn't change in the future, then any
While I've long advocated the use of Coefficient of Variation (CV) as a quick and dirty indicator of the forecastability of a time-series, its deficiencies are well recognized. It is true that any series with extremely low CV can be forecast quite accurately (using a moving average or simple exponential smoothing
"Forecastability" is a frequent topic of discussion on The BFD, and an essential consideration when evaluating the effectiveness of any forecasting process. A major critique of forecasting benchmarks is that they fail to take forecastability into consideration: An organization with "best in class" forecast accuracy may do so only because
SAS Forecast Server (release 12.3) is now shipping, and includes the new SAS Time Series Studio GUI. Time Series Studio (TSS), was released as "experimental" last August in 12.1, and is now in production. TSS provides an additional interface in Forecast Server, for time series data mining, exploration, and data preparation.
Mercifully, we have reached the final installment of Q&A from the June 20 Foresight-SAS webinar, "Forecast Value Added: A Reality Check on Forecasting Practices." As a reminder, a recording of the webinar is available for on-demand review, and the Foresight article (upon which the webinar was based) is available for free
If you are looking to find a job in business forecasting, or trying to fill one, there are many online resources available: Professional Organizations Institute of Business Forecasting & Planning - Very active searchable list of currently available jobs in forecasting and planning, from entry level to executive. Employers post
Q: Is the MAPE of the naive forecast the basis for understanding the forecastability of the behavior? Or are there other more in depth ways to measure the forecastability of a behavior? MAPE of the naive forecast indicates the worst you should be able to forecast the behavior. You can
Q: Company always try to forecast 12 or 24m ahead. Whether we should track accuracy of 1m/3m/ 6m or x month forecast, does that depend on lead time? How to determine out of these 12/24 months, which month should we track accuracy? Correct, forecast performance is usually evaluated against the
Q: What is a legitimate goal to expect from your FVA...5%, 10%? Q: How do we set Target FVA which Forecasters can drive towards? The appropriate goal is to do no worse than a naive model, that is FVA ≥ 0. Sometimes, especially over short periods of time, you may
Please enjoy a much-needed break from FVA Q&A with editor Len Tashman's preview of the Summer 2013 issue of Foresight: Enlightenment has been our guiding principle through this, our 30th issue of Foresight. Since the journal’s inception in 2005, our mission has been to help the forecasting profession come to
With this Q&A Part 3, we are about halfway through the questions submitted during the FVA webinar. We did over 15 minutes of live Q&A at the end of the webinar, and covered many of the submitted questions at that time, however I always prefer to issue complete written responses to
Q: Could you send me the presentation? With audio if possible. If you'd like a pdf of the slides, email me directly: mike.gilliland@sas.com For the audio, the webinar recording is available for free on-demand review: FVA: A Reality Check on Forecasting Practices Q: Can we get the case study referred here
As promised in yesterday's Foresight-SAS sponsored webinar on "Forecast Value Added: A Reality Check on Forecasting Practices," here is Part 1 of my written response to the over 25 questions that were submitted during the event. (Note: It may take a week or so to get through all of them.)
If an organization is spending time and money to have a forecasting process, is it not reasonable to expect the process to make the forecast more accurate and less biased (or at least not make it any worse!)? But how would we ever know what the process is accomplishing? To
Last time we saw two situations where you wouldn't bother trying to improve your forecast: When forecast accuracy is "good enough" and is not constraining organizational performance. When the costs and consequences of a less-than-perfect forecast are low. (Another situation was brought to my attention by Sean Schubert of
If the popularity of one's blog can be measured by the number of comments received, then The BFD has become quite popular. Many of the comments are quite flattering, such as: Hello, I check your blog like еvery week. Үour writing style is wittу, keep doing ωhat you're doing! Vеry
"Why Should I Trust Your Forecasts?" now available on-demand The SAS / Foresight webinar series had a rousing kickoff on April 24, with Paul Goodwin asking (and answering) the question, "Why Should I Trust Your Forecasts?" The webinar is now available for free on-demand review . Be sure to stick
Editor Len Tashman's Preview of Foresight For a look at articles in the Spring issue of Foresight: The International Journal of Applied Forecasting, here is editor Len Tashman's preview: Kevin Foley is an IIF–Certified Forecaster with over 15 years of consultant experience in defense and aerospace companies. Drawing on this
When you work at headquarters of the leader in advanced analytics software, you never know who you'll encounter in the lobby. It might be celebrity statistician (and New York Times FiveThirtyEight blogger) Nate Silver, of The Signal and the Noise and election forecasting fame. It might be Donald Wheeler, giant
This week Nate Silver, renowned election forecaster (fivethirtyeight blog) and top selling author (of the excellent The Signal and the Noise), spoke at an event here in my building on the SAS campus. Unfortunately, I wasn't considered a B enough of a FD to land an invite to Nate's presentation. However,
The one-number forecasting concept has been debated for years. Advocates argue that having different groups within the same organization working to different forecasts is insane. You can't have the supply chain building to X, the sales force selling to Y, and the financial folks counting on revenue of Z. This
There is a well recognized phenomenon that combining forecasts, derived from different methods using different sources of information, can improve forecast accuracy. This approach, sometimes called "ensemble forecasting," is available in SAS Forecast Server. Per Scott Armstrong's review of 57 studies on combining forecasts, "the combined forecast can be better
Back in 2012 I advocated man-dog love and introduced you to my foster dog Mikey. While an endearing little fellow, Mikey did have a bit of a fabric fetish, so his new family has neither curtains nor tablecloths. Mikey's fetish was discovered while still in foster, when he tore the skirt off