Forecasting using simple models

Slides:



Advertisements
Similar presentations
Chapter 9. Time Series From Business Intelligence Book by Vercellis Lei Chen, for COMP
Advertisements

Module 4. Forecasting MGS3100.
Decomposition Method.
Part II – TIME SERIES ANALYSIS C3 Exponential Smoothing Methods © Angel A. Juan & Carles Serrat - UPC 2007/2008.
Forecasting OPS 370.
Part II – TIME SERIES ANALYSIS C5 ARIMA (Box-Jenkins) Models
1 1 Slide © 2011 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole.
4-1 Operations Management Forecasting Chapter 4 - Part 2.
Time Series Analysis Autocorrelation Naive & Simple Averaging
Moving Averages Ft(1) is average of last m observations
Chapter 12 - Forecasting Forecasting is important in the business decision-making process in which a current choice or decision has future implications:
Data Sources The most sophisticated forecasting model will fail if it is applied to unreliable data Data should be reliable and accurate Data should be.
Forecasting.
Forecasting  Suppose your fraternity/sorority house consumed the following number of cases of beer for the last 6 weekends: 8, 5, 7, 3, 6, 9  How many.
MOVING AVERAGES AND EXPONENTIAL SMOOTHING
Chapter 3 Forecasting McGraw-Hill/Irwin
Chapter 13 Forecasting.
Part II – TIME SERIES ANALYSIS C2 Simple Time Series Methods & Moving Averages © Angel A. Juan & Carles Serrat - UPC 2007/2008.
Forecasting McGraw-Hill/Irwin Copyright © 2012 by The McGraw-Hill Companies, Inc. All rights reserved.
2. Forecasting. Forecasting  Using past data to help us determine what we think will happen in the future  Things typically forecasted Demand for products.
McGraw-Hill/Irwin Copyright © 2007 by The McGraw-Hill Companies, Inc. All rights reserved. 3 Forecasting.
Business Forecasting Chapter 5 Forecasting with Smoothing Techniques.
Slides 13b: Time-Series Models; Measuring Forecast Error
1 BABS 502 Moving Averages, Decomposition and Exponential Smoothing Revised March 11, 2011.
Constant process Separate signal & noise Smooth the data: Backward smoother: At any give T, replace the observation yt by a combination of observations.
Forecasting Chapter 15.
LSS Black Belt Training Forecasting. Forecasting Models Forecasting Techniques Qualitative Models Delphi Method Jury of Executive Opinion Sales Force.
1 1 Slide © 2011 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole.
IES 371 Engineering Management Chapter 13: Forecasting
Demand Management and Forecasting
Lecture 7: Forecasting: Putting it ALL together. The full model The model with seasonality, quadratic trend, and ARMA components can be written: Ummmm,
DSc 3120 Generalized Modeling Techniques with Applications Part II. Forecasting.
Copyright © 2011 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin Time Series Forecasting Chapter 16.
McGraw-Hill/Irwin Copyright © 2007 by The McGraw-Hill Companies, Inc. All rights reserved. Time Series Forecasting Chapter 13.
1 DSCI 3023 Forecasting Plays an important role in many industries –marketing –financial planning –production control Forecasts are not to be thought of.
DAVIS AQUILANO CHASE PowerPoint Presentation by Charlie Cook F O U R T H E D I T I O N Forecasting © The McGraw-Hill Companies, Inc., 2003 chapter 9.
Operations Management For Competitive Advantage 1Forecasting Operations Management For Competitive Advantage Chapter 11.
MBA.782.ForecastingCAJ Demand Management Qualitative Methods of Forecasting Quantitative Methods of Forecasting Causal Relationship Forecasting Focus.
Operations Research II Course,, September Part 6: Forecasting Operations Research II Dr. Aref Rashad.
Time Series Analysis and Forecasting
Forecasting Operations Management For Competitive Advantage.
Lesson 5.1 Forecasting Moving Averages Exponential Smoothing.
Forecasting. 預測 (Forecasting) A Basis of Forecasting In business, forecasts are the basis for budgeting and planning for capacity, sales, production and.
15-1 Copyright © 2010 Pearson Education, Inc. Publishing as Prentice Hall Forecasting Chapter 15.
4-1 Operations Management Forecasting Chapter 4 - Part 2.
Welcome to MM305 Unit 5 Seminar Prof Greg Forecasting.
Time Series Analysis and Forecasting. Introduction to Time Series Analysis A time-series is a set of observations on a quantitative variable collected.
15-1 Copyright © 2010 Pearson Education, Inc. Publishing as Prentice Hall Forecasting Chapter 15.
1 BABS 502 Moving Averages, Decomposition and Exponential Smoothing Revised March 14, 2010.
The Box-Jenkins (ARIMA) Methodology
MGS3100_03.ppt/Feb 11, 2016/Page 1 Georgia State University - Confidential MGS 3100 Business Analysis Time Series Forecasting Feb 11, 2016.
FORECASTING Introduction Quantitative Models Time Series.
Forecasting is the art and science of predicting future events.
Forecasting Demand. Problems with Forecasts Forecasts are Usually Wrong. Every Forecast Should Include an Estimate of Error. Forecasts are More Accurate.
1 Decision Making ADMI 6510 Forecasting Models Key Sources: Data Analysis and Decision Making (Albrigth, Winston and Zappe) An Introduction to Management.
Forecasting Production and Operations Management 3-1.
Demand Management and Forecasting Chapter 11 Portions Copyright © 2010 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin.
Forecast 2 Linear trend Forecast error Seasonal demand.
Chapter 15 Forecasting. Forecasting Methods n Forecasting methods can be classified as qualitative or quantitative. n Such methods are appropriate when.
Demand Forecasting Production and Operations Management Judit Uzonyi-Kecskés Ph.D. Student Department of Management and Corporate Economics Budapest University.
Chapter 11 – With Woodruff Modications Demand Management and Forecasting Copyright © 2010 by The McGraw-Hill Companies, Inc. All rights reserved.McGraw-Hill/Irwin.
Forecasting Quantitative Methods. READ FIRST Outline Define Forecasting The Three Time Frames of Forecasting Forms of Forecast Movement Forecasting Approaches.
TIME SERIES MODELS. Definitions Forecast is a prediction of future events used for planning process. Time Series is the repeated observations of demand.
Welcome to MM305 Unit 5 Seminar Dr. Bob Forecasting.
Forecasts.
Forecasting Methods Dr. T. T. Kachwala.
Exponential Smoothing
Demand Management and Forecasting
Forecasting Plays an important role in many industries
Presentation transcript:

Forecasting using simple models

Outline Basic forecasting models The basic ideas behind each model When each model may be appropriate Illustrate with examples Forecast error measures Automatic model selection Adaptive smoothing methods (automatic alpha adaptation) Ideas in model based forecasting techniques Regression Autocorrelation Prediction intervals

Basic Forecasting Models Moving average and weighted moving average First order exponential smoothing Second order exponential smoothing First order exponential smoothing with trends and/or seasonal patterns Croston’s method

M-Period Moving Average i.e. the average of the last M data points Basically assumes a stable (trend free) series How should we choose M? Advantages of large M? Average age of data = M/2

Weighted Moving Averages The Wi are weights attached to each historical data point Essentially all known (univariate) forecasting schemes are weighted moving averages Thus, don’t screw around with the general versions unless you are an expert

Simple Exponential Smoothing Pt+1(t) = Forecast for time t+1 made at time t Vt = Actual outcome at time t 0<<1 is the “smoothing parameter”

Two Views of Same Equation Pt+1(t) = Pt(t-1) + [Vt – Pt(t-1)] Adjust forecast based on last forecast error OR Pt+1(t) = (1- )Pt(t-1) + Vt Weighted average of last forecast and last Actual

Simple Exponential Smoothing Is appropriate when the underlying time series behaves like a constant + Noise Xt =  + Nt Or when the mean  is wandering around That is, for a quite stable process Not appropriate when trends or seasonality present

ES would work well here

Simple Exponential Smoothing We can show by recursive substitution that ES can also be written as: Pt+1(t) = Vt + (1-)Vt-1 + (1-)2Vt-2 + (1-)3Vt-3 +….. Is a weighted average of past observations Weights decay geometrically as we go backwards in time

Simple Exponential Smoothing Ft+1(t) = At + (1-)At-1 + (1-)2At-2 + (1-)3At-3 +….. Large  adjusts more quickly to changes Smaller  provides more “averaging” and thus lower variance when things are stable Exponential smoothing is intuitively more appealing than moving averages

Exponential Smoothing Examples

Zero Mean White Noise

Shifting Mean + Zero Mean White Noise

Automatic selection of  Using historical data Apply a range of  values For each, calculate the error in one-step-ahead forecasts e.g. the root mean squared error (RMSE) Select the  that minimizes RMSE

RMSE vs Alpha 1.45 1.4 1.35 RMSE 1.3 1.25 1.2 1.15 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 Alpha

Recommended Alpha Typically alpha should be in the range 0.05 to 0.3 If RMSE analysis indicates larger alpha, exponential smoothing may not be appropriate

Might look good, but is it?

Series and Forecast using Alpha=0.9 2 1.5 1 Forecast 0.5 -0.5 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 Period

Forecast RMSE vs Alpha 0.67 0.66 0.65 0.64 0.63 Forecast RMSE 0.62 Series1 0.61 0.6 0.59 0.58 0.57 0.2 0.4 0.6 0.8 1 Alpha

Forecast RMSE vs Alpha for Lake Huron Data 1.1 1.05 1 0.95 0.9 RMSE 0.85 0.8 0.75 0.7 0.65 0.6 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 Alpha

for Monthly Furniture Demand Data Forecast RMSE vs Alpha for Monthly Furniture Demand Data 45.6 40.6 35.6 30.6 25.6 RMSE 20.6 15.6 10.6 5.6 0.6 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 Alpha

Exponential smoothing will lag behind a trend Suppose Xt=b0+ b1t And St= (1- )St-1 + Xt Can show that

Double Exponential Smoothing Modifies exponential smoothing for following a linear trend i.e. Smooth the smoothed value

St Lags St[2] Lags even more

2St -St[2] doesn’t lag

Example

=0.2

Single Lags a trend

Double Over-shoots a change (must “re-learn” the slope) 6 5 4 Double Over-shoots a change (must “re-learn” the slope) 3 Trend 2 Series Data Single Smoothing 1 Double smoothing -1 1 6 11 16 21 26 31 36 41 46 51 56 61 66 71 76 81 86 91 96 101

Holt-Winters Trend and Seasonal Methods “Exponential smoothing for data with trend and/or seasonality” Two models, Multiplicative and Additive Models contain estimates of trend and seasonal components Models “smooth”, i.e. place greater weight on more recent data

Winters Multiplicative Model Xt = (b1+b2t)ct + t Where ct are seasonal terms and Note that the amplitude depends on the level of the series Once we start smoothing, the seasonal components may not add to L

Holt-Winters Trend Model Xt = (b1+b2t) + t Same except no seasonal effect Works the same as the trend + season model except simpler

Example:

(1+0.04t)

*150%

*50%

The seasonal terms average 100% (i.e. 1) Thus summed over a season, the ct must add to L Each period we go up or down some percentage of the current level value The amplitude increasing with level seems to occur frequently in practice

Recall Australian Red Wine Sales

Smoothing In Winters model, we smooth the “permanent component”, the “trend component” and the “seasonal component” We may have a different smoothing parameter for each (, , ) Think of the permanent component as the current level of the series (without trend)

Current Observation

Current Observation “deseasonalized”

Estimate of permanent component from last time = last level + slope*1

“observed” slope

“observed” slope “previous” slope

Extend the trend out  periods ahead

Use the proper seasonal adjustment

Winters Additive Method Xt = b1+ b2t + ct + t Where ct are seasonal terms and Similar to previous model except we “smooth” estimates of b1, b2, and the ct

Croston’s Method Can be useful for intermittent, erratic, or slow-moving demand e.g. when demand is zero most of the time (say 2/3 of the time) Might be caused by Short forecasting intervals (e.g. daily) A handful of customers that order periodically Aggregation of demand elsewhere (e.g. reorder points)

Typical situation Central spare parts inventory (e.g. military) Orders from manufacturer in batches (e.g. EOQ) periodically when inventory nearly depleted long lead times may also effect batch size

Example Demand each period follows a distribution that is usually zero

Example

Example Exponential smoothing applied (=0.2)

Using Exponential Smoothing: Forecast is highest right after a non-zero demand occurs Forecast is lowest right before a non-zero demand occurs

Croston’s Method Forecast = Separately Tracks Time between (non-zero) demands Demand size when not zero Smoothes both time between and demand size Combines both for forecasting Demand Size Forecast = Time between demands

Define terms V(t) = actual demand outcome at time t P(t) = Predicted demand at time t Z(t) = Estimate of demand size (when it is not zero) X(t) = Estimate of time between (non-zero) demands q = a variable used to count number of periods between non-zero demand

Forecast Update For a period with zero demand Z(t)=Z(t-1) X(t)=X(t-1) No new information about order size Z(t) time between orders X(t) q=q+1 Keep counting time since last order

Forecast Update q=1 For a period with non-zero demand Z(t)=Z(t-1) + (V(t)-Z(t-1)) X(t)=X(t-1) + (q - X(t-1)) q=1

Forecast Update q=1 Latest order size For a period with non-zero demand Z(t)=Z(t-1) + (V(t)-Z(t-1)) X(t)=X(t-1) + (q - X(t-1)) q=1 Update Size of order via smoothing Latest order size

Forecast Update q=1 Latest time between orders For a period with non-zero demand Z(t)=Z(t-1) + (V(t)-Z(t-1)) X(t)=X(t-1) + (q - X(t-1)) q=1 Update size of order via smoothing Update time between orders via smoothing Latest time between orders

Forecast Update q=1 Reset counter For a period with non-zero demand Z(t)=Z(t-1) + (V(t)-Z(t-1)) X(t)=X(t-1) + (q - X(t-1)) q=1 Update size of order via smoothing Update time between orders via smoothing Reset counter of time between orders Reset counter

Forecast P(t) = = Finally, our forecast is: Z(t) Non-zero Demand Size X(t) Time Between Demands

Recall example Exponential smoothing applied (=0.2)

Recall example Croston’s method applied (=0.2)

True average demand per period=0.176 What is it forecasting? Average demand per period True average demand per period=0.176

Behavior Forecast only changes after a demand Forecast constant between demands Forecast increases when we observe A large demand A short time between demands Forecast decreases when we observe A small demand A long time between demands

Croston’s Method Croston’s method assumes demand is independent between periods That is one period looks like the rest (or changes slowly)

Counter Example One large customer Orders using a reorder point The longer we go without an order The greater the chances of receiving an order In this case we would want the forecast to increase between orders Croston’s method may not work too well

Better Examples Demand is a function of intermittent random events Military spare parts depleted as a result of military actions Umbrella stocks depleted as a function of rain Demand depending on start of construction of large structure

Is demand Independent? If enough data exists we can check the distribution of time between demand Should “tail off” geometrically

Theoretical behavior

In our example:

Comparison

Counterexample Croston’s method might not be appropriate if the time between demands distribution looks like this:

Counterexample In this case, as time approaches 20 periods without demand, we know demand is coming soon. Our forecast should increase in this case

Error Measures Errors: The difference between actual and predicted (one period earlier) et = Vt – Pt(t-1) et =can be positive or negative Absolute error |et| Always positive Squared Error et2 The percentage error PEt = 100et / Vt Can be positive or negative

Bias and error magnitude Forecasts can be: Consistently too high or too low (bias) Right on average, but with large deviations both positive and negative (error magnitude) Should monitor both for changes

Error Measures Look at errors over time Cumulative measures summed or averaged over all data Error Total (ET) Mean Percentage Error (MPE) Mean Absolute Percentage Error (MAPE) Mean Squared Error (MSE) Root Mean Squared Error (RMSE) Smoothed measures reflects errors in the recent past Mean Absolute Deviation (MAD)

Error Measures Measure Bias Look at errors over time Cumulative measures summed or averaged over all data Error Total (ET) Mean Percentage Error (MPE) Mean Absolute Percentage Error (MAPE) Mean Squared Error (MSE) Root Mean Squared Error (RMSE) Smoothed measures reflects errors in the recent past Mean Absolute Deviation (MAD)

Error Measures Measure error magnitude Look at errors over time Cumulative measures summed or averaged over all data Error Total (ET) Mean Percentage Error (MPE) Mean Absolute Percentage Error (MAPE) Mean Squared Error (MSE) Root Mean Squared Error (RMSE) Smoothed measures reflects errors in the recent past Mean Absolute Deviation (MAD)

Error Total Sum of all errors Uses raw (positive or negative) errors ET can be positive or negative Measures bias in the forecast Should stay close to zero as we saw in last presentation

MPE Average of percent errors Can be positive or negative Measures bias, should stay close to zero

MSE Average of squared errors Always positive Measures “magnitude” of errors Units are “demand units squared”

RMSE Square root of MSE Always positive Measures “magnitude” of errors Units are “demand units” Standard deviation of forecast errors

MAPE Average of absolute percentage errors Always positive Measures magnitude of errors Units are “percentage”

Mean Absolute Deviation Smoothed absolute errors Always positive Measures magnitude of errors Looks at the recent past

Percentage or Actual units Often errors naturally increase as the level of the series increases Natural, thus no reason for alarm If true, percentage based measured preferred Actual units are more intuitive

Squared or Absolute Errors Absolute errors are more intuitive Standard deviation units less so 66% within  1 S.D. 95% within  2 S.D. When using measures for automatic model selection, there are statistical reasons for preferring measures based on squared errors

Ex-Post Forecast Errors Given A forecasting method Historical data Calculate (some) error measure using the historical data Some data required to initialize forecasting method. Rest of data (if enough) used to calculate ex-post forecast errors and measure

Automatic Model Selection For all possible forecasting methods (and possibly for all parameter values e.g. smoothing constants – but not in SAP?) Compute ex-post forecast error measure Select method with smallest error

Automatic  Adaptation Suppose an error measure indicates behavior has changed e.g. level has jumped up Slope of trend has changed We would want to base forecasts on more recent data Thus we would want a larger 

Tracking Signal (TS) Bias/Magnitude = “Standardized bias”

 Adaptation If TS increases, bias is increasing, thus increase  I don’t like these methods due to instability

Model Based Methods Find and exploit “patterns” in the data Trend and Seasonal Decomposition Time based regression Time Series Methods (e.g. ARIMA Models) Multiple Regression using leading indicators Assumes series behavior stays the same Requires analysis (no “automatic model generation”)

Univariate Time Series Models Based on Decomposition Vt = the time series to forecast Vt = Tt + St + Nt Where Tt is a deterministic trend component St is a deterministic seasonal/periodic component Nt is a random noise component

(Vt)=0.257

Simple Linear Regression Model: Vt=2.877174+0.020726t

Use Model to Forecast into the Future

Residuals = Actual-Predicted et = Vt-(2.877174+0.020726t)

Simple Seasonal Model Estimate a seasonal adjustment factor for each period within the season e.g. SSeptember

Sorted by season Season averages

Trend + Seasonal Model Vt=2.877174+0.020726t + Smod(t,3) Where

et = Vt - (2.877174 + 0.020726t + Smod(t,3)) (et)=0.145

Can use other trend models Vt= 0+ 1Sin(2t/k) (where k is period) Vt= 0+ 1t + 2t2 (multiple regression) Vt= 0+ 1ekt etc. Examine the plot, pick a reasonable model Test model fit, revise if necessary

Model: Vt = Tt + St + Nt After extracting trend and seasonal components we are left with “the Noise” Nt = Vt – (Tt + St) Can we extract any more predictable behavior from the “noise”? Use Time Series analysis Akin to signal processing in EE

Zero Mean, and Aperiodic: Is our best forecast ?

AR(1) Model This data was generated using the model Nt = 0.9Nt-1 + Zt Where Zt ~N(0,2) Thus to forecast Nt+1,we could use:

Time Series Models Examine the correlation of the time series to past values. This is called “autocorrelation” If Nt is correlated to Nt-1, Nt-2,….. Then we can forecast better than

Sample Autocorrelation Function

Back to our Demand Data

No Apparent Significant Autocorrelation

Multiple Linear Regression V= 0+ 1 X1 + 2 X2 +….+ p Xp +  Where V is the “independent variable” you want to predict The Xi‘s are the dependent variables you want to use for prediction (known) Model is linear in the i‘s

Examples of MLR in Forecasting Vt= 0+ 1t + 2t2 + 3Sin(2t/k) + 4ekt i.e a trend model, a function of t Vt= 0+ 1X1t + 2X2t Where X1t and X2t are leading indicators Vt= 0+ 1Vt-1+ 2Vt-2 + 12Vt-12 +13Vt-13 An Autoregressive model

Example: Sales and Leading Indicator

Example: Sales and Leading Indicator Sales(t) = -3.93+0.83Sales(t-3) -0.78Sales(t-2)+1.22Sales(t-1) -5.0Lead(t)