Lecture 7: Exponential Smoothing Methods Please read Chapter 4 and Chapter 2 of MWH Book

Similar documents
Economic Forecasting Lecture 9: Smoothing Methods

Time-series Forecasting - Exponential Smoothing

Forecasting: The First Step in Demand Planning

A time series is called strictly stationary if the joint distribution of every collection (Y t

Austrian Inflation Rate

Comparing the Univariate Modeling Techniques, Box-Jenkins and Artificial Neural Network (ANN) for Measuring of Climate Index

Multiplicative Winter s Smoothing Method

THE ROYAL STATISTICAL SOCIETY 2009 EXAMINATIONS SOLUTIONS GRADUATE DIPLOMA MODULAR FORMAT MODULE 3 STOCHASTIC PROCESSES AND TIME SERIES

DEPARTMENT OF QUANTITATIVE METHODS & INFORMATION SYSTEMS

Exponential Smoothing. INSR 260, Spring 2009 Bob Stine

STA 6104 Financial Time Series. Moving Averages and Exponential Smoothing

Forecasting. Al Nosedal University of Toronto. March 8, Al Nosedal University of Toronto Forecasting March 8, / 80

Probabilistic forecasting of solar radiation

Decision 411: Class 4

Forecasting. Simon Shaw 2005/06 Semester II

Operation and Supply Chain Management Prof. G. Srinivasan Department of Management Studies Indian Institute of Technology, Madras

TIMES SERIES INTRODUCTION INTRODUCTION. Page 1. A time series is a set of observations made sequentially through time

Time Series Analysis of United States of America Crude Oil and Petroleum Products Importations from Saudi Arabia

Lesson 2: Analysis of time series

Lecture 6a: Unit Root and ARIMA Models

Chapter 9: Forecasting

Econ 300/QAC 201: Quantitative Methods in Economics/Applied Data Analysis. 17th Class 7/1/10

at least 50 and preferably 100 observations should be available to build a proper model

Decision 411: Class 4

interval forecasting

FORECASTING FLUCTUATIONS OF ASPHALT CEMENT PRICE INDEX IN GEORGIA

Time Series Analysis -- An Introduction -- AMS 586

Lecture 8: ARIMA Forecasting Please read Chapters 7 and 8 of MWH Book

ECON/FIN 250: Forecasting in Finance and Economics: Section 3.2 Filtering Time Series

STAT 443 Final Exam Review. 1 Basic Definitions. 2 Statistical Tests. L A TEXer: W. Kong

Glossary. The ISI glossary of statistical terms provides definitions in a number of different languages:

data lam=36.9 lam=6.69 lam=4.18 lam=2.92 lam=2.21 time max wavelength modulus of max wavelength cycle

5 Transfer function modelling

INTRODUCTION TO FORECASTING (PART 2) AMAT 167

Foundations - 1. Time-series Analysis, Forecasting. Temporal Information Retrieval

Forecasting: Methods and Applications

Ch. 12: Workload Forecasting

CHLE Tank Test Results for Blended and NEI Fiber Beds with Aluminum Addition: Correlated Control Charts for Head Loss Response to Aluminum Addition

EXPONENTIAL SMOOTHING MODELING AND FORECASTING FOR INCIDENCE OF TUBERCULOSIS IN INDIA

Short-term electricity demand forecasting in the time domain and in the frequency domain

Econometrics of financial markets, -solutions to seminar 1. Problem 1

Forecasting Module 2. Learning Objectives. Trended Data. By Sue B. Schou Phone:

2. An Introduction to Moving Average Models and ARMA Models

Problem Set 2 Solution Sketches Time Series Analysis Spring 2010

CHAPTER 18. Time Series Analysis and Forecasting

Time Series Analysis. Smoothing Time Series. 2) assessment of/accounting for seasonality. 3) assessment of/exploiting "serial correlation"

FinQuiz Notes

Time Series and Forecasting

Forecasting. BUS 735: Business Decision Making and Research. exercises. Assess what we have learned

Notes on Time Series Modeling

Chapter 5: Forecasting

STATS24x7.com 2010 ADI NV, INC.

A new approach to forecasting based on exponential smoothing with independent regressors

Improved Holt Method for Irregular Time Series

RC Circuits. Lecture 13. Chapter 31. Physics II. Course website:

A. Recursively orthogonalized. VARs

Some Time-Series Models

Antti Salonen KPP Le 3: Forecasting KPP227

FORECASTING AND MODEL SELECTION

STAT 115: Introductory Methods for Time Series Analysis and Forecasting. Concepts and Techniques

Time Series in R: Forecasting and Visualisation. Forecast evaluation 29 May 2017

Multivariate forecasting with VAR models

Introduction to ARMA and GARCH processes

Automatic Forecasting

A State Space Framework For Automatic Forecasting Using Exponential Smoothing Methods

Exponentially weighted forecasts

Prof. Dr. Roland Füss Lecture Series in Applied Econometrics Summer Term Introduction to Time Series Analysis

Time Series Analysis

Chapter 2: simple regression model

Do we need Experts for Time Series Forecasting?

Forecasting Unemployment Rates in the UK and EU

Lecture 3: Statistical sampling uncertainty

IDENTIFICATION OF ARMA MODELS

WORD: EXAMPLE(S): COUNTEREXAMPLE(S): EXAMPLE(S): COUNTEREXAMPLE(S): WORD: EXAMPLE(S): COUNTEREXAMPLE(S): EXAMPLE(S): COUNTEREXAMPLE(S): WORD:

AUTO SALES FORECASTING FOR PRODUCTION PLANNING AT FORD

BUSI 460 Suggested Answers to Selected Review and Discussion Questions Lesson 7

Stat 248 Lab 2: Stationarity, More EDA, Basic TS Models

Time Series and Forecasting

Forecasting of Electric Consumption in a Semiconductor Plant using Time Series Methods

Stat 5100 Handout #12.e Notes: ARIMA Models (Unit 7) Key here: after stationary, identify dependence structure (and use for forecasting)

Time Series Forecasting Methods:

Characteristics of Time Series

Statistics 910, #5 1. Regression Methods

Decision 411: Class 9. HW#3 issues

Daily sales forecasting in foodservice: Developing a model for application in an expert system

The data was collected from the website and then converted to a time-series indexed from 1 to 86.

Suan Sunandha Rajabhat University

NATCOR. Forecast Evaluation. Forecasting with ARIMA models. Nikolaos Kourentzes

Basics: Definitions and Notation. Stationarity. A More Formal Definition

The Kalman Filter. Data Assimilation & Inverse Problems from Weather Forecasting to Neuroscience. Sarah Dance

Chapter 3: Regression Methods for Trends

Forecasting Automobile Sales using an Ensemble of Methods

Forecasting. Operations Analysis and Improvement Spring

Antti Salonen PPU Le 2: Forecasting 1

TMA4285 December 2015 Time series models, solution.

Vector autoregressions, VAR

A state space model for exponential smoothing with group seasonality

Quantitative Trendspotting. Rex Yuxing Du and Wagner A. Kamakura. Web Appendix A Inferring and Projecting the Latent Dynamic Factors

PPU411 Antti Salonen. Forecasting. Forecasting PPU Forecasts are critical inputs to business plans, annual plans, and budgets

Firstly, the dataset is cleaned and the years and months are separated to provide better distinction (sample below).

Transcription:

Lecture 7: Exponential Smoothing Methods Please read Chapter 4 and Chapter 2 of MWH Book 1

Big Picture 1. In lecture 6, smoothing (averaging) method is used to estimate the trend-cycle (decomposition) 2. Now, modified smoothing method is used to forecast future values. That means, in general, the averaging is one-sided, as opposed to two-sided 3. Another difference is, we focus on out-of-sample forecasting errors, other than the in-sample fitting errors (residuals) 4. Again, which method works best depends on whether trend or seasonality is present 2

Forecasting Scenario 1. The present is the t-th period 2. Past (in-sample) observations Y t,y t 1,...,Y 1 are used to estimate the model 3. Then forecasts F t+1,f t+2,... are computed for the future (out-of-sample) values Y t+1,y t+2,... 4. Do not confuse the out-of-sample forecasting errors Y t+1 F t+1,y t+2 F t+2,... with the in-sample fitting errors (residuals) Y t F t,...,y 1 F 1 3

Averaging Method Mean This method uses all past data available F t+1 = 1 t F t+2 = t Y i (1) i=1 t+1 1 t + 1 i=1 Y i (2) The mean method works well only when there is no trend or seasonality or change in the mean value. In other words, it assumes the underlying process has a constant mean (stationarity). 4

Averaging Method Moving Average Moving average, to some degree, can allow for time-varying mean value. It does so by including only the most recent k observations. More explicitly, the MA(k) forecast (not smoother) is F t+1 = 1 k F t+2 = 1 k t Y i (3) i=t k+1 t+1 Y i (4) i=t k+2 It follows that F t+2 = F t+1 + 1 k (Y t+1 Y t k+1 ) (5) So each new forecast F t+2 is an adjustment of the immediately preceding forecast F t+1. The adjustment term is 1 k (Y t+1 Y t k+1 ). Bigger k leads to smaller adjustment (or smoother forecasts). Figure 4-5 in the book clearly shows that the mean method cannot capture the change in the mean value, but the moving average can, though with some lags. 5

Exponential Smoothing Method Single Exponential Smoothing (SES) I We hope the forecast can catch up with the change in mean value more quickly. One way is to assign a bigger weight to the latest observation. Consider F t+1 = F t + α(y t F t ) (6) F t+2 = F t+1 + α(y t+1 F t+1 ) (7) F 1 = Y 1 (initialization) (8) where 0 < α < 1. There is substantial adjustment when α is close to one. In fact, the single exponential smoothing utilizes the idea of error correction we revise up the forecast, i.e., F t+1 > F t if F t underestimates the true value Y t F t > 0. The error correcting process is also called negative feedback. 6

Single Exponential Smoothing II Equation (6) can be rewritten as F t+1 = (1 α)f t + αy t (9) By using the lag operator, we can show that the single exponential smoothing is equivalent to MA( ) process: F t+1 = α[(1 (1 α)l] 1 Y t = αy t + α(1 α)y t 1 + α(1 α) 2 Y t 2 +... (10) Moreover, notice that the weights α,α(1 α),α(1 α) 2,... decay exponentially toward zero. 7

Mean Square Error (MSE) Different smoothing parameter α yields different forecast. We can evaluate or rank various forecasts using mean square error (MSE), which is the average of squared error (in-sample or out-of-sample) 1 MSE = total count (Y i F i ) 2 i We square the error in order to avoid cancellation. The forecast with the smallest MSE is the best one. This fact suggests that we can obtain the optimal smoothing parameter α by minimizing MSE. 8

Theil s U-Statistic An alternative way to rank forecasts is considering the Theil s U-statistic ( ) n 1 Yt+1 F 2 t+1 t=1 U = Y t where Y t+1 F t+1 Y t n 1 t=1 ( Yt+1 Y t ) 2 (11) Y t is the forecast relative change based on a particular method, while Y t+1 Y t Y t is the forecast relative change based on a naive method F t+1 = Y t. A U-statistic less than one indicates a method better than the naive method. The method with smallest U-statistic is the best one. 9

Autocorrelation of Forecasting Error The forecasting error should be serially uncorrelated if no pattern is left unexploited. So a method that produces a white noise forecasting error is better than the one with serially correlated forecasting error. 10

Single Exponential Smoothing III The equation (6) for SES is recursive. This suggests that we may use loop. The R code below, for example, computes the SES forecast, and in the end, compute the MSE for α = 0.5. alpha = 0.5 f = rep(0, length(y)) f[1] = y[1] tse = 0 for (j in 2:length(y)) { f[j] = f[(j-1)] + alpha*(y[j-1]-f[(j-1)]) tse = tse + } (y[j] - f[j])ˆ2 mse =tse/(length(y)-1) 11

Adaptive-Response-Rate SES (ARRSES) The single SES has three limitations: (i) it is arbitrary regarding which α to use; (2) it employs a unique α to compute all forecasts. By contrast, ARRSES allows the value of α to be determined in a data-driven fashion. That is, α is adjusted automatically if there is change in the pattern of data, which can be signalized by big forecasting error. Intuitively, we prefer bigger α when bigger forecasting error arises: F t+1 = (1 α t )F t + α t Y t (12) α t+1 = A t (13) M t A t = βe t + (1 β)a t 1 (14) M t = β E t + (1 β)m t 1 (15) E t = Y t F t (16) So E t is the forecasting error; M t is a smoothed estimate of absolute forecasting error, and A t is a smoothed estimate of forecasting error. Notice that it is α t+1, not α t, in (13). 12

Holt s Linear Method (HLM) The third limitation of SES is that it cannot catch up with trend quickly enough, while Holt s Linear Method (HLM) is able to account for trend. L t = αy t + (1 α)(l t 1 + b t 1 ) (17) b t = β(l t L t 1 ) + (1 β)b t 1 (18) F t+m = L t + b t m (19) L 1 = Y 1 (20) b 1 = Y 2 Y 1 (21) where 0 < α < 1,0 < β < 1, and m is the number of periods ahead to be forecasted. Here L t denotes the level of the series, and b t the slope of trend. HLM is also called double exponential smoothing. The HLM has the drawback of failing to account for seasonal effect. 13

Holt-Winter s Method (HWM) We use HWM to account for both trend and seasonality. HWM involves three equations: one for level, one for trend; and one for seasonality. L t = α Y t + (1 α)(l t 1 + b t 1 ) S t s (22) b t = β(l t L t 1 ) + (1 β)b t 1 (23) S t = γ Y t + (1 γ)s t s L t (24) F t+m = (L t + b t m)s t s+m (25) where s is the length of seasonality (e.g., the number of months in a year), and the seasonality has a multiplicative form. 14

Prediction Intervals We can use prediction intervals (PI) to illustrate the inherent uncertainty in the point forecast: 95% PI = F t+1 ± 1.96 MSE (26) where F t+1 is the point forecast, and MSE is the mean square error for the method that produces F t+1 15