A Beginner s Introduction. Box-Jenkins Models

Similar documents
Lecture 3: Autoregressive Moving Average (ARMA) Models and their Practical Applications

Lecture 3: Autoregressive Moving Average (ARMA) Models and their Practical Applications

Lecture 2: Univariate Time Series

2. An Introduction to Moving Average Models and ARMA Models

at least 50 and preferably 100 observations should be available to build a proper model

TIME SERIES ANALYSIS AND FORECASTING USING THE STATISTICAL MODEL ARIMA

Scenario 5: Internet Usage Solution. θ j

BJEST. Function: Usage:

Estimation of Parameters of Multiplicative Seasonal Autoregressive Integrated Moving Average Model Using Multiple Regression

MCMC analysis of classical time series algorithms.

Forecasting locally stationary time series

Suan Sunandha Rajabhat University

Reliability and Risk Analysis. Time Series, Types of Trend Functions and Estimates of Trends

Statistics 349(02) Review Questions

NANYANG TECHNOLOGICAL UNIVERSITY SEMESTER II EXAMINATION MAS451/MTH451 Time Series Analysis TIME ALLOWED: 2 HOURS

Exercises - Time series analysis

APPLIED ECONOMETRIC TIME SERIES 4TH EDITION

Lecture 5: Unit Roots, Cointegration and Error Correction Models The Spurious Regression Problem

FORECASTING SUGARCANE PRODUCTION IN INDIA WITH ARIMA MODEL

Econ 623 Econometrics II Topic 2: Stationary Time Series

AR, MA and ARMA models

COMPUTER SESSION 3: ESTIMATION AND FORECASTING.

A time series is called strictly stationary if the joint distribution of every collection (Y t

Time Series I Time Domain Methods

Autoregressive Moving Average (ARMA) Models and their Practical Applications

Circle the single best answer for each multiple choice question. Your choice should be made clearly.

Theoretical and Simulation-guided Exploration of the AR(1) Model

MODELING INFLATION RATES IN NIGERIA: BOX-JENKINS APPROACH. I. U. Moffat and A. E. David Department of Mathematics & Statistics, University of Uyo, Uyo

A SEASONAL TIME SERIES MODEL FOR NIGERIAN MONTHLY AIR TRAFFIC DATA

Box-Jenkins ARIMA Advanced Time Series

5 Autoregressive-Moving-Average Modeling

Week 9: An Introduction to Time Series

TRANSFER FUNCTION MODEL FOR GLOSS PREDICTION OF COATED ALUMINUM USING THE ARIMA PROCEDURE

10) Time series econometrics

Estimating Markov-switching regression models in Stata

5 Transfer function modelling

Modeling and forecasting global mean temperature time series

Firstly, the dataset is cleaned and the years and months are separated to provide better distinction (sample below).

Econ 424 Time Series Concepts

Univariate ARIMA Models

Empirical Market Microstructure Analysis (EMMA)

IDENTIFICATION OF ARMA MODELS

Some Time-Series Models

Stat 5100 Handout #12.e Notes: ARIMA Models (Unit 7) Key here: after stationary, identify dependence structure (and use for forecasting)

Circle a single answer for each multiple choice question. Your choice should be made clearly.

The ARIMA Procedure: The ARIMA Procedure

Time Series Analysis -- An Introduction -- AMS 586

INTRODUCTION TO TIME SERIES ANALYSIS. The Simple Moving Average Model

Applied Econometrics. Professor Bernard Fingleton

Module 3. Descriptive Time Series Statistics and Introduction to Time Series Models

Midterm Exam 1, section 1. Thursday, September hour, 15 minutes

Lecture 8: ARIMA Forecasting Please read Chapters 7 and 8 of MWH Book

THE SEASONAL UNIT ROOTS IN DEMOGRAPHIC TIME SERIES ANDTHE POSSIBILITIES OF ITS EXACT TESTING

{ } Stochastic processes. Models for time series. Specification of a process. Specification of a process. , X t3. ,...X tn }

TIME SERIES ANALYSIS. Forecasting and Control. Wiley. Fifth Edition GWILYM M. JENKINS GEORGE E. P. BOX GREGORY C. REINSEL GRETA M.

University of Oxford. Statistical Methods Autocorrelation. Identification and Estimation

STAT 520: Forecasting and Time Series. David B. Hitchcock University of South Carolina Department of Statistics

Study on Modeling and Forecasting of the GDP of Manufacturing Industries in Bangladesh

ξ t = Fξ t 1 + v t. then λ is less than unity in absolute value. In the above expression, A denotes the determinant of the matrix, A. 1 y t 1.

7. Forecasting with ARIMA models

Dynamic Time Series Regression: A Panacea for Spurious Correlations

Lecture 19 Box-Jenkins Seasonal Models

1 Linear Difference Equations

Problem Set 2: Box-Jenkins methodology

Basics: Definitions and Notation. Stationarity. A More Formal Definition

LECTURES 2-3 : Stochastic Processes, Autocorrelation function. Stationarity.

Minitab Project Report - Assignment 6

Analysis. Components of a Time Series

Review Session: Econometrics - CLEFIN (20192)

Solar irradiance forecasting for Chulalongkorn University location using time series models

Decision 411: Class 9. HW#3 issues

Week 5 Quantitative Analysis of Financial Markets Characterizing Cycles

TMA4285 December 2015 Time series models, solution.

Decision 411: Class 7

Autoregressive and Moving-Average Models

FE570 Financial Markets and Trading. Stevens Institute of Technology

Econometrics I. Professor William Greene Stern School of Business Department of Economics 25-1/25. Part 25: Time Series

Introduction to ARMA and GARCH processes

Rob J Hyndman. Forecasting using. 3. Autocorrelation and seasonality OTexts.com/fpp/2/ OTexts.com/fpp/6/1. Forecasting using R 1

Forecasting. Al Nosedal University of Toronto. March 8, Al Nosedal University of Toronto Forecasting March 8, / 80

Univariate ARIMA Forecasts (Theory)

Introduction to time series econometrics and VARs. Tom Holden PhD Macroeconomics, Semester 2

Econometrics I: Univariate Time Series Econometrics (1)

Homework 2 Solutions

Chapter 16: Correlation

Problem Set 1 Solution Sketches Time Series Analysis Spring 2010

7 Introduction to Time Series

ARIMA Models. Jamie Monogan. January 16, University of Georgia. Jamie Monogan (UGA) ARIMA Models January 16, / 27

Analysis of Violent Crime in Los Angeles County

Topic 4 Unit Roots. Gerald P. Dwyer. February Clemson University

Lab: Box-Jenkins Methodology - US Wholesale Price Indicator

1 Teaching notes on structural VARs.

Frequency Forecasting using Time Series ARIMA model

Forecasting 1 to h steps ahead using partial least squares

Read Section 1.1, Examples of time series, on pages 1-8. These example introduce the book; you are not tested on them.

Design of Time Series Model for Road Accident Fatal Death in Tamilnadu

Confidence Intervals for One-Way Repeated Measures Contrasts

Technical note on seasonal adjustment for Capital goods imports

Modelling using ARMA processes

Testing for Regime Switching in Singaporean Business Cycles

Transcription:

A Beginner s Introduction To Box-Jenkins Models I. Introduction In their seminal work, Time Series Analysis: Forecasting and Control (1970, Holden Day), Professors Box and Jenkins introduced a set of time series models, later called Box-Jenkins models, that revolutionized the analysis of time series data - observations on a variable observed over time at regular time intervals. An example would be real Gross Domestic Product produced by the Commerce Department of the U.S. Government that is published quarterly. See below for a graph of the real Gross Domestic Product observed from 1947Q1 to 01Q3. In a SAS program titled Real GDP.sas I have estimated a Box-Jenkins model for the quarterly, annualized growth rates of this data for the span of time 1980Q1 to 01Q3. Below you will find the

growth rate of Real GDP plotted for this period. The growth rate at time t is calculated as yy tt = (log(rrrrrrrr tt ) log(rrrrrrrr tt 1 )) 400. This is the quarterly annualized growth rate in percentage terms. Following the traditional approach of Box-Jenkins modeling, a very good model for this data is as follows: yy tt = 1.48 + 0.4yy tt 1 + aa tt (1) Equation (1) is obtained by using the growth data and applying least squares to the general Autoregressive Model of order one model, denoted AR(1), yy tt = 0 + 1 yy tt 1 + aa tt. () The quarterly per annum growth rate observed at time t is represented by yy tt while last period s growth rate is denoted by yy tt 1. Therefore, the least squares estimate of 0, denoted 0, is 1.48. This is the socalled constant term of the model. The least squares estimate of the autoregressive coefficient 1, denoted 1, is 0.4. The residuals of this model are represented by aa tt and represent the lack of perfect

fit of the model to the yy tt observations. The estimated mean of this data is 1.48/(1 0.4) =.56% quarterly per annum growth rate. If one were standing at the time 01Q3 and wanted to forecast 4 future growth rates using the Box-Jenkins approach and model (1) you would get Forecasts for variable grgdp Obs Forecast Std Error 95% Confidence Limits 13.771.770 -.6618 8.041 133.6495 3.0100-3.500 8.5489 134.5979 3.0506-3.381 8.5771 135.576 3.0579-3.417 8.5695 The four-step-ahead forecasts are.77%,.64%,.59%, and.57%..77% is the one-step-ahead (one quarter ahead) forecast while.57% is the four-step-ahead (four quarters ahead) forecast. Notice the pretty wide 95% confidence intervals. Future growth rates could either be positive or negative with the tendency being on the positive growth rate side and the best-guess growth rates being the numbers (point forecasts) listed above. The last available growth rate we had was that of 01Q3, namely, 3.0587%. Therefore, using Box-Jenkins methods the 4 forecasts above were generated by the formula yy tt+h =.56 + (3.0587.56)(0.4) h () where h denotes the forecast horizon and h = 1,, 3, and 4 in the current case. More generally, the forecasting equation for the AR(1) model is yy tt+h = μμ + (yy tt μμ )( 1 ) h (3) where μμ denotes the estimated mean of the yy tt data. Notice that these forecasts revert to the mean (.56) over time. This reversion to the mean is a hallmark characteristic of what we will eventually denote as a stationary Box-Jenkins model. The 95% prediction confidence intervals above are calculated as where yy tt+h ± 1.96ssss(yy tt+h ) (4) ssss(yy tt+h ) = σσ aa (1 + 1 + 1 4 + + 1 (h 1) ). (5) From our computer printout we have as the estimate of the sample variance of the errors, σσ aa = 7.68 and the autoregressive coefficient, 1 = 0.4. Therefore, as the above SAS Proc Arima output indicates ssss(yy tt+1 ) =.770, ssss(yy tt+ ) = 3.0100, ssss(yy tt+3 ) = 3.0506, and ssss(yy tt+4 ) = 3.0579. As we are dealing

with a stationary time series here we not that as h, ssss(yy tt+h ) approaches the unconditional standard deviation of the data, σσ aa /(1 1 ) = 7.68/(1 0.4 ) = 3.056. Of course, if one wants to convert these forecasts into the level of real GDP four periods (quarters) ahead we should use the following formulas: ll tt+1 = 1365.51 (1 + yy tt+1 /400) 0.5 = 13676.09 (01Q4 forecast) ll tt+ = ll tt+1 (1 + yy tt+ /400) 0.5 = 13698.68 (013Q1 forecast) ll tt+3 = ll tt+ (1 + yy tt+3 /400) 0.5 = 1370.87 (013Q forecast) ll tt+4 = ll tt+3 (1 + yy tt+4 /400) 0.5 = 1374.91 (013Q3 forecast) Here, 1365.51 is the last available real GDP figure as of 01Q3. So what is the logic of all of this number crunching? To get a sense of how Box-Jenkins modeling works we need to derive some results for the AR(1)model of equation (). II. The AR(1) Box-Jenkins Model To get a beginning understanding of Box-Jenkins models let us consider the AR(1) model of equation () but, for now, we will let 0 = 0. Thus, the model we are considering is yy tt = 1 yy tt 1 + aa tt. (6) The error terms of this model, aa tt, are assumed to have the following White Noise properties: EE(aa tt ) = 0 for all t (zero mean assumption) (6i) EE(aa tt ) = σσ tt for all t (constant variance assumption) (6ii) EE(aa ss aa tt ) = 0 for all s t (independence assumption) (6iii) aa tt is normally distributed for all t (Normality assumption) (6iv) In order for model (6) to have a constant mean, variance, and covariances we require that 1 < 1. The so-called moving average representation of model (6) is derived by recursive substitution as follows: yy tt = 1 ( 1 yy tt 1 + aa tt 1 ) + aa tt = 1 yy tt 1 + 1 aa tt 1 + aa tt. Substituting s times produces yy tt = ss+1 1 yy tt (ss+1) + aa tt + 1 aa tt 1 + 1 aa tt + + ss 1 aa tt ss.

Letting s = t 1 produces yy tt = 1 tt yy 0 + aa tt + 1 aa tt 1 + 1 aa tt + + 1 tt 1 aa 1 (7) where yy 0 denotes the initial value of the time series. Obviously, as time (t) goes to infinity, the initial value, yy 0, has a negligible effect on the series since 1 < 1. Therefore, without loss of generality, we can just set yy 0 = 0 and letting t be large we have the basic description of yy tt as yy tt = aa tt + 1 aa tt 1 + 1 aa tt + (8) Now we are, in turn, going to derive the mean, variance, covariances, and autocorrelations of the AR(1) process (6) above. The Mean Using the linearity of the expectation operator E(.) we have EE(yy tt ) = EE(aa tt + 1 aa tt 1 + 1 aa tt + + 1 tt 1 aa 1 ) = EE(aa tt ) + 1 EE(aa tt 1 ) + 1 EE(aa tt ) + + 1 tt 1 EE(aa 1 ) = 0 (9) because of the White Noise assumption (6i). Therefore, the mean of model (6) is zero. The Variance Using the White Noise independence property (6iii) it is straight forward to derive the variance of the process (6). Namely, VVVVVV(yy tt ) = EE(yy tt EE(yy tt )) = EE(yy tt ) = EE(aa tt + 1 aa tt 1 + 1 aa tt + ) = EE(aa tt ) + 1 EE(aa tt 1 ) + 4 1 EE(aa tt ) + = σσ aa + 1 σσ aa + 1 4 σσ aa + = σσ aa ( 1 + 1 + 1 4 + ) Then since the expression in the parentheses above represents a geometric progression with radius of 1 < 1, have as the unconditional variance of (6) VVVVVV(yy tt ) = σσ aa ( 1 + 1 + 1 4 + ) = The Auto-Covariance Function σσ aa 1 1. (10) Now let us consider the autocovariance between an observation at time t (yy tt ) and an observation the period before (yy tt 1 ). By definition we have the first-order autocorrelation as γγ 1 = EE yy tt EE(yy tt ) yy tt 1 EE(yy tt 1 ) = EE(yy tt yy tt 1 )

= EE[(aa tt + 1 aa tt 1 + 1 aa tt + )(aa tt 1 + 1 aa tt + 1 aa tt 3 + )] = EE 1 aa tt 1 + 3 1 aa tt + 5 1 aa tt 3 + = 1 EE(aa tt 1 ) + 3 1 EE(aa tt ) + 5 1 EE(aa tt 3 ) + = 1 σσ aa (1 + 1 + 1 4 + ) = 1 σσ aa /(1 1 ). (11) In general the j-th order autocovariance between an observation at time t (yy tt ) and an observation j- periods before (yy tt jj ) is γγ jj = EE yy tt EE(yy tt ) yy tt jj EE(yy tt 1 ) = EE(yy tt yy tt jj ) = EE[(aa tt + 1 aa tt 1 + 1 aa tt + ) aa tt jj + 1 aa tt jj 1 + 1 aa tt jj + ] 1 jj σσ aa (1 + 1 + 1 4 + ) = 1 jj σσ aa /(1 1 ). (1) The Autocorrelation Function The autocorrelation function is just a conversion of the autocovariance function into a measurement of association between time series observations that is not dependent on the scale of measurement of the time series (units, hundreds, thousands, millions, etc.). The autocorrelations between time series observations that are j periods apart is ρρ jj = CCCCCC(yy tt,yy tt jj ) vvvvvv(yy tt )VVVVVV(yy tt jj ) = γγ jj = jj VVVVVV(yy tt ) 1. (13) Consider, for example, when 1 = 0.8 in model (6) the autocorrelation function is

ACF for AR(1) Model with Phi1 = 0.8 Rho(j) 0.9 0.8 0.7 0.6 0.5 0.4 0.3 0. 0.1 0 1 3 4 5 6 7 8 9 10 Lag (j) But, consider when 1 = 0.8 in model (6). The autocorrelation function is ACF for AR(1) Model with Phi1 = -0.8 Rho(j) 0.8 0.6 0.4 0. 0-0. -0.4-0.6-0.8-1 1 3 4 5 6 7 8 9 10 Lag (j)