Estimating AR/MA models

Similar documents
University of Oxford. Statistical Methods Autocorrelation. Identification and Estimation

Empirical Market Microstructure Analysis (EMMA)

Lab: Box-Jenkins Methodology - US Wholesale Price Indicator

Chapter 8: Model Diagnostics

TIME SERIES ANALYSIS AND FORECASTING USING THE STATISTICAL MODEL ARIMA

Lesson 13: Box-Jenkins Modeling Strategy for building ARMA models

STAT Financial Time Series

Lecture 2: Univariate Time Series

Univariate ARIMA Models

Review Session: Econometrics - CLEFIN (20192)

Non-Stationary Time Series and Unit Root Testing

Econ 623 Econometrics II Topic 2: Stationary Time Series

Stationary Stochastic Time Series Models

Non-Stationary Time Series and Unit Root Testing

Autoregressive Moving Average (ARMA) Models and their Practical Applications

Forecasting using R. Rob J Hyndman. 2.4 Non-seasonal ARIMA models. Forecasting using R 1

Univariate linear models

Ch 8. MODEL DIAGNOSTICS. Time Series Analysis

4. MA(2) +drift: y t = µ + ɛ t + θ 1 ɛ t 1 + θ 2 ɛ t 2. Mean: where θ(l) = 1 + θ 1 L + θ 2 L 2. Therefore,

2. An Introduction to Moving Average Models and ARMA Models

A SARIMAX coupled modelling applied to individual load curves intraday forecasting

Univariate Time Series Analysis; ARIMA Models

Advanced Econometrics

ECON/FIN 250: Forecasting in Finance and Economics: Section 8: Forecast Examples: Part 1

Non-Stationary Time Series and Unit Root Testing

FE570 Financial Markets and Trading. Stevens Institute of Technology

Midterm Suggested Solutions

Problem Set 2: Box-Jenkins methodology

FORECASTING SUGARCANE PRODUCTION IN INDIA WITH ARIMA MODEL

Chapter 6: Model Specification for Time Series

Parameter estimation: ACVF of AR processes

Econometrics I: Univariate Time Series Econometrics (1)

MCMC analysis of classical time series algorithms.

EASTERN MEDITERRANEAN UNIVERSITY ECON 604, FALL 2007 DEPARTMENT OF ECONOMICS MEHMET BALCILAR ARIMA MODELS: IDENTIFICATION

Prof. Dr. Roland Füss Lecture Series in Applied Econometrics Summer Term Introduction to Time Series Analysis

G. S. Maddala Kajal Lahiri. WILEY A John Wiley and Sons, Ltd., Publication

Chapter 12: An introduction to Time Series Analysis. Chapter 12: An introduction to Time Series Analysis

CHAPTER 8 FORECASTING PRACTICE I

Suan Sunandha Rajabhat University

Estimation and application of best ARIMA model for forecasting the uranium price.

Forecasting. Simon Shaw 2005/06 Semester II

at least 50 and preferably 100 observations should be available to build a proper model

Lecture 1: Stationary Time Series Analysis

Introduction to Estimation Methods for Time Series models Lecture 2

Ch 6. Model Specification. Time Series Analysis

Module 3. Descriptive Time Series Statistics and Introduction to Time Series Models

BIO5312 Biostatistics Lecture 13: Maximum Likelihood Estimation

Lecture 4a: ARMA Model

Firstly, the dataset is cleaned and the years and months are separated to provide better distinction (sample below).

Applied time-series analysis

MODELING INFLATION RATES IN NIGERIA: BOX-JENKINS APPROACH. I. U. Moffat and A. E. David Department of Mathematics & Statistics, University of Uyo, Uyo

LATVIAN GDP: TIME SERIES FORECASTING USING VECTOR AUTO REGRESSION

AR, MA and ARMA models

Statistics 203: Introduction to Regression and Analysis of Variance Course review

Appendix 1 Model Selection: GARCH Models. Parameter estimates and summary statistics for models of the form: 1 if ɛt i < 0 0 otherwise

Box-Jenkins ARIMA Advanced Time Series

Elements of Multivariate Time Series Analysis

Basics: Definitions and Notation. Stationarity. A More Formal Definition

ARIMA Models. Jamie Monogan. January 16, University of Georgia. Jamie Monogan (UGA) ARIMA Models January 16, / 27

Modelling using ARMA processes

ARIMA Models. Richard G. Pierse

Forecasting Area, Production and Yield of Cotton in India using ARIMA Model

ARIMA Models. Jamie Monogan. January 25, University of Georgia. Jamie Monogan (UGA) ARIMA Models January 25, / 38

Lecture 7: Model Building Bus 41910, Time Series Analysis, Mr. R. Tsay

Economics 618B: Time Series Analysis Department of Economics State University of New York at Binghamton

Stat 5100 Handout #12.e Notes: ARIMA Models (Unit 7) Key here: after stationary, identify dependence structure (and use for forecasting)

MISCELLANEOUS TOPICS RELATED TO LIKELIHOOD. Copyright c 2012 (Iowa State University) Statistics / 30

Lecture 16: State Space Model and Kalman Filter Bus 41910, Time Series Analysis, Mr. R. Tsay

Chapter 3: Regression Methods for Trends

{ } Stochastic processes. Models for time series. Specification of a process. Specification of a process. , X t3. ,...X tn }

Dynamic Time Series Regression: A Panacea for Spurious Correlations

Lecture 1: Fundamental concepts in Time Series Analysis (part 2)

Lesson 2: Analysis of time series

5 Autoregressive-Moving-Average Modeling

Autoregressive and Moving-Average Models

Problem Set 2 Solution Sketches Time Series Analysis Spring 2010

Arma-Arch Modeling Of The Returns Of First Bank Of Nigeria

Lecture 3: Autoregressive Moving Average (ARMA) Models and their Practical Applications

Akaike criterion: Kullback-Leibler discrepancy

APPLIED ECONOMETRIC TIME SERIES 4TH EDITION

Introduction to Eco n o m et rics

ECONOMETRIA II. CURSO 2009/2010 LAB # 3

Stochastic Modelling Solutions to Exercises on Time Series

Model selection using penalty function criteria

Introduction to ARMA and GARCH processes

Lecture 1: Stationary Time Series Analysis

A Data-Driven Model for Software Reliability Prediction

Minitab Project Report - Assignment 6

9. Model Selection. statistical models. overview of model selection. information criteria. goodness-of-fit measures

Time Series Forecasting: A Tool for Out - Sample Model Selection and Evaluation

Time Series Analysis -- An Introduction -- AMS 586

ECON/FIN 250: Forecasting in Finance and Economics: Section 6: Standard Univariate Models

AR(p) + I(d) + MA(q) = ARIMA(p, d, q)

Seasonal Autoregressive Integrated Moving Average Model for Precipitation Time Series

MAT3379 (Winter 2016)

TMA4285 December 2015 Time series models, solution.

IDENTIFICATION OF ARMA MODELS

Using Analysis of Time Series to Forecast numbers of The Patients with Malignant Tumors in Anbar Provinc

Empirical Approach to Modelling and Forecasting Inflation in Ghana

Note: The primary reference for these notes is Enders (2004). An alternative and more technical treatment can be found in Hamilton (1994).

Transcription:

September 17, 2009

Goals The likelihood estimation of AR/MA models AR(1) MA(1) Inference Model specification for a given dataset

Why MLE? Traditional linear statistics is one methodology of estimating models standing on a set of assumptions that are rigidly defined. This yields a relative fixed set of models which can be estimated. One such assumption is the independence of the error term. Maximum Likelihood Estimation (MLE) appears a more complicated way of coming to the same answer, when looking for simple moment estimators (e.g. sample mean) or classical least squares. However, MLE permits us to go beyond simple problems. It offers a more generic way to deal with models of stochastic time series processes.

The likelihood approach For any model: y = f (x; θ), MLE involves: setting up the joint likelihood of observing the data finding the θ that maximises the likelihood of the data In non time-series problems, assume independence of y 1, y 2,..., y N L = f (y 1, y 2,..., y N θ) = f (y 1 θ).f (y 2 θ).....f (y N θ) In time series-problems, there is dependence in x 1, x 2,..., x T L = f (y 1, y 2,..., y N φ) = f (y 1 φ).f (y 2 y 1, φ).f (y 3 y 2, y 1, φ).....f (y N y N 1,..., y 1, φ) Here we need to use the joint probability of conditional probabilities.

MLE setup for AR(1) estimation

The model The AR(1) process is Y t = c + φy t 1 + ɛ t where ɛ t i.i.d.n(0, σ 2 ) We know E(Y t ) = µ = c/(1 φ) and E(Y t µ) 2 = σ 2 /(1 φ 2 ) Now we need to setup the Likelihood of the data set: Y 1, Y 2,..., Y T

Likelihood for AR(1) Probability of the 1st observation is: f (y 1 ; θ) = f (y 1 ; c, φ, σ 2 ) ( ) 1 1 = 2π σ 2 /(1 φ 2 ) exp {y 1 (c/(1 φ))} 2 2σ 2 /(1 φ 2 )

The second observation Y 2 = c + φy 1 + ɛ 2 Conditioning on Y 1, i.e. treating Y 1 as a constant y 1, Y 2 (Y 1 = y 1 ) N(c + φy 1, σ 2 ) Conditional mean of Y 2 = c + φy 1 Conditional variance of Y 2 = E(Y 2 E(Y 2 )) 2 = E(ɛ 2 ) 2 = σ 2. Conditional density of Y 2 is: [ 1 f (Y 2 Y 1 ) = exp (Y2 c φy 1 ) 2 ] 2πσ 2 2σ 2 = ] 1 [ ɛ 2 exp 2 2πσ 2 2σ 2

Putting it together The joint of 1 and 2 is the product of these two elements: f Y1,Y 2 (y 1, y 2 ; θ) = f Y1 (y 1 ; θ)f Y2 Y 1 (y 2 y 1 ; θ) The conditional for observation 3 is [ 1 f (Y 3 Y 2 ) = exp (y3 c φy 2 ) 2 ] 2πσ 2 2σ 2 In this fashion we can setup all the conditionals, and multiply them together to get the joint.

Final Log likelihood The objective function would be to maximise L or minimise log L: logl = T 1 2 2πσ2 T t=2 ɛ 2 t σ 2 πσ2 (1 φ 2 ) (y 1 c 1 φ )2 2σ 2 (1 φ 2 )

Exact vs. Conditional likelihood The above strategy yields the exact MLE : This is because L includes the probability of the first observation, y 1. Suppose we just ignore observation 1. Then all other observations have an identical and familiar form it s just an sum of squared errors, SSE. This becomes equivalent to running OLS on the dataset, with Y t as the LHS and the lagged values Y t 1 as the RHS in the equation. When the probability of the first observation in an AR(1) model is not included, the MLE is called the conditional MLE.

Conditional likelihood for AR(1) It is the same as earlier, except for the f (Y 1 θ) term. logl = (T 1)πσ 2 T t=2 ɛ 2 t σ 2 When T is very large, the exact and the conditional MLE estimates have the same distribution. This is true when the series is stationary. When the series is non-stationary, φ > 1, the conditional MLE gives consistent estimates. But the exact MLE does not. Thus, for most AR estimations, OLS is used to estimate the parameters of the model.

MLE setup for MA(1) estimation

The MA(1) model The model Y t = µ + ɛ t + θɛ t 1 ɛ t iidn(0, σ 2 ) In this case, the exact likelihood is harder. So we estimate using a conditional MLE.

Conditional MLE for MA(1) Suppose we knew that ɛ 0 = 0 exactly. Then (Y 1 ɛ 0 = 0) N(µ, σ 2 ) Once Y 1 is observed, we know ɛ 1 = Y 1 µ 1 exactly. Then: f Y2 Y 1,ɛ 0 =0(y 2 y 1, ɛ 0 = 0; θ) = [ 1 exp (y2 µ θɛ 1 ) 2 ] 2πσ 2 2σ 2

Conditional likelihood of MA(1) In this fashion, we can go forward, iterating on ɛ t = y t µ θɛ t 1. This gives us L(θ) = log f YT,Y T 1,...,Y 1 ɛ 0 =0(y T, y T 1,..., y 1 ɛ 0 = 0; θ) = T 2 log(2π) T T ɛ 2 2 log(σ2 t ) 2σ 2 L is different here from the AR(1) process: we need to calculate the L by an iterative process. Here, OLS cannot be applied to estimate an MA(1) model. t=1

Summarising MLE for ARMA models There are two likelihood functions that can be used for the maximisation of the MLE: Exact MLE: where the probabilities of the first p observations of an AR(p) model or the first q observations of an MA(q) model are explicitly included. Conditional MLE: These are assumed to be known with certainty and are included as inputs in the estimation. An AR process can be estimated using OLS under the conditional MLE setup. All MA processes have to be estimated using MLE.

Inference

Inference for MLE parameters Inference of the estimated model parameters is based on the observed Fischer information. var(ˆθ mle ) = 1 T I 1 I is the informaiton matrix and can be estimated either as: 1 The second derivative estimate: 2 The first derivative estimate: Î = T 1 frac 2 L(θ) θ θ Î = T 1 T [ t=1 log L(θ) θ log L(θ) θ ] Both estimated at θ = ˆθ If T is large enough, then a standard t test can be performed using ˆθ mle and var(ˆθ mle ).

Time series model specification

Specification = formulation + selection Formulation is done based on a mixture of prior, theoretical knowledge about the problem and diagnostic, exploratory tests of the data. Selection is based on estimation and hypothesis tests. The Box Jenkins methodology of forecasting: seperate the identification of the model from the estimation of the model.

Formulating a model Examples of prior knowledge driving model formulation: Monthly series of agricultural produce will have a seasonal behaviour for the kharif and rabi crop. Daily data on the call money rates will have a fortnightly pattern because of banks having report their capital requirements to the central bank every fortnight. The time series of prices of a single futures contract will have a steadily decreasing trend as the contract comes close to expiration. Diagnostic tests: the Box Jenkins methodology of a priori identification.

The Box Jenkins identification methodology Graphs of the raw data: To pick out the possible existence of a trend, seasonality, etc. These are only indicative. The question of how the seasonality or trend affects the time series dyanamics whether as an additive component of f (t), or as part of the polynomial structure of g(l) depends upon more rigorous tests. ACFs, PACFs: More subjective measures of whether there is a stochastic trend, or a seasonal pattern. A plot of the autocorrelation function is also useful to detect the manner of time dependance whether it is an AR, or MA, or a mixed ARMA process, and how many lags are likely to be required to describe the DGP.

Statistical inference for the ACF, PACF The statistical significance of each correlation coefficient is tested as a t-test, where the σ of the coefficient is given by Bartlett(1928). Typically, we test against the null of white noise, φ = θ = 0. Here, the Bartlett s formula approximates to var(ˆρ k ) = 1/T Another test is the Portmanteau test of significance of the sum of a set of k autocorrelation coefficients, Q k. Q k = T k i=1 ˆρ 2 k χ 2 (k p q)

Problems of underestimating the significance of ˆρ k We neither know the true model nor the true model parameters. In this case, our bound is typically an over estimate of the true σ. For example, an AR(1) model will have var(ˆρ 1 ) = φ 2 /T If the model is stationary, then 1 < φ < 1, and φ 2 /T << 1/T. Therefore, we end up underestimating the presence of temporal dependance when using var(ˆρ k ) = 1/T.

Model estimation Once the form and the order of the temporal dependance has been approximately identified, the outcome is a set of possible ARMA models that should be estimated. Estimation is done using MLE/OLS depending upon whether it has an MA term or not.

Tests for model selection We use one of the standard MLE tests to do a first brush selection of a model. The standard tests are: 1 Likelihood Ratio (LR) 2 Wald 3 Lagrange Multiplier Tests that incorporate a penality for over-parameterisation are: 1 Akaike Information Criteria (AIC): (2 log L)/T + (k 2)/T 2 Schwarz Bayes Criteria (SBC): (2 log(l))/t + (k log T )/T where k is the number of parameters in the model, and T is the number of obsrevations.

Tests for model selection These tests are superior to simple hypothesis testing for parameters because: They give numerical values. In some cases, they can be used to compare non nested models. With these tests, one model is being tested against the other, whereas hypothesis testing requires a null of a true model.

Box Jenkin s a posteriori identification The last stage of the modelling process is checking whether the model chosen using the processes listed above is a suitable approximation to the true DGP or not. 1 A model must be consistent with the prior/theoretical knowledge and properties of the data. 2 Apply in sample checks, residual analysis: Use the model to calculate the residuals, and analyse the properties of the residuals for consistency with prior assumptions/knowledge. 3 Apply out of sample checks, forecast bias: The dataset used for estimation must be a subset of the total dataset. Once the model is estimated, it can be used for forecasting future values the data not used in estimation should be used to check the quality of the forecasts from the model.