Parameter estimation: ACVF of AR processes

Similar documents
STA 6857 Estimation ( 3.6)

STAT Financial Time Series

STAT 443 Final Exam Review. 1 Basic Definitions. 2 Statistical Tests. L A TEXer: W. Kong

Introduction to Time Series Analysis. Lecture 11.

Akaike criterion: Kullback-Leibler discrepancy

6.3 Forecasting ARMA processes

Introduction to Time Series Analysis. Lecture 12.

University of Oxford. Statistical Methods Autocorrelation. Identification and Estimation

MAT3379 (Winter 2016)

Contents. 1 Time Series Analysis Introduction Stationary Processes State Space Modesl Stationary Processes 8

MAT 3379 (Winter 2016) FINAL EXAM (PRACTICE)

STAT 720 sp 2019 Lec 06 Karl Gregory 2/15/2019

MAT 3379 (Winter 2016) FINAL EXAM (SOLUTIONS)

Autoregressive Moving Average (ARMA) Models and their Practical Applications

Levinson Durbin Recursions: I

Levinson Durbin Recursions: I

STAT 443 (Winter ) Forecasting

Midterm Suggested Solutions

ITSM-R Reference Manual

Financial Time Series Analysis Week 5

Multivariate Time Series

System Identification

Figure 29: AR model fit into speech sample ah (top), the residual, and the random sample of the model (bottom).

Lecture 1: Stationary Time Series Analysis

Covariances of ARMA Processes

Some Time-Series Models

Next tool is Partial ACF; mathematical tools first. The Multivariate Normal Distribution. e z2 /2. f Z (z) = 1 2π. e z2 i /2

Akaike criterion: Kullback-Leibler discrepancy

BIO5312 Biostatistics Lecture 13: Maximum Likelihood Estimation

Automatic Autocorrelation and Spectral Analysis

Lecture 1: Stationary Time Series Analysis

Estimating AR/MA models

Lecture 1: Fundamental concepts in Time Series Analysis (part 2)

STAT STOCHASTIC PROCESSES. Contents

MEI Exam Review. June 7, 2002

STOR 356: Summary Course Notes

Residual Bootstrap for estimation in autoregressive processes

Marcel Dettling. Applied Time Series Analysis SS 2013 Week 05. ETH Zürich, March 18, Institute for Data Analysis and Process Design

STAD57 Time Series Analysis. Lecture 8

COMPUTER ALGEBRA DERIVATION OF THE BIAS OF LINEAR ESTIMATORS OF AUTOREGRESSIVE MODELS

Econ 623 Econometrics II Topic 2: Stationary Time Series

Stationary Stochastic Time Series Models

Recall that the AR(p) model is defined by the equation

Dynamic models. Dependent data The AR(p) model The MA(q) model Hidden Markov models. 6 Dynamic models

Calculation of ACVF for ARMA Process: I consider causal ARMA(p, q) defined by

6.3.3 Parameter Estimation

ARMA Estimation Recipes

Notes on the Multivariate Normal and Related Topics

Statistics of stochastic processes

Time Series: Theory and Methods

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A.

Problem Selected Scores

Elements of Multivariate Time Series Analysis

CHANGE DETECTION IN TIME SERIES

Nonlinear time series

Univariate Time Series Analysis; ARIMA Models

Review Session: Econometrics - CLEFIN (20192)

Reliability and Risk Analysis. Time Series, Types of Trend Functions and Estimates of Trends

UC San Diego Recent Work

Model Identification for Time Series with Dependent Innovations

Chapter 6: Model Specification for Time Series

Model selection using penalty function criteria

Robust Autocorrelation Estimation

13.2 Example: W, LM and LR Tests

Lecture 2: Univariate Time Series

Introduction to ARMA and GARCH processes

Time Series 2. Robert Almgren. Sept. 21, 2009

Classical Decomposition Model Revisited: I

Stat 248 Lab 2: Stationarity, More EDA, Basic TS Models

Applied time-series analysis

Ph.D. Qualifying Exam Friday Saturday, January 6 7, 2017

3. ARMA Modeling. Now: Important class of stationary processes

TIME SERIES ANALYSIS. Forecasting and Control. Wiley. Fifth Edition GWILYM M. JENKINS GEORGE E. P. BOX GREGORY C. REINSEL GRETA M.

MS&E 226: Small Data. Lecture 11: Maximum likelihood (v2) Ramesh Johari

Econometrics II Heij et al. Chapter 7.1

COMPUTER SESSION 3: ESTIMATION AND FORECASTING.

Time Series I Time Domain Methods

Introduction to Estimation Methods for Time Series models Lecture 2

Problem Set 6 Solution

STOR 356: Summary Course Notes Part III

Econometric Forecasting

REGRESSION WITH SPATIALLY MISALIGNED DATA. Lisa Madsen Oregon State University David Ruppert Cornell University

Time Series Examples Sheet

Generalized Linear Models. Kurt Hornik

Moving Average (MA) representations

Econometrics I: Univariate Time Series Econometrics (1)

Differencing Revisited: I ARIMA(p,d,q) processes predicated on notion of dth order differencing of a time series {X t }: for d = 1 and 2, have X t

Time Series Analysis

3 Theory of stationary random processes

An algorithm for robust fitting of autoregressive models Dimitris N. Politis

Efficient inference for periodic autoregressive coefficients with polynomial spline smoothing approach

18.S096 Problem Set 4 Fall 2013 Time Series Due Date: 10/15/2013

Exam 2. Jeremy Morris. March 23, 2006

Random vectors X 1 X 2. Recall that a random vector X = is made up of, say, k. X k. random variables.

Stationary Time Series

Autoregressive Approximation in Nonstandard Situations: Empirical Evidence

Ch 6. Model Specification. Time Series Analysis

Module 3. Descriptive Time Series Statistics and Introduction to Time Series Models

Lecture 15. Hypothesis testing in the linear model

AR, MA and ARMA models

Transcription:

Parameter estimation: ACVF of AR processes Yule-Walker s for AR processes: a method of moments, i.e. µ = x and choose parameters so that γ(h) = ˆγ(h) (for h small ). 12 novembre 2013 1 / 8

Parameter estimation: ACVF of AR processes Yule-Walker s for AR processes: a method of moments, i.e. µ = x and choose parameters so that γ(h) = ˆγ(h) (for h small ). Multiplying X t = ϕ 1 X t 1 + + ϕ p X t p + Z t by X t h (h = 1... p) and taking expectations, one has Γ p φ = γ p with (Γ p ) i,j = γ(i j), i.e. γ(0) γ(1) γ(2) γ(p 1) γ(1) γ(0) γ(1). γ(1) Γ p =. γ(2) γ(1) γ(0).. γ(2), γ p =........ γ(p) γ(p 1) γ(1) γ(0) and multiplying by X t one obtains σ 2 = γ(0) φ t γ p. 12 novembre 2013 1 / 8

Parameter estimation: Yule-Walker s method ˆϕ 1 ˆϕ 2 Hence estimate ˆφ =. by solving ˆΓ p ˆφ = ˆγ p. (1) ˆϕ p Moreover ˆσ 2 = ˆγ(0) ˆφ t ˆγ p. (2) 12 novembre 2013 2 / 8

Parameter estimation: Yule-Walker s method Remarks ˆϕ 1 ˆϕ 2 Hence estimate ˆφ =. by solving ˆΓ p ˆφ = ˆγ p. (1) ˆϕ p Moreover ˆσ 2 = ˆγ(0) ˆφ t ˆγ p. (2) 12 novembre 2013 2 / 8

Parameter estimation: Yule-Walker s method Remarks ˆϕ 1 ˆϕ 2 Hence estimate ˆφ =. by solving ˆΓ p ˆφ = ˆγ p. (1) ˆϕ p 1 We proved ˆΓ p non-negative definite. Moreover ˆσ 2 = ˆγ(0) ˆφ t ˆγ p. (2) 12 novembre 2013 2 / 8

Parameter estimation: Yule-Walker s method Remarks ˆϕ 1 ˆϕ 2 Hence estimate ˆφ =. by solving ˆΓ p ˆφ = ˆγ p. (1) ˆϕ p Moreover ˆσ 2 = ˆγ(0) ˆφ t ˆγ p. (2) 1 We proved ˆΓ p non-negative definite. If ˆγ(0) > 0 (i.e. data are not identical), is positive definite, hence invertible = ˆφ well-defined. 12 novembre 2013 2 / 8

Parameter estimation: Yule-Walker s method Remarks ˆϕ 1 ˆϕ 2 Hence estimate ˆφ =. by solving ˆΓ p ˆφ = ˆγ p. (1) ˆϕ p Moreover ˆσ 2 = ˆγ(0) ˆφ t ˆγ p. (2) 1 We proved ˆΓ p non-negative definite. If ˆγ(0) > 0 (i.e. data are not identical), is positive definite, hence invertible = ˆφ well-defined. 2 One can divide (1) by γ(0) to obtain ˆR p ˆφ = ˆρ p with (R p ) i,j = ρ(i j), ρ i = ρ(i). 12 novembre 2013 2 / 8

Parameter estimation: Yule-Walker s method Remarks ˆϕ 1 ˆϕ 2 Hence estimate ˆφ =. by solving ˆΓ p ˆφ = ˆγ p. (1) ˆϕ p Moreover ˆσ 2 = ˆγ(0) ˆφ t ˆγ p. (2) 1 We proved ˆΓ p non-negative definite. If ˆγ(0) > 0 (i.e. data are not identical), is positive definite, hence invertible = ˆφ well-defined. 2 One can divide (1) by γ(0) to obtain ˆR p ˆφ = ˆρ p with (R p ) i,j = ρ(i j), ρ i = ρ(i). 3 One can compute φ iteratively, applying Durbin-Levinson to ˆγ. 12 novembre 2013 2 / 8

Asymptotic properties of Yule-Walker Theorem If {X t } is the causal AR process satisfying X t = ϕ 1 X t 1 + + ϕ p X t p + Z t {Z t } IID(0, σ 2 ) then n 1/2 ( ˆφ φ) = N(0, σ 2 Γ 1 p ) and ˆσ2 P σ 2. 12 novembre 2013 3 / 8

Asymptotic properties of Yule-Walker Theorem If {X t } is the causal AR process satisfying X t = ϕ 1 X t 1 + + ϕ p X t p + Z t {Z t } IID(0, σ 2 ) then n 1/2 ( ˆφ φ) = N(0, σ 2 Γ 1 p ) and ˆσ2 P σ 2. An approximate γ-confidence interval for ϕ i, i = 1... p is then ( ˆϕ i ˆσ [(ˆΓ p ) 1] 1/2 ii where P( N(0, 1) z γ ) = γ. z γ, ˆϕ i + ˆσ [(ˆΓ p ) 1] 1/2 n ii z γ n ) 12 novembre 2013 3 / 8

Comparison with linear regression One could write the problem as Y = X φ + Z with X n X n 1 X n p X n 1 Y =. X = X n 2 X n p 1.. and Z IID(0, σ2 ). X p+1 X p X 1 Least square estimation asks for minimizing Y X φ 2 and results into X t Y = n t=p+1 n t=p+1 X t X t 1. X t X t p ˆφ = (X t X ) 1 X t Y. p i nˆγ p (X t X ) ij = t=p i+1 X t X t+i j n(ˆγ) ij. 12 novembre 2013 4 / 8

Variations on Yule-Walker: Burg s algorithm Yule-Walker s method finds ˆφ as the coefficients of the best linear predictor of X p+1 using X 1,..., X p, assuming that the ACF coincides with the sample ACF for i = 1,... p. 12 novembre 2013 5 / 8

Variations on Yule-Walker: Burg s algorithm Yule-Walker s method finds ˆφ as the coefficients of the best linear predictor of X p+1 using X 1,..., X p, assuming that the ACF coincides with the sample ACF for i = 1,... p. Burg s algorithm considers actual forward and backward linear predictors. Define u p (t) (p t < n) as the difference between X n t+p and the best linear prediction using the previous p values, i.e. u p (t) = X n t+p P L(Xn t,...,x n t+p 1 )X n t+p and v p (t) (1 t < n) be the difference between the value of X n t and the best linear prediction using the following p values, i.e. v p (t) = X n t P L(Xn t+1,...,x n t+p )X n t. 12 novembre 2013 5 / 8

Variations on Yule-Walker: Burg s algorithm Yule-Walker s method finds ˆφ as the coefficients of the best linear predictor of X p+1 using X 1,..., X p, assuming that the ACF coincides with the sample ACF for i = 1,... p. Burg s algorithm considers actual forward and backward linear predictors. Define u p (t) (p t < n) as the difference between X n t+p and the best linear prediction using the previous p values, i.e. u p (t) = X n t+p P L(Xn t,...,x n t+p 1 )X n t+p and v p (t) (1 t < n) be the difference between the value of X n t and the best linear prediction using the following p values, i.e. v p (t) = X n t P L(Xn t+1,...,x n t+p )X n t. The coefficients are found by minimizing n 1 t=p (u2 p(t) + v 2 p (t)). Details on the algorithm in an exercise. 12 novembre 2013 5 / 8

Order of the auto-regressive process Assume the true model is AR(p) and we fit a model of order m. 12 novembre 2013 6 / 8

Order of the auto-regressive process Assume the true model is AR(p) and we fit a model of order m. If m < p (i.e. the model is wrong), we should see this from the residuals. p X t ˆX t = X t ϕ j X t j = Z t WN(0, σ 2 ) for t > p. 12 novembre 2013 6 / 8

Order of the auto-regressive process Assume the true model is AR(p) and we fit a model of order m. If m < p (i.e. the model is wrong), we should see this from the residuals. p X t ˆX t = X t ϕ j X t j = Z t WN(0, σ 2 ) for t > p. Ŵ t = X t p ˆϕ j X t j are not exactly WN, but close for n large. 12 novembre 2013 6 / 8

Order of the auto-regressive process Assume the true model is AR(p) and we fit a model of order m. If m < p (i.e. the model is wrong), we should see this from the residuals. p X t ˆX t = X t ϕ j X t j = Z t WN(0, σ 2 ) for t > p. Ŵ t = X t p ˆϕ j X t j are not exactly WN, but close for n large. If m < p and we compute (Ŵ m ) t = X t m ( ˆϕ m ) j X t j, we expect to find them correlated. 12 novembre 2013 6 / 8

Order of the auto-regressive process Assume the true model is AR(p) and we fit a model of order m. If m < p (i.e. the model is wrong), we should see this from the residuals. p X t ˆX t = X t ϕ j X t j = Z t WN(0, σ 2 ) for t > p. Ŵ t = X t p ˆϕ j X t j are not exactly WN, but close for n large. If m < p and we compute (Ŵ m ) t = X t m ( ˆϕ m ) j X t j, we expect to find them correlated. Significant ACF of Ŵ m suggests m < p. 12 novembre 2013 6 / 8

Order of the auto-regressive process Assume the true model is AR(p) and we fit a model of order m. If m < p (i.e. the model is wrong), we should see this from the residuals. p X t ˆX t = X t ϕ j X t j = Z t WN(0, σ 2 ) for t > p. Ŵ t = X t p ˆϕ j X t j are not exactly WN, but close for n large. If m < p and we compute (Ŵ m ) t = X t m ( ˆϕ m ) j X t j, we expect to find them correlated. Significant ACF of Ŵ m suggests m < p. If m p, n 1/2 ( ˆφ m φ) = N(0, σ 2 Γ 1 m ) (Theorem). In particular for m > p, ˆϕ m N(0, 1/n). 12 novembre 2013 6 / 8

Order of the auto-regressive process Assume the true model is AR(p) and we fit a model of order m. If m < p (i.e. the model is wrong), we should see this from the residuals. p X t ˆX t = X t ϕ j X t j = Z t WN(0, σ 2 ) for t > p. Ŵ t = X t p ˆϕ j X t j are not exactly WN, but close for n large. If m < p and we compute (Ŵ m ) t = X t m ( ˆϕ m ) j X t j, we expect to find them correlated. Significant ACF of Ŵ m suggests m < p. If m p, n 1/2 ( ˆφ m φ) = N(0, σ 2 Γ 1 m ) (Theorem). In particular for m > p, ˆϕ m N(0, 1/n). Finding ˆϕ m within ±1.96n 1/2 suggests that the true value of ϕ m may be 0, and to fit a model of lower order. 12 novembre 2013 6 / 8

Maximum likelihood estimation A standard method of parameter estimate is maximum likelihood estimation (MLE). 12 novembre 2013 7 / 8

Maximum likelihood estimation A standard method of parameter estimate is maximum likelihood estimation (MLE). Likelihood of ARMA processes is computed assuming {Z t } IIDN(0, σ 2 ). Then {X t } is also normally distributed. 12 novembre 2013 7 / 8

Maximum likelihood estimation A standard method of parameter estimate is maximum likelihood estimation (MLE). Likelihood of ARMA processes is computed assuming {Z t } IIDN(0, σ 2 ). Then {X t } is also normally distributed. Example of AR(1): X t = µ + ϕ(x t 1 µ) + Z t, ϕ < 1. Then X t N(µ, σ 2 1 ϕ 2 ). 12 novembre 2013 7 / 8

Maximum likelihood estimation A standard method of parameter estimate is maximum likelihood estimation (MLE). Likelihood of ARMA processes is computed assuming {Z t } IIDN(0, σ 2 ). Then {X t } is also normally distributed. Example of AR(1): X t = µ + ϕ(x t 1 µ) + Z t, ϕ < 1. σ Then X t N(µ, 2 ). 1 ϕ 2 Joint distributions can be obtained by conditional distributions: conditional on the value of X t 1, X t N(µ + ϕ(x t 1 µ), σ 2 ). 12 novembre 2013 7 / 8

Maximum likelihood estimation A standard method of parameter estimate is maximum likelihood estimation (MLE). Likelihood of ARMA processes is computed assuming {Z t } IIDN(0, σ 2 ). Then {X t } is also normally distributed. Example of AR(1): X t = µ + ϕ(x t 1 µ) + Z t, ϕ < 1. σ Then X t N(µ, 2 ). 1 ϕ 2 Joint distributions can be obtained by conditional distributions: conditional on the value of X t 1, X t N(µ + ϕ(x t 1 µ), σ 2 ). Hence likelihood of data (x 1, x 2,..., x n ) can be obtained as f (x 1 )f (x 2 x 1 ) f (x n x n 1 ) where f ( ) is the (unconditional) density of X t, while f ( x) is the density of X t conditional on X t 1 = x. 12 novembre 2013 7 / 8

Computing likelihood for AR(1) f (x) = (2π σ2 (x 1 ϕ 2 ) 1/2 µ)2 exp{ 2σ 2 /(1 ϕ 2 ) } ( ) 1 ϕ 2 1/2 = 2πσ 2 exp{ (1 ϕ2 )(x µ) 2 2σ 2 } f (y, x) = (2πσ 2 ) 1/2 (y µ ϕ(x µ))2 exp{ 2σ 2 }. 12 novembre 2013 8 / 8

Computing likelihood for AR(1) f (x) = (2π σ2 (x 1 ϕ 2 ) 1/2 µ)2 exp{ 2σ 2 /(1 ϕ 2 ) } ( ) 1 ϕ 2 1/2 = 2πσ 2 exp{ (1 ϕ2 )(x µ) 2 2σ 2 } f (y, x) = (2πσ 2 ) 1/2 (y µ ϕ(x µ))2 exp{ 2σ 2 }. Parameters of the model are µ, ϕ, σ 2 : L(µ, ϕ, σ 2 ) = (1 ϕ 2 ) 1/2 (2πσ 2 ) n/2 exp{ (1 ϕ2 )(x 1 µ) 2 2σ 2 } n exp{ (x j µ ϕ(x j 1 µ)) 2 2σ 2 } j=2 12 novembre 2013 8 / 8

Computing likelihood for AR(1) f (x) = (2π σ2 (x 1 ϕ 2 ) 1/2 µ)2 exp{ 2σ 2 /(1 ϕ 2 ) } ( ) 1 ϕ 2 1/2 = 2πσ 2 exp{ (1 ϕ2 )(x µ) 2 2σ 2 } f (y, x) = (2πσ 2 ) 1/2 (y µ ϕ(x µ))2 exp{ 2σ 2 }. Parameters of the model are µ, ϕ, σ 2 : L(µ, ϕ, σ 2 ) = (1 ϕ 2 ) 1/2 (2πσ 2 ) n/2 exp{ (1 ϕ2 )(x 1 µ) 2 2σ 2 } n exp{ (x j µ ϕ(x j 1 µ)) 2 2σ 2 } j=2 [ = (1 ϕ2 ) 1/2 1 n (2πσ 2 exp{ ) n/2 2σ 2 (1 ϕ 2 )(x 1 µ) 2 + (x t µ ϕ(x t 1 µ)) ]} 2 t=2 12 novembre 2013 8 / 8