Analyzing Aggregated AR(1) Processes

Similar documents
Economics Department LSE. Econometrics: Timeseries EXERCISE 1: SERIAL CORRELATION (ANALYTICAL)

Unit Root and Cointegration

Chapter 12 - Lecture 2 Inferences about regression coefficient

Lesson 7: Estimation of Autocorrelation and Partial Autocorrela

Problem Set 1 Solution Sketches Time Series Analysis Spring 2010

Nonlinear time series

Pure Random process Pure Random Process or White Noise Process: is a random process {X t, t 0} which has: { σ 2 if k = 0 0 if k 0

Empirical Market Microstructure Analysis (EMMA)

α i ξ t i, where ξ t i.i.d. (0, 1),

STAT Financial Time Series

Review of probability and statistics 1 / 31

On optimal stopping of autoregressive sequences

STAT 443 Final Exam Review. 1 Basic Definitions. 2 Statistical Tests. L A TEXer: W. Kong

Questions and Answers on Unit Roots, Cointegration, VARs and VECMs

Ec402 Econometrics. Suitable for all candidates. Summer 2012 (part of) examination. Instructions to candidates. Time allowed: 3 hours

Journal of Statistical Research 2007, Vol. 41, No. 1, pp Bangladesh

Stat 248 Lab 2: Stationarity, More EDA, Basic TS Models

Multivariate Time Series Analysis and Its Applications [Tsay (2005), chapter 8]

Introduction to ARMA and GARCH processes

Time Series Analysis. Solutions to problems in Chapter 5 IMM

STA205 Probability: Week 8 R. Wolpert

18.S096 Problem Set 4 Fall 2013 Time Series Due Date: 10/15/2013

Bootstrapping high dimensional vector: interplay between dependence and dimensionality

Review Session: Econometrics - CLEFIN (20192)

Introduction to Stochastic processes

Research Article Least Squares Estimators for Unit Root Processes with Locally Stationary Disturbance

Module 3. Descriptive Time Series Statistics and Introduction to Time Series Models

Econ 424 Time Series Concepts

UC San Diego Recent Work

Lecture 3: Autoregressive Moving Average (ARMA) Models and their Practical Applications

Stochastic Processes

Solutions to Homework Set #6 (Prepared by Lele Wang)

Multiresolution Models of Time Series

If we want to analyze experimental or simulated data we might encounter the following tasks:

Semester , Example Exam 1

Cointegration, Stationarity and Error Correction Models.

6.3 Forecasting ARMA processes

ACE 562 Fall Lecture 2: Probability, Random Variables and Distributions. by Professor Scott H. Irwin

ECON3327: Financial Econometrics, Spring 2016

Characteristics of Time Series

Modeling and testing long memory in random fields

Linear Model Under General Variance Structure: Autocorrelation

On the robustness of cointegration tests when series are fractionally integrated

Time Series Analysis -- An Introduction -- AMS 586

Econometría 2: Análisis de series de Tiempo

MAT 3379 (Winter 2016) FINAL EXAM (SOLUTIONS)

Key Words: first-order stationary autoregressive process, autocorrelation, entropy loss function,

A Study on the Correlation of Bivariate And Trivariate Normal Models

Standard Testing Procedures for White Noise and Heteroskedasticity

A time series is called strictly stationary if the joint distribution of every collection (Y t

MAT 3379 (Winter 2016) FINAL EXAM (PRACTICE)

Recursive Estimation of Time-Average Variance Constants Through Prewhitening

F9 F10: Autocorrelation

Financial Econometrics Review Session Notes 3

Multivariate Time Series

STAT Financial Time Series

A TIME SERIES PARADOX: UNIT ROOT TESTS PERFORM POORLY WHEN DATA ARE COINTEGRATED

3. ARMA Modeling. Now: Important class of stationary processes

CHANGE DETECTION IN TIME SERIES

Lecture 1: Fundamental concepts in Time Series Analysis (part 2)

5: MULTIVARATE STATIONARY PROCESSES

FE570 Financial Markets and Trading. Stevens Institute of Technology

Prof. Dr. Roland Füss Lecture Series in Applied Econometrics Summer Term Introduction to Time Series Analysis

7. MULTIVARATE STATIONARY PROCESSES

Inference in VARs with Conditional Heteroskedasticity of Unknown Form

Introduction to Regression Analysis. Dr. Devlina Chatterjee 11 th August, 2017

Ch 5. Models for Nonstationary Time Series. Time Series Analysis

Applied Statistics I

MAT3379 (Winter 2016)

The Role of "Leads" in the Dynamic Title of Cointegrating Regression Models. Author(s) Hayakawa, Kazuhiko; Kurozumi, Eiji

Time Series Analysis. James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY

STAT 804: Lecture 1. Today: Plots of some time series. Discuss series using some of the jargon we will study. Basic classes of models.

Variance reduction. Michel Bierlaire. Transport and Mobility Laboratory. Variance reduction p. 1/18

Next tool is Partial ACF; mathematical tools first. The Multivariate Normal Distribution. e z2 /2. f Z (z) = 1 2π. e z2 i /2

Probability Distributions Columns (a) through (d)

Serial Correlation. Edps/Psych/Stat 587. Carolyn J. Anderson. Fall Department of Educational Psychology

ESTIMATION AND OUTPUT ANALYSIS (L&K Chapters 9, 10) Review performance measures (means, probabilities and quantiles).

STAT 248: EDA & Stationarity Handout 3

γ 0 = Var(X i ) = Var(φ 1 X i 1 +W i ) = φ 2 1γ 0 +σ 2, which implies that we must have φ 1 < 1, and γ 0 = σ2 . 1 φ 2 1 We may also calculate for j 1

Stationary Stochastic Time Series Models

Spring 2012 Math 541B Exam 1

Simple Linear Regression

Introduction to Computational Finance and Financial Econometrics Probability Review - Part 2

Nonlinear Error Correction Model and Multiple-Threshold Cointegration May 23, / 31

The Number of Bootstrap Replicates in Bootstrap Dickey-Fuller Unit Root Tests

Introduction to Signal Processing

Economics 620, Lecture 13: Time Series I

Econ 623 Econometrics II Topic 2: Stationary Time Series

STT 843 Key to Homework 1 Spring 2018

STAT Financial Time Series

Part II. Time Series

1 Presessional Probability

Model Identification for Time Series with Dependent Innovations

Minitab Project Report Assignment 3

Problem Set 2 Solution Sketches Time Series Analysis Spring 2010

Time Series Analysis. James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY

Predictive Regressions: A Reduced-Bias. Estimation Method

Nonstationary Time Series:

A Functional Central Limit Theorem for an ARMA(p, q) Process with Markov Switching

Lecture 25: Review. Statistics 104. April 23, Colin Rundel

Transcription:

Analyzing Aggregated AR(1) Processes Jon Gunnip Supervisory Committee Professor Lajos Horváth (Committee Chair) Professor Davar Koshnevisan Professor Paul Roberts Analyzing Aggregated AR(1) Processes p.1

What is an AR(1) Process? Let with ǫ i be i.i.d. Eǫ = 0, Varǫ = σ 2. For some constant ρ, < ρ <, and for all i Z, let X i = ρx i 1 + ǫ i. This is an autoregressive process of order 1. Analyzing Aggregated AR(1) Processes p.2

AR(1) Process Example ρ =.5, ǫ N(0, 1) Analyzing Aggregated AR(1) Processes p.3

Aggregated AR(1) Processes What if a financial statistic for N companies each followed an AR(1) process? X (j) i = ρ (j) X (j) i 1 + ǫ(j) i, 1 j N, 1 i Analyzing Aggregated AR(1) Processes p.4

Aggregated AR(1) Processes What if a financial statistic for N companies each followed an AR(1) process? X (j) i = ρ (j) X (j) i 1 + ǫ(j) i, 1 j N, 1 i Only summary statistics might be reported: Y i = 1 N N j=1 X(j) i, 1 i n Analyzing Aggregated AR(1) Processes p.4

Aggregated AR(1) Processes What if a financial statistic for N companies each followed an AR(1) process? X (j) i = ρ (j) X (j) i 1 + ǫ(j) i, 1 j N, 1 i Only summary statistics might be reported: Y i = 1 N N j=1 X(j) i, 1 i n Is it plausible to consider Y 1,...,Y n as an AR(1) process? Y i = ρ Y i 1 + ǫ i, 1 i n Analyzing Aggregated AR(1) Processes p.4

Agenda Discuss some elementary facts about AR(1) processes Analyzing Aggregated AR(1) Processes p.5

Agenda Discuss some elementary facts about AR(1) processes Derive an estimator ˆρ for ρ in an AR(1) process and analyze the distribution of n(ˆρ ρ) Analyzing Aggregated AR(1) Processes p.5

Agenda Discuss some elementary facts about AR(1) processes Derive an estimator ˆρ for ρ in an AR(1) process and analyze the distribution of n(ˆρ ρ) Consider aggregations of AR(1) processes where ρ is a random variable and use simulations to test two estimators for Eρ that rely on the aggregated data Analyzing Aggregated AR(1) Processes p.5

Elementary Facts about AR(1) Processes Analyzing Aggregated AR(1) Processes p.6

Stationary, Predictable Solutions A solution to an AR(1) process is weakly stationary if EX i is independent of i and Cov(X i+h,x i ) is independent of i for each integer h Analyzing Aggregated AR(1) Processes p.7

Stationary, Predictable Solutions A solution to an AR(1) process is weakly stationary if EX i is independent of i and Cov(X i+h,x i ) is independent of i for each integer h A solution is predictable if X i is a function of ǫ i,ǫ i 1,... Analyzing Aggregated AR(1) Processes p.7

Solutions to AR(1) Processes An AR(1) process has a unique, stationary, predictable solution if and only if ρ < 1 Analyzing Aggregated AR(1) Processes p.8

Solutions to AR(1) Processes An AR(1) process has a unique, stationary, predictable solution if and only if ρ < 1 Assume ρ < 1. Using X i i = ρx i 2 + ǫ i 1, recursively expand X i = ρx i 1 + ǫ i to get Y i = k=0 ρk ǫ i k = ǫ i + ρǫ i 1 + ρ 2 ǫ i 2 +... as solution Analyzing Aggregated AR(1) Processes p.8

Solutions to AR(1) Processes An AR(1) process has a unique, stationary, predictable solution if and only if ρ < 1 Assume ρ < 1. Using X i i = ρx i 2 + ǫ i 1, recursively expand X i = ρx i 1 + ǫ i to get Y i = k=0 ρk ǫ i k = ǫ i + ρǫ i 1 + ρ 2 ǫ i 2 +... as solution Solution is predictable. It is also defined with probability 1 and that it satisfies X i = ρx i 1 + ǫ i Analyzing Aggregated AR(1) Processes p.8

Solutions to AR(1) Processes (cont d) Mean function is µ Y (i) = ( 0 and covariance function is γ Y (h) = ρ h so solution is stationary ) σ 2 1 ρ 2 Analyzing Aggregated AR(1) Processes p.9

Solutions to AR(1) Processes (cont d) Mean function is µ Y (i) = ( 0 and covariance function is γ Y (h) = ρ h so solution is stationary ) σ 2 1 ρ 2 For ρ > 1, there is a unique, stationary, non-predictable solution Analyzing Aggregated AR(1) Processes p.9

Solutions to AR(1) Processes - Conclusion Since we have a unique, stationary, predictable solution if and only if ρ < 1, we assume ρ < 1 throughout the rest of the presentation Analyzing Aggregated AR(1) Processes p.10

Solutions to AR(1) Processes - Conclusion Since we have a unique, stationary, predictable solution if and only if ρ < 1, we assume ρ < 1 throughout the rest of the presentation Next step is to have a way to estimate ρ given data X 1,...,X n from an AR(1) process Analyzing Aggregated AR(1) Processes p.10

Deriving ˆρ and Analyzing n(ˆρ ρ) Analyzing Aggregated AR(1) Processes p.11

Estimating ρ in AR(1) Process Least squares estimation: using ǫ k = X k ρx k 1 and minimizing n k=2 (X k ρx k 1 ) 2 yields ˆρ = n X k X k 1 k=2 n k=2 X 2 k 1 Analyzing Aggregated AR(1) Processes p.12

Estimating ρ in AR(1) Process Least squares estimation: using ǫ k = X k ρx k 1 and minimizing n k=2 (X k ρx k 1 ) 2 yields ˆρ = n X k X k 1 k=2 n k=2 X 2 k 1 Agrees with maximum liklihood estimator for ǫ N(0,σ 2 ) Analyzing Aggregated AR(1) Processes p.12

Properties of n(ˆρ ρ) By substituting ρx k 1 + ǫ k for X k in ˆρ we derive n n X k 1 ǫ k X k 1 ǫ k ˆρ ρ = k=2 n k=2 X 2 k 1 k=2 nex 2 0 Analyzing Aggregated AR(1) Processes p.13

Properties of n(ˆρ ρ) By substituting ρx k 1 + ǫ k for X k in ˆρ we derive n n X k 1 ǫ k X k 1 ǫ k ˆρ ρ = k=2 n k=2 X 2 k 1 k=2 nex 2 0 Predictability of X k implies EX k 1 ǫ k = 0 Analyzing Aggregated AR(1) Processes p.13

Properties of n(ˆρ ρ) By substituting ρx k 1 + ǫ k for X k in ˆρ we derive n n X k 1 ǫ k X k 1 ǫ k ˆρ ρ = k=2 n k=2 X 2 k 1 k=2 nex 2 0 Predictability of X k implies EX k 1 ǫ k = 0 Thus, E n(ˆρ ρ) 0 Analyzing Aggregated AR(1) Processes p.13

Properties of n(ˆρ ρ) (cont d) Similarly we can show Var n(ˆρ ρ) σ2 EX 2 0 Analyzing Aggregated AR(1) Processes p.14

Properties of n(ˆρ ρ) (cont d) Similarly we can show Var n(ˆρ ρ) σ2 EX 2 0 If n(ˆρ ρ) is normally distributed we would expect it to be approximately N(0, σ 2 EX 2 0 ) Analyzing Aggregated AR(1) Processes p.14

Properties of n(ˆρ ρ) (cont d) Similarly we can show Var n(ˆρ ρ) σ2 EX 2 0 If n(ˆρ ρ) is normally distributed we would expect it to be approximately N(0, σ 2 EX 2 0 ) We examine this proposition through simulations using several combinations of ρ and ǫ Analyzing Aggregated AR(1) Processes p.14

Properties of n(ˆρ ρ) (cont d) ρ =.1, ǫ N(0, 1) Analyzing Aggregated AR(1) Processes p.15

Properties of n(ˆρ ρ) (cont d) ρ =.5, ǫ N(0, 1) Analyzing Aggregated AR(1) Processes p.16

Properties of n(ˆρ ρ) (cont d) ρ =.9, ǫ N(0, 1) Analyzing Aggregated AR(1) Processes p.17

Properties of n(ˆρ ρ) (cont d) ρ =.99, ǫ N(0, 1) Analyzing Aggregated AR(1) Processes p.18

Properties of n(ˆρ ρ) (cont d) ρ =.5, ǫ DE(1, 0) f(x) = 1 2 e x Analyzing Aggregated AR(1) Processes p.19

Properties of n(ˆρ ρ) (cont d) ρ =.5, ǫ CAU(1, 0) f(x) = 1 π(1+x 2 ) Analyzing Aggregated AR(1) Processes p.20

Properties of n(ˆρ ρ) - Conclusion When ǫ is distributed as N(0, 1) or DE(1, 0), n(ˆρ ρ) is distributed approximately as N(0, σ 2 EX 2 0 ) Analyzing Aggregated AR(1) Processes p.21

Properties of n(ˆρ ρ) - Conclusion When ǫ is distributed as N(0, 1) or DE(1, 0), n(ˆρ ρ) is distributed approximately as N(0, σ 2 EX 2 0 ) We can create confidence intervals or do hypothesis testing for ρ under these circumstances Analyzing Aggregated AR(1) Processes p.21

Properties of n(ˆρ ρ) - Conclusion When ǫ is distributed as N(0, 1) or DE(1, 0), n(ˆρ ρ) is distributed approximately as N(0, σ 2 EX 2 0 ) We can create confidence intervals or do hypothesis testing for ρ under these circumstances We will use ǫ distributed as N(0, 1) and DE(1, 0) for our simulations of aggregrated AR(1) processes Analyzing Aggregated AR(1) Processes p.21

Aggregated AR(1) Processes Analyzing Aggregated AR(1) Processes p.22

Aggregated AR(1) Processes Assume a financial statistic for each of N companies follows an AR(1) process with ρ (j) randomly distributed X (j) i = ρ (j) X (j) i 1 + ǫ(j) i, 1 j N, 1 i Analyzing Aggregated AR(1) Processes p.23

Aggregated AR(1) Processes Assume a financial statistic for each of N companies follows an AR(1) process with ρ (j) randomly distributed X (j) i = ρ (j) X (j) i 1 + ǫ(j) i, 1 j N, 1 i We only have summary statistics and we want to estimate Eρ Y i = 1 N N j=1 X (j) i, 1 i n Analyzing Aggregated AR(1) Processes p.23

Using ˆρ LSE to estimate Eρ If we consider Y 1,...,Y n as an AR(1) process, Y i = ρ Y i 1 + ǫ i, 1 i n, our initial estimator for Eρ would be the least squares estimator ˆρ LSE = n i=2 n i=2 Y i Y i 1 Y 2 i 1 Analyzing Aggregated AR(1) Processes p.24

Using ˆρ LSE to est. Eρ - Granger and Morris Important results in the aggregation of stationary time series were achieved by Granger and Morris (1976). In particular, they showed that the sum of two AR(1) processes is not an AR(1) process, but is a more complex autoregressive moving average (ARMA) process. Analyzing Aggregated AR(1) Processes p.25

Using ˆρ LSE to est. Eρ - Horváth and Leipus Results of Horváth and Leipus (2005) Analyzing Aggregated AR(1) Processes p.26

Using ˆρ LSE to est. Eρ - Horváth and Leipus Results of Horváth and Leipus (2005) If P(0 < ρ < 1) = 1 or P( 1 < ρ < 0) = 1, ˆρ LSE converges in probability as the number of companies or time periods increase but the limit is not Eρ. Analyzing Aggregated AR(1) Processes p.26

Using ˆρ LSE to est. Eρ - Horváth and Leipus Results of Horváth and Leipus (2005) If P(0 < ρ < 1) = 1 or P( 1 < ρ < 0) = 1, ˆρ LSE converges in probability as the number of companies or time periods increase but the limit is not Eρ. If the distribution of ρ is symmetric around 0, ˆρ LSE converges to Eρ in probability as the number of companies or time periods goes to infinity. Analyzing Aggregated AR(1) Processes p.26

Using ˆρ LSE to estimate Eρ (cont d) We run simulations to test ˆρ LSE for ρ U(0, 1),ρ U( 1, 1) ǫ N(0, 1),ǫ DE(1, 0) N {100, 200, 300} n {10, 20, 30, 40, 50} Analyzing Aggregated AR(1) Processes p.27

Simulation Results for ρ U(0, 1) Analyzing Aggregated AR(1) Processes p.28

Simulation Results for ρ U( 1, 1) Analyzing Aggregated AR(1) Processes p.29

Using ˆρ MLSE to estimate Eρ Horváth and Leipus suggest the following modification of ˆρ LSE which should converge to Eρ for ρ U(0, 1) ˆρ MLSE = n i=2 Y i Y i 1 n i=1 Y 2 i n i=4 n i=3 Y i Y i 3 Y i Y i 2 Analyzing Aggregated AR(1) Processes p.30

Using ˆρ MLSE to estimate Eρ Horváth and Leipus suggest the following modification of ˆρ LSE which should converge to Eρ for ρ U(0, 1) ˆρ MLSE = n i=2 Y i Y i 1 n i=1 Y 2 i n i=4 n i=3 Y i Y i 3 Y i Y i 2 We repeat previous simulations using ˆρ MLSE instead of ˆρ LSE Analyzing Aggregated AR(1) Processes p.30

Simulation Results for ρ U(0, 1) Analyzing Aggregated AR(1) Processes p.31

Simulation Results for ρ U( 1, 1) Analyzing Aggregated AR(1) Processes p.32

Aggregated AR(1) Processes - Final Thoughts Did not fully replicate results of Horváth and Leipus with ˆρ MLSE for ρ U(0, 1) Analyzing Aggregated AR(1) Processes p.33

Aggregated AR(1) Processes - Final Thoughts Did not fully replicate results of Horváth and Leipus with ˆρ MLSE for ρ U(0, 1) Examined limited range of values for the number of companies and time periods Analyzing Aggregated AR(1) Processes p.33

Aggregated AR(1) Processes - Final Thoughts Did not fully replicate results of Horváth and Leipus with ˆρ MLSE for ρ U(0, 1) Examined limited range of values for the number of companies and time periods The AR(1) processes for each company were started with an initial value of 0 Resulted in non-stationary process? Throwing away a number of initial values in the series might have overcome this issue Analyzing Aggregated AR(1) Processes p.33

Questions? Analyzing Aggregated AR(1) Processes p.34

Thank You! Analyzing Aggregated AR(1) Processes p.35