Nonlinear time series

Similar documents
Applied time-series analysis

STAT 248: EDA & Stationarity Handout 3

Introduction to ARMA and GARCH processes

Econometría 2: Análisis de series de Tiempo

Introduction to Stochastic processes

Some Time-Series Models

TIME SERIES AND FORECASTING. Luca Gambetti UAB, Barcelona GSE Master in Macroeconomic Policy and Financial Markets

Lecture 3: Autoregressive Moving Average (ARMA) Models and their Practical Applications

Econometric Forecasting

3. ARMA Modeling. Now: Important class of stationary processes

Discrete time processes

STAT 443 Final Exam Review. 1 Basic Definitions. 2 Statistical Tests. L A TEXer: W. Kong

LECTURES 2-3 : Stochastic Processes, Autocorrelation function. Stationarity.

Stat 248 Lab 2: Stationarity, More EDA, Basic TS Models

Stochastic processes: basic notions

Stochastic Processes: I. consider bowl of worms model for oscilloscope experiment:

Time Series Models and Inference. James L. Powell Department of Economics University of California, Berkeley

Problem Set 1 Solution Sketches Time Series Analysis Spring 2010

Econ 424 Time Series Concepts

Prof. Dr. Roland Füss Lecture Series in Applied Econometrics Summer Term Introduction to Time Series Analysis

Single Equation Linear GMM with Serially Correlated Moment Conditions

Lesson 7: Estimation of Autocorrelation and Partial Autocorrela

STAT STOCHASTIC PROCESSES. Contents

Time Series 3. Robert Almgren. Sept. 28, 2009

Time Series Analysis -- An Introduction -- AMS 586

Next tool is Partial ACF; mathematical tools first. The Multivariate Normal Distribution. e z2 /2. f Z (z) = 1 2π. e z2 i /2

Autoregressive Moving Average (ARMA) Models and their Practical Applications

MAT 3379 (Winter 2016) FINAL EXAM (SOLUTIONS)

Econometría 2: Análisis de series de Tiempo

Lecture 2: Univariate Time Series

Time Series 2. Robert Almgren. Sept. 21, 2009

Module 3. Descriptive Time Series Statistics and Introduction to Time Series Models

Econometrics I: Univariate Time Series Econometrics (1)

Class 1: Stationary Time Series Analysis

Lecture 1: Fundamental concepts in Time Series Analysis (part 2)

Covariances of ARMA Processes

Stationary Stochastic Time Series Models

Stochastic process for macro

A SARIMAX coupled modelling applied to individual load curves intraday forecasting

FE570 Financial Markets and Trading. Stevens Institute of Technology

Chapter 4: Models for Stationary Time Series

Part III Example Sheet 1 - Solutions YC/Lent 2015 Comments and corrections should be ed to

Ch. 14 Stationary ARMA Process

ECON 616: Lecture 1: Time Series Basics

Permanent Income Hypothesis (PIH) Instructor: Dmytro Hryshko

The autocorrelation and autocovariance functions - helpful tools in the modelling problem

Review Session: Econometrics - CLEFIN (20192)

at least 50 and preferably 100 observations should be available to build a proper model

5: MULTIVARATE STATIONARY PROCESSES

EC402: Serial Correlation. Danny Quah Economics Department, LSE Lent 2015

AR, MA and ARMA models

MAT 3379 (Winter 2016) FINAL EXAM (PRACTICE)

Time Series Analysis. James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY

1 Class Organization. 2 Introduction

1. Fundamental concepts

Covariance Stationary Time Series. Example: Independent White Noise (IWN(0,σ 2 )) Y t = ε t, ε t iid N(0,σ 2 )

Econ 623 Econometrics II Topic 2: Stationary Time Series

Time Series Analysis. James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY

STAT Financial Time Series

Multivariate Time Series

Basic concepts and terminology: AR, MA and ARMA processes

Nonlinear Time Series Modeling

Review of Statistics

6. The econometrics of Financial Markets: Empirical Analysis of Financial Time Series. MA6622, Ernesto Mordecki, CityU, HK, 2006.

Empirical Market Microstructure Analysis (EMMA)

Lesson 9: Autoregressive-Moving Average (ARMA) models

7. MULTIVARATE STATIONARY PROCESSES

Contents. 1 Time Series Analysis Introduction Stationary Processes State Space Modesl Stationary Processes 8

Generalised AR and MA Models and Applications

1 Introduction to Generalized Least Squares

Lecture 2: ARMA(p,q) models (part 2)

Lecture 3 Stationary Processes and the Ergodic LLN (Reference Section 2.2, Hayashi)

Econometrics II Heij et al. Chapter 7.1

γ 0 = Var(X i ) = Var(φ 1 X i 1 +W i ) = φ 2 1γ 0 +σ 2, which implies that we must have φ 1 < 1, and γ 0 = σ2 . 1 φ 2 1 We may also calculate for j 1

Marcel Dettling. Applied Time Series Analysis SS 2013 Week 05. ETH Zürich, March 18, Institute for Data Analysis and Process Design

EASTERN MEDITERRANEAN UNIVERSITY ECON 604, FALL 2007 DEPARTMENT OF ECONOMICS MEHMET BALCILAR ARIMA MODELS: IDENTIFICATION

Long-range dependence

1. Stochastic Processes and Stationarity

A Data-Driven Model for Software Reliability Prediction

Time Series I Time Domain Methods

1 Linear Difference Equations

STA 6857 Autocorrelation and Cross-Correlation & Stationary Time Series ( 1.4, 1.5)

Econometrics of financial markets, -solutions to seminar 1. Problem 1

A time series is called strictly stationary if the joint distribution of every collection (Y t

Time Series Analysis

Chapter 6: Model Specification for Time Series

Circle the single best answer for each multiple choice question. Your choice should be made clearly.

Lecture 3: Autoregressive Moving Average (ARMA) Models and their Practical Applications

Stochastic Processes

ARMA (and ARIMA) models are often expressed in backshift notation.

Statistics of stochastic processes

Note: The primary reference for these notes is Enders (2004). An alternative and more technical treatment can be found in Hamilton (1994).

Module 4. Stationary Time Series Models Part 1 MA Models and Their Properties

STOR 356: Summary Course Notes

ELEG 3143 Probability & Stochastic Process Ch. 6 Stochastic Process

Time Series: Theory and Methods

Ch 4. Models For Stationary Time Series. Time Series Analysis

Identifiability, Invertibility

Time Series Examples Sheet

GARCH Models. Eduardo Rossi University of Pavia. December Rossi GARCH Financial Econometrics / 50

Transcription:

Based on the book by Fan/Yao: Nonlinear Time Series Robert M. Kunst robert.kunst@univie.ac.at University of Vienna and Institute for Advanced Studies Vienna October 27, 2009

Outline Characteristics of Time Series Threshold models ARCH and GARCH models Bilinear models

What is a nonlinear time series? Formal definition: a nonlinear process is any stochastic process that is not linear. To this aim, a linear process must be defined. Realizations of time-series processes are called time series but the word is also often applied to the generating processes. Intuitive definition: nonlinear time series are generated by nonlinear dynamic equations. They display features that cannot be modelled by linear processes: time-changing variance, asymmetric cycles, higher-moment structures, thresholds and breaks.

Definition of a linear process Definition A stochastic process (X t, t Z) is said to be a linear process if for every t Z X t = a j ε t j, where a 0 = 1, (ε t, t Z) is iid with Eε t = 0, Eε 2 t <, and j=0 a j <. j=0

For comparison: the Wold theorem Theorem (Wold s Theorem) Any covariance-stationary process (X t ) has a unique representation as the sum of a purely deterministic component and an infinite sum of white-noise terms, in symbols X t = δ t + a j ε t j, with a 0 = 1, j=0 a2 j <, and the terms ε t defined as the linear innovations X t E (X t H t 1 ), where E denotes the linear expectation or projection on the space H t 1 that is generated by the observations X s, s t 1. j=0

Linear processes and the Wold representation There are many covariance-stationary processes that are not linear: either the innovations are not independent (though white noise) or the absolute coefficients do not converge. If it is just the absolute coefficients, the processes are long memory. Those are not really nonlinear, and we will not handle them here. Many nonlinear processes have a Wold representation that looks linear. It is correct but it just describes the autocovariance structure. It is an incomplete representation. If Eε 2 t < is violated, (X t ) becomes an infinite-variance linear process, conditions on coefficient series must be adjusted. Outside the scope here.

A simple example: an AR-ARCH model The dynamic generating law X t = 0.9X t 1 + u t, (1) u t = ht 0.5 ε t, (2) h t = 1 + 0.9ut 1, 2 ε t NID(0, 1), (3) defines a stable AR-ARCH process. It is an AR(1) model with Wold innovations u t, which are white noise but not independent.

A time series plot of 1000 observations 30 20 10 0 10 0 200 400 600 800 1000 Impression: not too nonlinear, but outlier patches point to fat-tailed distributions: variable has no finite kurtosis.

Correlograms 0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.2 0.4 0.6 0.8 1.0 0 5 10 15 20 25 30 Lag 0 5 10 15 20 25 30 Lag Impression: AR-ARCH on the left inconspicuous, fairly identical to a standard AR(1) correlogram on the right.

Plots versus lags 20 10 0 10 20 6 4 2 0 2 4 6 y x 20 10 0 10 20 y( 1) 6 4 2 0 2 4 6 x( 1) Impression: a bit more dispersed than the standard AR(1) plot on the right: leptokurtosis. Basically linear (as should be).

Plots of squares versus lagged squares y² 0 100 200 300 400 x² 0 10 20 30 40 0 100 200 300 400 y²( 1) 0 10 20 30 40 x²( 1) Impression: even this device, suggested by Andrew A. Weiss allows no reliable discrimination to the standard AR case on the right.

I. Characteristics of Time Series The meaning of this section: This section corresponds to Section 2 of the book by Fan & Yao and is meant to review the basic concepts of (mostly linear) time-series analysis.

Stationarity Definition A time series (X t, t Z) is (weakly, covariance) stationary if (a) EX 2 t <, (b) EX t = µ t, and (c) cov(x t, X t+k ) is independent of t k. Remark. For k = 0, this definition yields time-constant finite variance. Definition A time series (X t, t Z) is strictly stationary if (X 1,...,X n ) and (X 1+k,...,X n+k ) have the same distribution for any n 1 and n, k Z. Remark. For nonlinear time series, often strict stationarity is the more natural concept.

ARMA processes The ARMA(p, q) model with p, q N X t = b 1 X t 1 +... + b p X t p + ε t + a 1 ε t 1 +... + a q ε t q, for (ε t ) white noise, is technically rewritten as b(b)x t = a(b)ε t, with polynomials a(z) and b(z) and the backshift operator B defined by B k X t = X t k. This is the most popular linear time-series model. Often, the expression ARMA process is reserved for stable polynomials.

Stationarity of ARMA processes Theorem The process defined by the ARMA(p, q) model is stationary if b(z) 0 for all z C with z 1, assuming that a(z) and b(z) have no common factors. Remark. Note that t Z. If t N, the ARMA process can be started from arbitrary conditions, and the usual distinction of stable and stationary applies. Clearly, here a pure MA process (p = 0) is always stationary.

Causal time series Definition A time series (X t ) is causal if for all t X t = with white noise (ε t ). d j ε t j, d j <, j=0 j=0 Remark. A causal process is always stationary. X t = 2X t 1 + ε t violates the ARMA stability conditions and nonetheless has a stationary non-causal solution. These are at odds with intuition and will be excluded.

Stationary Gaussian processes A process (X t ) is called Gaussian if all its finite-dimensional marginal distributions are normal. For Gaussian processes, the Wold representation holds with iid innovations. The purely nondeterministic part of a Gaussian process is linear. A Gaussian MA(q) process is q dependent, i.e. X t and X t+q+k are independent for all k 1. For a Gaussian AR process, X t is independent of X t k, k > p given X t 1,...,X t p.

Autocovariance and autocorrelation Definition Let (X t ) be stationary. The autocovariance function (ACVF) of (X t ) is γ(k) = cov(x t+k, X t ), k Z. The autocorrelation function (ACF) of (X t ) is ρ(k) = γ(k)/γ(0) = corr(x t+k, X t ), k Z. It follows that γ( k) = γ(k) and ρ( k) = ρ(k) (even functions).

Characterization of the ACVF Theorem A function γ(.) : Z R is the ACVF of a stationary time series if and only if it is even and nonnegative definite in the sense that n i=1 j=1 n a i a j γ(i j) 0 for all integer n 1 and arbitrary real a 1,...,a n. Remark. One direction is easy to show, as it uses the properties of a covariance matrix of X t,...,x t n+1. The reverse direction is hard to prove and needs a Theorem by Kolmogorov.

ACF of stationary ARMA processes Any ARMA process has an MA( ) representation X t = a j ε t j, j=0 with j=0 a j <. It follows that γ(k) = σ 2 a j a j+k, ρ(k) = j=0 j=0 j=0 a ja j+k. j=0 a2 j Clearly, for pure MA(q) processes, these expressions become 0 for k > q.

Properties of the ACF for ARMA processes Proposition 1. For causal ARMA processes, ρ(k) 0 like c k for c < 1 as k (exponential); 2. for MA(q) processes, ρ(k) = 0 for k > q. This proposition does not warrant a clear distinction between AR and general ARMA processes.

Estimating the ACF The sample ACF (the correlogram) is defined by ˆρ(k) = ˆγ(k)/ˆγ(0), where ˆγ(k) = 1 T T k (X t X T )(X t+k X T ), t=1 for small k, for example k T/4 or k 2 T, with X T = T 1 T t=1 X t.

Statistical properties of the mean estimate Theorem Let (X t ) is a linear stationary process defined by X t = µ + j=0 a jε t j, with (ε t ) iid(0, σ 2 ) and j=0 a j <. If j=0 a j 0, T( X T µ) N(0, ν1 2 ), where ν 2 1 = j= γ(j) = σ 2 j=0 a j 2. Remark. The variance of the mean estimate depends on the spectrum at 0 and may be called the long-run variance.

The long-run variance A stationary time-series process X t = a j ε t j has the variance varx t = γ(0) = σε 2 aj 2, j=0 which usually differs from the long-run variance σε( 2 a j ) 2. The long-run variance may also be seen as the spectrum at 0, j= γ(j), or as the limit variance n lim n n 1 var X t. j=0 t=0 For white noise (X t ), variance and long-run variance coincide.

Statistical properties of the variance estimate Theorem Let (X t ) is a linear stationary process defined by X t = µ + j=0 a jε t j, with (ε t ) iid(0, σ 2 ) and j=0 a j <. If Eε 4 t <, T {ˆγ(0) γ(0)} N(0, ν2 2 ), where ν 2 2 = 2σ 2 (1 + 2 ρ(j) 2 ). j=1 Remark. This is reminiscent of the portmanteau Q.

Statistical properties of the correlogram Theorem Let (X t ) is a linear stationary process defined by X t = µ + j=0 a jε t j, with (ε t ) iid(0, σ 2 ) and j=0 a j <. If Eε 4 t <, T {ˆ (k) (k)} N(0,W), where w ij = {ρ(t + i)ρ(t + j) + ρ(t i)ρ(t + j) + 2ρ(i)ρ(j)ρ(t) 2 t= 2ρ(i)ρ(t)ρ(t + j) 2ρ(j)ρ(t)ρ(t + i)}, where (k) = (ρ(1),..., ρ(k)). Remark. This is Bartlett s formula, impressive but not immediately useful. In simple cases, it allows determining confidence bands for the correlogram. White noise yields T ˆρ k N(0, 1) for k 0.

Partial autocorrelation function Definition (X t ) is a stationary process with EX t = 0. The partial autocorrelation function (PACF) π : N [ 1, 1] is defined by π(1) = ρ(1) and π(k) = corr(r 1 2,...,k, R k+1 2,...,k ), for k 2, where R j 2,...,k denotes residuals from regressing X j on X 2,...,X k by least squares. Remark. For non-gaussian processes, the thus defined PACF does not necessarily correspond to partial correlations, if partial correlations are defined via conditional expectations.

Properties of the PACF Proposition 1. For any stationary process, π(k) is a function of the ACVF values γ(1),...,γ(k); 2. For an AR(p) process, π(k) = 0 for k > p, i.e. the PACF cuts off at p. Remark. Mathematically, the PACF does not provide any new information on top of the ACVF. The sample PACF facilitates visual pattern recognition and may indicate AR models and their lag order.

Mixing White noise is usually insufficient for ergodic theorems such as laws of large numbers (LLN) or central limit theorems (CLT). For linear processes with iid innovations, some moment conditions will suffice. For nonlinear processes, more is needed. Generally, mixing conditions guarantee that X t and X t+h are more or less independent for large h. The metaphor is mixing drinks: a drop of a liquid will not remain close to its origin. If we mix two separate glasses, the outcome will be stationary but not mixing.

Five mixing coefficients α(n) = β(n) = E ρ(n) = φ(n) = ψ(n) = sup A F 0,B F n { sup B F n P(A)P(B) P(AB), P(B) P(B X 0, X 1, X 2,...) sup corr(x, Y ), X L 2 (F 0 ),Y L2 (Fn ) sup A F 0,B F n,p(a)>0 P(B) P(B A), sup 1 P(B A)/P(B). A F 0,B F n,p(a)p(b)>0 Here, F l k is the σ algebra generated by X t,k t l, and L 2 are square integrable functions. Good mixing coefficients converge to 0 as n. },

Mixing processes Definition A process (X t ) is α mixing ( strong mixing ) if α(n) 0 as n, and similarly for β mixing etc. A basic property is α(k) 1 4 ρ(k) 1 2 φ(k), and generally φ mixing implies ρ mixing (and β mixing), and ρ mixing (or β mixing) implies α mixing. ψ mixing implies φ mixing and thus all others. Even for simple examples, the mixing coefficients cannot be evaluated directly.

Some properties of mixing processes 1. If (X t ) is mixing (any definition) and m(.) is a measurable function, then (m(x t )) is again mixing. The property is hereditary. 2. If (X t ) is a linear ARMA process and ε t has a density, it is β mixing with β(n) 0 exponentially. 3. A strictly stationary process on Z is mixing iff its restriction to N is mixing (any definition). 4. Finite-dependent (such as strict MA) or independent processes are mixing. 5. Deterministic processes are not mixing.

A LLN for α mixing processes Proposition Assume (X t ) is strictly stationary and α mixing, and E X t <. Then, as n, S n /n EX t a.s., where S n = n t=1 X t. A comparable LLN with slightly stronger conditions guarantees convergence of n 1 vars t to the long-run variance.

A CLT for α mixing processes Theorem Assume (X t ) is strictly stationary with EX t = 0, the long-run variance σ 2 is positive, and one of the following conditions hold 1. E X t δ < and j=0 α(j)1 2/δ < for some δ > 2; 2. P( X t < c) = 1 for some c > 0 and j=1 α(j) <. Then, S n / n N(0, σ 2 ), with S n = n t=1 X t. Remark. The theorem cares for processes with bounded support (case 2) as well as for some quite fat-tailed ones (case 1 for δ = 2 + ǫ), which however need fast decrease of their mixing coefficients.