Gaussian processes. Basic Properties VAG002-

Similar documents
Time Series: Theory and Methods

Some Time-Series Models

Time Series Analysis -- An Introduction -- AMS 586

Notes on Time Series Modeling

Econ 424 Time Series Concepts

Time Series Models and Inference. James L. Powell Department of Economics University of California, Berkeley

Autoregressive Moving Average (ARMA) Models and their Practical Applications

Module 3. Descriptive Time Series Statistics and Introduction to Time Series Models

A time series is called strictly stationary if the joint distribution of every collection (Y t

Introduction to Stochastic processes

Stochastic Processes: I. consider bowl of worms model for oscilloscope experiment:

Lecture 2: Univariate Time Series

7. Forecasting with ARIMA models

{ } Stochastic processes. Models for time series. Specification of a process. Specification of a process. , X t3. ,...X tn }

Time Series Analysis. James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY

ARMA Estimation Recipes

6.3 Forecasting ARMA processes

LECTURE 10 LINEAR PROCESSES II: SPECTRAL DENSITY, LAG OPERATOR, ARMA. In this lecture, we continue to discuss covariance stationary processes.

For a stochastic process {Y t : t = 0, ±1, ±2, ±3, }, the mean function is defined by (2.2.1) ± 2..., γ t,

Covariance Stationary Time Series. Example: Independent White Noise (IWN(0,σ 2 )) Y t = ε t, ε t iid N(0,σ 2 )

Stochastic Processes

Contents. 1 Time Series Analysis Introduction Stationary Processes State Space Modesl Stationary Processes 8

Stat 248 Lab 2: Stationarity, More EDA, Basic TS Models

Chapter 4: Models for Stationary Time Series

LECTURES 2-3 : Stochastic Processes, Autocorrelation function. Stationarity.

Statistics of stochastic processes

STAT 248: EDA & Stationarity Handout 3

AR, MA and ARMA models

Time Series Analysis. James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY

Residuals in Time Series Models

Applied time-series analysis

Time Series I Time Domain Methods

white noise Time moving average

Time Series 2. Robert Almgren. Sept. 21, 2009

3. ARMA Modeling. Now: Important class of stationary processes

Introduction to Time Series Analysis. Lecture 11.

data lam=36.9 lam=6.69 lam=4.18 lam=2.92 lam=2.21 time max wavelength modulus of max wavelength cycle

Exponential decay rate of partial autocorrelation coefficients of ARMA and short-memory processes

Prof. Dr. Roland Füss Lecture Series in Applied Econometrics Summer Term Introduction to Time Series Analysis

1 Linear Difference Equations

Lecture 3: Autoregressive Moving Average (ARMA) Models and their Practical Applications

Multivariate Time Series

Linear Processes in Function Spaces

EC402: Serial Correlation. Danny Quah Economics Department, LSE Lent 2015

STAT 443 Final Exam Review. 1 Basic Definitions. 2 Statistical Tests. L A TEXer: W. Kong

TIME SERIES AND FORECASTING. Luca Gambetti UAB, Barcelona GSE Master in Macroeconomic Policy and Financial Markets

Parametric Inference on Strong Dependence

Lecture 3: Autoregressive Moving Average (ARMA) Models and their Practical Applications

Discrete time processes

Econometría 2: Análisis de series de Tiempo

3 ARIMA Models. 3.1 Introduction

ECON 616: Lecture 1: Time Series Basics

Introduction to ARMA and GARCH processes

1. Stochastic Processes and Stationarity

1 Teaching notes on structural VARs.

at least 50 and preferably 100 observations should be available to build a proper model

An algorithm for robust fitting of autoregressive models Dimitris N. Politis

Elements of Multivariate Time Series Analysis

7. MULTIVARATE STATIONARY PROCESSES

Difference equations. Definitions: A difference equation takes the general form. x t f x t 1,,x t m.

11. Further Issues in Using OLS with TS Data

Econ 623 Econometrics II Topic 2: Stationary Time Series

STA 6857 Autocorrelation and Cross-Correlation & Stationary Time Series ( 1.4, 1.5)

Part II. Time Series

γ 0 = Var(X i ) = Var(φ 1 X i 1 +W i ) = φ 2 1γ 0 +σ 2, which implies that we must have φ 1 < 1, and γ 0 = σ2 . 1 φ 2 1 We may also calculate for j 1

A SARIMAX coupled modelling applied to individual load curves intraday forecasting

Ch. 15 Forecasting. 1.1 Forecasts Based on Conditional Expectations

5: MULTIVARATE STATIONARY PROCESSES

TIME SERIES ANALYSIS. Forecasting and Control. Wiley. Fifth Edition GWILYM M. JENKINS GEORGE E. P. BOX GREGORY C. REINSEL GRETA M.

Nonlinear time series

Class 1: Stationary Time Series Analysis

Levinson Durbin Recursions: I

Week 5 Quantitative Analysis of Financial Markets Characterizing Cycles

Automatic Autocorrelation and Spectral Analysis

Chapter 3 - Temporal processes

STAT 520: Forecasting and Time Series. David B. Hitchcock University of South Carolina Department of Statistics

The autocorrelation and autocovariance functions - helpful tools in the modelling problem

Empirical Market Microstructure Analysis (EMMA)

Fitting Linear Statistical Models to Data by Least Squares: Introduction

On 1.9, you will need to use the facts that, for any x and y, sin(x+y) = sin(x) cos(y) + cos(x) sin(y). cos(x+y) = cos(x) cos(y) - sin(x) sin(y).

Time Series 3. Robert Almgren. Sept. 28, 2009

Empirical Macroeconomics

covariance function, 174 probability structure of; Yule-Walker equations, 174 Moving average process, fluctuations, 5-6, 175 probability structure of

Univariate Time Series Analysis; ARIMA Models

Chapter 3. ARIMA Models. 3.1 Autoregressive Moving Average Models

University of Oxford. Statistical Methods Autocorrelation. Identification and Estimation

COMPUTER ALGEBRA DERIVATION OF THE BIAS OF LINEAR ESTIMATORS OF AUTOREGRESSIVE MODELS

Levinson Durbin Recursions: I

ARIMA Modelling and Forecasting

Marcel Dettling. Applied Time Series Analysis SS 2013 Week 05. ETH Zürich, March 18, Institute for Data Analysis and Process Design

Vector Time-Series Models

Multivariate Time Series Analysis and Its Applications [Tsay (2005), chapter 8]

Large Sample Properties of Estimators in the Classical Linear Regression Model

9. Multivariate Linear Time Series (II). MA6622, Ernesto Mordecki, CityU, HK, 2006.

GARCH Models. Eduardo Rossi University of Pavia. December Rossi GARCH Financial Econometrics / 50

ON VARIANCE COVARIANCE COMPONENTS ESTIMATION IN LINEAR MODELS WITH AR(1) DISTURBANCES. 1. Introduction

1. Fundamental concepts

Next tool is Partial ACF; mathematical tools first. The Multivariate Normal Distribution. e z2 /2. f Z (z) = 1 2π. e z2 i /2

LINEAR STOCHASTIC MODELS

Chapter 9: Forecasting

Transcription:

Gaussian processes The class of Gaussian processes is one of the most widely used families of stochastic processes for modeling dependent data observed over time, or space, or time and space. The popularity of such processes stems primarily from two essential properties. First, a Gaussian process is completely determined by its mean and covariance functions. This property facilitates model fitting as only the first- and second-order moments of the process require specification. Second, solving the prediction problem is relatively straightforward. The best predictor of a Gaussian process at an unobserved location is a linear function of the observed values and, in many cases, these functions can be computed rather quickly using recursive formulas. The fundamental characterization, as described below, of a Gaussian process is that all the finitedimensional distributions have a multivariate normal (or Gaussian) distribution. In particular the distribution of each observation must be normally distributed. There are many applications, however, where this assumption is not appropriate. For example, consider observations x 1,...,x n,wherex t denotes a 1 or 0, depending on whether or not the air pollution on the tth day at a certain site exceeds a government standard. A model for there data should only allow the values of 0 and 1 for each daily observation thereby precluding the normality assumption imposed by a Gaussian model. Nevertheless, Gaussian processes can still be used as building blocks to construct more complex models that are appropriate for non-gaussian data. See [3 5] for more on modeling non-gaussian data. Basic Properties A real-valued stochastic process fx t,t2 Tg, where T is an index set, is a Gaussian process if all the finite-dimensional distributions have a multivariate normal distribution. That is, for any choice of distinct values t 1,...,t k 2 T, the random vector X D X t1,...,x tk 0 has a multivariate normal distribution with mean vector m D EX and covariance matrix 6 D cov X, X, which will be denoted by X ¾ N m,6 Provided the covariance matrix 6 is nonsingular, the random vector X has a Gaussian probability density function given by f X x D 2 n/2 det 6 1/2 ð exp 1 2 x m 0 6 1 x m 1 In environmental applications, the subscript t will typically denote a point in time, or space, or space and time. For simplicity, we shall restrict attention to the case of time series for which t represents time. In such cases, the index set T is usually [0, 1 for time series recorded continuously or f0, 1,...,g for time series recorded at equally spaced time units. The mean and covariance functions of a Gaussian process are defined by and t D EX t s, t D cov X s,x t 2 3 respectively. While Gaussian processes depend only on these two quantities, modeling can be difficult without introducing further simplifications on the form of the mean and covariance functions. The assumption of stationarity frequently provides the proper level of simplification without sacrificing much generalization. Moreover, after applying elementary transformations to the data, the assumption of stationarity of the transformed data is often quite plausible. A Gaussian time series fx t g is said to be stationary if 1. m t D EX t D is independent of t, and 2. t C h, t D cov X tch,x t is independent of t for all h. For stationary processes, it is conventional to express the covariance function as a function on T instead of on T ð T. That is, we define h D cov X tch,x t and call it the autocovariance function of the process. For stationary Gaussian processes fx t g,wehave 3. X t ¾ N, 0 for all t, and 4. X tch,x t 0 has a bivariate normal distribution with covariance matrix [ ] 0 h h 0

2 Gaussian processes for all t and h. A general stochastic process fx t g satisfying conditions 1 and 2 is said to be weakly or second-order stationary. The first- and second-order moments of weakly stationary processes are invariant with respect to time translations. A stochastic process fx t g is strictly stationary if the distribution of X t1,...,x tn is the same as X t1cs,...,x tncs for any s. Inother words, the distributional properties of the time series are the same under any time translation. For Gaussian time series, the concepts of weak and strict stationarity coalesce. This result follows immediately from the fact that for weakly stationary processes, X t1,...,x tn and X t1cs,...,x tncs have the same mean vector and covariance matrix. Since each of the two vectors has a multivariate normal distribution, they must be identically distributed. Properties of the Autocovariance Function An autocovariance function Ð has the properties: 1. 0 ½ 0, 2. j h j 0 for all h, 3. h D h, i.e. Ð is an even function. Autocovariances have another fundamental property, namely that of non-negative definiteness, n a i t i t j a j ½ 0 4 i,jd1 for all positive integers n, real numbers a 1,...,a n, and t 1,...,t n 2 T. Note that the expression on the left of (4) is merely the variance of a 1 X t1 CÐÐÐC a n X tn and hence must be non-negative. Conversely, if a function Ð is non-negative definite and even, then it must be an autocovariance function of some stationary Gaussian process. Gaussian Linear Processes If fx t,td 0, š1, š2,...,g is a stationary Gaussian process with mean 0, then the Wold decomposition allows X t to be expressed as a sum of two independent components, 1 X t D jz t j C V t 5 jd0 where fz t g is a sequence of independent and identically distributed (iid) normal random variables with mean 0 and variance 2, f j g is a sequence of square summable coefficients with 0 D 1, and fv t g is a deterministic process that is independent of fz t g.the Z t are referred to as innovations and are defined by Z t D X t E X t jx t 1,X t 2,...). A process fv t g is deterministic if V t is completely determined by its past history fv s,s<tg. An example of such a processes is the random sinusoid, V t D A cos t C 2, where A and 2 are independent random variables with A ½ 0and2 distributed uniformly on [0, 2). In this case, V 2 is completely determined by the values of V 0 and V 1. In most time series modeling applications, the deterministic component of a time series is either not present or easily removed. Purely nondeterministic Gaussian processes do not possess a deterministic component and can be represented as a Gaussian linear processes, X t D 1 jd0 jz t j The autocovariance of fx t g has the form h D 1 jd0 j jch 6 7 The class of autoregressive (AR) processes, and its extensions, autoregressive moving-average (ARMA) processes, are dense in the class of Gaussian linear processes. A Gaussian AR(p) process satisfies the recursions X t D 1 X t 1 CÐÐÐC p X t p C Z t 8 where fz t g is an iid sequence of N 0, 2 random variables, and the polynomial z D 1 1 z ÐÐÐ p z p has no zeros inside or on the unit circle. The AR(p) process has a linear representation (6) where the coefficients are found as functions of the j (see [2]). Now for any Gaussian linear process, there exists an AR(p) process such that the difference in the two autocovariance functions can be made arbitrarily small for all lags. In fact, the autocovariances can be matched up perfectly for the first p lags.

Gaussian processes 3 Prediction Recall that if two random vectors X 1 and X 2 have a joint normal distribution, i.e. [ ] ([ ] [ ]) X1 m1 611 6 ¾ N, 12 X 2 m 2 6 21 6 22 and 6 22 is nonsingular, then the conditional distribution of X 1 given X 2 has a multivariate normal distribution with mean m X1 jx 2 D m 1 C 6 12 6 1 22 X 2 m 2 9 and covariance matrix 6 X1 jx 2 D 6 11 6 12 6 1 22 6 21 10 The key observation here is that the best mean square error predictor of X 1 in terms of X 2 (i.e. the multivariate function g X 2 that minimizes EjjX 1 g X 2 jj 2,wherejj Ð jj is Euclidean distance) is E X 1 jx 2 D m X1 jx 2 which is a linear function of X 2. Also, the covariance matrix of prediction error, 6 X1 jx 2, does not depend on the value of X 2.These results extend directly to the prediction problem for Gaussian processes. Suppose fx t,td 1, 2,...g is a stationary Gaussian process with mean and autocovariance function Ð and that based on the random vector consisting of the first n observations, X n D X 1,...,X n 0,we wish to predict the next observation X nc1. Prediction for other lead times is analogous to this special case. Applying the formula in (9), the best one-step-ahead predictor of X nc1 is given by X nc1 : D E X nc1 jx 1,...,X n D C n1 X n where CÐÐÐC nn X 1 11 n1,..., nn 0 D 6 1 n n 12 6 n D cov X n, X n, and n D cov X nc1, X n D 1,..., n 0. The mean square error of prediction is given by v n D 0 0 n 6 1 n n 13 These formulas assume that 6 n is nonsingular. If 6 n is singular, then there is a linear relationship among X 1,...,X n and the prediction problem can then be recast by choosing a generating prediction subset consisting of linear independent variables. The covariance matrix of this prediction subset will be nonsingular. A mild and easily verifiable condition for ensuring nonsingularity of 6 n for all n is that h! 0ash!1with 0 >0 (see [1]). While (12) and (13) completely solve the prediction problem, these equations require the inversion of an n ð n covariance matrix which may be difficult and time consuming for large n. The Durbin Levinson algorithm (see [1]) allows one to compute the coefficient vector n D n1,..., nn 0 and the one-step prediction errors v n recursively from n 1, v n 1, and the autocovariance function. The Durbin Levinson Algorithm The coefficients n in the calculation of the one-step prediction error (11) and the mean square error of prediction (13) can be computed recursively from the equations n,1. n,n 1 nn D n D 1 n n 1,j n 1 jd1 n 1,1 n 1,n 1. nn. n 1,n 1 n 1,1 v n D v n 1 1 2 nn v 1 n 1 14 where 11 D 1 / 0 and v 0 D 0. If fx t g follows that AR(p) process in (8), then the recursions simplify a great deal. In particular, for n>p, the coefficients nj D j for j D 1,...,p and nj D 0forj>pgiving X nc1 D 1 X n CÐÐÐC p X n p 15 with v n D 2. The sequence of coefficients f jj,j½ 1g is called the partial autocorrelation function and is a useful tool for model identification. The partial autocorrelation at lag j is interpreted as the correlation between X 1 and X jc1 after correcting for the intervening observations X 2,...,X j. Specifically, jj is the correlation of the two residuals obtained by regression of X 1 and X jc1 on the intermediate observations X 2,...,X j. Of particular interest is the relationship between nn and the reduction in the one-step mean square error as the number of predictors is increased from n 1

4 Gaussian processes to n. The one-step prediction error has the following decomposition in terms of the partial autocorrelation function: v n D 0 1 2 11 ÐÐÐ 1 2 nn 16 For a Gaussian process, X nc1 X nc1 is normally distributed with mean 0 and variance v n. Thus, X nc1 š z 1 /2 v 1/2 n constitute (1 ) 100% prediction bounds for the observation X nc1, where z 1 /2 is the (1 /2) quantile of the standard normal distribution. In other words, X nc1 lies between the bounds X nc1 š z 1 /2 v 1/2 n with probability 1. Estimation for Gaussian Processes One of the advantages of Gaussian models is that an explicit and closed form of the likelihood is readily available. Suppose that fx t,t D 1, 2,...,g is a stationary Gaussian time series with mean and autocovariance function Ð. Denote the data vector by X n D X 1,...,X n and the vector of onestep predictors by X n D X 1,..., X n 0,where X 1 D and X j D E X j jx 1,...,X j 1 for j ½ 2. If 6 n denotes the covariance matrix of X n, which we assume is nonsingular, then the likelihood of X n is L 6 n, D 2 n/2 det 6 n 1/2 ( ) ð exp 1 2 X n 1 0 6 1 n X n 1 17 where 1 D 1,...,1 0. Typically, 6 n will be expressible in terms of a finite number of unknown parameters, ˇ1,...,ˇr, so that the maximum likelihood estimator of these parameters and arethosevalues that maximize L for the given dataset. Under mild regularity assumptions, the resulting maximum likelihood estimators are approximately normally distributed with covariance matrix given by the inverse of the Fisher information. In most settings, direct-closed-form maximization of L with respect to the parameter set is not achievable. In order to maximize L using numerical methods, either derivatives or repeated calculation of the function are required. For moderate to large sample sizes n, calculation of both the determinant of 6 n and the quadratic form in the exponential of L can be difficult and time consuming. On the other hand, there is a useful representation of the likelihood in terms of the one-step prediction errors and their mean square errors. By the form of X n, we can write X n X n D A n X n 18 where A n is a lower triangular square matrix with ones on the diagonal. Inverting this expression, we have X n D C n X n X n 19 where C n is also lower triangular with ones on the diagonal. Since X j E X j jx 1,...,X j 1 is uncorrelated with X 1,...,X j 1, it follows that the vector X n X n consists of uncorrelated, and hence independent, normal random variables with mean 0 and variance v j 1, j D 1,...,n. Taking covariances on both sides of (19) and setting D n D diagfv 0,...,v n 1 g,wefindthat and 6 n D C n D n C 0 n 20 X n 1 0 6 1 n X n 1 D X n X n 0 Dn 1 X n X n X j n D X j 2 21 v jd1 j 1 It follows that det 6 n D v 0 v 1...v n 1 so that the likelihood reduces to L 6 n, D 2 n/2 v 0 v 1...v n 1 1/2 exp 1 n X j X j 2 2 v jd1 j 1 22 The calculation of the one-step prediction errors and their mean square errors required in the computation of L based on (22) can be simplified further for a variety of time series models such as ARMA processes. We illustrate this for an AR process. Gaussian Likelihood for an AR(p) Process If fx t g is the AR(p) process specified in (8) with mean, then one can take advantage of the simple form for the one-step predictors and associated mean square errors. The likelihood becomes L 1,..., p,, 2 D 2 n p /2 n p

Gaussian processes 5 ð exp 1 n X j X j 2 2 2 2 p/2 jdpc1 ð v 0 v 1...v p 1 1/2 exp 1 p X j X j 2 2 v jd1 j 1 23 where, for j>p, X j D C 1 X j 1 CÐÐÐC p X j p 1 are the one-step predictors. The likelihood is a product of two terms, the conditional density of X n given X p and the density of X p.often, just the conditional maximum likelihood estimator is computed which is found by maximizing the first term. For the AR process, the conditional maximum 20 19 Temperature 18 17 16 15 1900 1920 1940 1960 1980 Figure 1 Average maximum temperature, 1885 1993. Regression line is 16.83 C 0.008 45t 2 Temperature for september 1 0 1 2 3 2 1 0 1 2 Quantiles of standard normal Figure 2 QQ plot for normality of the innovations

6 Gaussian processes likelihood estimator can be computed in closed form. Example This example consists of the average maximum temperature over the month of September for the years 1895 1993 in an area of the US whose vegetation is characterized as tundra. The time series x 1,...,x 99 is plotted in Figure 1. Here we investigate the possibility of the data exhibiting a slight linear trend. After inspecting the residuals from fitting a least squares regression line to the data, we entertain a time series model of the form X t D ˇ0 C ˇ1t C W t where fw t g is the Gaussian AR(1), W t D 1 W t 1 C Z t 24 25 and fz t g is a sequence of iid N 0, 2 random variables. After maximizing the Gaussian likelihood over the parameters ˇ0,ˇ1, 1,and 2, we find that the maximum likelihood estimate of the mean function is 16.83 C 0.008 45t. The maximum likelihood parameters of 1 and 2 are estimated by 0.1536 and 1.3061, respectively. The maximum likelihood estimates of ˇ0 and ˇ1 can be viewed as generalized least squares estimates assuming that the residual process follows the estimated AR(1) model. The resulting standard errors of these estimates are 0.277 81 and 0.004 82, respectively, which provides some doubt about the significance of a nonzero slope of the line. Without modeling the dependence in the residuals, the slope would have been deemed significant using classical inference procedures. By modeling the dependence in the residuals, the evidence in favor of a nonzero slope has diminished somewhat. The QQ plot of the estimated innovations is displayed in Figure 2. This plot shows that the AR(1) model is not far from being Gaussian. Further details about inference procedures for regression models with time series errors can be found in [2, Chapter 6]. References [1] Brockwell, P.J. & Davis, R.A. (1991). Time Series: Theory and Methods, 2nd Edition, Springer-Verlag, New York. [2] Brockwell, P.J. & Davis, R.A. (1996). Introduction to Time Series and Forecasting, Springer-Verlag, New York. [3] Diggle, Peter J., Liang, Kung-Yee & Zeger, Scott L. (1996). Analysis of Longitudinal Data, Clarendon Press, Oxford. [4] Fahrmeir, L. & Tutz, G. (1994). Multivariate Statistical Modeling Based on Generalized Linear Models, Springer-Verlag, New York. [5] Rosenblatt, M. (2000). Gaussian and Non-Gaussian Linear Time Series and Random Fields, Springer-Verlag, New York. RICHARD A. DAVIS