Some basic properties of cross-correlation functions of n-dimensional vector time series

Similar documents
IV. Covariance Analysis

{ } Stochastic processes. Models for time series. Specification of a process. Specification of a process. , X t3. ,...X tn }

Estimation of Parameters of Multiplicative Seasonal Autoregressive Integrated Moving Average Model Using Multiple Regression

LECTURES 2-3 : Stochastic Processes, Autocorrelation function. Stationarity.

Comparative Analysis of Linear and Bilinear Time Series Models

Poincaré Plots for Residual Analysis

Computational Data Analysis!

On 1.9, you will need to use the facts that, for any x and y, sin(x+y) = sin(x) cos(y) + cos(x) sin(y). cos(x+y) = cos(x) cos(y) - sin(x) sin(y).

MULTIVARIATE TIME SERIES ANALYSIS AN ADAPTATION OF BOX-JENKINS METHODOLOGY Joseph N Ladalla University of Illinois at Springfield, Springfield, IL

16.584: Random (Stochastic) Processes

Multiplicative Sarima Modelling Of Nigerian Monthly Crude Oil Domestic Production

ARIMA Models. Richard G. Pierse

Multiresolution Models of Time Series

Notes on Random Processes

STAT 248: EDA & Stationarity Handout 3

For a stochastic process {Y t : t = 0, ±1, ±2, ±3, }, the mean function is defined by (2.2.1) ± 2..., γ t,

Large Sample Properties of Estimators in the Classical Linear Regression Model

Residuals in Time Series Models

TIME SERIES AND FORECASTING. Luca Gambetti UAB, Barcelona GSE Master in Macroeconomic Policy and Financial Markets

POWER AND TYPE I ERROR RATE COMPARISON OF MULTIVARIATE ANALYSIS OF VARIANCE

Small-Area Population Forecasting Using a Spatial Regression Approach

SOME COMMENTS ON THE THEOREM PROVIDING STATIONARITY CONDITION FOR GSTAR MODELS IN THE PAPER BY BOROVKOVA et al.

Communication Theory II

Stochastic Processes. A stochastic process is a function of two variables:

Ch.10 Autocorrelated Disturbances (June 15, 2016)

Economic modelling and forecasting

Multivariate Time Series

COMPUTER ALGEBRA DERIVATION OF THE BIAS OF LINEAR ESTIMATORS OF AUTOREGRESSIVE MODELS

EXAMINATION OF RELATIONSHIPS BETWEEN TIME SERIES BY WAVELETS: ILLUSTRATION WITH CZECH STOCK AND INDUSTRIAL PRODUCTION DATA

INDIAN INSTITUTE OF SCIENCE STOCHASTIC HYDROLOGY. Lecture -33 Course Instructor : Prof. P. P. MUJUMDAR Department of Civil Engg., IISc.

STAT 520: Forecasting and Time Series. David B. Hitchcock University of South Carolina Department of Statistics

white noise Time moving average

The Bayesian Approach to Multi-equation Econometric Model Estimation

State-space Model. Eduardo Rossi University of Pavia. November Rossi State-space Model Fin. Econometrics / 53

Definition of a Stochastic Process

Coregionalization by Linear Combination of Nonorthogonal Components 1

MULTIVARIATE PROBABILITY DISTRIBUTIONS

Time-Varying Moving Average Model for Autocovariance Nonstationary Time Series

Geostatistics for Gaussian processes

Stochastic Processes

Modern Navigation. Thomas Herring

Chapter 3 - Temporal processes

Stat 248 Lab 2: Stationarity, More EDA, Basic TS Models

Generalized cross-covariances and their estimation

International Journal of Advancement in Physical Sciences, Volume 4, Number 2, 2012

New York University Department of Economics. Applied Statistics and Econometrics G Spring 2013

Non-stationary Cross-Covariance Models for Multivariate Processes on a Globe

Econometría 2: Análisis de series de Tiempo

Deccan Education Society s FERGUSSON COLLEGE, PUNE (AUTONOMOUS) SYLLABUS UNDER AUTOMONY. SECOND YEAR B.Sc. SEMESTER - III

Lecture 1: Brief Review on Stochastic Processes

ADVANCED ECONOMETRICS I (INTRODUCTION TO TIME SERIES ECONOMETRICS) Ph.D. FAll 2014

DEPARTMENT OF ECONOMETRICS AND BUSINESS STATISTICS

ON VARIANCE COVARIANCE COMPONENTS ESTIMATION IN LINEAR MODELS WITH AR(1) DISTURBANCES. 1. Introduction

Wavelet Methods for Time Series Analysis. Motivating Question

Dynamic Time Series Regression: A Panacea for Spurious Correlations

State-space Model. Eduardo Rossi University of Pavia. November Rossi State-space Model Financial Econometrics / 49

Wavelet Methods for Time Series Analysis. Part IX: Wavelet-Based Bootstrapping

A TIME SERIES PARADOX: UNIT ROOT TESTS PERFORM POORLY WHEN DATA ARE COINTEGRATED

Statistics 349(02) Review Questions

Lecture 3 Stationary Processes and the Ergodic LLN (Reference Section 2.2, Hayashi)

Time Series Forecasting: A Tool for Out - Sample Model Selection and Evaluation

Single Equation Linear GMM with Serially Correlated Moment Conditions

Stochastic Processes. M. Sami Fadali Professor of Electrical Engineering University of Nevada, Reno

Use of the Autocorrelation Function for Frequency Stability Analysis

Nonlinear Time Series Modeling

Algorithm 808: ARfit A Matlab package for the estimation of parameters and eigenmodes of multivariate autoregressive models

A Hybrid Method to Improve Forecasting Accuracy An Application to the Processed Cooked Rice

1 Teaching notes on structural VARs.

Prof. Dr. Roland Füss Lecture Series in Applied Econometrics Summer Term Introduction to Time Series Analysis

Journal of Asian Scientific Research COMBINED PARAMETERS ESTIMATION METHODS OF LINEAR REGRESSION MODEL WITH MULTICOLLINEARITY AND AUTOCORRELATION

Long-range dependence

EAS 305 Random Processes Viewgraph 1 of 10. Random Processes

1 Random walks and data

Karhunen-Loeve Expansion and Optimal Low-Rank Model for Spatial Processes

A Non-Parametric Approach of Heteroskedasticity Robust Estimation of Vector-Autoregressive (VAR) Models

Fundamentals of Applied Probability and Random Processes

Design of Time Series Model for Road Accident Fatal Death in Tamilnadu

FORECASTING SUGARCANE PRODUCTION IN INDIA WITH ARIMA MODEL

Study on Modeling and Forecasting of the GDP of Manufacturing Industries in Bangladesh

Simulating Uniform- and Triangular- Based Double Power Method Distributions

SEQUENTIAL ESTIMATION OF THE STEADY-STATE VARIANCE IN DISCRETE EVENT SIMULATION

Stochastic Simulation Introduction Bo Friis Nielsen

Chapter 2 Random Processes

An introduction to volatility models with indices

CONTROL SYSTEMS, ROBOTICS, AND AUTOMATION - Vol. V - Prediction Error Methods - Torsten Söderström

Deterministic. Deterministic data are those can be described by an explicit mathematical relationship

Time series analysis of activity and temperature data of four healthy individuals

On Consistency of Tests for Stationarity in Autoregressive and Moving Average Models of Different Orders

On prediction and density estimation Peter McCullagh University of Chicago December 2004

Forecasting locally stationary time series

On studying relations between time series in climatology

Module 9: Stationary Processes

COMPUTER ALGEBRA DERIVATION OF THE BIAS OF LINEAR ESTIMATORS OF AUTOREGRESSIVE MODELS

Time Series Analysis -- An Introduction -- AMS 586

Lecture Note 1: Probability Theory and Statistics

Ross Bettinger, Analytical Consultant, Seattle, WA

Forecasting Data Streams: Next Generation Flow Field Forecasting

Econ 623 Econometrics II Topic 2: Stationary Time Series

Reliability and Risk Analysis. Time Series, Types of Trend Functions and Estimates of Trends

Econometría 2: Análisis de series de Tiempo

Transcription:

Journal of Statistical and Econometric Methods, vol.4, no.1, 2015, 63-71 ISSN: 2241-0384 (print), 2241-0376 (online) Scienpress Ltd, 2015 Some basic properties of cross-correlation functions of n-dimensional vector time series Anthony E. Usoro 1 Abstract In this work, cross-correlation function of multivariate time series was the interest. The design of cross-correlation function at different lags was presented. γ Xit+k X jt+l is the matrix of the cross-covariance functions, γ Xit and γ Xjt are the variances of X it and X jt vectors respectively. Vector cross-correlation function was derived as ρ Xit+k,X jt+l = γ X it+k X jt+l γ X it γ Xjt. A statistical package was used to verify the vector cross correlation functions, with trivariate analysis as a special case. From the results, some properties of vector cross-correlation functions were established. Keywords: Vector time series; cross-covariance function and cross-correlation function 1 Department of Mathematics and Statistics, Akwa Ibom State University, Mkpat Enin, Akwa Ibom State. Email: toskila2@yahoo.com Article Info: Received : January 17, 2015. Revised : February 28, 2015. Published online : March 31, 2015.

64 Cross-correlation functions of n-dimensional vector time series 1 Introduction In statistics, the term cross-covariance is sometimes used to refer to the covariance corr(x,y) between two random vectors X and Y, (where X = X 1, X 2,, X n and Y = Y 1, Y 2,, Y n ). In signal processing, the cross-covariance is often called cross-correlation and is a measure of similarity of two signals, commonly used to find features in an unknown signal by comparing it to a known one. It is a function of the relative time between the signals, and it is sometimes called the sliding dot product. In univariate time series, the autocorrelation of a random process describes the correlation between values of the process at different points in time, as a function of the two times or of the time difference. Let X be some repeatable process, and i be some point in time after the start of that process. (i may be an integer for a discrete-time process or a real number for a continuous-time process.) Then X i is the value (or realization) produced by a given run of the process at time i. Suppose that the process is further known to have defined values for mean μ i and variance σ i 2 for all times i. Then the definition of the autocorrelation between times s and t is ρ( s,t ) = E{(X t μ t )(X s μ s )] σ t σ s, where E is the expected value operator. It is required to note that the above expression is not well-defined for all time series or processes, because the variance may be zero. If the function ρ is well-defined, its value must lie in the range [-1,1], with 1 indicating perfect correlation and -1 indicating perfect anticorrelation. If X is a second-order stationary process then the mean μ and the variance σ 2 are time-independent, and further the autocorrelation depends only on the difference between t and s: the correlation depends only on the time-distance between the pair of values but not on their position in time. This further implies that the autocorrelation can be expressed as a function of the time-lag, and that

Anthony E. Usoro 65 this would be an even function of the lag k = s t, which implies s = t + k. This gives the more familiar form, ρ k = E[(X t μ)(x t+k μ)] = E[(X t μ)(x t+k μ)] E[(X t μ) 2 ]E[(X t+k μ) 2 ] σ 2 where X t and X t+k are time series process at lag k time difference. Hence, autocovariance coefficient γ k at lag k, measures the covariance between two values Z t and Z t+k, a distance k apart. The autocorrelation coefficient ρ k is defined as the autocovariance γ k at lag k divided by variance γ 0(k=0). The plot of γ k against lag k is called the autocovariance function (γ k ), while the plot of ρ k against lag k is called the autocorrelation function (Box and Jenkins 1976). In multivariate time series, cross-correlation or covariance involves more than one process. For instance, X t and Y t are two processes of which X t could be cross-correlated with Y t at lag k. The lag k value return by ccf(x, Y) estimates the correlation between X(t + k) and Y(t), Venables and Ripley (2002). Storch and Zwiers (2001) described cross-correlation in signal processing and time series. In signal processing, cross-correlation is a measure of similarity of two waveforms as a function of a time lag applied to one of them. This is also known as a sliding dot product or sliding inner-product. It is commonly used for searching a -duration signal for a shorter known feature. It also has application in pattern recognition, signal particle analysis, electron tomographic averaging, cryptanalysis and neurophysiology. In autocorrelation, which is the cross-correlation of a signal with itself, there is always a peak at a lag of zero unless the signal is a trivial zero signal. In probability theory and Statistics, correlation is always used to include a standardising factor in such a way that correlations have values between -1 and 1. Let (X t, Y t ) represent a pair of stochastic process that are jointly wide sense stationary. Then the cross covariance given by Box et al (1984) is γ xy (τ) = E[(X t μ x ) Y t+τ μ y ],

66 Cross-correlation functions of n-dimensional vector time series where μ x and μ y are the means of X t and Y t respectively. The cross-correlation function ρ xy is the normalized cross-covariance function. Therefore, ρ xy (τ) = γ xy(τ) σ x σ y where σ x and σ y are the standard deviation of processes X t and Y t respectively. If X t = Y t for all t, then the cross-correlation function is simply the autocorrelation function for a discrete process of length n defined as {X 1,, X n } which known mean and variance, an estimate of the autocorrelation may be obtained as n k 1 R (k) = (n k)σ 2 (X t μ)(x t+k μ) t=1 for any positive integer k<n, Patrick (2005). When the true mean μ and variance σ 2 are known, the estimate is unbiased. If the true mean, this estimate is unbiased. If the true mean and variance of the process are not known, there are several probabilities: i. if μ and σ 2 are replaced by the standard formulas for sample mean and sample variance, then this is a biased estimate. ii. if n-k in the above formula is replaced with n, the estimate is biased. However, it usually has a smaller mean square error, Priestly (1982) and Donald and Walden (1993). iii. if X t is stationary process, then the following are true μ t = μ s = μ, for all t,s and C xx(t,s) = C xx(s t) = C xx(t), where T=s-t, is the lag time or the moment of time by which the signal has been shifted. As a result, the autocovariance becomes C xx (T) = E[ X (t) μ X (t+t) μ ] = E X (t) X (t+t) μ 2 = R xx(t) μ 2, where R xx represents the autocorrelation in the signal processing sense. R xx (T) = C xx(t), Hoel (1984). σ2

Anthony E. Usoro 67 For X t and Y t, the following properties hold: 1. ρ xy(h) 1 2. ρ xy(h) = ρ xy( h) 3. ρ xy(0) 1 4. ρ xy(h) = γ xy(h) γ x(0) γ y(0) Mardia and Goodall (1993) defined separable cross-correlation function as C ij (X 1, X 2 ) = ρ(x 1, X 2 )a ij, where A = [a ij ] is a p p positive definite matrix and ρ(.,. ) is a valid correlation function. Goulard & Voltz (1992); Wackernage (2003); Ver Hoef and Barry (1998) implied that the cross- covariance function is r C ij (X 1 X 2 ) = k=1 ρ k (X 1 X 2 )a ik a jk, for an integer 1 r p, where A = [a ij ] is a p r full rank matrix and ρ k(.) are valid stationary correlation functions. Apanasovich and Genton (2010) constructed valid parametric cross-covariance functions. Apanasovich and Genton proposed a simple methodology based on latent dimensions and existing covariance models for univariate covariance, to develop flexible, interpretable and computationally feasible classes of cross-covariance functions in closed forms. They discussed estimation of the models and performed a small simulation study to demonstrate the models. The interest in this work is to extend cross-correlation functions beyond a-two variable case, present the multivariate design of vector crosscovariance and correlation functions and therefore establish some basic properties of vector cross-correlation functions from the analysis of vector cross-correlation functions.

68 Cross-correlation functions of n-dimensional vector time series 2 The design of cross-covariance cross-correlation functions The matrix of cross-covariance functions is as shown below: γ X(1t+k),X (1t+l) γ X(1t+k),X (2t+l) γ X(1t+k),X (3t+l)... γ X(1t+k),X (nt+l) γ X(2t+k),X (1t+l) γ X(2t+k),X (2t+l) γ X(2t+k),X (3t+l)... γ X(2t+k),X (nt+l) γ X(3t+k),X (1t+l) γ X(3t+k),X (2t+l) γ X(3t+k),X (3t+l)... γ X(3t+k),X (nt+l)...... γ X(mt+k),X (1t+l) γ X(mt+k),X (2t+l) γ X(mt+k),X (3t+l)... γ X(mt+k),X (nt+l) where k = 0,, a, l = 0,, b. The above matrix is a square matrix, and could be reduced to the form, γ X(it+k),X (jt+l) where i = 1,, m, j = 1,, n, k = 0,, a, l = 0,, b, (n = m). From the above cross-covariance matrix, ρ Xit+k,X jt+l = γ X it+k X jt+l γ X it γ Xjt, where, γ Xit+k X jt+l is the matrix of the cross-covariance functions, γ Xit and γ Xjt are the variances of X it and X jt vectors respectively. Given the above matrix, it is required to note that two vector processes X it+k and X jt+l can only be crosscorrelated at different lags, if either k lag of X it or l lag of X jt has a fixed value zero. That is X it can be cross-correlated with X jt+l (l ± 1,2,, b), or X jt can be cross-correlated with X it+k (k ± 1,2,, a).

Anthony E. Usoro 69 3 Analysis of the cross-correlation functions Given two processes X 1t and X 2t, ρ (X1t,x 2t+k ) is the cross-correlation between X 1t and X 2t at lag k, while, ρ (x2t,x 1t+k ) is the cross-correlation between X 2t and X 1t at lag k, Box et al (1984). In this work, three vector processes X 1t, X 2t and X 3t are used to carry out the cross-correlation analysis. For k = 0, ± 1, 2,,4, the following results were obtained with a software: Lag ρ (x1t,x 2t+k ) ρ (x2t,x 1t+k ) ρ (x1t,x 3t+k ) ρ (x3t,x 1t+k ) ρ (x2t,x 3t+k ) ρ (x3t,x 2t+k ) k -4-0.172 0.572-0.102 0.643-0.427-0.350-3 -0.517 0.405-0.501 0.410-0.076 0.042-2 -0.611 0.098-0.662 0.067 0.327 0.399-1 -0.605-0.290-0.674-0.303 0.659 0.697 0-0.506-0.506-0.578-0.578 0.900 0.900 1-0.290-0.605-0.303-0.674 0.697 0.659 2 0.098-0.611 0.067-0.662 0.399 0.327 3 0.405-0.517 0.410-0.501 0.042-0.076 4 0.572-0.172 0.643-0.102-0.350-0.427 From the above analysis, the following properties were established: 1. a. ρ Xit+k, X jt+l 1, for k = 0, l = ±1,, ±b, i j, b. ρ Xit+k, X jt+l 1, for l = 0, k = ±1,, ±a, i j. 2. a. ρ Xit+k, X jt+l ρ Xit+k,X jt l, for k = 0, l = 1,, b, i j, b. ρ Xit+k, X jt+l ρ Xit k,x jt+l, for l = 0, k = 1,, a, i j. 3. a. ρ Xit+k, X jt+l = ρ Xjt+k,X it l, for k = 0, l = 1,, b, i j, b. ρ Xit+k, X jt+l = ρ Xjt k,x it+l, for l = 0, k = 1,, a, i j.

70 Cross-correlation functions of n-dimensional vector time series 4 Conclusion The motivation behind this research work was to carry out crosscorrelation functions of multivariate time series. Ordinarily, cross-correlation compares two series by shifting one of them relative to the other. In the case of X and Y variables, the variable X may be cross-correlated at different lags of Y, and vice versa. In this work, X it and X jt were used as vector time series, using trivariate as a special case of multivariate cross-correlation functions. The design of the cross-covariance functions has been displayed in a matrix form. Estimates obtained revealed some basic properties of vector cross-correlation functions. References [1] Tatiyana V. Apanasovich and Marc G. Genton, Cross-Covariance functions for multivariate random fields based on latent dimensions, Biometrika, 97(1), (2010), 15-30. [2] G.E.P. Box and G.M. Jenkins, Time Series Analysis; Forecasting and Control, 1 st Edition, Holden-day, San Francisco, 1976. [3] G.E.P. Box, G.M. Jenkins and G.C. Reinsel, Time Series Analysis: Forecasting and Control, 3 rd ed. Upper Saddle River, NJ: Prentice-Hall, 1994. [4] M. Goulard and M. VolTz, Linear Coregionalization Model: tools for estimation and choice of croo-variogram matrix, Math. Geol., 24, (1992), 269-282. [5] P.G. Hoel, Mathematical Statistics, Wiley, New York, 1984. [6] K.V. Mardia and C.R. Goodall, Spartial-temporal analysis of multivariate environmental monitoring data. Multivariate environmental Statistics, Northytolland Series in Statistics & Probability 6, Amsterdam, North- Holland, 1993.

Anthony E. Usoro 71 [7] Pattrick F. Dunn, Measurement and Data Analysis for Engineering and Science, New York, McGraw-Hill, 2005. [8] Donald B. Percival and Andrew T. Walden, Spectral Analysis for Physical Applications: Multitaper and Conventional Univariate Techniques, Cambridge University Press, pp. 190-195, 1993. [9] M.B. Priestly, Spectral Analysis and Time Series, Academic press, London, 1982. [10] H.F. Storch and W. Zwiers, Statistical Analysis in climate research, Cambridge University Press, 2001. [11] W.N. Venables and B.D. Ripley, Auto and Cross-Covariance and Correlation Function Estimatiom. Modern Applied Statistics, Fourth Edition, Springer- Verlag, 2002. [12] J.M. Ver Hoef and R.P. Barry, Constructing and Fitting models for Cokriging and multivariate spatial prediction, Journal Statistics Plan Inference, 69, (1998), 275-294. [13] H. Wackernagel, Multivariate Geostatistics: An introduction with Application, 2 nd ed Berlin, Springer, 2003.