Stat 248 Lab 2: Stationarity, More EDA, Basic TS Models

Size: px
Start display at page:

Download "Stat 248 Lab 2: Stationarity, More EDA, Basic TS Models"

Transcription

1 Stat 248 Lab 2: Stationarity, More EDA, Basic TS Models Tessa L. Childers-Day February 8, Introduction Today s section will deal with topics such as: the mean function, the auto- and cross-covariance function, stationarity, and the sample acf. This material is drawn from chapters 1 and 2 from Shumway and Stoffer s book, which is available for free online through the university library system (there is a link on the section website). 2 Stochastic Processes and Stationarity A stochastic process is a collection of random variables {X i } i I. The set I over which the index ranges is called the index set. A time series is a type of stochastic process, where the index set I is a subset of the integers; i.e., time series data is of the form {X 1, X 2, X 3,...}. The index set I of a stochastic process may be quite general though such as continuous time (where I R) or spatial processes (where I R 2 or R 3 ). We often describe a time series (or stochastic process) by its first and second moments, the mean function and the covariance function. You can think of the covariance function as the average cross-product relative to the joint distribution f(x r, X s ). Definition: [mean function] The mean function is defined as: µ Xt = E(X t ) = xf t (x)dx (2.1) where F t (x) = P (X t x) and f t (x) = Ft(x) x. Note that F t(x) is a cdf, while f t (x) is a pdf. Definition: [autocovariance function (acvf)] If {X t, t T } is a process s.t. V ar(x t ) < for each t T, then its autocovariance function γ X (.,.) is defined as: γ X (r, s) = Cov(X r, X s ) = E[(X r E(X r ))(X s E(X s ))] r, s T (2.2) The autocovariance function measures the linear dependence between two points on the same series observed at different times. If X s and X r are independent then γ X (r, s) = 0. However, if γ X (r, s) = 0 then X s and X r are NOT necessarily independent. Remark: What happens when s=r? In theory, we observe many draws from the same distribution and can use these draws to make inferences about the underlying distribution. However, we typically only observe one realization of a time series. Stationarity is a key property that allows us to draw statistical inferences from a single stochastic process. Roughly, stationarity says that the distribution of {X t } looks the same at different locations t; this allows us to do statistical inference by treating the stochastic process as multiple correlated draws from the same distribution. 1

2 Definition: [strong stationarity] A stochastic process is called strong stationary or strictly stationary if the joint distributions of (X t1,..., X tk ) and (X t1+h,..., X tk +h) are the same for all positive integers k and for all t 1,..., t k, h Z. Strict stationarity means intuitively that the graphs over two equal-length time intervals of a realization of the time series should exhibit similar statistical characteristics. For example, the proportion of ordinates not exceeding a given level x should be roughly the same for both intervals. Definition: [(weak) stationarity]: The time series {X t, t Z} with index set Z = {0, ±1, ±2,...} is said to be (weakly) stationary if: 1. E X t 2 < t Z 2. E(X t ) = µ t Z 3. γ X (r, s) = γ X (r + t, s + t) r, s, t Z Intuitively a time series is stationary if it has a finite variance such that the mean value is constant and doesn t depend on time. The covariance function depends on s and t only through the difference s t. Remark: A strictly stationary process with finite second moments is stationary. true. The inverse is not necessarily Stationary processes play a crucial role in the analysis of time series. Of course many observed time series are decidedly non-stationary in appearance. Frequently such data sets can be transformed into series which can reasonably be modeled as realizations of some stationary process. The theory of stationary processes is then used for the analysis, fitting and prediction of the resulting series. In all of this the autocovariance function is a primary tool. Remark 1: Stationarity as defined is frequently referred to in the literature as weak stationarity, covariance stationarity, stationarity in the wide sense or second-order stationarity. For us when we refer to stationarity we will think about the three properties mentioned in the definition above. Remark 2: If {X t, t Z} is stationary then γ X (r, s) = γ X (r s, 0) for all r, s Z. It is therefore convenient to redefine the autocovariance function of a stationary process as the function of just one variable: γ X (h) = γ x (h, 0) = Cov(X t+h, X t ) t, h Z The function γ X (.) will be referred as the autocovariance function of {X t } and γ X (h) as its value at lag h. Note that γ X (s, t) = γ X (t, s) for all points s and t. Elementary properties of the covariance function. If γ(.) is the autocovariance function of a stationary process {X t, t Z}, then: γ(0) 0 γ(h) γ(0) γ(h) = γ( h) h Z h Z Definition: [autocorrelation function (acf)] The autocorrelation function of a stationary time series is the function whose value at lag h is: ρ X (h) = γ x(h) γ x (0) = Corr(X t+h, X t ) t, h Z (2.3) The Cauchy-Schwarz inequality shows that 1 ρ(h) 1 for all h. Further, ρ X (h) = 0 if X t and X t+h are not correlated, and ρ X (h) = ±1 if X t+h = α 0 + α 1 X t. The value ρ X (h) is a rough measure of the ability to forecast the series at time t + h from the value at time t. 2

3 Definition: [cross-covariance function (ccvf)] If {X t, t T } and {Y t, t T } are processes s.t. V ar(x t ) < & V ar(y t ) < for each t T, then the cross-covariance function γ XY (.,.) is defined as: γ XY (r, s) = Cov(X r, Y s ) = E[(X r E(X r ))(Y s E(Y s ))] r, s T (2.4) Of course there is also a scaled version of the cross-covariance function: Definition: [cross-correlation function (ccf)] The cross-correlation function of {X t, t T } and {Y t, t T } is defined as: ρ XY (s, t) = γ XY (s, t) γx (s, s)γ Y (t, t) s, t Z (2.5) 3 Estimation of Mean and Correlation Function As we just have one realization of our time series, the assumption of stationarity becomes critical. Somehow, we must use averages over this single realization to estimate the population means and covariance functions. If a time series is stationary, the mean function is constant µ t = µ and can be estimated by the sample mean. Definition: The sample mean is defined as: ˆµ = x t = 1 n n x t (3.1) Assuming stationarity, the autocovariance and autocorrelation function can be estimated using: t=1 Definition: The sample autocovariance function is defined as: n h ˆγ(h) = n 1 ((x t+h x)(x t x)) (3.2) t=1 with ˆγ( h) = ˆγ(h) for h = 0, 1,..., n 1. Dividing by n ensures that the function is non-negative definite (allowing for linear combinations of data to have non-negative variances), although it leads to a biased estimate of γ(h). Definition: The sample autocorrelation function is defined, analogously, as: ˆρ(h) = ˆγ(h) ˆγ(0) (3.3) The sample autocorrelation function has a sampling distribution that allows us to assess whether the data comes from a completely random or white series, or whether correlations are statistically significant at some lags Large sample distribution of the acf Under general conditions, if x t is white noise, then for large n, the sample acf, ˆρ X (h) for h = 1, 2,..., H, where H is fixed but arbitrary, is approximately normally distributed with zero mean and standard deviation given by: σˆρx(h) = 1 n Remark 1 Based on this result, we obtain a rough method of assessing whether peaks in ˆρ(h) are significant by determining whether the observed peak is outside the interval ±2/ n; for a white noise sequence, approximately 95% 3

4 of the sample acf s should be within these limits. After trying to reduce a time series to a white noise series the acf s of the residuals should then lie roughly within the limits given above. A plot of these values and the corresponding intervals is often called a correlogram. Remark 2 The sample autocovariance and autocorrelation functions can be computed for non-stationary process. For data containing a trend, ˆρ(h) will exhibit slow decay as h increases, and for data with a substantial deterministic periodic, ˆρ(h) will exhibit similar behavior with the same periodicity. Remark 3 The sample cross-covariance and sample cross-correlation function are defined analogously to the sample autocovariance and sample autocorrelation function. Correlogram of 'white noise' Correlogram of 'white noise with trend' Figure 1: Left: Sample acf of Gaussian white noise. Right: Sample acf of the series generated by X t = t + Z t, where Z t is Gaussian white noise (i.e. X t is white noise with a deterministic trend t). > white.noise = as.ts(rnorm(100)) > acf(white.noise, main = "Correlogram of white noise ") > t = 1:100 > Xt = t + white.noise > acf(xt, main = "Correlogram of white noise with trend ") 4 Some Basic Series Models 4.1 White Noise A sequence of uncorrelated random variables, {Z t ; t {1,..., n}}, with mean 0 and finite variance σz 2 is called a white noise process. It can be used as a model for noise in some engineering applications. The term white noise comes from the fact that a frequency analysis of this model shows that all frequencies enter equally, something that is also the case if you investigate the frequencies contained in white light. A white noise process will be denoted as Z t WN(0, σz 2 ). 4

5 Gaussian white noise (SD = 20) Correlogram white noise Zt Figure 2: Simulation (N=1000) of Gaussian white noise with σ Z = 20. > Zt = as.ts(rnorm(1000, sd = 20)) > par(mfrow = c(1,2)) > plot(zt, main = "Gaussian white noise (SD = 20)") > acf(zt, main = "Correlogram") Note that while the definition above fixes the first two moments of the random variables Z t, the rest of their distributions are left unspecified. It is even possible that Z i and Z j have different higher moments (and hence different distributions). However, sometimes we will require the random variables Z t, besides independent, also to be identically distributed with mean 0 and variance σz 2. In that case, we will denote the series as Z IID t (0, σ 2 Z ). A wide class of stationary processes can be generated by using white noise as the noise term in a set of linear difference equations. These processes are called autoregressive-moving average (ARMA), and will be discussed next week. 4.2 Moving Average If we replace the white noise series Z t by a moving average that smoothes the series, the resulting process is called a moving average. For example, let us consider a random variable V t that will average the current value with its immediate neighbors in the past and the future. That is, V t = 1 3 (Z t 1 + Z t + Z t+1 ) for t = 2, 3,..., n 1 (4.1) This is a 3-point moving average of the series Z t. If we look at the moving average it shows a smoother version of the first series, reflecting that the slower oscillations are more apparent and some of the faster oscillations are taken out. Note that in this example all involved points have equal weights. The use of different weights is possible as well, an example of that is the general MA(q)-model which will be discussed next week. 5

6 Moving 5 point average of above white noise series Correlogram 5 point moving average Vt Moving 21 point average of above white noise series Correlogram 21 point moving average Vt Figure 3: Moving averages of the white noise process of figure 1. Above: 5-points moving average. Below: 21-points moving average. > par(mfrow = c(2,2)) > Vt = filter(zt, rep(1/5, 5), sides = 2) > plot(vt, main = "Moving 5-point average of above white noise series") > acf(vt, main="correlogram 5-point moving average", lag.max=40, na.action=na.pass) > Vt = filter(zt, rep(1/21, 21), sides = 2) > plot(vt, main = "Moving 21-point average of above white noise series") > acf(vt, main="correlogram 21-point moving average", lag.max=40, na.action=na.pass) 4.3 Auto regressions Now, let us consider the process generated by the following equation: X t = φ 1 X t 1 + φ 2 X t 2 + Z t for t = 1, 2,..., n (4.2) Such a process is called a (second order) auto regression. Here, the current value of this series depends on the actual values at different time points. Actually, in the above example, the value X t depends on all previous shocks. (Why?) This equation represents a regression or prediction of the current value X t of a time series as a function of the past two values of the series. Note that along with above equation we should also provide startup values X 0 and X 1. 6

7 First order auto regression (phi = 0.9) Correlogram first order autoregression Xt Second order auto regression (phi_1 = 1, phi_2 = 0.8) Correlogram second order autoregression Xt Figure 4: Auto regressions with the white noise process of figure 1 as input, X 0 and X 1 are set to 0. Above: φ 1 = 0.8 and φ 2 = 0. Below: φ 1 = 0.8 and φ 2 = 0. > Xt1 = filter(zt, 0.9, method = "recursive") > Xt2 = filter(zt, c(1, -0.8), method = "recursive") > par(mfrow = c(2,2)) > plot(xt1, main = "First order auto regression (phi = 0.9)") > acf(xt1, main = "Correlogram first order autoregression") > plot(xt2, main = "Second order auto regression (phi1 = 1, phi2 = -0.8)") > acf(xt2, main = "Correlogram second order autoregression") 4.4 Random Walk If we take a first order auto regression with φ 1 = 1, the resulting process is called a random walk: Y t = Y t 1 + Z t for t = 1, 2,..., n, with initial condition Y 0 = 0 (4.3) The term random walk comes from the fact that the value of the time series at time t is the value of the series at time t 1 plus a completely random movement determined by Z t. We may rewrite it as a cumulative sum of white noise variates: Y t = t Z j for t = 1, 2,..., n (4.4) A variant of the random walk model is the random walk with drift, given by: j=1 Y t = δ + Y t 1 + Z t for t = 1, 2,..., n, with initial condition Y 0 = 0 (4.5) The constant δ is called the drift. This model can also be rewritten as a cumulative sum: Y t = δt + t Z j for t = 1, 2,..., n (4.6) j=1 7

8 The random walk with drift is clearly not stationary (why?), but what about the pure random walk process? This one is also not stationarity, which can be seen by calculating its variance, which turns out to be a function of the time t: > RW1 = as.ts(cumsum(zt)) V ar(y t ) = V ar( t Z j ) = tσz 2 for t = 1, 2,..., n (4.7) j=1 Random walk Correlogram random walk RW Random walk with drift (delta = 0.2) Correlogram random walk with drift RW Figure 5: Above: Simulated random walk. Below: Simulated random walk with drift δ = 0.2. > RW2 = 0.2*(1:length(Zt)) + as.ts(cumsum(zt)) > par(mfrow = c(2,2)) > plot(rw1, main = "Random walk") > acf(rw1, main = "Correlogram random walk") > plot(rw2, main = "Random walk with drift (delta = 0.2)") > acf(rw2, main = "Correlogram random walk with drift") 5 R functions mean(), acf(), acf(type = "covariance"), ccf(), ccf(type = "covariance"), filter(). 6 Bibliography This handout has benefitted by handouts prepared by Jack Kamm, Gido van de Ven, and Irma Hernandez-Magallanes, previous GSIs for this course. Additional sources that are used, and that could be useful for you: Series: Data Analysis and Theory by David R. Brillinger Series: Theory and Methods by Peter Brockwell & Richard Davis Series Analysis and Its Applications: With R Examples by Robert Schumway & David Stoffer 8

STAT 248: EDA & Stationarity Handout 3

STAT 248: EDA & Stationarity Handout 3 STAT 248: EDA & Stationarity Handout 3 GSI: Gido van de Ven September 17th, 2010 1 Introduction Today s section we will deal with the following topics: the mean function, the auto- and crosscovariance

More information

LECTURES 2-3 : Stochastic Processes, Autocorrelation function. Stationarity.

LECTURES 2-3 : Stochastic Processes, Autocorrelation function. Stationarity. LECTURES 2-3 : Stochastic Processes, Autocorrelation function. Stationarity. Important points of Lecture 1: A time series {X t } is a series of observations taken sequentially over time: x t is an observation

More information

STA 6857 Autocorrelation and Cross-Correlation & Stationary Time Series ( 1.4, 1.5)

STA 6857 Autocorrelation and Cross-Correlation & Stationary Time Series ( 1.4, 1.5) STA 6857 Autocorrelation and Cross-Correlation & Stationary Time Series ( 1.4, 1.5) Outline 1 Announcements 2 Autocorrelation and Cross-Correlation 3 Stationary Time Series 4 Homework 1c Arthur Berg STA

More information

Characteristics of Time Series

Characteristics of Time Series Characteristics of Time Series Al Nosedal University of Toronto January 12, 2016 Al Nosedal University of Toronto Characteristics of Time Series January 12, 2016 1 / 37 Signal and Noise In general, most

More information

Time Series Analysis -- An Introduction -- AMS 586

Time Series Analysis -- An Introduction -- AMS 586 Time Series Analysis -- An Introduction -- AMS 586 1 Objectives of time series analysis Data description Data interpretation Modeling Control Prediction & Forecasting 2 Time-Series Data Numerical data

More information

Some Time-Series Models

Some Time-Series Models Some Time-Series Models Outline 1. Stochastic processes and their properties 2. Stationary processes 3. Some properties of the autocorrelation function 4. Some useful models Purely random processes, random

More information

Part II. Time Series

Part II. Time Series Part II Time Series 12 Introduction This Part is mainly a summary of the book of Brockwell and Davis (2002). Additionally the textbook Shumway and Stoffer (2010) can be recommended. 1 Our purpose is to

More information

Econometría 2: Análisis de series de Tiempo

Econometría 2: Análisis de series de Tiempo Econometría 2: Análisis de series de Tiempo Karoll GOMEZ kgomezp@unal.edu.co http://karollgomez.wordpress.com Segundo semestre 2016 II. Basic definitions A time series is a set of observations X t, each

More information

A time series is called strictly stationary if the joint distribution of every collection (Y t

A time series is called strictly stationary if the joint distribution of every collection (Y t 5 Time series A time series is a set of observations recorded over time. You can think for example at the GDP of a country over the years (or quarters) or the hourly measurements of temperature over a

More information

Econ 424 Time Series Concepts

Econ 424 Time Series Concepts Econ 424 Time Series Concepts Eric Zivot January 20 2015 Time Series Processes Stochastic (Random) Process { 1 2 +1 } = { } = sequence of random variables indexed by time Observed time series of length

More information

Statistics of stochastic processes

Statistics of stochastic processes Introduction Statistics of stochastic processes Generally statistics is performed on observations y 1,..., y n assumed to be realizations of independent random variables Y 1,..., Y n. 14 settembre 2014

More information

Stochastic Processes: I. consider bowl of worms model for oscilloscope experiment:

Stochastic Processes: I. consider bowl of worms model for oscilloscope experiment: Stochastic Processes: I consider bowl of worms model for oscilloscope experiment: SAPAscope 2.0 / 0 1 RESET SAPA2e 22, 23 II 1 stochastic process is: Stochastic Processes: II informally: bowl + drawing

More information

ELEG 3143 Probability & Stochastic Process Ch. 6 Stochastic Process

ELEG 3143 Probability & Stochastic Process Ch. 6 Stochastic Process Department of Electrical Engineering University of Arkansas ELEG 3143 Probability & Stochastic Process Ch. 6 Stochastic Process Dr. Jingxian Wu wuj@uark.edu OUTLINE 2 Definition of stochastic process (random

More information

Lecture 2: Univariate Time Series

Lecture 2: Univariate Time Series Lecture 2: Univariate Time Series Analysis: Conditional and Unconditional Densities, Stationarity, ARMA Processes Prof. Massimo Guidolin 20192 Financial Econometrics Spring/Winter 2017 Overview Motivation:

More information

Reliability and Risk Analysis. Time Series, Types of Trend Functions and Estimates of Trends

Reliability and Risk Analysis. Time Series, Types of Trend Functions and Estimates of Trends Reliability and Risk Analysis Stochastic process The sequence of random variables {Y t, t = 0, ±1, ±2 } is called the stochastic process The mean function of a stochastic process {Y t} is the function

More information

white noise Time moving average

white noise Time moving average 1.3 Time Series Statistical Models 13 white noise w 3 1 0 1 0 100 00 300 400 500 Time moving average v 1.5 0.5 0.5 1.5 0 100 00 300 400 500 Fig. 1.8. Gaussian white noise series (top) and three-point moving

More information

TIME SERIES AND FORECASTING. Luca Gambetti UAB, Barcelona GSE Master in Macroeconomic Policy and Financial Markets

TIME SERIES AND FORECASTING. Luca Gambetti UAB, Barcelona GSE Master in Macroeconomic Policy and Financial Markets TIME SERIES AND FORECASTING Luca Gambetti UAB, Barcelona GSE 2014-2015 Master in Macroeconomic Policy and Financial Markets 1 Contacts Prof.: Luca Gambetti Office: B3-1130 Edifici B Office hours: email:

More information

Nonlinear time series

Nonlinear time series Based on the book by Fan/Yao: Nonlinear Time Series Robert M. Kunst robert.kunst@univie.ac.at University of Vienna and Institute for Advanced Studies Vienna October 27, 2009 Outline Characteristics of

More information

Time Series I Time Domain Methods

Time Series I Time Domain Methods Astrostatistics Summer School Penn State University University Park, PA 16802 May 21, 2007 Overview Filtering and the Likelihood Function Time series is the study of data consisting of a sequence of DEPENDENT

More information

On 1.9, you will need to use the facts that, for any x and y, sin(x+y) = sin(x) cos(y) + cos(x) sin(y). cos(x+y) = cos(x) cos(y) - sin(x) sin(y).

On 1.9, you will need to use the facts that, for any x and y, sin(x+y) = sin(x) cos(y) + cos(x) sin(y). cos(x+y) = cos(x) cos(y) - sin(x) sin(y). On 1.9, you will need to use the facts that, for any x and y, sin(x+y) = sin(x) cos(y) + cos(x) sin(y). cos(x+y) = cos(x) cos(y) - sin(x) sin(y). (sin(x)) 2 + (cos(x)) 2 = 1. 28 1 Characteristics of Time

More information

STAT 443 Final Exam Review. 1 Basic Definitions. 2 Statistical Tests. L A TEXer: W. Kong

STAT 443 Final Exam Review. 1 Basic Definitions. 2 Statistical Tests. L A TEXer: W. Kong STAT 443 Final Exam Review L A TEXer: W Kong 1 Basic Definitions Definition 11 The time series {X t } with E[X 2 t ] < is said to be weakly stationary if: 1 µ X (t) = E[X t ] is independent of t 2 γ X

More information

Prof. Dr. Roland Füss Lecture Series in Applied Econometrics Summer Term Introduction to Time Series Analysis

Prof. Dr. Roland Füss Lecture Series in Applied Econometrics Summer Term Introduction to Time Series Analysis Introduction to Time Series Analysis 1 Contents: I. Basics of Time Series Analysis... 4 I.1 Stationarity... 5 I.2 Autocorrelation Function... 9 I.3 Partial Autocorrelation Function (PACF)... 14 I.4 Transformation

More information

Econ 623 Econometrics II Topic 2: Stationary Time Series

Econ 623 Econometrics II Topic 2: Stationary Time Series 1 Introduction Econ 623 Econometrics II Topic 2: Stationary Time Series In the regression model we can model the error term as an autoregression AR(1) process. That is, we can use the past value of the

More information

3. ARMA Modeling. Now: Important class of stationary processes

3. ARMA Modeling. Now: Important class of stationary processes 3. ARMA Modeling Now: Important class of stationary processes Definition 3.1: (ARMA(p, q) process) Let {ɛ t } t Z WN(0, σ 2 ) be a white noise process. The process {X t } t Z is called AutoRegressive-Moving-Average

More information

Applied time-series analysis

Applied time-series analysis Robert M. Kunst robert.kunst@univie.ac.at University of Vienna and Institute for Advanced Studies Vienna October 18, 2011 Outline Introduction and overview Econometric Time-Series Analysis In principle,

More information

For a stochastic process {Y t : t = 0, ±1, ±2, ±3, }, the mean function is defined by (2.2.1) ± 2..., γ t,

For a stochastic process {Y t : t = 0, ±1, ±2, ±3, }, the mean function is defined by (2.2.1) ± 2..., γ t, CHAPTER 2 FUNDAMENTAL CONCEPTS This chapter describes the fundamental concepts in the theory of time series models. In particular, we introduce the concepts of stochastic processes, mean and covariance

More information

1 Linear Difference Equations

1 Linear Difference Equations ARMA Handout Jialin Yu 1 Linear Difference Equations First order systems Let {ε t } t=1 denote an input sequence and {y t} t=1 sequence generated by denote an output y t = φy t 1 + ε t t = 1, 2,... with

More information

Stochastic Processes

Stochastic Processes Stochastic Processes Stochastic Process Non Formal Definition: Non formal: A stochastic process (random process) is the opposite of a deterministic process such as one defined by a differential equation.

More information

STOR 356: Summary Course Notes

STOR 356: Summary Course Notes STOR 356: Summary Course Notes Richard L. Smith Department of Statistics and Operations Research University of North Carolina Chapel Hill, NC 7599-360 rls@email.unc.edu February 19, 008 Course text: Introduction

More information

Covariances of ARMA Processes

Covariances of ARMA Processes Statistics 910, #10 1 Overview Covariances of ARMA Processes 1. Review ARMA models: causality and invertibility 2. AR covariance functions 3. MA and ARMA covariance functions 4. Partial autocorrelation

More information

STAT Financial Time Series

STAT Financial Time Series STAT 6104 - Financial Time Series Chapter 2 - Probability Models Chun Yip Yau (CUHK) STAT 6104:Financial Time Series 1 / 34 Agenda 1 Introduction 2 Stochastic Process Definition 1 Stochastic Definition

More information

{ } Stochastic processes. Models for time series. Specification of a process. Specification of a process. , X t3. ,...X tn }

{ } Stochastic processes. Models for time series. Specification of a process. Specification of a process. , X t3. ,...X tn } Stochastic processes Time series are an example of a stochastic or random process Models for time series A stochastic process is 'a statistical phenomenon that evolves in time according to probabilistic

More information

Introduction to Stochastic processes

Introduction to Stochastic processes Università di Pavia Introduction to Stochastic processes Eduardo Rossi Stochastic Process Stochastic Process: A stochastic process is an ordered sequence of random variables defined on a probability space

More information

Autoregressive Moving Average (ARMA) Models and their Practical Applications

Autoregressive Moving Average (ARMA) Models and their Practical Applications Autoregressive Moving Average (ARMA) Models and their Practical Applications Massimo Guidolin February 2018 1 Essential Concepts in Time Series Analysis 1.1 Time Series and Their Properties Time series:

More information

Multivariate Time Series

Multivariate Time Series Multivariate Time Series Notation: I do not use boldface (or anything else) to distinguish vectors from scalars. Tsay (and many other writers) do. I denote a multivariate stochastic process in the form

More information

Lesson 4: Stationary stochastic processes

Lesson 4: Stationary stochastic processes Dipartimento di Ingegneria e Scienze dell Informazione e Matematica Università dell Aquila, umberto.triacca@univaq.it Stationary stochastic processes Stationarity is a rather intuitive concept, it means

More information

Class 1: Stationary Time Series Analysis

Class 1: Stationary Time Series Analysis Class 1: Stationary Time Series Analysis Macroeconometrics - Fall 2009 Jacek Suda, BdF and PSE February 28, 2011 Outline Outline: 1 Covariance-Stationary Processes 2 Wold Decomposition Theorem 3 ARMA Models

More information

Classic Time Series Analysis

Classic Time Series Analysis Classic Time Series Analysis Concepts and Definitions Let Y be a random number with PDF f Y t ~f,t Define t =E[Y t ] m(t) is known as the trend Define the autocovariance t, s =COV [Y t,y s ] =E[ Y t t

More information

STAT 153, FALL 2015 HOMEWORK 1 SOLUTIONS

STAT 153, FALL 2015 HOMEWORK 1 SOLUTIONS STAT 153, FALL 015 HOMEWORK 1 SOLUTIONS Problem 1 (a) A solution in R is provided in Figure 1. We conclude that if δ = 0 the process is a random walk without drift. When δ 0 there is a linear drift with

More information

Time Series 2. Robert Almgren. Sept. 21, 2009

Time Series 2. Robert Almgren. Sept. 21, 2009 Time Series 2 Robert Almgren Sept. 21, 2009 This week we will talk about linear time series models: AR, MA, ARMA, ARIMA, etc. First we will talk about theory and after we will talk about fitting the models

More information

Discrete time processes

Discrete time processes Discrete time processes Predictions are difficult. Especially about the future Mark Twain. Florian Herzog 2013 Modeling observed data When we model observed (realized) data, we encounter usually the following

More information

1. Stochastic Processes and Stationarity

1. Stochastic Processes and Stationarity Massachusetts Institute of Technology Department of Economics Time Series 14.384 Guido Kuersteiner Lecture Note 1 - Introduction This course provides the basic tools needed to analyze data that is observed

More information

AR, MA and ARMA models

AR, MA and ARMA models AR, MA and AR by Hedibert Lopes P Based on Tsay s Analysis of Financial Time Series (3rd edition) P 1 Stationarity 2 3 4 5 6 7 P 8 9 10 11 Outline P Linear Time Series Analysis and Its Applications For

More information

Stochastic Processes. A stochastic process is a function of two variables:

Stochastic Processes. A stochastic process is a function of two variables: Stochastic Processes Stochastic: from Greek stochastikos, proceeding by guesswork, literally, skillful in aiming. A stochastic process is simply a collection of random variables labelled by some parameter:

More information

Module 9: Stationary Processes

Module 9: Stationary Processes Module 9: Stationary Processes Lecture 1 Stationary Processes 1 Introduction A stationary process is a stochastic process whose joint probability distribution does not change when shifted in time or space.

More information

ECON 616: Lecture 1: Time Series Basics

ECON 616: Lecture 1: Time Series Basics ECON 616: Lecture 1: Time Series Basics ED HERBST August 30, 2017 References Overview: Chapters 1-3 from Hamilton (1994). Technical Details: Chapters 2-3 from Brockwell and Davis (1987). Intuition: Chapters

More information

Time Series Models and Inference. James L. Powell Department of Economics University of California, Berkeley

Time Series Models and Inference. James L. Powell Department of Economics University of California, Berkeley Time Series Models and Inference James L. Powell Department of Economics University of California, Berkeley Overview In contrast to the classical linear regression model, in which the components of the

More information

1 Class Organization. 2 Introduction

1 Class Organization. 2 Introduction Time Series Analysis, Lecture 1, 2018 1 1 Class Organization Course Description Prerequisite Homework and Grading Readings and Lecture Notes Course Website: http://www.nanlifinance.org/teaching.html wechat

More information

ESSE Mid-Term Test 2017 Tuesday 17 October :30-09:45

ESSE Mid-Term Test 2017 Tuesday 17 October :30-09:45 ESSE 4020 3.0 - Mid-Term Test 207 Tuesday 7 October 207. 08:30-09:45 Symbols have their usual meanings. All questions are worth 0 marks, although some are more difficult than others. Answer as many questions

More information

13. Time Series Analysis: Asymptotics Weakly Dependent and Random Walk Process. Strict Exogeneity

13. Time Series Analysis: Asymptotics Weakly Dependent and Random Walk Process. Strict Exogeneity Outline: Further Issues in Using OLS with Time Series Data 13. Time Series Analysis: Asymptotics Weakly Dependent and Random Walk Process I. Stationary and Weakly Dependent Time Series III. Highly Persistent

More information

Module 3. Descriptive Time Series Statistics and Introduction to Time Series Models

Module 3. Descriptive Time Series Statistics and Introduction to Time Series Models Module 3 Descriptive Time Series Statistics and Introduction to Time Series Models Class notes for Statistics 451: Applied Time Series Iowa State University Copyright 2015 W Q Meeker November 11, 2015

More information

If we want to analyze experimental or simulated data we might encounter the following tasks:

If we want to analyze experimental or simulated data we might encounter the following tasks: Chapter 1 Introduction If we want to analyze experimental or simulated data we might encounter the following tasks: Characterization of the source of the signal and diagnosis Studying dependencies Prediction

More information

Advanced Econometrics

Advanced Econometrics Advanced Econometrics Marco Sunder Nov 04 2010 Marco Sunder Advanced Econometrics 1/ 25 Contents 1 2 3 Marco Sunder Advanced Econometrics 2/ 25 Music Marco Sunder Advanced Econometrics 3/ 25 Music Marco

More information

Introduction to Economic Time Series

Introduction to Economic Time Series Econometrics II Introduction to Economic Time Series Morten Nyboe Tabor Learning Goals 1 Give an account for the important differences between (independent) cross-sectional data and time series data. 2

More information

Gaussian processes. Basic Properties VAG002-

Gaussian processes. Basic Properties VAG002- Gaussian processes The class of Gaussian processes is one of the most widely used families of stochastic processes for modeling dependent data observed over time, or space, or time and space. The popularity

More information

Problem Set 1 Solution Sketches Time Series Analysis Spring 2010

Problem Set 1 Solution Sketches Time Series Analysis Spring 2010 Problem Set 1 Solution Sketches Time Series Analysis Spring 2010 1. Construct a martingale difference process that is not weakly stationary. Simplest e.g.: Let Y t be a sequence of independent, non-identically

More information

2. An Introduction to Moving Average Models and ARMA Models

2. An Introduction to Moving Average Models and ARMA Models . An Introduction to Moving Average Models and ARMA Models.1 White Noise. The MA(1) model.3 The MA(q) model..4 Estimation and forecasting of MA models..5 ARMA(p,q) models. The Moving Average (MA) models

More information

ENSC327 Communications Systems 19: Random Processes. Jie Liang School of Engineering Science Simon Fraser University

ENSC327 Communications Systems 19: Random Processes. Jie Liang School of Engineering Science Simon Fraser University ENSC327 Communications Systems 19: Random Processes Jie Liang School of Engineering Science Simon Fraser University 1 Outline Random processes Stationary random processes Autocorrelation of random processes

More information

Lecture 4a: ARMA Model

Lecture 4a: ARMA Model Lecture 4a: ARMA Model 1 2 Big Picture Most often our goal is to find a statistical model to describe real time series (estimation), and then predict the future (forecasting) One particularly popular model

More information

Lecture 3: Autoregressive Moving Average (ARMA) Models and their Practical Applications

Lecture 3: Autoregressive Moving Average (ARMA) Models and their Practical Applications Lecture 3: Autoregressive Moving Average (ARMA) Models and their Practical Applications Prof. Massimo Guidolin 20192 Financial Econometrics Winter/Spring 2018 Overview Moving average processes Autoregressive

More information

Lecture 1: Fundamental concepts in Time Series Analysis (part 2)

Lecture 1: Fundamental concepts in Time Series Analysis (part 2) Lecture 1: Fundamental concepts in Time Series Analysis (part 2) Florian Pelgrin University of Lausanne, École des HEC Department of mathematics (IMEA-Nice) Sept. 2011 - Jan. 2012 Florian Pelgrin (HEC)

More information

Topic 4 Unit Roots. Gerald P. Dwyer. February Clemson University

Topic 4 Unit Roots. Gerald P. Dwyer. February Clemson University Topic 4 Unit Roots Gerald P. Dwyer Clemson University February 2016 Outline 1 Unit Roots Introduction Trend and Difference Stationary Autocorrelations of Series That Have Deterministic or Stochastic Trends

More information

Time Series: Theory and Methods

Time Series: Theory and Methods Peter J. Brockwell Richard A. Davis Time Series: Theory and Methods Second Edition With 124 Illustrations Springer Contents Preface to the Second Edition Preface to the First Edition vn ix CHAPTER 1 Stationary

More information

Minitab Project Report Assignment 3

Minitab Project Report Assignment 3 3.1.1 Simulation of Gaussian White Noise Minitab Project Report Assignment 3 Time Series Plot of zt Function zt 1 0. 0. zt 0-1 0. 0. -0. -0. - -3 1 0 30 0 50 Index 0 70 0 90 0 1 1 1 1 0 marks The series

More information

Week 5 Quantitative Analysis of Financial Markets Characterizing Cycles

Week 5 Quantitative Analysis of Financial Markets Characterizing Cycles Week 5 Quantitative Analysis of Financial Markets Characterizing Cycles Christopher Ting http://www.mysmu.edu/faculty/christophert/ Christopher Ting : christopherting@smu.edu.sg : 6828 0364 : LKCSB 5036

More information

11. Further Issues in Using OLS with TS Data

11. Further Issues in Using OLS with TS Data 11. Further Issues in Using OLS with TS Data With TS, including lags of the dependent variable often allow us to fit much better the variation in y Exact distribution theory is rarely available in TS applications,

More information

Chapter 4: Models for Stationary Time Series

Chapter 4: Models for Stationary Time Series Chapter 4: Models for Stationary Time Series Now we will introduce some useful parametric models for time series that are stationary processes. We begin by defining the General Linear Process. Let {Y t

More information

SOME BASICS OF TIME-SERIES ANALYSIS

SOME BASICS OF TIME-SERIES ANALYSIS SOME BASICS OF TIME-SERIES ANALYSIS John E. Floyd University of Toronto December 8, 26 An excellent place to learn about time series analysis is from Walter Enders textbook. For a basic understanding of

More information

Homework 2 Solutions

Homework 2 Solutions Math 506 Spring 201 Homework 2 Solutions 1. Textbook Problem 2.7: Since {Y t } is stationary, E[Y t ] = µ Y for all t and {Y t } has an autocovariance function γ Y. Therefore, (a) E[W t ] = E[Y t Y t 1

More information

Chapter 12: An introduction to Time Series Analysis. Chapter 12: An introduction to Time Series Analysis

Chapter 12: An introduction to Time Series Analysis. Chapter 12: An introduction to Time Series Analysis Chapter 12: An introduction to Time Series Analysis Introduction In this chapter, we will discuss forecasting with single-series (univariate) Box-Jenkins models. The common name of the models is Auto-Regressive

More information

1. Fundamental concepts

1. Fundamental concepts . Fundamental concepts A time series is a sequence of data points, measured typically at successive times spaced at uniform intervals. Time series are used in such fields as statistics, signal processing

More information

Time Series Analysis. tsaez. Using the R Statistical Package.

Time Series Analysis. tsaez. Using the R Statistical Package. Time Series Analysis Using the R Statistical Package tsaez http://www.stat.pitt.edu/stoffer/tsa4/ Copyright 2017 by R.H. Shumway & D.S. Stoffer Published by free dog publishing This work is licensed under

More information

Lecture 3 Stationary Processes and the Ergodic LLN (Reference Section 2.2, Hayashi)

Lecture 3 Stationary Processes and the Ergodic LLN (Reference Section 2.2, Hayashi) Lecture 3 Stationary Processes and the Ergodic LLN (Reference Section 2.2, Hayashi) Our immediate goal is to formulate an LLN and a CLT which can be applied to establish sufficient conditions for the consistency

More information

Introduction to ARMA and GARCH processes

Introduction to ARMA and GARCH processes Introduction to ARMA and GARCH processes Fulvio Corsi SNS Pisa 3 March 2010 Fulvio Corsi Introduction to ARMA () and GARCH processes SNS Pisa 3 March 2010 1 / 24 Stationarity Strict stationarity: (X 1,

More information

White Noise Processes (Section 6.2)

White Noise Processes (Section 6.2) White Noise Processes (Section 6.) Recall that covariance stationary processes are time series, y t, such. E(y t ) = µ for all t. Var(y t ) = σ for all t, σ < 3. Cov(y t,y t-τ ) = γ(τ) for all t and τ

More information

STAT 520: Forecasting and Time Series. David B. Hitchcock University of South Carolina Department of Statistics

STAT 520: Forecasting and Time Series. David B. Hitchcock University of South Carolina Department of Statistics David B. University of South Carolina Department of Statistics What are Time Series Data? Time series data are collected sequentially over time. Some common examples include: 1. Meteorological data (temperatures,

More information

Stationary Stochastic Time Series Models

Stationary Stochastic Time Series Models Stationary Stochastic Time Series Models When modeling time series it is useful to regard an observed time series, (x 1,x,..., x n ), as the realisation of a stochastic process. In general a stochastic

More information

Lecture 2: ARMA(p,q) models (part 2)

Lecture 2: ARMA(p,q) models (part 2) Lecture 2: ARMA(p,q) models (part 2) Florian Pelgrin University of Lausanne, École des HEC Department of mathematics (IMEA-Nice) Sept. 2011 - Jan. 2012 Florian Pelgrin (HEC) Univariate time series Sept.

More information

STAT 443 (Winter ) Forecasting

STAT 443 (Winter ) Forecasting Winter 2014 TABLE OF CONTENTS STAT 443 (Winter 2014-1141) Forecasting Prof R Ramezan University of Waterloo L A TEXer: W KONG http://wwkonggithubio Last Revision: September 3, 2014 Table of Contents 1

More information

Introduction to Signal Processing

Introduction to Signal Processing to Signal Processing Davide Bacciu Dipartimento di Informatica Università di Pisa bacciu@di.unipi.it Intelligent Systems for Pattern Recognition Signals = Time series Definitions Motivations A sequence

More information

Basics: Definitions and Notation. Stationarity. A More Formal Definition

Basics: Definitions and Notation. Stationarity. A More Formal Definition Basics: Definitions and Notation A Univariate is a sequence of measurements of the same variable collected over (usually regular intervals of) time. Usual assumption in many time series techniques is that

More information

Notes on Random Processes

Notes on Random Processes otes on Random Processes Brian Borchers and Rick Aster October 27, 2008 A Brief Review of Probability In this section of the course, we will work with random variables which are denoted by capital letters,

More information

Stochastic process for macro

Stochastic process for macro Stochastic process for macro Tianxiao Zheng SAIF 1. Stochastic process The state of a system {X t } evolves probabilistically in time. The joint probability distribution is given by Pr(X t1, t 1 ; X t2,

More information

University of Oxford. Statistical Methods Autocorrelation. Identification and Estimation

University of Oxford. Statistical Methods Autocorrelation. Identification and Estimation University of Oxford Statistical Methods Autocorrelation Identification and Estimation Dr. Órlaith Burke Michaelmas Term, 2011 Department of Statistics, 1 South Parks Road, Oxford OX1 3TG Contents 1 Model

More information

Introduction. Spatial Processes & Spatial Patterns

Introduction. Spatial Processes & Spatial Patterns Introduction Spatial data: set of geo-referenced attribute measurements: each measurement is associated with a location (point) or an entity (area/region/object) in geographical (or other) space; the domain

More information

7 Introduction to Time Series Time Series vs. Cross-Sectional Data Detrending Time Series... 15

7 Introduction to Time Series Time Series vs. Cross-Sectional Data Detrending Time Series... 15 Econ 495 - Econometric Review 1 Contents 7 Introduction to Time Series 3 7.1 Time Series vs. Cross-Sectional Data............ 3 7.2 Detrending Time Series................... 15 7.3 Types of Stochastic

More information

Ch. 14 Stationary ARMA Process

Ch. 14 Stationary ARMA Process Ch. 14 Stationary ARMA Process A general linear stochastic model is described that suppose a time series to be generated by a linear aggregation of random shock. For practical representation it is desirable

More information

Part III Example Sheet 1 - Solutions YC/Lent 2015 Comments and corrections should be ed to

Part III Example Sheet 1 - Solutions YC/Lent 2015 Comments and corrections should be  ed to TIME SERIES Part III Example Sheet 1 - Solutions YC/Lent 2015 Comments and corrections should be emailed to Y.Chen@statslab.cam.ac.uk. 1. Let {X t } be a weakly stationary process with mean zero and let

More information

Time Series Analysis. tsaez. Using the R Statistical Package.

Time Series Analysis. tsaez. Using the R Statistical Package. Time Series Analysis Using the R Statistical Package tsaez http://www.stat.pitt.edu/stoffer/tsa4/ Copyright 2018 by R.H. Shumway & D.S. Stoffer Published by free dog publishing live free or bark This work

More information

Chapter 2. Some basic tools. 2.1 Time series: Theory Stochastic processes

Chapter 2. Some basic tools. 2.1 Time series: Theory Stochastic processes Chapter 2 Some basic tools 2.1 Time series: Theory 2.1.1 Stochastic processes A stochastic process is a sequence of random variables..., x 0, x 1, x 2,.... In this class, the subscript always means time.

More information

Contents. 1 Time Series Analysis Introduction Stationary Processes State Space Modesl Stationary Processes 8

Contents. 1 Time Series Analysis Introduction Stationary Processes State Space Modesl Stationary Processes 8 A N D R E W T U L L O C H T I M E S E R I E S A N D M O N T E C A R L O I N F E R E N C E T R I N I T Y C O L L E G E T H E U N I V E R S I T Y O F C A M B R I D G E Contents 1 Time Series Analysis 5

More information

ECON3327: Financial Econometrics, Spring 2016

ECON3327: Financial Econometrics, Spring 2016 ECON3327: Financial Econometrics, Spring 2016 Wooldridge, Introductory Econometrics (5th ed, 2012) Chapter 11: OLS with time series data Stationary and weakly dependent time series The notion of a stationary

More information

Lesson 7: Estimation of Autocorrelation and Partial Autocorrela

Lesson 7: Estimation of Autocorrelation and Partial Autocorrela Lesson 7: Estimation of Autocorrelation and Partial Autocorrelation Function Dipartimento di Ingegneria e Scienze dell Informazione e Matematica Università dell Aquila, umberto.triacca@ec.univaq.it Estimation

More information

Marcel Dettling. Applied Time Series Analysis SS 2013 Week 05. ETH Zürich, March 18, Institute for Data Analysis and Process Design

Marcel Dettling. Applied Time Series Analysis SS 2013 Week 05. ETH Zürich, March 18, Institute for Data Analysis and Process Design Marcel Dettling Institute for Data Analysis and Process Design Zurich University of Applied Sciences marcel.dettling@zhaw.ch http://stat.ethz.ch/~dettling ETH Zürich, March 18, 2013 1 Basics of Modeling

More information

Empirical Market Microstructure Analysis (EMMA)

Empirical Market Microstructure Analysis (EMMA) Empirical Market Microstructure Analysis (EMMA) Lecture 3: Statistical Building Blocks and Econometric Basics Prof. Dr. Michael Stein michael.stein@vwl.uni-freiburg.de Albert-Ludwigs-University of Freiburg

More information

Time Series Examples Sheet

Time Series Examples Sheet Lent Term 2001 Richard Weber Time Series Examples Sheet This is the examples sheet for the M. Phil. course in Time Series. A copy can be found at: http://www.statslab.cam.ac.uk/~rrw1/timeseries/ Throughout,

More information

Chapter 6. Random Processes

Chapter 6. Random Processes Chapter 6 Random Processes Random Process A random process is a time-varying function that assigns the outcome of a random experiment to each time instant: X(t). For a fixed (sample path): a random process

More information

EASTERN MEDITERRANEAN UNIVERSITY ECON 604, FALL 2007 DEPARTMENT OF ECONOMICS MEHMET BALCILAR ARIMA MODELS: IDENTIFICATION

EASTERN MEDITERRANEAN UNIVERSITY ECON 604, FALL 2007 DEPARTMENT OF ECONOMICS MEHMET BALCILAR ARIMA MODELS: IDENTIFICATION ARIMA MODELS: IDENTIFICATION A. Autocorrelations and Partial Autocorrelations 1. Summary of What We Know So Far: a) Series y t is to be modeled by Box-Jenkins methods. The first step was to convert y t

More information

7 Introduction to Time Series

7 Introduction to Time Series Econ 495 - Econometric Review 1 7 Introduction to Time Series 7.1 Time Series vs. Cross-Sectional Data Time series data has a temporal ordering, unlike cross-section data, we will need to changes some

More information

Forecasting. Simon Shaw 2005/06 Semester II

Forecasting. Simon Shaw 2005/06 Semester II Forecasting Simon Shaw s.c.shaw@maths.bath.ac.uk 2005/06 Semester II 1 Introduction A critical aspect of managing any business is planning for the future. events is called forecasting. Predicting future

More information