Homework 2 Solutions
|
|
- Daniel Stewart
- 5 years ago
- Views:
Transcription
1 Math 506 Spring 201 Homework 2 Solutions 1. Textbook Problem 2.7: Since {Y t } is stationary, E[Y t ] = µ Y for all t and {Y t } has an autocovariance function γ Y. Therefore, (a) E[W t ] = E[Y t Y t 1 ] = E[Y t ] E[Y t 1 ] = µ Y µ Y = 0. (1) Cov(W t, W t+h ) = Cov(Y t Y t 1, Y t+h Y t+h 1 ) = Cov(Y t, Y t+h ) Cov(Y t, Y t+h 1 ) Cov(Y t 1, Y t+h ) + Cov(Y t 1, Y t+h 1 ) = γ Y (h) γ Y (h 1) γ Y (h + 1) + γ Y (h) = 2γ Y (h) γ Y (h 1) γ Y (h + 1), which depends on h but not on t, so the autocorrelation function of {W t } is γ W (h) = 2γ Y (h) γ Y (h 1) γ Y (h + 1). (2) Equations (1) and (2) imply that {W t } is stationary. (b) We showed in (a) that if {Y t } is stationary, so is { Y t }. Therefore, since by (a), { Y t } is stationary, so is {U t } = { ( Y t )} = { 2 Y t }. 2. Textbook Problem 2.2: Since {X t } is stationary, we can write E[X t ] = µ X and Cov(X t, X t+h ) = γ X (h). We first note that E[Y t ] = E[X t + e t ] = E[X t ] + E[e t ] = E[X t ] + 0 = E[X t ] = µ X. (3) Moreover, since {X t } and {e t } are independent processes (which means that X t is independent of e s for all t, s), we have γ Y (h) = Cov(Y t, Y t+h ) = Cov(X t + e t, X t+h + e t+h ) = Cov(X t, X t+h ) + Cov(X t, e t+h ) + Cov(e t, X t+h ) + Cov(e t, e t+h ) = γ X (h) γ e (h) = γ X (h) + γ e (h) = γ X (h) + σ 2 eδ(h), () where the last equality follows from the fact that {e t } W N(0, σe). 2 Note that this last expression in () depends only on h, not on t, which, together with (3) implies that {Y t } is stationary. All the above implies that if we write γ X (0) = σx 2, then for h 1, ρ Y (h) = γ Y (h) γ Y (0) = γ X(h) γ X (0) + σ 2 e = γ X(h)/γ X (0) 1 + σe/γ 2 X (0) = ρ X(h). 1 + σ2 e σx 2
2 3. Since X t = φx t 1 + Z t, we have X 2 = φx 1 + Z 2 X 3 = φx 2 + Z 3 = φ 2 X 1 + φz 2 + Z 3 X = φx 3 + Z = φ 3 X 1 + φ 2 Z 2 + φz 3 + Z, so that X 1 + X 2 + X 3 + X = 1 ( ) (1 + φ + φ 2 + φ 3 )X 1 + (1 + φ + φ 2 )Z 2 + (1 + φ)z 3 + Z. Therefore, since X 1, Z 2, Z 3, and Z are independent, we get ( ) X1 + X 2 + X 3 + X Var = 1 ( (1 + φ + φ 2 + φ 3 ) 2 Var X (1 + φ + φ 2 ) 2 Var Z 2 + (1 + φ) 2 Var Z 3 + Var Z ). Using the fact (derived in class) that Var X i = σ2 and that Var Z 1 φ 2 i = σ 2, we get ( ) ( ) X1 + X 2 + X 3 + X Var = σ2 (1 + φ + φ 2 + φ 3 ) 2 + (1 + φ + φ 2 ) 2 + (1 + φ) φ 2 If σ 2 = 1 and φ = 0.9, we get ( ) X1 + X 2 + X 3 + X Var =.6375 and if σ 2 = 1 and φ = 0.9, we get ( ) X1 + X 2 + X 3 + X Var When φ = 0.9, the lag-1 auto-correlation is 0.9, so if X t is positive, one is more likely to see a positive value of X t 1 as well. This means that such AR processes have more variability in the sense that some realizations will spend longer periods of time above the horizontal axis (respectively, below the horizontal axis) than if φ = 0.9 when the process will have a tendency to oscillate between positive and negative values. The sample mean of such an oscillating process (corresponding to negative φ) will typically be close to 0, while the sample mean might be farther from 0 (either positive or negative) when there are runs of positive or negative values, implying more variability in the sample mean in the latter case. This is reflected by the larger value obtained above in the case where φ = 0.9 than when φ = If {X t } and {Y t } are stationary, it means that there exist constants µ X and µ Y such that for any t Z, E[X t ] = µ X and E[Y t ] = µ Y. Also, there exist two functions γ X and γ Y such that for all t, h Z, γ X (h) = Cov(X t+h, X t ) and γ Y (h) = Cov(Y t+h, Y t ). Therefore, for any t Z, E[X t + Y t ] = E[X t ] + E[Y t ] = µ X + µ Y, (5) which is a constant, and since {X t } and {Y t } are uncorrelated, Cov(X t+h, Y t ) = Cov(Y t+h, X t ) = 0 for any t, h Z, so that Cov(X t+h + Y t+h, X t + Y t ) = Cov(X t+h, X t ) + Cov(X t+h, Y t ) Cov(Y t+h, X t ) + Cov(Y t+h, Y t ) = γ X (h) γ Y (h) = γ X (h) + γ Y (h), which depends only on h. This together with (5) implies that {X t + Y t } is stationary.
3 5. Textbook Problem.5: We can draw these ACFs of the four AR processes using the commands below (one might need to fiddle around a bit to find an appropriate maximal lag). Note that the command abline(h=0) adds a horizontal line at height 0, which makes the graphs a bit easier to understand. Note also that when using the command ARMAacf, R draws the ACF shifted to the right by one unit (starting to index at 1, as opposed to 0). One way around this is to shift the graph of the ACF drawn by R to the left by one unit using the command ACF1[-1] as below. (a) > ACF1=ARMAacf(ar=0.6,lag.max=12) > plot(x=1:12,y=acf1[-1],type= h ); abline(h=0) ACF1[-1] :12 (b) > ACF2=ARMAacf(ar=-0.6,lag.max=12) > plot(x=1:12,y=acf2[-1],type= h ); abline(h=0) ACF2[-1] :12 (c) > ACF3=ARMAacf(ar=0.95,lag.max=20) > plot(x=1:20,y=acf3[-1],type= h ); abline(h=0) ACF3[-1] :20
4 (d) > ACF=ARMAacf(ar=0.3,lag.max=8) > plot(x=1:8,y=acf[-1],type= h ); abline(h=0) ACF[-1] :8 6. Textbook Problem.9: (a) We are considering the time series defined by Y t = 0.6Y t Y t 2 + Z t with characteristic equation 1 0.6z 0.3z 2 = 0. We can compute the solutions of this equation using R: > a=c(1,-0.6,-0.3) > polyroot(a) [1] i i > abs(polyroot(a)) [1] This shows that both roots, and , are real and have modulus greater than 1. To generate the ACF of this time series using the recursive equations ρ(0) = 1, ρ(1) = φ 1 1 φ 2, and for i 2, ρ(i) = φ 1 ρ(i 1) + φ 2 ρ(i 2), we use the following code: > phi1=0.6 > phi2=0.3 > rho[1]=1 > rho[2]=phi1/(1-phi2) > for (i in 3:30) rho[i]=phi1*rho[i-1]+phi2*rho[i-2] This gives the following data > rho The graph of the ACF is then obtained using the command > plot(rho[-1],type= h );abline(h=0)
5 rho[-1] Index (b) We are considering the time series defined by Y t = 0.Y t Y t 2 + Z t with characteristic equation z 0.5z 2 = 0. We can compute the solutions of this equation using R: > b=c(1,0.,-0.5) > polyroot(b) [1] i i > abs(polyroot(b)) [1] This shows that both roots, and , are real and have modulus greater than 1. To generate the ACF of this time series using the recursive equations, we use the same code as in (a) but define > phi1=-0. > phi2=0.5 and get the data set and the graph rho[-1] Index
6 (e) We are considering the time series defined by Y t = 0.5Y t 1 0.9Y t 2 + Z t with characteristic equation 1 0.5z + 0.9z 2 = 0. We can compute the solutions of this equation using R: > e=c(1,-0.5,0.9) > polyroot(e) [1] i i > abs(polyroot(e)) [1] This shows that both roots are complex and have modulus greater than 1. To generate the ACF of this time series using the recursive equations, we use the same code as in (a) but define > phi1=-0. > phi2=0.5 and get the data set and the graph rho[-1] Index To see a bit more clearly the shape of the damped sine wave, on can use the following command: > plot(rho[-1],type= l );abline(h=0) rho[-1] Index
7 Since the roots of the characteristic polynomial in this problem are complex, we also look for the damping factor R and the frequency Θ: R = φ 2 = ( ) Also, cos Θ = φ 1 2 Φ 2 = Θ = arccos Note that one can verify that the expression at the end of Lecture 10 in the notes is correct using the following code and comparing the data with that obtained in the table above: > R=sqrt(0.9) > Theta=acos(0.5/(2*R)) > Phi=atan(tan(Theta)*19) > X=0:30 > Rho=RˆX*sin(Theta*X+Phi)/sin(Phi) A third way of obtaining the same values is by simply using the command > ARMAacf(ar=c(0.5,-0.9),lag.max=30) 7. Textbook Problem.12: We consider the MA(2) process defined by (a) We know that if the process is as defined above, ρ 1 = θ 1 + θ 1 θ 2, ρ 1 + θ1 2 + θ2 2 2 = Therefore, if θ 1 = θ 2 = 1, we get 6 ρ 1 = Y t = Z t θ 1 Z t 1 θ 2 Z t 1. (6) θ 2, and ρ 1 + θ1 2 + θ2 2 k = 0 for k = 3,, 5,.... 1/6 + 1/ /36 + 1/36 = 5/36 38/36 = /6 ρ 2 = 1 + 1/36 + 1/36 = 6/36 38/36 = Moreover, if θ 1 = 1, θ 2 = 6, we get ρ 1 = = ρ 2 = = 6 38 = Since for both processes, ρ k = 0 for k = 3,,..., we see that they have the same ACF. (b) The characteristic equation for the time series defined in (6) is 1 θ 1 z θ 2 z 2 = 0, so if θ 1 = θ 2 = 1, the equation is z 1 6 z2 = 0 z 2 + z 6 = 0 (z + 3)(z 2) = 0 z = 3 or z = 2 and if θ 1 = 1, θ 2 = 6, the equation is 1 + z 6z 2 = 0 z 2 z/6 1/6 = 0 z = 1/6 ± ( 1/6) 2 1 ( 1/6) 2 z = 1 12 ± 1/36 + 2/36 2 z = 1 12 ± 5 12 z = 1 3 or z = 1 2.
8 Note that this shows that while the ACF is the same for both sets of parameters, for the second set of parameters, the solutions of the characteristic equation both have modulus less than 1, which means that the second of the two MA processes is not invertible.
Ch 4. Models For Stationary Time Series. Time Series Analysis
This chapter discusses the basic concept of a broad class of stationary parametric time series models the autoregressive moving average (ARMA) models. Let {Y t } denote the observed time series, and {e
More informationSTAD57 Time Series Analysis. Lecture 8
STAD57 Time Series Analysis Lecture 8 1 ARMA Model Will be using ARMA models to describe times series dynamics: ( B) X ( B) W X X X W W W t 1 t1 p t p t 1 t1 q tq Model must be causal (i.e. stationary)
More information18.S096 Problem Set 4 Fall 2013 Time Series Due Date: 10/15/2013
18.S096 Problem Set 4 Fall 2013 Time Series Due Date: 10/15/2013 1. Covariance Stationary AR(2) Processes Suppose the discrete-time stochastic process {X t } follows a secondorder auto-regressive process
More informationChapter 4: Models for Stationary Time Series
Chapter 4: Models for Stationary Time Series Now we will introduce some useful parametric models for time series that are stationary processes. We begin by defining the General Linear Process. Let {Y t
More informationUNIVERSITY OF TORONTO SCARBOROUGH Department of Computer and Mathematical Sciences Midterm Test, March 2014
UNIVERSITY OF TORONTO SCARBOROUGH Department of Computer and Mathematical Sciences Midterm Test, March 2014 STAD57H3 Time Series Analysis Duration: One hour and fifty minutes Last Name: First Name: Student
More informationProblem Set 1 Solution Sketches Time Series Analysis Spring 2010
Problem Set 1 Solution Sketches Time Series Analysis Spring 2010 1. Construct a martingale difference process that is not weakly stationary. Simplest e.g.: Let Y t be a sequence of independent, non-identically
More information3. ARMA Modeling. Now: Important class of stationary processes
3. ARMA Modeling Now: Important class of stationary processes Definition 3.1: (ARMA(p, q) process) Let {ɛ t } t Z WN(0, σ 2 ) be a white noise process. The process {X t } t Z is called AutoRegressive-Moving-Average
More informationStat 248 Lab 2: Stationarity, More EDA, Basic TS Models
Stat 248 Lab 2: Stationarity, More EDA, Basic TS Models Tessa L. Childers-Day February 8, 2013 1 Introduction Today s section will deal with topics such as: the mean function, the auto- and cross-covariance
More informationHomework 4. 1 Data analysis problems
Homework 4 1 Data analysis problems This week we will be analyzing a number of data sets. We are going to build ARIMA models using the steps outlined in class. It is also a good idea to read section 3.8
More informationStatistics of stochastic processes
Introduction Statistics of stochastic processes Generally statistics is performed on observations y 1,..., y n assumed to be realizations of independent random variables Y 1,..., Y n. 14 settembre 2014
More informationDiscrete time processes
Discrete time processes Predictions are difficult. Especially about the future Mark Twain. Florian Herzog 2013 Modeling observed data When we model observed (realized) data, we encounter usually the following
More informationγ 0 = Var(X i ) = Var(φ 1 X i 1 +W i ) = φ 2 1γ 0 +σ 2, which implies that we must have φ 1 < 1, and γ 0 = σ2 . 1 φ 2 1 We may also calculate for j 1
4.2 Autoregressive (AR) Moving average models are causal linear processes by definition. There is another class of models, based on a recursive formulation similar to the exponentially weighted moving
More informationSTA 6857 Autocorrelation and Cross-Correlation & Stationary Time Series ( 1.4, 1.5)
STA 6857 Autocorrelation and Cross-Correlation & Stationary Time Series ( 1.4, 1.5) Outline 1 Announcements 2 Autocorrelation and Cross-Correlation 3 Stationary Time Series 4 Homework 1c Arthur Berg STA
More informationLecture 4a: ARMA Model
Lecture 4a: ARMA Model 1 2 Big Picture Most often our goal is to find a statistical model to describe real time series (estimation), and then predict the future (forecasting) One particularly popular model
More informationSpectral Analysis (Theory)
Spectral Analysis (Theory) Al Nosedal University of Toronto Winter 2016 Al Nosedal University of Toronto Spectral Analysis (Theory) Winter 2016 1 / 28 Idea: Decompose a stationary time series {X t } into
More informationCovariances of ARMA Processes
Statistics 910, #10 1 Overview Covariances of ARMA Processes 1. Review ARMA models: causality and invertibility 2. AR covariance functions 3. MA and ARMA covariance functions 4. Partial autocorrelation
More informationIntroduction to Stochastic processes
Università di Pavia Introduction to Stochastic processes Eduardo Rossi Stochastic Process Stochastic Process: A stochastic process is an ordered sequence of random variables defined on a probability space
More informationEconometría 2: Análisis de series de Tiempo
Econometría 2: Análisis de series de Tiempo Karoll GOMEZ kgomezp@unal.edu.co http://karollgomez.wordpress.com Segundo semestre 2016 III. Stationary models 1 Purely random process 2 Random walk (non-stationary)
More informationFor a stochastic process {Y t : t = 0, ±1, ±2, ±3, }, the mean function is defined by (2.2.1) ± 2..., γ t,
CHAPTER 2 FUNDAMENTAL CONCEPTS This chapter describes the fundamental concepts in the theory of time series models. In particular, we introduce the concepts of stochastic processes, mean and covariance
More informationExamination paper for Solution: TMA4285 Time series models
Department of Mathematical Sciences Examination paper for Solution: TMA4285 Time series models Academic contact during examination: Håkon Tjelmeland Phone: 4822 1896 Examination date: December 7th 2013
More informationMultivariate Time Series
Multivariate Time Series Notation: I do not use boldface (or anything else) to distinguish vectors from scalars. Tsay (and many other writers) do. I denote a multivariate stochastic process in the form
More informationMarcel Dettling. Applied Time Series Analysis SS 2013 Week 05. ETH Zürich, March 18, Institute for Data Analysis and Process Design
Marcel Dettling Institute for Data Analysis and Process Design Zurich University of Applied Sciences marcel.dettling@zhaw.ch http://stat.ethz.ch/~dettling ETH Zürich, March 18, 2013 1 Basics of Modeling
More informationReview Session: Econometrics - CLEFIN (20192)
Review Session: Econometrics - CLEFIN (20192) Part II: Univariate time series analysis Daniele Bianchi March 20, 2013 Fundamentals Stationarity A time series is a sequence of random variables x t, t =
More informationCharacteristics of Time Series
Characteristics of Time Series Al Nosedal University of Toronto January 12, 2016 Al Nosedal University of Toronto Characteristics of Time Series January 12, 2016 1 / 37 Signal and Noise In general, most
More information1 Class Organization. 2 Introduction
Time Series Analysis, Lecture 1, 2018 1 1 Class Organization Course Description Prerequisite Homework and Grading Readings and Lecture Notes Course Website: http://www.nanlifinance.org/teaching.html wechat
More informationTMA4285 December 2015 Time series models, solution.
Norwegian University of Science and Technology Department of Mathematical Sciences Page of 5 TMA4285 December 205 Time series models, solution. Problem a) (i) The slow decay of the ACF of z t suggest that
More informationTime Series Solutions HT 2009
Time Series Solutions HT 2009 1. Let {X t } be the ARMA(1, 1) process, X t φx t 1 = ɛ t + θɛ t 1, {ɛ t } WN(0, σ 2 ), where φ < 1 and θ < 1. Show that the autocorrelation function of {X t } is given by
More informationWe will only present the general ideas on how to obtain. follow closely the AR(1) and AR(2) cases presented before.
ACF and PACF of an AR(p) We will only present the general ideas on how to obtain the ACF and PACF of an AR(p) model since the details follow closely the AR(1) and AR(2) cases presented before. Recall that
More information1 Linear Difference Equations
ARMA Handout Jialin Yu 1 Linear Difference Equations First order systems Let {ε t } t=1 denote an input sequence and {y t} t=1 sequence generated by denote an output y t = φy t 1 + ε t t = 1, 2,... with
More informationNANYANG TECHNOLOGICAL UNIVERSITY SEMESTER II EXAMINATION MAS451/MTH451 Time Series Analysis TIME ALLOWED: 2 HOURS
NANYANG TECHNOLOGICAL UNIVERSITY SEMESTER II EXAMINATION 2012-2013 MAS451/MTH451 Time Series Analysis May 2013 TIME ALLOWED: 2 HOURS INSTRUCTIONS TO CANDIDATES 1. This examination paper contains FOUR (4)
More informationMAT 3379 (Winter 2016) FINAL EXAM (SOLUTIONS)
MAT 3379 (Winter 2016) FINAL EXAM (SOLUTIONS) 15 April 2016 (180 minutes) Professor: R. Kulik Student Number: Name: This is closed book exam. You are allowed to use one double-sided A4 sheet of notes.
More information3 Theory of stationary random processes
3 Theory of stationary random processes 3.1 Linear filters and the General linear process A filter is a transformation of one random sequence {U t } into another, {Y t }. A linear filter is a transformation
More informationProf. Dr. Roland Füss Lecture Series in Applied Econometrics Summer Term Introduction to Time Series Analysis
Introduction to Time Series Analysis 1 Contents: I. Basics of Time Series Analysis... 4 I.1 Stationarity... 5 I.2 Autocorrelation Function... 9 I.3 Partial Autocorrelation Function (PACF)... 14 I.4 Transformation
More informationLECTURES 2-3 : Stochastic Processes, Autocorrelation function. Stationarity.
LECTURES 2-3 : Stochastic Processes, Autocorrelation function. Stationarity. Important points of Lecture 1: A time series {X t } is a series of observations taken sequentially over time: x t is an observation
More informationECON/FIN 250: Forecasting in Finance and Economics: Section 6: Standard Univariate Models
ECON/FIN 250: Forecasting in Finance and Economics: Section 6: Standard Univariate Models Patrick Herb Brandeis University Spring 2016 Patrick Herb (Brandeis University) Standard Univariate Models ECON/FIN
More informationEconometría 2: Análisis de series de Tiempo
Econometría 2: Análisis de series de Tiempo Karoll GOMEZ kgomezp@unal.edu.co http://karollgomez.wordpress.com Segundo semestre 2016 II. Basic definitions A time series is a set of observations X t, each
More informationSTAT 520: Forecasting and Time Series. David B. Hitchcock University of South Carolina Department of Statistics
David B. University of South Carolina Department of Statistics What are Time Series Data? Time series data are collected sequentially over time. Some common examples include: 1. Meteorological data (temperatures,
More informationA time series is called strictly stationary if the joint distribution of every collection (Y t
5 Time series A time series is a set of observations recorded over time. You can think for example at the GDP of a country over the years (or quarters) or the hourly measurements of temperature over a
More informationTime Series Analysis. Solutions to problems in Chapter 5 IMM
Time Series Analysis Solutions to problems in Chapter 5 IMM Solution 5.1 Question 1. [ ] V [X t ] = V [ǫ t + c(ǫ t 1 + ǫ t + )] = 1 + c 1 σǫ = The variance of {X t } is not limited and therefore {X t }
More informationCovariance Stationary Time Series. Example: Independent White Noise (IWN(0,σ 2 )) Y t = ε t, ε t iid N(0,σ 2 )
Covariance Stationary Time Series Stochastic Process: sequence of rv s ordered by time {Y t } {...,Y 1,Y 0,Y 1,...} Defn: {Y t } is covariance stationary if E[Y t ]μ for all t cov(y t,y t j )E[(Y t μ)(y
More informationPart II. Time Series
Part II Time Series 12 Introduction This Part is mainly a summary of the book of Brockwell and Davis (2002). Additionally the textbook Shumway and Stoffer (2010) can be recommended. 1 Our purpose is to
More informationClass 1: Stationary Time Series Analysis
Class 1: Stationary Time Series Analysis Macroeconometrics - Fall 2009 Jacek Suda, BdF and PSE February 28, 2011 Outline Outline: 1 Covariance-Stationary Processes 2 Wold Decomposition Theorem 3 ARMA Models
More informationMinitab Project Report Assignment 3
3.1.1 Simulation of Gaussian White Noise Minitab Project Report Assignment 3 Time Series Plot of zt Function zt 1 0. 0. zt 0-1 0. 0. -0. -0. - -3 1 0 30 0 50 Index 0 70 0 90 0 1 1 1 1 0 marks The series
More informationPart III Example Sheet 1 - Solutions YC/Lent 2015 Comments and corrections should be ed to
TIME SERIES Part III Example Sheet 1 - Solutions YC/Lent 2015 Comments and corrections should be emailed to Y.Chen@statslab.cam.ac.uk. 1. Let {X t } be a weakly stationary process with mean zero and let
More informationChapter 3. ARIMA Models. 3.1 Autoregressive Moving Average Models
Chapter 3 ARIMA Models Classical regression is often insu cient for explaining all of the interesting dynamics of a time series. For example, the ACF of the residuals of the simple linear regression fit
More informationEmpirical Market Microstructure Analysis (EMMA)
Empirical Market Microstructure Analysis (EMMA) Lecture 3: Statistical Building Blocks and Econometric Basics Prof. Dr. Michael Stein michael.stein@vwl.uni-freiburg.de Albert-Ludwigs-University of Freiburg
More informationMATH 5075: Time Series Analysis
NAME: MATH 5075: Time Series Analysis Final For the entire test {Z t } WN(0, 1)!!! 1 1) Let {Y t, t Z} be a stationary time series with EY t = 0 and autocovariance function γ Y (h). Assume that a) Show
More informationEcon 623 Econometrics II Topic 2: Stationary Time Series
1 Introduction Econ 623 Econometrics II Topic 2: Stationary Time Series In the regression model we can model the error term as an autoregression AR(1) process. That is, we can use the past value of the
More informationHomework 3 Solutions
Math 4506 Sprig 04 Homework 3 Solutios. a The ACF of a MA process has a o-zero value oly at lags, 0, ad. Problem 4.3 from the textbook which you did t do, so I did t expect you to metio this shows that
More informationSTOR 356: Summary Course Notes
STOR 356: Summary Course Notes Richard L. Smith Department of Statistics and Operations Research University of North Carolina Chapel Hill, NC 7599-360 rls@email.unc.edu February 19, 008 Course text: Introduction
More informationSTAT 443 Final Exam Review. 1 Basic Definitions. 2 Statistical Tests. L A TEXer: W. Kong
STAT 443 Final Exam Review L A TEXer: W Kong 1 Basic Definitions Definition 11 The time series {X t } with E[X 2 t ] < is said to be weakly stationary if: 1 µ X (t) = E[X t ] is independent of t 2 γ X
More informationMAT3379 (Winter 2016)
MAT3379 (Winter 2016) Assignment 4 - SOLUTIONS The following questions will be marked: 1a), 2, 4, 6, 7a Total number of points for Assignment 4: 20 Q1. (Theoretical Question, 2 points). Yule-Walker estimation
More informationEconometrics I. Professor William Greene Stern School of Business Department of Economics 25-1/25. Part 25: Time Series
Econometrics I Professor William Greene Stern School of Business Department of Economics 25-1/25 Econometrics I Part 25 Time Series 25-2/25 Modeling an Economic Time Series Observed y 0, y 1,, y t, What
More informationReliability and Risk Analysis. Time Series, Types of Trend Functions and Estimates of Trends
Reliability and Risk Analysis Stochastic process The sequence of random variables {Y t, t = 0, ±1, ±2 } is called the stochastic process The mean function of a stochastic process {Y t} is the function
More informationTAKEHOME FINAL EXAM e iω e 2iω e iω e 2iω
ECO 513 Spring 2015 TAKEHOME FINAL EXAM (1) Suppose the univariate stochastic process y is ARMA(2,2) of the following form: y t = 1.6974y t 1.9604y t 2 + ε t 1.6628ε t 1 +.9216ε t 2, (1) where ε is i.i.d.
More informationLecture 2: ARMA(p,q) models (part 2)
Lecture 2: ARMA(p,q) models (part 2) Florian Pelgrin University of Lausanne, École des HEC Department of mathematics (IMEA-Nice) Sept. 2011 - Jan. 2012 Florian Pelgrin (HEC) Univariate time series Sept.
More informationUnivariate ARIMA Models
Univariate ARIMA Models ARIMA Model Building Steps: Identification: Using graphs, statistics, ACFs and PACFs, transformations, etc. to achieve stationary and tentatively identify patterns and model components.
More informationTime Series Analysis
Time Series Analysis hm@imm.dtu.dk Informatics and Mathematical Modelling Technical University of Denmark DK-2800 Kgs. Lyngby 1 Outline of the lecture Chapter 9 Multivariate time series 2 Transfer function
More informationModule 9: Stationary Processes
Module 9: Stationary Processes Lecture 1 Stationary Processes 1 Introduction A stationary process is a stochastic process whose joint probability distribution does not change when shifted in time or space.
More informationEconomics Department LSE. Econometrics: Timeseries EXERCISE 1: SERIAL CORRELATION (ANALYTICAL)
Economics Department LSE EC402 Lent 2015 Danny Quah TW1.10.01A x7535 : Timeseries EXERCISE 1: SERIAL CORRELATION (ANALYTICAL) 1. Suppose ɛ is w.n. (0, σ 2 ), ρ < 1, and W t = ρw t 1 + ɛ t, for t = 1, 2,....
More informationWhite Noise Processes (Section 6.2)
White Noise Processes (Section 6.) Recall that covariance stationary processes are time series, y t, such. E(y t ) = µ for all t. Var(y t ) = σ for all t, σ < 3. Cov(y t,y t-τ ) = γ(τ) for all t and τ
More informationDifference equations. Definitions: A difference equation takes the general form. x t f x t 1,,x t m.
Difference equations Definitions: A difference equation takes the general form x t fx t 1,x t 2, defining the current value of a variable x as a function of previously generated values. A finite order
More informationAPPLIED ECONOMETRIC TIME SERIES 4TH EDITION
APPLIED ECONOMETRIC TIME SERIES 4TH EDITION Chapter 2: STATIONARY TIME-SERIES MODELS WALTER ENDERS, UNIVERSITY OF ALABAMA Copyright 2015 John Wiley & Sons, Inc. Section 1 STOCHASTIC DIFFERENCE EQUATION
More informationNext tool is Partial ACF; mathematical tools first. The Multivariate Normal Distribution. e z2 /2. f Z (z) = 1 2π. e z2 i /2
Next tool is Partial ACF; mathematical tools first. The Multivariate Normal Distribution Defn: Z R 1 N(0,1) iff f Z (z) = 1 2π e z2 /2 Defn: Z R p MV N p (0, I) if and only if Z = (Z 1,..., Z p ) (a column
More informationUnivariate Time Series Analysis; ARIMA Models
Econometrics 2 Fall 24 Univariate Time Series Analysis; ARIMA Models Heino Bohn Nielsen of4 Outline of the Lecture () Introduction to univariate time series analysis. (2) Stationarity. (3) Characterizing
More informationStochastic Modelling Solutions to Exercises on Time Series
Stochastic Modelling Solutions to Exercises on Time Series Dr. Iqbal Owadally March 3, 2003 Solutions to Elementary Problems Q1. (i) (1 0.5B)X t = Z t. The characteristic equation 1 0.5z = 0 does not have
More informationTIME SERIES AND FORECASTING. Luca Gambetti UAB, Barcelona GSE Master in Macroeconomic Policy and Financial Markets
TIME SERIES AND FORECASTING Luca Gambetti UAB, Barcelona GSE 2014-2015 Master in Macroeconomic Policy and Financial Markets 1 Contacts Prof.: Luca Gambetti Office: B3-1130 Edifici B Office hours: email:
More informationLecture 3: Autoregressive Moving Average (ARMA) Models and their Practical Applications
Lecture 3: Autoregressive Moving Average (ARMA) Models and their Practical Applications Prof. Massimo Guidolin 20192 Financial Econometrics Winter/Spring 2018 Overview Moving average processes Autoregressive
More informationFE570 Financial Markets and Trading. Stevens Institute of Technology
FE570 Financial Markets and Trading Lecture 5. Linear Time Series Analysis and Its Applications (Ref. Joel Hasbrouck - Empirical Market Microstructure ) Steve Yang Stevens Institute of Technology 9/25/2012
More informationEcon 427, Spring Problem Set 3 suggested answers (with minor corrections) Ch 6. Problems and Complements:
Econ 427, Spring 2010 Problem Set 3 suggested answers (with minor corrections) Ch 6. Problems and Complements: 1. (page 132) In each case, the idea is to write these out in general form (without the lag
More information3 ARIMA Models. 3.1 Introduction
3 ARIMA Models 3. Introduction In Chapters and, we introduced autocorrelation and cross-correlation functions (ACFs and CCFs) as tools for clarifying relations that may occur within and between time series
More informationNotes on Random Processes
otes on Random Processes Brian Borchers and Rick Aster October 27, 2008 A Brief Review of Probability In this section of the course, we will work with random variables which are denoted by capital letters,
More informationSTAT 153, FALL 2015 HOMEWORK 1 SOLUTIONS
STAT 153, FALL 015 HOMEWORK 1 SOLUTIONS Problem 1 (a) A solution in R is provided in Figure 1. We conclude that if δ = 0 the process is a random walk without drift. When δ 0 there is a linear drift with
More informationA Beginner s Introduction. Box-Jenkins Models
A Beginner s Introduction To Box-Jenkins Models I. Introduction In their seminal work, Time Series Analysis: Forecasting and Control (1970, Holden Day), Professors Box and Jenkins introduced a set of time
More informationApplied time-series analysis
Robert M. Kunst robert.kunst@univie.ac.at University of Vienna and Institute for Advanced Studies Vienna October 18, 2011 Outline Introduction and overview Econometric Time-Series Analysis In principle,
More informationIntroduction to ARMA and GARCH processes
Introduction to ARMA and GARCH processes Fulvio Corsi SNS Pisa 3 March 2010 Fulvio Corsi Introduction to ARMA () and GARCH processes SNS Pisa 3 March 2010 1 / 24 Stationarity Strict stationarity: (X 1,
More informationMAT 3379 (Winter 2016) FINAL EXAM (PRACTICE)
MAT 3379 (Winter 2016) FINAL EXAM (PRACTICE) 15 April 2016 (180 minutes) Professor: R. Kulik Student Number: Name: This is closed book exam. You are allowed to use one double-sided A4 sheet of notes. Only
More informationStochastic process for macro
Stochastic process for macro Tianxiao Zheng SAIF 1. Stochastic process The state of a system {X t } evolves probabilistically in time. The joint probability distribution is given by Pr(X t1, t 1 ; X t2,
More informationSTAT 248: EDA & Stationarity Handout 3
STAT 248: EDA & Stationarity Handout 3 GSI: Gido van de Ven September 17th, 2010 1 Introduction Today s section we will deal with the following topics: the mean function, the auto- and crosscovariance
More informationTime Series 2. Robert Almgren. Sept. 21, 2009
Time Series 2 Robert Almgren Sept. 21, 2009 This week we will talk about linear time series models: AR, MA, ARMA, ARIMA, etc. First we will talk about theory and after we will talk about fitting the models
More informationCh3. TRENDS. Time Series Analysis
3.1 Deterministic Versus Stochastic Trends The simulated random walk in Exhibit 2.1 shows a upward trend. However, it is caused by a strong correlation between the series at nearby time points. The true
More informationStatistics 349(02) Review Questions
Statistics 349(0) Review Questions I. Suppose that for N = 80 observations on the time series { : t T} the following statistics were calculated: _ x = 10.54 C(0) = 4.99 In addition the sample autocorrelation
More informationThe autocorrelation and autocovariance functions - helpful tools in the modelling problem
The autocorrelation and autocovariance functions - helpful tools in the modelling problem J. Nowicka-Zagrajek A. Wy lomańska Institute of Mathematics and Computer Science Wroc law University of Technology,
More information6.3 Forecasting ARMA processes
6.3. FORECASTING ARMA PROCESSES 123 6.3 Forecasting ARMA processes The purpose of forecasting is to predict future values of a TS based on the data collected to the present. In this section we will discuss
More informationProperties of Summation Operator
Econ 325 Section 003/004 Notes on Variance, Covariance, and Summation Operator By Hiro Kasahara Properties of Summation Operator For a sequence of the values {x 1, x 2,..., x n, we write the sum of x 1,
More informationMath 253 Homework due Wednesday, March 9 SOLUTIONS
Math 53 Homework due Wednesday, March 9 SOLUTIONS 1. Do Section 8.8, problems 11,, 15, 17 (these problems have to do with Taylor s Inequality, and they are very similar to what we did on the last homework.
More informationAPPLIED OPTICS POLARIZATION
A. La Rosa Lecture Notes APPLIED OPTICS POLARIZATION Linearly-polarized light Description of linearly polarized light (using Real variables) Alternative description of linearly polarized light using phasors
More information1 Introduction to Generalized Least Squares
ECONOMICS 7344, Spring 2017 Bent E. Sørensen April 12, 2017 1 Introduction to Generalized Least Squares Consider the model Y = Xβ + ɛ, where the N K matrix of regressors X is fixed, independent of the
More informationX t = a t + r t, (7.1)
Chapter 7 State Space Models 71 Introduction State Space models, developed over the past 10 20 years, are alternative models for time series They include both the ARIMA models of Chapters 3 6 and the Classical
More informationTime Series 3. Robert Almgren. Sept. 28, 2009
Time Series 3 Robert Almgren Sept. 28, 2009 Last time we discussed two main categories of linear models, and their combination. Here w t denotes a white noise: a stationary process with E w t ) = 0, E
More informationCh 8. MODEL DIAGNOSTICS. Time Series Analysis
Model diagnostics is concerned with testing the goodness of fit of a model and, if the fit is poor, suggesting appropriate modifications. We shall present two complementary approaches: analysis of residuals
More information6 NONSEASONAL BOX-JENKINS MODELS
6 NONSEASONAL BOX-JENKINS MODELS In this section, we will discuss a class of models for describing time series commonly referred to as Box-Jenkins models. There are two types of Box-Jenkins models, seasonal
More informationLevinson Durbin Recursions: I
Levinson Durbin Recursions: I note: B&D and S&S say Durbin Levinson but Levinson Durbin is more commonly used (Levinson, 1947, and Durbin, 1960, are source articles sometimes just Levinson is used) recursions
More informationEnd-Semester Examination MA 373 : Statistical Analysis on Financial Data
End-Semester Examination MA 373 : Statistical Analysis on Financial Data Instructor: Dr. Arabin Kumar Dey, Department of Mathematics, IIT Guwahati Note: Use the results in Section- III: Data Analysis using
More informationECE302 Spring 2015 HW10 Solutions May 3,
ECE32 Spring 25 HW Solutions May 3, 25 Solutions to HW Note: Most of these solutions were generated by R. D. Yates and D. J. Goodman, the authors of our textbook. I have added comments in italics where
More informationEconometrics II Heij et al. Chapter 7.1
Chapter 7.1 p. 1/2 Econometrics II Heij et al. Chapter 7.1 Linear Time Series Models for Stationary data Marius Ooms Tinbergen Institute Amsterdam Chapter 7.1 p. 2/2 Program Introduction Modelling philosophy
More informationChapter 3 - Temporal processes
STK4150 - Intro 1 Chapter 3 - Temporal processes Odd Kolbjørnsen and Geir Storvik January 23 2017 STK4150 - Intro 2 Temporal processes Data collected over time Past, present, future, change Temporal aspect
More information8.2 Harmonic Regression and the Periodogram
Chapter 8 Spectral Methods 8.1 Introduction Spectral methods are based on thining of a time series as a superposition of sinusoidal fluctuations of various frequencies the analogue for a random process
More informationChapter 9: Forecasting
Chapter 9: Forecasting One of the critical goals of time series analysis is to forecast (predict) the values of the time series at times in the future. When forecasting, we ideally should evaluate the
More informationTime Series Examples Sheet
Lent Term 2001 Richard Weber Time Series Examples Sheet This is the examples sheet for the M. Phil. course in Time Series. A copy can be found at: http://www.statslab.cam.ac.uk/~rrw1/timeseries/ Throughout,
More information