Time Series (Part II)

Size: px
Start display at page:

Download "Time Series (Part II)"

Transcription

1 Time Series (Part II) Luca Gambetti UAB IDEA Winter

2 Contacts Prof.: Luca Gambetti Office: B Edifici B Office hours: luca.gambetti@uab.es webpage: 2

3 Goal of the course The main objective of the course is to provide the students with the knowledge of a comprehensive set of tools necessary for empirical research with time series data. Description This is the second part of an introductory 40-hours course in Time Series Analysis with applications in macroeconomics. This part focuses on the theory of multivariate time series models. 3

4 Contents 1. VAR: preliminaries, representation, stationarity, second moments, specification (1-2). 2. VAR: estimation and hypothesis testing, Granger causality (1-2). 3. VAR: forecasting and impulse response functions (1). 4. Structural VAR: theory (1). 5. Structural VAR: applications (1). 6. Nonstationary data - Cointegration (1). 7. Structural Factor models and FAVARs (1). 8. Bayesian VARs (1). 4

5 References 1. P. J.Brockwell, and R. A. Davis, (2009), Time Series: Theory and Methods, Springer-Verlag: Berlin 2. F. Canova (2007), Methods for Applied Macroeconomic Research, Princeton University Press: Princeton 3. J. D. Hamilton (1994), Time Series Analysis, Princeton University Press: Princeton 4. H. Lutkepohl (2005), New Introduction to Multiple Time Series, Springer- Verlag: Berlin 5

6 Grades 40% problem sets 60% final take-home exam. Econometric Software GRETL, MATLAB. 6

7 1. STATIONARY VECTOR PROCESSES 1 1 This part is partly based on the Hamilton textbook and Marco Lippi s notes. 7

8 1 Some Preliminary Definitions and Results Random Vector: A vector X = (X 1,..., X n ) whose components are scalarvalued random variables on a probability space. Vector Random Process: A family of random vectors {X t, t T } defined on a probability space, where T is a set of time points. Typically T = R, T = Z or T = N, the sets or real, integer and natural numbers, respectively. Time Series Vector: A particular realization of a vector random process. 8

9 1.1 The Lag operator The lag operator L maps a sequence {X t } into a sequence {Y t } such that Y t = LX t = X t 1, for all t. If we apply L repeatedly on a process, for instance L(L(LX t )), we will use the convention L(L(LX t )) = L 3 X t = X t 3. If we apply L to a constant c, Lc = c. Inversion: L 1 is the inverse of L, such that L 1 (L)X t = X t. 9

10 1.2 Polynomials in the lag operator We can form polynomials; α(l) = 1 + α 1 L + α 2 L α p L p is a polynomial in the lag operator of order p and is such that α(l)x t = X t +α 1 X t α p X t p. Lag polynomials can also be inverted. For a polynomial φ(l), we are looking for the values of the coefficients α i of φ(l) 1 = α 0 + α 1 L + α 2 L such that φ(l) 1 φ(l) = 1. Case 1: p = 1. Let φ(l) = (1 φl) with φ < 1. To find the inverse write (1 φl)(α 0 + α 1 L + α 2 L ) 1 = 1 note that all the coefficients of the non-zero powers of L must be equal to zero. This gives α 0 = 1 φ + α 1 = 0 α 1 = φ α 1 = φ φα 1 + α 2 = 0 α 2 = φ 2 φα 2 + α 3 = 0 α 3 = φ 3 10

11 and so on. In general α k = φ k, so (1 φl) 1 = j=0 φj L j provided that φ < 1. It is easy to check this because so (1 φl)(1 + φl + φ 2 L φ k L k ) = 1 φ k+1 L k+1 (1 + φl + φ 2 L φ k L k ) = 1 φk+1 L k+1 and k k j=0 φj L j 1 (1 φl). (1 φl) 11

12 Case 2: p = 2. Let φ(l) = (1 φ 1 L φ 2 L 2 ). To find the inverse it is useful to factor the polynomial in the following way (1 φ 1 L φ 2 L 2 ) = (1 λ 1 L)(1 λ 2 L) where the λ 1, λ 2 are the reciprocal of the roots of the above left-hand side polynomial or equivalently the eigenvalues of the matrix ( ) φ1 φ Suppose λ 1, λ 2 < 0 and λ 1 λ 2. We have that (1 φ 1 L φ 2 L 2 ) 1 = (1 λ 1 L) 1 (1 λ 2 L) 1. Therefore we can use what we have seen above for the case p = 1. We can write [ (1 λ 1 L) 1 (1 λ 2 L) 1 = (λ 1 λ 2 ) 1 λ 1 1 λ 1 L λ ] 2 1 λ 2 L λ 1 [ = 1 + λ1 L + λ 1 L ] λ 1 λ 2 λ 2 λ 1 λ 2 [ 1 + λ2 L + λ 2 L ] = (c 1 + c 2 ) + (c 1 λ 1 + c 2 λ 2 )L + (c 1 λ c 2 λ 2 2)L

13 where c 1 = λ 1 /(λ 1 λ 2 ), c 2 = λ 2 /(λ 1 λ 2 ) Matrix of polynomial in the lag operator: A(L) if its elements are polynomial in the lag operator, i.e. ( ) ( ) ( ) 1 L A(L) = = + L L We also define and A(0) = A(1) = ( ) ( )

14 1.3 Covariance Stationarity Let Y t be a n-dimensional random vector, Y t = [Y 1t,..., Y nt ]. Then Y t is covariance (weakly) stationary if E(Y t ) = µ, and the autocovariance matrix E(Y t µ)(y t j µ) = Γ j for all t, j, that is are independent of t and both finite. Stationarity of each of the components of Y t does not imply stationarity of the vector Y t. Stationarity in the vector case requires that the components of the vector are stationary and costationary. Although γ j = γ j for a scalar process, the same is not true for a vector process. The correct relation is Γ j = Γ j 14

15 1.4 Convergence of random variables We refresh three concepts of stochastic convergence in the univariate case and then we extend the three cases to the multivariate framework. Let {x T, T = 1, 2,...} be a sequence of random variables. Convergence in probability The sequence {x T } converges in probability to the random variable x if for every ɛ > 0 lim P r( x T x > ɛ) = 0 T When the above condition is satisfied, the variable c is called probability limit or plim of the sequence {x T }, indicated as plimx T = x Convergence in mean square. The sequence of random variables {x T } converges in mean square to x, denoted by x T c, m.s. if lim E(x T x) 2 = 0 T 15

16 Convergence in distribution Let F T denote the cumulative distribution function of x T and F the cumulative distribution of the scalar x. The sequence is said to converge in distribution written as x T d x (or xt L x), if for all real numbers c for which F is continuous We will equivalently write lim F T(c) = F (c). n X T d N(µ, σ 2 ) The concepts of stochastic convergence can be generalized to the multivariate setting. Suppose {X T } T =1 is a sequence of n-dimensional random vectors and X is a n-dimensional random vector. Then 1. X T p X if XkT p Xk for k = 1,..., K 2. X T m.s. X if lim E(X T X) (X t X) = 0 3. X T d X if limn F T (c) = F (c) where F and F T are the joint distribution of X and X T. 16

17 Proposition C.1 L. Suppose {X T } T =1 is a sequence of n-dimensional random vectors. Then the following relations hold: (a) X T m.s. X X T p X XT d X (c) (Slutsky s Theorem) If g : R n R m is a continuous function, then X T p X g(xt ) p g(x) (i.e. plimg(x T ) = g(plimx T )) X T d X g(xt ) d g(x) d Example Suppose that X T N(0, 1). Then X 2 T converges in distribution to the square of N(0, 1), i.e. XT 2 d χ

18 Proposition C.2 L. Suppose {X T } T =1 and {Y T } T =1 are sequences of n 1 random vectors and A T is a sequence of n n random matrices, x s a n 1 random vector, c s a fixed n 1 vector, and A is a fixed n n matrix. 1. If plimx T, plimy T and plima T exist then (a) plim(x T +Y T ) = plimx T +plimy T, plim(x T Y T ) = plimx T plimy T, (b) plimc X T = c plimx T (c) plimx T Y T = (plimx T ) (plimy T ) (d) plima T X T = (plima T )(plimx T ) 2. If X T d X and plim(xt Y T ) = 0 then Y T d X. 3. If X T d X and plimyt = c, then (a) X t + Y T d X + c (b) Y T X t d c X 4. If X T d X and plimat = A then A T X T d AX 5. If X T d X and plimat = 0 then plima T X T = 0 18

19 d Example Let {X T } be a sequence of n 1 random vectors with X T N(µ, Ω), p and Let {Y T } be a sequence of n 1 random vectors with Y T C. Then by 3.(b) Y T X d T N(C µ, CΩC ). 19

20 1.5 Limit Theorems The Law of Large Numbers and the Central Limit Theorem are the most important results for computing the limits of sequences of random variables. There are many versions of LLN and CLT that differ on the assumptions about the dependence of the variables. Proposition C.12 L (Weak law of large numbers) 1. (iid sequences) Let {Y t } be an i.i.d sequence of random variables with finite mean µ. Then T Ȳ T = T 1 p Y t µ 2. (independent sequences) Let {Y t } be a sequence of independent random variables with E(X t ) = µ < and E X t 1+ɛ c < for some ɛ > 0 and a finite constant c. Then T 1 T t=1 Y p t µ. 3. (uncorrelated sequences) Let {Y t } be a sequence of uncorrelated random variables E(X t ) = µ < and V ar(x t ) c < for some finite constant 20 t=1

21 c. Then T 1 T t=1 Y t p µ. 4. (stationary processes) Let Y t be a convariance stationary process with finite E(Y t ) = µ and E[(Y t µ)(y t j µ)] = γ j with absolutely summable autocovariances j=0 γ j <. Then ȲT m.s. µ hence ȲT p µ. Weak stationarity and absolutely summable covariances are sufficient conditions for a law of large numbers to hold. Proposition C.13L (Central limit theorem) 1. (i.i.d. sequences) Let {Y T } be a sequence of n-dimensional iid(µ, Σ) random variables then T ( Ȳ T µ) d N(0, Σ) 2. (stationary processes) Let Y t = µ + j=0 Φ jε t j be a n-dimensional stationary random process with ε t a i.i.d vector, E(Y t ) = µ <, j=0 Φ j <. Then T (ȲT µ) d N(0, j= Γ j) where Γ j is the autcovariance matrix at lag j. 21

22 2 Some stationary processes 2.1 White Noise (WN) A n-dimensional vector white noise ɛ t = [ɛ 1t,..., ɛ nt ] W N(0, Ω) is such if E(ɛ t ) = 0 and Γ k = Ω (Ω a symmetric positive definite matrix) if k = 0 and 0 if k 0. If ɛ t, ɛ τ are independent the process is an independent vector White Noise (i.i.d). If also ɛ t N the process is a Gaussian WN. Important: A vector whose components are white noise is not necessarily a white noise. Example: ( let u t be a scalar white noise and define ɛ t = (u t, u t 1 ). Then σ 2 ) ( ) E(ɛ t ɛ t) = u σu 2 and E(ɛ t ɛ t 1) = σu

23 2.2 Vector Moving Average (VMA) Given the n-dimensional vector White Noise ɛ t a vector moving average of order q is defined as Y t = µ + ɛ t + C 1 ɛ t C q ɛ t q where C j are n n matrices of coefficients and µ is the mean of Y t. The VMA(1) Let us consider the VMA(1) Y t = µ + ɛ t + C 1 ɛ t 1 with ɛ t W N(0, Ω). The variance of the process is given by with autocovariances Γ 0 = E[(Y t µ)(y t µ) ] = Ω + C 1 ΩC 1 Γ 1 = C 1 Ω, Γ 1 = ΩC 1, Γ j = 0 for j > 1 23

24 The VMA(q) Let us consider the VMA(q) Y t = µ + ɛ t + C 1 ɛ t C q ɛ t q with ɛ t W N(0, Ω), µ is the mean of Y t. The variance of the process is given by with autocovariances Γ 0 = E[(Y t µ)(y t µ) ] = Ω + C 1 ΩC 1 + C 2 ΩC C q ΩC q Γ j = C j Ω + C j+1 ΩC 1 + C j+2 ΩC C q ΩC q j for j = 1, 2,..., q Γ j = ΩC j + C 1 ΩC j+1 + C 2 ΩC j C q+j ΩC q for j = 1, 2,..., q Γ j = 0 for j > q 24

25 The VMA( ) A useful process, as we will see, is the VMA( ) Y t = µ + C j ε t j (1) j=0 A very important result is that if the sequence {C j } is absolutely summable (i.e. j=0 C j < where C j = m n c mn,j, or equivalenty each sequence formed by the elements of the matrix is absolutely summable) then infinite sequence above generates a well defined (mean square convergent) process (see for instance proposition C.10L). Proposition (10.2H). Let Y t be an n 1 vector satisfying Y t = µ + C j ε t j j=0 where ε t is a vector WN with E(ε t ) = 0 and E(ε t ε t j ) = Ω for j = 0 and zero otherwise and {C j } j=0 is absolutely summable. Let Y it denote the ith element of Y t and µ i the ith element of µ. Then 25

26 (a) The autocovariance between the ith variable at time t and the jth variable at time s periods earlier, E(Y it µ i )(Y jt s µ j ) exists and is given by the row i column j element of Γ s = C s+v ΩC v v=0 for s = 0, 1, 2,... (b) The sequence of matrices {Γ s } s=0 is absolutely summable. If furthermore {ε t } t= is an i.i.d. sequence with E ε i1 tε i2 tε i3 tε i4 t for i 1, i 2, i 3, i 4 = 1, 2,..., n, then also (c) E Y i1 t 1 Y i2 t 2 Y i3 t 3 Y i4 t 4 for all t 1, t 2, t 3, t 4 (d) (1/T ) T t=1 Y ity jt s p E(Yit Y jt s ), for i, j = 1, 2,..., n and for all s Implications: 1. Result (a) implies that the second moments of a MA( ) with absolutely summable coefficients can be found by taking the limit of the autocovariance of an MA(q). 26

27 2. Result (b) ensures ergodicity for the mean 3. Result (c) says that Y t has bounded fourth moments 4. Result (d) says that Y t is ergodic for second moments 27

28 2.3 Invertibility and fundamentalness The VMA is invertible if and only if the determinant of C(L) vanishes only outside the unit circle, i.e. if det(c(z)) 0 for all z 1. If the process is invertible it possesses a unique VAR representation (clear later on). Example Consider the process ( Y1t Y 2t ) = ( 1 L ) ( ) ε1t 0 θ L ε 2t det(c(z)) = θ z which is zero for z = θ. Obviously the process is invertible if and only if θ > 1. The VMA is fundamental if and only if the det(c(z)) 0 for all z < 1. In the previous example the process is fundamental if and only if θ 1. In the case θ = 1 the process is fundamental but noninvertible. 28

29 Provided that θ > 1 the MA process can be inverted and the shock can be obtained as a combination of present and past values of Y t. In fact ( 1 L ) ( ) θ L Y1t 1 0 θ L Y 2t = ( ε1t Notice that for any noninvertible process with determinant that does not vanish on the unit circle there is an invertible process with identical autocovariance structure. ε 2t ) Example: univariate MA. Consider the MA(1) y t = u t + mu t 1 with u t W N, and m > 1 i.e. y t is noninvertible. The autocovariances are E(y 2 t ) = (1 + m) 2 σ 2 u, E(y t y t j ) = mσ 2 u for j = 1, 1 and E(y t y t j ) = 0 for j > 1. Now consider this alternative representation y t = v t + 1 m v t 1 29

30 which is invertible and ( v t = ) 1 m L y t = ( m L ) 1 (1 + ml)u t where v t is indeed a white noise process with variance m 2 σ 2 u. 30

31 2.4 Wold Decomposition Any zero-mean stationary vector process Y t admits the following representation Y t = C(L)ε t + µ t (2) where C(L)ɛ t is the stochastic component with C(L) = i=0 C il i and µ t the purely deterministic component, the one perfectly forecastable using linear combinations of past Y t. If µ t = 0 the process is said regular. Here we only consider regular processes. (2) represents the Wold representation of Y t which is unique and for which the following properties hold: (a) ɛ t is the innovation for Y t, i.e. ɛ t = Y t Proj(Y t Y t 1, Y t 1,...). (b) ɛ t is White noise, Eɛ t = 0, Eɛ t ɛ τ = 0, for t τ, Eɛ t ɛ t = Ω (c) The coefficients are square summable j=0 C j 2 <. (d) C 0 = I 31

32 The result is very powerful since holds for any covariance stationary process. However the theorem does not implies that (2) is the true representation of the process. For instance the process could be stationary but non-linear or noninvertible. 32

33 2.5 Other fundamental MA( ) Representations It is easy to extend the Wold representation to the general class of fundamental MA( ) representations. For any non singular matrix R of constant we have Y t = C(L)Ru t = D(L)u t where u t W N(0, R 1 ΩR 1 ) and u t = R 1 ɛ t. Fundamentalness is ensured since u t is a linear combination of the Wold shocks. The roots of the deteminant of D(L) will coincide with those of C(L). 33

34 3 VAR: representations If the MA matrix of lag polynomials is invertible, then for a vector stationary process a Vector Autoregressive (VAR) representation exists. We define C(L) 1 as an (n n) lag polynomial such that C(L) 1 C(L) = I; i.e. when these lag polynomial matrices are matrix-multiplied, all the lag terms cancel out. This operation in effect converts lags of the errors into lags of the vector of dependent variables. Thus we move from MA coefficient to VAR coefficients. Define A(L) = C(L) 1. Then given the (invertible) MA coefficients, it is easy to map these into the VAR coefficients: Y t = C(L)ɛ t A(L)Y t = ɛ t (3) where A(L) = A 0 L 0 + A 1 L 1 + A 2 L and A j for all j are (n n) matrices of coefficients. 34

35 To show that this matrix lag polynomial exists and how it maps into the coefficients in C(L), note that by assumption we have the identity (A 0 + A 1 L 1 + A 2 L )(I + C 1 L 1 + C 2 L ) = I After distributing, the identity implies that coefficients on the lag operators must be zero, which implies the following recursive solution for the VAR coefficients: A 0 = I A 1 = A 0 C 1 A k = A 0 C k A 1 C k... A k 1 C 1 As noted, the VAR is of infinite order (i.e. infinite number of lags required to fully represent joint density). In practice, the VAR is usually restricted for estimation by truncating the lag-length. The pth-order vector autoregression, denoted VAR(p) is given by Y t = A 1 Y t 1 + A 2 Y t A p Y t p + ɛ t (4) 35

36 Note: Here we are considering zero mean processes. In case the mean of Y t is not zero we should add a constant in the VAR equations. VAR(1) representation Any VAR(p) can be rewritten as a VAR(1). To form a VAR(1) from the general model we define: e t = [ɛ, 0,..., 0], Y t = [Y t, Y t 1,..., Y t p+1] A 1 A 2... A p 1 A p I n A = 0 I n I n 0 Therefore we can rewrite the VAR(p) as a VAR(1) Y t = AY t 1 + e t This is also known as the companion form of the VAR(p). 36

37 SUR representation The VAR(p) can be stacked as Y = XΓ + u where X = [X 1,..., X T ], X t = [Y [ɛ 1,..., ɛ T ] and Γ = [A 1,..., A p ] t 1, Y t 2..., Y t p] Y = [Y 1,..., Y T ], u = Vec representation Let vec denote the stacking columns operator, i.e X = X 11 X 12 X 21 X 22 then vec(x) = X 31 X 32 X 11 X 21 X 31 X 12 X 22 X 32 Let γ = vec(γ), then the VAR can be rewritten as Y t = (I n X t)γ + ɛ t 37

38 4 VAR: Stationarity 4.1 Stability and stationarity Consider the VAR(1) Y t = µ + AY t 1 + ε t Substituting backward we obtain Y t = µ + AY t 1 + ε t = µ + A(µ + AY t 2 + ε t 1 ) + ε t = (I + A)µ + A 2 Y t 2 + Aε t 1 + ε t. Y t = (I + A A j )µ + A j Y t j + j 1 i=0 A i ε t i If all the eigenvalues of A are smaller than one in modulus then 1. A j = P Λ j P the sequence A i, i = 0, 1,... is absolutely summable. 38

39 3. the infinite sum j 1 i=0 Ai ε t i exists in mean square (see e.g. proposition C.10L); 4. (I + A A j )µ (I A) 1 and A j 0 as j goes to infinity. Therefore if the eigenvalues are smaller than one in modulus then Y t has the following representation Y t = (I A) 1 + A i ε t i i=0 Note that the eigenvalues (λ) of A satisfy det(iλ A) = 0. Therefore the eigenvalues correspond to the reciprocal of the roots of the determinant of A(z) = I Az. A VAR(1) is called stable if det(i Az) 0 for z 1. Equivalently stability requires that all the eigenvalues of A are smaller than one in absoulte value. 39

40 For a VAR(p) the stability condition also requires that all the eigenvalues of A (the AR matrix of the companion form of Y t ) are smaller than one in modulus or all the roots larger than one. Therefore we have that a VAR(p) is called stable if det(i A 1 z A 2 z 2,..., A p z p ) 0 for z 1. A condition for stationarity: A stable VAR process is stationary. Notice that the converse is not true. An unstable process can be stationary. Notice that the vector M A( ) representation of a stationary VAR satisfies the absolute summability condition so that assumptions of 10.2H hold. 40

41 4.2 Back the Wold representation How can we find it the Wold representation starting from a VAR? We need to invert the VAR representation. So let us rewrite the VAR(p) as a VAR(1). Substituting backward in the companion form we have Y t = A j Y t j + A j 1 e t j A 1 e t e t If conditions for stationarity are satisfied, the series i=1 Aj converges and Y t has an VMA( ) representation in terms of the Wold shock e t given by Y t = (I AL) 1 e t = A j e t j i=1 = C(L)e t where C 0 = A 0 = I, C 1 = A 1, C 2 = A 2,..., C k = A k. The coefficients of the Wold representation of Y t, C j, will be the n n upper left matrix of C j. 41

42 Example A stationary VAR(1) ( ) ( ) ( ) ( ) Y1t Y1t 1 ɛ1t = + Y 2t Y 2t 1 ɛ 2t ( ) ( ) E(ɛ t ɛ t) = Ω = λ = Figure 1: Blu: Y 1, green Y 2. 42

43 5 VAR: second moments Let us consider the companion form of a stationary (zero mean for simplicity) VAR(p) defined earlier Y t = AY t 1 + e t (5) The variance of Y t is given by Σ = E(Y t Y t) = AΣA + Ω (6) a closed form solution to (6) can be obtained in terms of the vec operator. Let A, B, C be matrices such that the product ABC exists. A property of the vec operator is that vec(abc) = (C A)vec(B) Applying the vec operator to both sides of (7) we have vec(σ) = (A A)vec(Σ) + vec(ω) If we define A = (A A) then we have vec(σ) = (I A) 1 vec(ω) 43

44 The jth autocovariance of Y t (denoted Γ j ) can be found by post multiplying (6) by Y t j and taking expectations: Thus or E(Y t Y t j ) = AE(Y t Y t j ) + E(e t Y t j ) Γ j = AΓ j 1 Γ j = A j Γ The variance Σ and the jth autocovariance Γ j of the original series Y t is given by the first n rows and columns of Σ and Γ j respectively. 44

45 6 VAR: specification Specification of the VAR is key for empirical analysis. We have to decide about the following: 1. Number of lags p. 2. Which variables. 3. Type of transformations. 45

46 6.1 Number of lags As in the univariate case, care must be taken to account for all systematic dynamics in multivariate models. In VAR models, this is usually done by choosing a sufficient number of lags to ensure that the residuals in each of the equations are white noise. AIC: Akaike information criterion Choosing the p that minimizes the following AIC(p) = T ln ˆΩ + 2(n 2 p) BIC: Bayesian information criterionchoosing the p that minimizes the following BIC(p) = T ln ˆΩ + (n 2 p) ln T HQ: Hannan- Quinn information criterion Choosing the p that minimizes the following HQ(p) = T ln ˆΩ + (n 2 p) ln T T 46

47 ˆp obtained using BIC and HQ are consistent while with AIC it is not. AIC overestimate the true order with positive probability and underestimate the true order with zero probability. Suppose a VAR(p) is fitted to Y 1,..., Y T small sample the following relations hold: (Y t not necessarily stationary). In ˆp BIC ˆp AIC if T 8 ˆp BIC ˆp HQ for all T ˆp HQ ˆp AIC if T 16 47

48 6.2 Type of variables Variables selection is a key step in the specification of the model. VAR models are small scale models so usually 2 to 8 variables are used. Variables selection depends on the particular application and in general should be included all the variable conveying relevant information. We will see several examples. 48

49 6.3 Type of transformations Problem: many economic time series display trend over time and are clearly non stationary (mean not constant). Trend-stationary series Y t = µ + bt + ε t, ε t W N. Difference-stationary series Y t = µ + Y t 1 + ε t, ε t W N. 49

50 Figure 1: blu: log(gdp). green: log(cpi) 50

51 These series can be thought as generated by some nonstationary process. Here there are some examples Example: Difference stationary ( ) ( ) ( ) ( Y1t Y1t 1 = Y 2t Y 2t 1 ) ( ) ɛ1t +, Ω = ɛ 2t ( ), λ = ( )

52 Example: Trend stationary ( ) ( ) ( ) ( Y1t Y1t 1 = t Y 2t Y 2t 1 ) ( ) ɛ1t + ɛ 2t Ω = ( ), λ = ( )

53 So: 1) How do I know whether the series are stationary? 2) What to do if they are non stationary? 53

54 Dickey-Fuller test In 1979, Dickey and Fuller have proposed the following test for stationarity 1. Estimate with OLS the following equation x t = b + γx t + ε t 2. Test the null γ = 0 against the alternative γ < If the null is not rejected then which is a random walk with drift. x t = b + x t + ε t 4. On the contrary if γ t < 0, then x t is a stationary AR with a = 1 + γ < 1. x t = b + ax t + ε t An alternative is to specify the equation augmented by a deterministic trend x t = b + γx t + ct + ε t 54

55 With this specification under the alternative the preocess is stationary with a deterministic linear trend. Augmented Dickey-Fuller test. In the augmented version of the test p lags of the lags of x t can be added, i.e. or A(L) x t = b + γx t + ε t A(L) x t = b + γx t + ct + ε t If the test statistic is smaller than (negative) the critical value, then the null hypothesis of unit root is rejected. 55

56 Transformations I: first differences Let = 1 L be the first differences filter, i.e. a filter such that Y t = Y t Y t 1 and let us consider the simple case of a random walk with drift Y t = µ + Y t 1 + ɛ t where ɛ t is WN. By applying the first differences filter (1 L) the process is transformed into a stationary process Y t = µ + ɛ t Let us now consider a process with deterministic trend By differencing the process we obtain Y t = µ + δt + ɛ t Y t = δ + ɛ t which is a stationary process but is not invertible because it contains a unit root in the MA part. 56

57 log(gdp) and log(cpi) in first differences 57

58 Transformations II: removing deterministic trend Removing a deterministic trend (linear or quadratic) from a process from a trend stationary variable is ok. However this is not enough if the process is a unit root with drift. To see this consider again the process this can be writen as Y t = µ + Y t 1 + ɛ t Y t = µt + Y 0 + By removing the deterministic trend the mean of the process becomes constant but the variance grows over time so the process is not stationary. t j=0 ɛ j 58

59 log(gdp) and log(cpi) linearly detrended 59

60 Transformations of trending variables: Hodrick-Prescott filter The filter separates the trend from the cyclical component of a scalar time series. Suppose y t = g t + c t, where g t is the trend component and c t is the cycle. The trend is obtained by solving the following minimization problem min {g t } T t=1 T T 1 c 2 t + λ [(g t+1 g t ) (g t g t 1 )] 2 t=1 t=2 The parameter λ is a positive number (quarterly data usually =1600) which penalizes variability in the growth component series while the first part is the penalty to the cyclical component. The larger λ the smoother the trend component. 60

5: MULTIVARATE STATIONARY PROCESSES

5: MULTIVARATE STATIONARY PROCESSES 5: MULTIVARATE STATIONARY PROCESSES 1 1 Some Preliminary Definitions and Concepts Random Vector: A vector X = (X 1,..., X n ) whose components are scalarvalued random variables on the same probability

More information

7. MULTIVARATE STATIONARY PROCESSES

7. MULTIVARATE STATIONARY PROCESSES 7. MULTIVARATE STATIONARY PROCESSES 1 1 Some Preliminary Definitions and Concepts Random Vector: A vector X = (X 1,..., X n ) whose components are scalar-valued random variables on the same probability

More information

TIME SERIES AND FORECASTING. Luca Gambetti UAB, Barcelona GSE Master in Macroeconomic Policy and Financial Markets

TIME SERIES AND FORECASTING. Luca Gambetti UAB, Barcelona GSE Master in Macroeconomic Policy and Financial Markets TIME SERIES AND FORECASTING Luca Gambetti UAB, Barcelona GSE 2014-2015 Master in Macroeconomic Policy and Financial Markets 1 Contacts Prof.: Luca Gambetti Office: B3-1130 Edifici B Office hours: email:

More information

Topic 4 Unit Roots. Gerald P. Dwyer. February Clemson University

Topic 4 Unit Roots. Gerald P. Dwyer. February Clemson University Topic 4 Unit Roots Gerald P. Dwyer Clemson University February 2016 Outline 1 Unit Roots Introduction Trend and Difference Stationary Autocorrelations of Series That Have Deterministic or Stochastic Trends

More information

Prof. Dr. Roland Füss Lecture Series in Applied Econometrics Summer Term Introduction to Time Series Analysis

Prof. Dr. Roland Füss Lecture Series in Applied Econometrics Summer Term Introduction to Time Series Analysis Introduction to Time Series Analysis 1 Contents: I. Basics of Time Series Analysis... 4 I.1 Stationarity... 5 I.2 Autocorrelation Function... 9 I.3 Partial Autocorrelation Function (PACF)... 14 I.4 Transformation

More information

Vector Auto-Regressive Models

Vector Auto-Regressive Models Vector Auto-Regressive Models Laurent Ferrara 1 1 University of Paris Nanterre M2 Oct. 2018 Overview of the presentation 1. Vector Auto-Regressions Definition Estimation Testing 2. Impulse responses functions

More information

VAR Models and Applications

VAR Models and Applications VAR Models and Applications Laurent Ferrara 1 1 University of Paris West M2 EIPMC Oct. 2016 Overview of the presentation 1. Vector Auto-Regressions Definition Estimation Testing 2. Impulse responses functions

More information

Time Series Analysis. James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY

Time Series Analysis. James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY Time Series Analysis James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY & Contents PREFACE xiii 1 1.1. 1.2. Difference Equations First-Order Difference Equations 1 /?th-order Difference

More information

3. ARMA Modeling. Now: Important class of stationary processes

3. ARMA Modeling. Now: Important class of stationary processes 3. ARMA Modeling Now: Important class of stationary processes Definition 3.1: (ARMA(p, q) process) Let {ɛ t } t Z WN(0, σ 2 ) be a white noise process. The process {X t } t Z is called AutoRegressive-Moving-Average

More information

Cointegrated VAR s. Eduardo Rossi University of Pavia. November Rossi Cointegrated VAR s Financial Econometrics / 56

Cointegrated VAR s. Eduardo Rossi University of Pavia. November Rossi Cointegrated VAR s Financial Econometrics / 56 Cointegrated VAR s Eduardo Rossi University of Pavia November 2013 Rossi Cointegrated VAR s Financial Econometrics - 2013 1 / 56 VAR y t = (y 1t,..., y nt ) is (n 1) vector. y t VAR(p): Φ(L)y t = ɛ t The

More information

ECON 616: Lecture 1: Time Series Basics

ECON 616: Lecture 1: Time Series Basics ECON 616: Lecture 1: Time Series Basics ED HERBST August 30, 2017 References Overview: Chapters 1-3 from Hamilton (1994). Technical Details: Chapters 2-3 from Brockwell and Davis (1987). Intuition: Chapters

More information

Time Series Analysis. James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY

Time Series Analysis. James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY Time Series Analysis James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY PREFACE xiii 1 Difference Equations 1.1. First-Order Difference Equations 1 1.2. pth-order Difference Equations 7

More information

Ch. 14 Stationary ARMA Process

Ch. 14 Stationary ARMA Process Ch. 14 Stationary ARMA Process A general linear stochastic model is described that suppose a time series to be generated by a linear aggregation of random shock. For practical representation it is desirable

More information

Vector autoregressive Moving Average Process. Presented by Muhammad Iqbal, Amjad Naveed and Muhammad Nadeem

Vector autoregressive Moving Average Process. Presented by Muhammad Iqbal, Amjad Naveed and Muhammad Nadeem Vector autoregressive Moving Average Process Presented by Muhammad Iqbal, Amjad Naveed and Muhammad Nadeem Road Map 1. Introduction 2. Properties of MA Finite Process 3. Stationarity of MA Process 4. VARMA

More information

Econometría 2: Análisis de series de Tiempo

Econometría 2: Análisis de series de Tiempo Econometría 2: Análisis de series de Tiempo Karoll GOMEZ kgomezp@unal.edu.co http://karollgomez.wordpress.com Segundo semestre 2016 IX. Vector Time Series Models VARMA Models A. 1. Motivation: The vector

More information

Week 5 Quantitative Analysis of Financial Markets Characterizing Cycles

Week 5 Quantitative Analysis of Financial Markets Characterizing Cycles Week 5 Quantitative Analysis of Financial Markets Characterizing Cycles Christopher Ting http://www.mysmu.edu/faculty/christophert/ Christopher Ting : christopherting@smu.edu.sg : 6828 0364 : LKCSB 5036

More information

Covariance Stationary Time Series. Example: Independent White Noise (IWN(0,σ 2 )) Y t = ε t, ε t iid N(0,σ 2 )

Covariance Stationary Time Series. Example: Independent White Noise (IWN(0,σ 2 )) Y t = ε t, ε t iid N(0,σ 2 ) Covariance Stationary Time Series Stochastic Process: sequence of rv s ordered by time {Y t } {...,Y 1,Y 0,Y 1,...} Defn: {Y t } is covariance stationary if E[Y t ]μ for all t cov(y t,y t j )E[(Y t μ)(y

More information

ECONOMETRICS Part II PhD LBS

ECONOMETRICS Part II PhD LBS ECONOMETRICS Part II PhD LBS Luca Gambetti UAB, Barcelona GSE February-March 2014 1 Contacts Prof.: Luca Gambetti email: luca.gambetti@uab.es webpage: http://pareto.uab.es/lgambetti/ Description This is

More information

Empirical Macroeconomics

Empirical Macroeconomics Empirical Macroeconomics Francesco Franco Nova SBE April 5, 2016 Francesco Franco Empirical Macroeconomics 1/39 Growth and Fluctuations Supply and Demand Figure : US dynamics Francesco Franco Empirical

More information

Lecture 1: Stationary Time Series Analysis

Lecture 1: Stationary Time Series Analysis Syllabus Stationarity ARMA AR MA Model Selection Estimation Lecture 1: Stationary Time Series Analysis 222061-1617: Time Series Econometrics Spring 2018 Jacek Suda Syllabus Stationarity ARMA AR MA Model

More information

1 Class Organization. 2 Introduction

1 Class Organization. 2 Introduction Time Series Analysis, Lecture 1, 2018 1 1 Class Organization Course Description Prerequisite Homework and Grading Readings and Lecture Notes Course Website: http://www.nanlifinance.org/teaching.html wechat

More information

Chapter 2: Unit Roots

Chapter 2: Unit Roots Chapter 2: Unit Roots 1 Contents: Lehrstuhl für Department Empirische of Wirtschaftsforschung Empirical Research and undeconometrics II. Unit Roots... 3 II.1 Integration Level... 3 II.2 Nonstationarity

More information

Trend-Cycle Decompositions

Trend-Cycle Decompositions Trend-Cycle Decompositions Eric Zivot April 22, 2005 1 Introduction A convenient way of representing an economic time series y t is through the so-called trend-cycle decomposition y t = TD t + Z t (1)

More information

1 Introduction to Multivariate Models

1 Introduction to Multivariate Models 1 Introduction to Multivariate Models Beyond univariate models... Consider the following AR(2) process for inflation (y t ) Y t = a 1 Y t 1 + a 2 Y t 2 + ε t ε t W N In this course we study multivariate

More information

Empirical Macroeconomics

Empirical Macroeconomics Empirical Macroeconomics Francesco Franco Nova SBE April 21, 2015 Francesco Franco Empirical Macroeconomics 1/33 Growth and Fluctuations Supply and Demand Figure : US dynamics Francesco Franco Empirical

More information

2.5 Forecasting and Impulse Response Functions

2.5 Forecasting and Impulse Response Functions 2.5 Forecasting and Impulse Response Functions Principles of forecasting Forecast based on conditional expectations Suppose we are interested in forecasting the value of y t+1 based on a set of variables

More information

Multivariate Time Series: VAR(p) Processes and Models

Multivariate Time Series: VAR(p) Processes and Models Multivariate Time Series: VAR(p) Processes and Models A VAR(p) model, for p > 0 is X t = φ 0 + Φ 1 X t 1 + + Φ p X t p + A t, where X t, φ 0, and X t i are k-vectors, Φ 1,..., Φ p are k k matrices, with

More information

11. Further Issues in Using OLS with TS Data

11. Further Issues in Using OLS with TS Data 11. Further Issues in Using OLS with TS Data With TS, including lags of the dependent variable often allow us to fit much better the variation in y Exact distribution theory is rarely available in TS applications,

More information

Time Series Models and Inference. James L. Powell Department of Economics University of California, Berkeley

Time Series Models and Inference. James L. Powell Department of Economics University of California, Berkeley Time Series Models and Inference James L. Powell Department of Economics University of California, Berkeley Overview In contrast to the classical linear regression model, in which the components of the

More information

Multivariate Time Series Analysis and Its Applications [Tsay (2005), chapter 8]

Multivariate Time Series Analysis and Its Applications [Tsay (2005), chapter 8] 1 Multivariate Time Series Analysis and Its Applications [Tsay (2005), chapter 8] Insights: Price movements in one market can spread easily and instantly to another market [economic globalization and internet

More information

1 Linear Difference Equations

1 Linear Difference Equations ARMA Handout Jialin Yu 1 Linear Difference Equations First order systems Let {ε t } t=1 denote an input sequence and {y t} t=1 sequence generated by denote an output y t = φy t 1 + ε t t = 1, 2,... with

More information

Class 1: Stationary Time Series Analysis

Class 1: Stationary Time Series Analysis Class 1: Stationary Time Series Analysis Macroeconometrics - Fall 2009 Jacek Suda, BdF and PSE February 28, 2011 Outline Outline: 1 Covariance-Stationary Processes 2 Wold Decomposition Theorem 3 ARMA Models

More information

Questions and Answers on Unit Roots, Cointegration, VARs and VECMs

Questions and Answers on Unit Roots, Cointegration, VARs and VECMs Questions and Answers on Unit Roots, Cointegration, VARs and VECMs L. Magee Winter, 2012 1. Let ɛ t, t = 1,..., T be a series of independent draws from a N[0,1] distribution. Let w t, t = 1,..., T, be

More information

Advanced Econometrics

Advanced Econometrics Based on the textbook by Verbeek: A Guide to Modern Econometrics Robert M. Kunst robert.kunst@univie.ac.at University of Vienna and Institute for Advanced Studies Vienna May 2, 2013 Outline Univariate

More information

A TIME SERIES PARADOX: UNIT ROOT TESTS PERFORM POORLY WHEN DATA ARE COINTEGRATED

A TIME SERIES PARADOX: UNIT ROOT TESTS PERFORM POORLY WHEN DATA ARE COINTEGRATED A TIME SERIES PARADOX: UNIT ROOT TESTS PERFORM POORLY WHEN DATA ARE COINTEGRATED by W. Robert Reed Department of Economics and Finance University of Canterbury, New Zealand Email: bob.reed@canterbury.ac.nz

More information

Notes on Time Series Modeling

Notes on Time Series Modeling Notes on Time Series Modeling Garey Ramey University of California, San Diego January 17 1 Stationary processes De nition A stochastic process is any set of random variables y t indexed by t T : fy t g

More information

Empirical Market Microstructure Analysis (EMMA)

Empirical Market Microstructure Analysis (EMMA) Empirical Market Microstructure Analysis (EMMA) Lecture 3: Statistical Building Blocks and Econometric Basics Prof. Dr. Michael Stein michael.stein@vwl.uni-freiburg.de Albert-Ludwigs-University of Freiburg

More information

Univariate Nonstationary Time Series 1

Univariate Nonstationary Time Series 1 Univariate Nonstationary Time Series 1 Sebastian Fossati University of Alberta 1 These slides are based on Eric Zivot s time series notes available at: http://faculty.washington.edu/ezivot Introduction

More information

Title. Description. var intro Introduction to vector autoregressive models

Title. Description. var intro Introduction to vector autoregressive models Title var intro Introduction to vector autoregressive models Description Stata has a suite of commands for fitting, forecasting, interpreting, and performing inference on vector autoregressive (VAR) models

More information

Discrete time processes

Discrete time processes Discrete time processes Predictions are difficult. Especially about the future Mark Twain. Florian Herzog 2013 Modeling observed data When we model observed (realized) data, we encounter usually the following

More information

Introduction to Stochastic processes

Introduction to Stochastic processes Università di Pavia Introduction to Stochastic processes Eduardo Rossi Stochastic Process Stochastic Process: A stochastic process is an ordered sequence of random variables defined on a probability space

More information

Econ 423 Lecture Notes: Additional Topics in Time Series 1

Econ 423 Lecture Notes: Additional Topics in Time Series 1 Econ 423 Lecture Notes: Additional Topics in Time Series 1 John C. Chao April 25, 2017 1 These notes are based in large part on Chapter 16 of Stock and Watson (2011). They are for instructional purposes

More information

Lecture 1: Stationary Time Series Analysis

Lecture 1: Stationary Time Series Analysis Syllabus Stationarity ARMA AR MA Model Selection Estimation Forecasting Lecture 1: Stationary Time Series Analysis 222061-1617: Time Series Econometrics Spring 2018 Jacek Suda Syllabus Stationarity ARMA

More information

Lecture 2: Univariate Time Series

Lecture 2: Univariate Time Series Lecture 2: Univariate Time Series Analysis: Conditional and Unconditional Densities, Stationarity, ARMA Processes Prof. Massimo Guidolin 20192 Financial Econometrics Spring/Winter 2017 Overview Motivation:

More information

It is easily seen that in general a linear combination of y t and x t is I(1). However, in particular cases, it can be I(0), i.e. stationary.

It is easily seen that in general a linear combination of y t and x t is I(1). However, in particular cases, it can be I(0), i.e. stationary. 6. COINTEGRATION 1 1 Cointegration 1.1 Definitions I(1) variables. z t = (y t x t ) is I(1) (integrated of order 1) if it is not stationary but its first difference z t is stationary. It is easily seen

More information

Advanced Econometrics

Advanced Econometrics Advanced Econometrics Marco Sunder Nov 04 2010 Marco Sunder Advanced Econometrics 1/ 25 Contents 1 2 3 Marco Sunder Advanced Econometrics 2/ 25 Music Marco Sunder Advanced Econometrics 3/ 25 Music Marco

More information

Vector autoregressions, VAR

Vector autoregressions, VAR 1 / 45 Vector autoregressions, VAR Chapter 2 Financial Econometrics Michael Hauser WS17/18 2 / 45 Content Cross-correlations VAR model in standard/reduced form Properties of VAR(1), VAR(p) Structural VAR,

More information

Booth School of Business, University of Chicago Business 41914, Spring Quarter 2013, Mr. Ruey S. Tsay. Midterm

Booth School of Business, University of Chicago Business 41914, Spring Quarter 2013, Mr. Ruey S. Tsay. Midterm Booth School of Business, University of Chicago Business 41914, Spring Quarter 2013, Mr. Ruey S. Tsay Midterm Chicago Booth Honor Code: I pledge my honor that I have not violated the Honor Code during

More information

Autoregressive Moving Average (ARMA) Models and their Practical Applications

Autoregressive Moving Average (ARMA) Models and their Practical Applications Autoregressive Moving Average (ARMA) Models and their Practical Applications Massimo Guidolin February 2018 1 Essential Concepts in Time Series Analysis 1.1 Time Series and Their Properties Time series:

More information

Ch. 15 Forecasting. 1.1 Forecasts Based on Conditional Expectations

Ch. 15 Forecasting. 1.1 Forecasts Based on Conditional Expectations Ch 15 Forecasting Having considered in Chapter 14 some of the properties of ARMA models, we now show how they may be used to forecast future values of an observed time series For the present we proceed

More information

Chapter 2. Some basic tools. 2.1 Time series: Theory Stochastic processes

Chapter 2. Some basic tools. 2.1 Time series: Theory Stochastic processes Chapter 2 Some basic tools 2.1 Time series: Theory 2.1.1 Stochastic processes A stochastic process is a sequence of random variables..., x 0, x 1, x 2,.... In this class, the subscript always means time.

More information

Econ 623 Econometrics II Topic 2: Stationary Time Series

Econ 623 Econometrics II Topic 2: Stationary Time Series 1 Introduction Econ 623 Econometrics II Topic 2: Stationary Time Series In the regression model we can model the error term as an autoregression AR(1) process. That is, we can use the past value of the

More information

ECON/FIN 250: Forecasting in Finance and Economics: Section 7: Unit Roots & Dickey-Fuller Tests

ECON/FIN 250: Forecasting in Finance and Economics: Section 7: Unit Roots & Dickey-Fuller Tests ECON/FIN 250: Forecasting in Finance and Economics: Section 7: Unit Roots & Dickey-Fuller Tests Patrick Herb Brandeis University Spring 2016 Patrick Herb (Brandeis University) Unit Root Tests ECON/FIN

More information

Lecture 1: Fundamental concepts in Time Series Analysis (part 2)

Lecture 1: Fundamental concepts in Time Series Analysis (part 2) Lecture 1: Fundamental concepts in Time Series Analysis (part 2) Florian Pelgrin University of Lausanne, École des HEC Department of mathematics (IMEA-Nice) Sept. 2011 - Jan. 2012 Florian Pelgrin (HEC)

More information

Some Time-Series Models

Some Time-Series Models Some Time-Series Models Outline 1. Stochastic processes and their properties 2. Stationary processes 3. Some properties of the autocorrelation function 4. Some useful models Purely random processes, random

More information

Statistics of stochastic processes

Statistics of stochastic processes Introduction Statistics of stochastic processes Generally statistics is performed on observations y 1,..., y n assumed to be realizations of independent random variables Y 1,..., Y n. 14 settembre 2014

More information

Cointegrated VAR s. Eduardo Rossi University of Pavia. November Rossi Cointegrated VAR s Fin. Econometrics / 31

Cointegrated VAR s. Eduardo Rossi University of Pavia. November Rossi Cointegrated VAR s Fin. Econometrics / 31 Cointegrated VAR s Eduardo Rossi University of Pavia November 2014 Rossi Cointegrated VAR s Fin. Econometrics - 2014 1 / 31 B-N decomposition Give a scalar polynomial α(z) = α 0 + α 1 z +... + α p z p

More information

ECON 4160, Spring term Lecture 12

ECON 4160, Spring term Lecture 12 ECON 4160, Spring term 2013. Lecture 12 Non-stationarity and co-integration 2/2 Ragnar Nymoen Department of Economics 13 Nov 2013 1 / 53 Introduction I So far we have considered: Stationary VAR, with deterministic

More information

FE570 Financial Markets and Trading. Stevens Institute of Technology

FE570 Financial Markets and Trading. Stevens Institute of Technology FE570 Financial Markets and Trading Lecture 5. Linear Time Series Analysis and Its Applications (Ref. Joel Hasbrouck - Empirical Market Microstructure ) Steve Yang Stevens Institute of Technology 9/25/2012

More information

AR, MA and ARMA models

AR, MA and ARMA models AR, MA and AR by Hedibert Lopes P Based on Tsay s Analysis of Financial Time Series (3rd edition) P 1 Stationarity 2 3 4 5 6 7 P 8 9 10 11 Outline P Linear Time Series Analysis and Its Applications For

More information

Non-Stationary Time Series and Unit Root Testing

Non-Stationary Time Series and Unit Root Testing Econometrics II Non-Stationary Time Series and Unit Root Testing Morten Nyboe Tabor Course Outline: Non-Stationary Time Series and Unit Root Testing 1 Stationarity and Deviation from Stationarity Trend-Stationarity

More information

ECON 4160, Lecture 11 and 12

ECON 4160, Lecture 11 and 12 ECON 4160, 2016. Lecture 11 and 12 Co-integration Ragnar Nymoen Department of Economics 9 November 2017 1 / 43 Introduction I So far we have considered: Stationary VAR ( no unit roots ) Standard inference

More information

MA Advanced Econometrics: Applying Least Squares to Time Series

MA Advanced Econometrics: Applying Least Squares to Time Series MA Advanced Econometrics: Applying Least Squares to Time Series Karl Whelan School of Economics, UCD February 15, 2011 Karl Whelan (UCD) Time Series February 15, 2011 1 / 24 Part I Time Series: Standard

More information

Single Equation Linear GMM with Serially Correlated Moment Conditions

Single Equation Linear GMM with Serially Correlated Moment Conditions Single Equation Linear GMM with Serially Correlated Moment Conditions Eric Zivot October 28, 2009 Univariate Time Series Let {y t } be an ergodic-stationary time series with E[y t ]=μ and var(y t )

More information

Non-Stationary Time Series and Unit Root Testing

Non-Stationary Time Series and Unit Root Testing Econometrics II Non-Stationary Time Series and Unit Root Testing Morten Nyboe Tabor Course Outline: Non-Stationary Time Series and Unit Root Testing 1 Stationarity and Deviation from Stationarity Trend-Stationarity

More information

Problem Set 1 Solution Sketches Time Series Analysis Spring 2010

Problem Set 1 Solution Sketches Time Series Analysis Spring 2010 Problem Set 1 Solution Sketches Time Series Analysis Spring 2010 1. Construct a martingale difference process that is not weakly stationary. Simplest e.g.: Let Y t be a sequence of independent, non-identically

More information

Single Equation Linear GMM with Serially Correlated Moment Conditions

Single Equation Linear GMM with Serially Correlated Moment Conditions Single Equation Linear GMM with Serially Correlated Moment Conditions Eric Zivot November 2, 2011 Univariate Time Series Let {y t } be an ergodic-stationary time series with E[y t ]=μ and var(y t )

More information

7. Integrated Processes

7. Integrated Processes 7. Integrated Processes Up to now: Analysis of stationary processes (stationary ARMA(p, q) processes) Problem: Many economic time series exhibit non-stationary patterns over time 226 Example: We consider

More information

Multivariate Time Series: Part 4

Multivariate Time Series: Part 4 Multivariate Time Series: Part 4 Cointegration Gerald P. Dwyer Clemson University March 2016 Outline 1 Multivariate Time Series: Part 4 Cointegration Engle-Granger Test for Cointegration Johansen Test

More information

Cointegration Lecture I: Introduction

Cointegration Lecture I: Introduction 1 Cointegration Lecture I: Introduction Julia Giese Nuffield College julia.giese@economics.ox.ac.uk Hilary Term 2008 2 Outline Introduction Estimation of unrestricted VAR Non-stationarity Deterministic

More information

New Introduction to Multiple Time Series Analysis

New Introduction to Multiple Time Series Analysis Helmut Lütkepohl New Introduction to Multiple Time Series Analysis With 49 Figures and 36 Tables Springer Contents 1 Introduction 1 1.1 Objectives of Analyzing Multiple Time Series 1 1.2 Some Basics 2

More information

Econometrics II Heij et al. Chapter 7.1

Econometrics II Heij et al. Chapter 7.1 Chapter 7.1 p. 1/2 Econometrics II Heij et al. Chapter 7.1 Linear Time Series Models for Stationary data Marius Ooms Tinbergen Institute Amsterdam Chapter 7.1 p. 2/2 Program Introduction Modelling philosophy

More information

LINEAR STOCHASTIC MODELS

LINEAR STOCHASTIC MODELS LINEAR STOCHASTIC MODELS Let {x τ+1,x τ+2,...,x τ+n } denote n consecutive elements from a stochastic process. If their joint distribution does not depend on τ, regardless of the size of n, then the process

More information

Basic concepts and terminology: AR, MA and ARMA processes

Basic concepts and terminology: AR, MA and ARMA processes ECON 5101 ADVANCED ECONOMETRICS TIME SERIES Lecture note no. 1 (EB) Erik Biørn, Department of Economics Version of February 1, 2011 Basic concepts and terminology: AR, MA and ARMA processes This lecture

More information

ECON 616: Lecture Two: Deterministic Trends, Nonstationary Processes

ECON 616: Lecture Two: Deterministic Trends, Nonstationary Processes ECON 616: Lecture Two: Deterministic Trends, Nonstationary Processes ED HERBST September 11, 2017 Background Hamilton, chapters 15-16 Trends vs Cycles A commond decomposition of macroeconomic time series

More information

Applied time-series analysis

Applied time-series analysis Robert M. Kunst robert.kunst@univie.ac.at University of Vienna and Institute for Advanced Studies Vienna October 18, 2011 Outline Introduction and overview Econometric Time-Series Analysis In principle,

More information

Chapter 5. Analysis of Multiple Time Series. 5.1 Vector Autoregressions

Chapter 5. Analysis of Multiple Time Series. 5.1 Vector Autoregressions Chapter 5 Analysis of Multiple Time Series Note: The primary references for these notes are chapters 5 and 6 in Enders (2004). An alternative, but more technical treatment can be found in chapters 10-11

More information

Multivariate Time Series

Multivariate Time Series Multivariate Time Series Fall 2008 Environmental Econometrics (GR03) TSII Fall 2008 1 / 16 More on AR(1) In AR(1) model (Y t = µ + ρy t 1 + u t ) with ρ = 1, the series is said to have a unit root or a

More information

Review Session: Econometrics - CLEFIN (20192)

Review Session: Econometrics - CLEFIN (20192) Review Session: Econometrics - CLEFIN (20192) Part II: Univariate time series analysis Daniele Bianchi March 20, 2013 Fundamentals Stationarity A time series is a sequence of random variables x t, t =

More information

10. Time series regression and forecasting

10. Time series regression and forecasting 10. Time series regression and forecasting Key feature of this section: Analysis of data on a single entity observed at multiple points in time (time series data) Typical research questions: What is the

More information

Introduction to Modern Time Series Analysis

Introduction to Modern Time Series Analysis Introduction to Modern Time Series Analysis Gebhard Kirchgässner, Jürgen Wolters and Uwe Hassler Second Edition Springer 3 Teaching Material The following figures and tables are from the above book. They

More information

E 4160 Autumn term Lecture 9: Deterministic trends vs integrated series; Spurious regression; Dickey-Fuller distribution and test

E 4160 Autumn term Lecture 9: Deterministic trends vs integrated series; Spurious regression; Dickey-Fuller distribution and test E 4160 Autumn term 2016. Lecture 9: Deterministic trends vs integrated series; Spurious regression; Dickey-Fuller distribution and test Ragnar Nymoen Department of Economics, University of Oslo 24 October

More information

Econ 424 Time Series Concepts

Econ 424 Time Series Concepts Econ 424 Time Series Concepts Eric Zivot January 20 2015 Time Series Processes Stochastic (Random) Process { 1 2 +1 } = { } = sequence of random variables indexed by time Observed time series of length

More information

Multivariate forecasting with VAR models

Multivariate forecasting with VAR models Multivariate forecasting with VAR models Franz Eigner University of Vienna UK Econometric Forecasting Prof. Robert Kunst 16th June 2009 Overview Vector autoregressive model univariate forecasting multivariate

More information

7. Integrated Processes

7. Integrated Processes 7. Integrated Processes Up to now: Analysis of stationary processes (stationary ARMA(p, q) processes) Problem: Many economic time series exhibit non-stationary patterns over time 226 Example: We consider

More information

Vector error correction model, VECM Cointegrated VAR

Vector error correction model, VECM Cointegrated VAR 1 / 58 Vector error correction model, VECM Cointegrated VAR Chapter 4 Financial Econometrics Michael Hauser WS17/18 2 / 58 Content Motivation: plausible economic relations Model with I(1) variables: spurious

More information

Econometrics of Panel Data

Econometrics of Panel Data Econometrics of Panel Data Jakub Mućk Meeting # 9 Jakub Mućk Econometrics of Panel Data Meeting # 9 1 / 22 Outline 1 Time series analysis Stationarity Unit Root Tests for Nonstationarity 2 Panel Unit Root

More information

Cointegration, Stationarity and Error Correction Models.

Cointegration, Stationarity and Error Correction Models. Cointegration, Stationarity and Error Correction Models. STATIONARITY Wold s decomposition theorem states that a stationary time series process with no deterministic components has an infinite moving average

More information

Bonn Summer School Advances in Empirical Macroeconomics

Bonn Summer School Advances in Empirical Macroeconomics Bonn Summer School Advances in Empirical Macroeconomics Karel Mertens Cornell, NBER, CEPR Bonn, June 2015 In God we trust, all others bring data. William E. Deming (1900-1993) Angrist and Pischke are Mad

More information

LECTURE 10 LINEAR PROCESSES II: SPECTRAL DENSITY, LAG OPERATOR, ARMA. In this lecture, we continue to discuss covariance stationary processes.

LECTURE 10 LINEAR PROCESSES II: SPECTRAL DENSITY, LAG OPERATOR, ARMA. In this lecture, we continue to discuss covariance stationary processes. MAY, 0 LECTURE 0 LINEAR PROCESSES II: SPECTRAL DENSITY, LAG OPERATOR, ARMA In this lecture, we continue to discuss covariance stationary processes. Spectral density Gourieroux and Monfort 990), Ch. 5;

More information

Labor-Supply Shifts and Economic Fluctuations. Technical Appendix

Labor-Supply Shifts and Economic Fluctuations. Technical Appendix Labor-Supply Shifts and Economic Fluctuations Technical Appendix Yongsung Chang Department of Economics University of Pennsylvania Frank Schorfheide Department of Economics University of Pennsylvania January

More information

TIME SERIES ANALYSIS. Forecasting and Control. Wiley. Fifth Edition GWILYM M. JENKINS GEORGE E. P. BOX GREGORY C. REINSEL GRETA M.

TIME SERIES ANALYSIS. Forecasting and Control. Wiley. Fifth Edition GWILYM M. JENKINS GEORGE E. P. BOX GREGORY C. REINSEL GRETA M. TIME SERIES ANALYSIS Forecasting and Control Fifth Edition GEORGE E. P. BOX GWILYM M. JENKINS GREGORY C. REINSEL GRETA M. LJUNG Wiley CONTENTS PREFACE TO THE FIFTH EDITION PREFACE TO THE FOURTH EDITION

More information

Non-Stationary Time Series and Unit Root Testing

Non-Stationary Time Series and Unit Root Testing Econometrics II Non-Stationary Time Series and Unit Root Testing Morten Nyboe Tabor Course Outline: Non-Stationary Time Series and Unit Root Testing 1 Stationarity and Deviation from Stationarity Trend-Stationarity

More information

1. Stochastic Processes and Stationarity

1. Stochastic Processes and Stationarity Massachusetts Institute of Technology Department of Economics Time Series 14.384 Guido Kuersteiner Lecture Note 1 - Introduction This course provides the basic tools needed to analyze data that is observed

More information

Estimating and Identifying Vector Autoregressions Under Diagonality and Block Exogeneity Restrictions

Estimating and Identifying Vector Autoregressions Under Diagonality and Block Exogeneity Restrictions Estimating and Identifying Vector Autoregressions Under Diagonality and Block Exogeneity Restrictions William D. Lastrapes Department of Economics Terry College of Business University of Georgia Athens,

More information

Stationary Stochastic Time Series Models

Stationary Stochastic Time Series Models Stationary Stochastic Time Series Models When modeling time series it is useful to regard an observed time series, (x 1,x,..., x n ), as the realisation of a stochastic process. In general a stochastic

More information

ECON 4160: Econometrics-Modelling and Systems Estimation Lecture 7: Single equation models

ECON 4160: Econometrics-Modelling and Systems Estimation Lecture 7: Single equation models ECON 4160: Econometrics-Modelling and Systems Estimation Lecture 7: Single equation models Ragnar Nymoen Department of Economics University of Oslo 25 September 2018 The reference to this lecture is: Chapter

More information

Nonstationary Time Series:

Nonstationary Time Series: Nonstationary Time Series: Unit Roots Egon Zakrajšek Division of Monetary Affairs Federal Reserve Board Summer School in Financial Mathematics Faculty of Mathematics & Physics University of Ljubljana September

More information

Lecture 3: Autoregressive Moving Average (ARMA) Models and their Practical Applications

Lecture 3: Autoregressive Moving Average (ARMA) Models and their Practical Applications Lecture 3: Autoregressive Moving Average (ARMA) Models and their Practical Applications Prof. Massimo Guidolin 20192 Financial Econometrics Winter/Spring 2018 Overview Moving average processes Autoregressive

More information

Variable Targeting and Reduction in High-Dimensional Vector Autoregressions

Variable Targeting and Reduction in High-Dimensional Vector Autoregressions Variable Targeting and Reduction in High-Dimensional Vector Autoregressions Tucker McElroy (U.S. Census Bureau) Frontiers in Forecasting February 21-23, 2018 1 / 22 Disclaimer This presentation is released

More information