Estimation and Testing for Common Cycles

Similar documents
Cointegration Lecture I: Introduction

The Role of "Leads" in the Dynamic Title of Cointegrating Regression Models. Author(s) Hayakawa, Kazuhiko; Kurozumi, Eiji

ECON 4160, Spring term Lecture 12

Multivariate Time Series: Part 4

Reduced rank regression in cointegrated models

(Y jz) t (XjZ) 0 t = S yx S yz S 1. S yx:z = T 1. etc. 2. Next solve the eigenvalue problem. js xx:z S xy:z S 1

Econometrics II - EXAM Answer each question in separate sheets in three hours

Y t = ΦD t + Π 1 Y t Π p Y t p + ε t, D t = deterministic terms

ECON 4160, Lecture 11 and 12

Cointegrated VAR s. Eduardo Rossi University of Pavia. November Rossi Cointegrated VAR s Financial Econometrics / 56

Unit roots in vector time series. Scalar autoregression True model: y t 1 y t1 2 y t2 p y tp t Estimated model: y t c y t1 1 y t1 2 y t2

Numerical Distribution Functions of. Likelihood Ratio Tests for Cointegration. James G. MacKinnon. Alfred A. Haug. Leo Michelis

Bootstrapping the Grainger Causality Test With Integrated Data

Title. Description. var intro Introduction to vector autoregressive models

VAR Models and Cointegration 1

Econ 423 Lecture Notes: Additional Topics in Time Series 1

Modelling of Economic Time Series and the Method of Cointegration

A Test of Cointegration Rank Based Title Component Analysis.

Vector Auto-Regressive Models

VAR Models and Applications

Massachusetts Institute of Technology Department of Economics Time Series Lecture 6: Additional Results for VAR s

Residual Autocorrelation Testing for Vector Error Correction Models 1

Online Appendix. j=1. φ T (ω j ) vec (EI T (ω j ) f θ0 (ω j )). vec (EI T (ω) f θ0 (ω)) = O T β+1/2) = o(1), M 1. M T (s) exp ( isω)

THE UNIVERSITY OF CHICAGO Booth School of Business Business 41914, Spring Quarter 2013, Mr. Ruey S. Tsay

Econometrics II - EXAM Outline Solutions All questions have 25pts Answer each question in separate sheets

Testing Restrictions and Comparing Models

Numerically Stable Cointegration Analysis

New Introduction to Multiple Time Series Analysis

Analytical derivates of the APARCH model

A PANIC Attack on Unit Roots and Cointegration. July 31, Preliminary and Incomplete

Cointegrated VAR s. Eduardo Rossi University of Pavia. November Rossi Cointegrated VAR s Fin. Econometrics / 31

Vector error correction model, VECM Cointegrated VAR

Time Series Analysis. James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY

Cointegration. Example 1 Consider the following model: x t + βy t = u t (1) x t + αy t = e t (2) u t = u t 1 + ε 1t (3)

Multivariate Time Series: VAR(p) Processes and Models

Time Series Analysis. James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY

Multivariate Time Series

Vector Autoregressive Model. Vector Autoregressions II. Estimation of Vector Autoregressions II. Estimation of Vector Autoregressions I.

Time-Varying Vector Autoregressive Models with Structural Dynamic Factors

Darmstadt Discussion Papers in Economics

This chapter reviews properties of regression estimators and test statistics based on

Generalized Method of Moment

Why Segregating Cointegration Test?

1 Procedures robust to weak instruments

Test for Structural Change in Vector Error Correction Models 1

Advanced Econometrics

Nonstationary Time-Series Modeling versus Structural Equation Modeling: With an Application to Japanese Money Demand

Multivariate Time Series Analysis and Its Applications [Tsay (2005), chapter 8]

B y t = γ 0 + Γ 1 y t + ε t B(L) y t = γ 0 + ε t ε t iid (0, D) D is diagonal

Introduction to Algorithmic Trading Strategies Lecture 3

Asymptotic Properties of White s Test for Heteroskedasticity and the Jarque-Bera Test for Normality

The Effects of Ignoring Level Shifts on Systems Cointegration Tests

Tests of the Co-integration Rank in VAR Models in the Presence of a Possible Break in Trend at an Unknown Point

4.1 Order Specification

CANONICAL CORRELATION ANALYSIS AND REDUCED RANK REGRESSION IN AUTOREGRESSIVE MODELS. BY T. W. ANDERSON Stanford University

Cointegrated VARIMA models: specification and. simulation

(θ θ ), θ θ = 2 L(θ ) θ θ θ θ θ (θ )= H θθ (θ ) 1 d θ (θ )

On the Error Correction Model for Functional Time Series with Unit Roots

Hypothesis Testing Based on the Maximum of Two Statistics from Weighted and Unweighted Estimating Equations

GARCH Models Estimation and Inference

Structural Vector Autoregressions with Markov Switching. Markku Lanne University of Helsinki. Helmut Lütkepohl European University Institute, Florence

Vector autoregressions, VAR

Volume 03, Issue 6. Comparison of Panel Cointegration Tests

Comparing Forecast Accuracy of Different Models for Prices of Metal Commodities

The outline for Unit 3

Likelihood Ratio Tests of Restrictions on Common Trends Loading Matrices in I(2) VAR Systems

Economics 536 Lecture 7. Introduction to Specification Testing in Dynamic Econometric Models

Purchasing power parity: A nonlinear multivariate perspective. Abstract

Central Limit Theorem ( 5.3)

Stochastic Trends & Economic Fluctuations

Matrix Differential Calculus with Applications in Statistics and Econometrics

A Non-Parametric Approach of Heteroskedasticity Robust Estimation of Vector-Autoregressive (VAR) Models

Research Statement. Zhongwen Liang

Cointegration between Trends and Their Estimators in State Space Models and Cointegrated Vector Autoregressive Models

EC821: Time Series Econometrics, Spring 2003 Notes Section 9 Panel Unit Root Tests Avariety of procedures for the analysis of unit roots in a panel

Are US Output Expectations Unbiased? A Cointegrated VAR Analysis in Real Time

Generalized Reduced Rank Tests using the Singular Value Decomposition

Department of Economics, UCSD UC San Diego

Statistics and econometrics

Financial Econometrics

UNIVERSITY OF CALIFORNIA, SAN DIEGO DEPARTMENT OF ECONOMICS

Testing Error Correction in Panel data

Codependence and Cointegration

Economics 583: Econometric Theory I A Primer on Asymptotics

TESTING FOR CO-INTEGRATION

GARCH Models Estimation and Inference. Eduardo Rossi University of Pavia

ORTHOGONALITY TESTS IN LINEAR MODELS SEUNG CHAN AHN ARIZONA STATE UNIVERSITY ABSTRACT

Generalized Method Of Moments (GMM)

Recent Advances in Cointegration Analysis

Cointegration Analysis in the Case of I(2) General Overview

QED. Queen s Economics Department Working Paper No Improved Likelihood Ratio Tests for Cointegration Rank in the VAR Model

Review of Classical Least Squares. James L. Powell Department of Economics University of California, Berkeley

Cointegration. Overview and Development

Economic modelling and forecasting

by Søren Johansen Department of Economics, University of Copenhagen and CREATES, Aarhus University

Fractional integration and cointegration: The representation theory suggested by S. Johansen

Ph.D. Qualifying Exam Friday Saturday, January 3 4, 2014

Simulating Properties of the Likelihood Ratio Test for a Unit Root in an Explosive Second Order Autoregression

References. Press, New York

Estimation of Threshold Cointegration

Transcription:

Estimation and esting for Common Cycles Anders Warne February 27, 2008 Abstract: his note discusses estimation and testing for the presence of common cycles in cointegrated vector autoregressions A simple two-stage estimator is considered where the cointegration vectors are estimated in the first stage and the remaining parameters, subject to common cycles, in the second stage he latter stage is equivalent to using reduced rank regression in a stationary framework Simple procedures for iterating on these two stages are discussed with emphasis on estimating the cointegration space conditional on the common cycles restriction It is shown that the two-stage estimator of the parameters describing the dynamics is asymptotically Gaussian and efficient Furthermore, an estimator of the co-feature matrix is examined and its asymptotic properties are derived Finally, two asymptotically equivalent methods for computing the likelihood ratio test for the null of s versus s g common cycles are presented along with the limiting behavior of the tests Keywords: Asymptotics, Cointegration, Common Cycles, Likelihood Ratio est JEL Classification Numbers: C32 1 Setup he model we shall consider is a standard cointegrated VAR model written on error correction form as: k 1 x t Γ i x t i αβ x t 1 ε t, t 1,,, (1) i 1 where x t is p dimensional, the parameters α and β are p r with full column rank r {0, 1,,p 1}, x 0,,x 1 k are fixed, while ε t is iid Gaussian with zero mean and positive definite covariance matrix Ω For simplicity, but without loss of generality, I have excluded all deterministic variables from the model since the common cycles hypothesis does not impose any restrictions on the parameters of such variables Following Vahid and Engle (1993) we say that x t has s common cycles if (and only if) ] Θ Γ 1 Γ k 1 α ξη, (2) where ξ is a full rank p s matrix and η is a full rank p k 1 r s matrix for s {r,,p 1} Given that x t has s common cycles it follows that there exists a p p s matrix ξ such that ξ x t is serially uncorrelated (since ξ ξ 0 by construction) he matrix ξ is called the co-feature matrix 1 In the event that k 1, s is always equal to r, the cointegration rank, and ξ α his means that if we wish to test for the presence of s common cycles under k 1 it is equivalent to testing for the rank of the p p matrix Π αβ Once we have established a proper rank for this matrix, there is no need to pursue any further estimation or testing regarding common cycles When the lag order is greater than 1, however, then s canbeanyintegerbetweenr and p 1 It is clear that s cannot be smaller than r since that would violate the rank assumption for α Furthermore, the case s p is uninteresting since this implies that Θ is a full rank matrix, Remarks: Copyright c 2004 2008 Anders Warne 1 Hecq, Palm and Urbain (2000) suggest two forms of reduced rank for Θ he strong form is equivalent to equation (2) being satisfied, while their weak form only requires that the Γ i matrices can be written as functions of ξ, ie, that ξ x t αβ x t 1 ξ ε t he weak form is not discussed in this note since, eg, it is not invariant to the specification of the error correction form, ie, if x appears in levels on the right hand side of (1) for t 1orfort k

ie, common cycles would not impose any restrictions on Θ and the co-feature matrix would be empty Rather, the case s p maybeusedasonealternativehypothesiswhenweareinterested in testing the null hypothesis that Θ has reduced rank s<p Assuming that x t is subject to s common cycles, it follows that Θ has less than p p k 1 r unique parameters Since ξη ξmm 1 η ξ η for any full rank s s matrix M we know that, eg, only the space spanned by the columns of η can identified o uniquely determine η (and ξ) let us therefore select a basis for the space spanned by its columns, expressed as: η h h ψ, (3) where h is a known p k 1 r s matrix such that, for simplicity, h h I s, h is p k 1 r p k 1 r s of full rank such that h h 0andh h I p k 1 r s,whileψ is p k 1 r s s contains the free parameters of η From this discussion it directly follows that if x t has s common cycles, this imposes exactly q p s p k 1 r s restrictions on the Γ i and α parameters of the cointegrated VAR model relative to the case when Θ has full rank p Generally, suppose we wish to test the hypothesis that Θ has rank s versus the alternative that it has rank s g, the number of restrictions on Γ i and α is q g pk r 2s g Against all g {1,p s}, the number of restrictions is positive when k 2 Again, for k 1 we cannot impose any additional restrictions on α based on the common cycles hypothesis since s r by assumption 2 Estimation: A wo-stage Algorithm As discussed above, the assumption of s common cycles imposes a second reduced rank condition on the cointegrated VAR model whenever the lag order of the VAR, k, is greater than or equal to 2 Since k 1 is trivial, we shall assume that k 2 First of all, suppose that s r In this case the two reduced rank conditions can be merged into one reduced rank condition since ξ αm forsomefullrankr r matrix M his means that ] x t α η 2 β x t 1 x t k 1 x t 1 ε t αβ Z 1t ε t, where Γ 1 Γ k 1 α η 2 his means that Θ α η 2 I r so that h 0 I r Hence, ξ α and ψ η 2, while the co-feature matrix ξ spans the space spanned by α,asinthecasewhen k 1 he parameters α and β can be estimated by maximum likelihood through reduced rank regression In case β should satisfy some restrictions, eg, exact identification, procedures analysed by Johansen (1996) can be used directly Second, suppose s r so that the two reduced rank conditions cannot necessarily be merged as above One approach to estimation of the cointegrated VAR with s common cycles, is to use a procedure similar to the one considered for the I(2) model by Johansen (1995a); see also Paruolo (2000) 2 he first stage is to determine the cointegration rank, r, and obtain an estimate of β using, eg, reduced rank regression, as in Johansen (1996) his stage may also involve estimation of β subject to linear restrictions 2 Maximum likelihood estimation may be considered using an algorithm similar to the one studied by Johansen (1997) when s>r Alternatively, one may consider the iterative Gaussian reduced rank estimator suggested by Ahn (1997), where β, ξ, ψ (and Ω) are computed through an approximate Newton-Raphson algorithm; see also Ahn and Reinsel (1988) Furthermore, if β isknownthentheestimatorof ξ, ψ, Ω discussed in this section is the maximum likelihood estimator 2

In stage two, we condition on the estimated β from stage one and estimate ξ and η using reduced rank regression; see Hecq, Palm and Urbain (2001) for an alternative two-stage approach hat is, let z t x t 1 x t k 1 β x t 1 so that x t ξη z t ε t (4) From the reduced rank regression procedure discussed in, eg, Johansen (1996, Chap 6) we obtain estimates of ξ and η such that 1/ t 1 η z t z t η I s,whileξ η ξη Since h h I s it follows that the estimate of η, which satisfies (3), can also be computed from η η h η 1 while ψ h η3 Accordingly, the estimate of ξ must satisfy ξ ξ η hgiven ˆΘ ˆξˆη we compute ˆΩ 1/ i 1 ˆε tˆε t,whereˆε t x t ˆΘz t 4 Since the estimated Θ is asymptotically independent of the estimated β, and the latter converges weakly at a faster rate than the former, we may stop here In practise, however, it may be advantagenous to consider iteration over the two stages, but where the first stage is slighly modified In case β is generically identified in the sense of Johansen (1995b) we can write it as a function of its unique parameters Let β i be the i:th column of β with β i m i M i ϕ i, i {1,,r}, (5) where m i and M i are known and ϕ i contains the free parameters for β i Letting vec be the column stacking operator, we can express the identified β as m 1 M 1 0 0 ϕ 1 m 2 0 M 2 0 ϕ 2 vec β (6) m r 0 0 M r ϕ r m Mϕ he gradient vector for the log-likelihood function in the direction of ϕ can be expressed as: ln L ϕ ( vec α ) I r Ω 1( M 01 αβ ) ] M 11 ΓM 21 M, (7) where is the Kronecker product, Γ Γ 1 Γ k 1, M ij 1/ t 1 Z it Z jt, Z 0t x t, Z 1t x t 1,andZ 2t x t 1 x t k 1 Plugging the estimated parameters and product moment matrices into (7) we first check if the gradient is sufficiently close to zero If it is, we are done If not, we can solve for ϕ conditional on Θ and Ω by setting the gradient vector to zero his provides us with ϕ M ( α Ω 1 ) ] 1M α M 11 M vec ( M 10 M 12 Γ Ω 1 α ) α Ω 1 ] ] α M 11 m (8) Conditional on the new β we can then re-estimate ξ, η, andω as in stage two Given the outcome of this stage, we can compute the gradient vector in (7) and check if it is sufficiently close to zero If it is the iterations are finished, if not we re-estimate β using (8), and so on until we have convergence 5 3 o show this, notice that for η η h η 1 the following p k 1 r equations hold: h η I s h h η h η 1 Since the inverse of the matrix which is multiplies by η is given by h h, premultiplication of both sides by this inverse gives us η h h h η h η 1 h h ψifhwere such that h h I s, the above relations would have to be changed to account for this 4 he vector z t is actually an estimate of the unknown vector z t whenever an estimate of β is used 5 It may be noted that the estimator of β could also be based on, eg, cross-equation restrictions; see, eg, Pesaran and Shin (2002) What is important here is that β is identified and that vec β m Mϕ 3

Suppose instead that β has not been generically identified as in (6) but has been obtained through reduced rank regression as in Johansen (1996, Chap 6) Let us decompose η such that Γ ξη 2 and α ξη 1 his means that (1) can be expressed as: Z 0t ξη 2 Z 2t ξη 1 β Z 1t ε t (9) o determine if re-estimation of β is useful in the first place we can use the following expression (which we may think of as a gradient matrix ), with β evaluated at its previous value: α Ω 1( M 01 ΓM 21 αβ ) M 11 η1 ξ Ω 1( M 01 ξη 2 M 21 ξη ) 1 β M 11 (10) If this matrix is sufficiently close to zero we do not need to re-estimate β If it is not close to zero, we may proceed by setting (10) to zero and solve for β conditional on ξ, η, andω his gives us: β M 1 ( 11 M10 M 12 η 2 ξ ) Ω 1 ξη 1( η1 ξ Ω 1 ξη 1) 1 (11) his represents the modified stage one for the iterations when β is not generically identified As above, we can re-estimate ξ, η, andω as in stage two, but conditional on the new β Giventhe outcome of this stage we can compute the right hand side of (10) and check if it is sufficiently close to zero If it is, we are done, if not we continue until the algorithm has converged 3 Asymptotic Properties of the Estimators Suppose first that β is known his means that the estimator of ξ, η, Ω is the maximum likelihood estimator and is equivalent to reduced rank regression using, eg, the algorithm presented by Johansen (1996, Chap 6); see, also, Anderson (1951) Accordingly, the estimated parameters solve the first order conditions for the log-likelihood function Let θ vec Θ and γ vec ξ vec ψ his means that the partial derivatives of θ with respect to the free parameters γ is: G γ θ η γ ] Ip h ξ ] ] K p k 1 r s,s, (12) where K m,n is the mn mn commutation matrix; see Magnus and Neudecker (1988) hat is, vec ψ K p k 1 r s,s vec ψ If there are exactly s common cycles, then G γ has full column rank he first order condition for γ is equal to the partial derivatives of the loglikelihood function with respect to θ times the matrix G γ LetM zz 1/ t 1 z t z t while M xz 1/ t 1 x tz t 6 We then have that the gradient vector is given by ln L γ ( vec M xz ΘM zz ) Ip k 1 r Ω 1] G γ (13) Furthermore, since our estimator of γ is a maximum likelihood estimator, z t is stationary, and the log-likelihood function satisfies standard regularity conditions (being the Gaussian loglikelihood), we know that ˆγ is a consistent estimator of γ and, furthermore, that ˆγ γ converges in distribution to a Gaussian with zero mean and asymptotic covariance matrix given by the inverse of the information matrix for γ Furthermore, ˆΩ is a consistent estimator of Ω, it is asymptotically independent of ˆγ (which is the reason why the asymptotic covariance matrix for ˆγ is given by the inverse of the information matrix for γ), and converges in distribution at the rate to a Gaussian Specifically, by a aylor expansion of the gradient in (13) and some standard rearrangement we have that: ) 1G γ (ˆγ γ G γ M zz Ω 1] G γ ] ( I Ω 1] 1 ) vec ε t z t o p 1, (14) t 1 6 Notive that M xz M 02 M 01 β, while M zz M 22 β M 12 M 21 β β M 11 β 4

where Since M zz 2 ln L γ γ G γ M zz Ω 1] G γ, (15) p µ, a positive definite matrix, we find from, eg, Lütkepohl (1991, Chap 3) that ( 1 ) vec ε t z d t N ( 0,µ Ω ), t 1 where p denotes convergence in probability and d convergence in distribution It then follows that ) d (ˆγ γ N 0, Σγ, (16) where 1, Σ γ G γ µ Ω 1] G γ ] is the inverse of the information matrix Hence, the estimator of γ is asymptotically efficient since its asymptotic covariance matrix is equal to the Cramér-Rao lower bound Furthermore, the asymptotic covariance matrix is equal to the one presented in Ahn (1997, heorem 1) for the approximate Newton-Raphson algorithm for estimating all parameters jointly; see, also, Section 4 Now, as noted above, θ is a nonlinear and continuously differentiable function of γ 7 Based on well known results for such functions (see, eg, Serfling, 1980, Chap 3) we therefore have that ) ) (ˆθ θ G γ (ˆγ γ op 1 d N ( 0,G γ Σ γ G γ ) he asymptotic covariance matrix for θ therefore has reduced rank, and its rank is equal to the rank of Σ γ, ie, s pk r s Furthermore, and following the results in, eg, Magnus and Neudecker (1988), the asymptotic behavior of the maximum likelihood estimator of Ω can directly be derived Let ω vech Ω, where vech is the column stacking operator which takes only the values on and below the diagonal It can easily be shown (see, eg, Magnus and Neudecker, 1988, or Lütkepohl, 1991) that ˆω and ˆγ are asymptotically independent by checking the matrix with second order cross partials Furthermore, let D p be the p 2 p 1 p/2 duplication matrix which equates vec Ω D p ω,whiledp D pd p 1 D p is its Moore-Penrose inverse We then find that ( ) d ( ˆω ω N 0, 2D p Ω Ω Dp ) (18) Aagain, the asymptotic covariance matrix for ˆω is equal to the inverse of the information matrix for ω and the estimator is therefore asymptotically efficient It may also be of interest to estimate the co-feature matrix ξ and determine its asymptotic properties Given a value of ξ we know the space spanned by the columns of ξ However, we need to select a basis for this space to uniquely determine the elements of ξ and to derive its asymptotic behavior o this end, let ( ξ H H ξ ) 1ξ H H, (19) where H is a known p p s matrix with full rank, and H is p s such that ξ H has rank s and H H 0 It is easily seen that ξ ξ 0 for this choice of basis for ξ Wethenfindthat the partial derivatives of the elements of ξ with respect to the elements of ξ is a p p s ps (17) 7 he relationship between these parameters can actually be written as: θ h 1 ] ] 1 2 h ψ I p 2 h ξ ]K p k 1 r s,s γ QG γ γ, where Q is a diagonal matrix with some elements equal to 1 and others equal to 1/2 5

matrix given by vec ξ vec ξ ξ H ( ξ H ) 1 ] K p,s, (20) since H H ξ H ξ 1 H ξ Sinceξ does not depend on ψ it follows that: F γ vec ξ ] vec ξ γ vec ξ 0 (21) Let the estimator of the co-feature matrix ξ be denoted as ˆξ and suppose that its obtained by inserting the estimator of ξ into equation (19) We then find that ) ) (ˆξ ξ F γ (ˆγ γ op 1 d N ( 0,F γ Σ γ F γ ) (22) Since F γ has rank p s s it follows that the asymptotic covariance matrix for the estimator of the unique elements of ξ has full rank (where the latter can be represented by the s p s matrix H ξ ) Moreover, since ˆγ is asymptotically efficient ˆξ is efficient in this sense as well he above results are derived for the assumption that β is known However, all results also hold when β is estimated and the estimator satisfies ) p (ˆβ β 0 Since, eg, the Gaussian maximum likelihood estimator of β satisfies this super-consistency requirement, the above asymptotic results are valid for the two-stage approach to estimating a cointegrated VAR with common cycles restrictions o see this, notice first that the ẑ t regressors, based on the estimated β, is related to the z t regressors, using the true value of β, through x t 1 0 ẑ t z x t t k 1 0 ˆβ x t 1 ˆβ β x t 1 It then follows that M zz based on ẑ t converges in probability to µ, justlikem zz based on knowing β Furthermore, and importantly for convergence in distribution, let ˆε t x t Θẑ t hen, 1 t 1 ˆε t ẑ t 1 ε t z t o p 1 (23) Hence, the limiting behavior of 1/ t 1 ˆε t ẑ t is equivalent to the limiting behavior of 1/ t 1 ε tz t At the same time, using an estimator of β instead of a known matrix introduces a bias, which although asymptotically negligable, may still be relevant in small samples t 1 4 Estimation: A One-Stage Algorithm Ahn (1997) has suggested that we can estimate ϕ and γ simultaneously through an iterative approximate Newton-Raphson algorithm Hence, the algorithm is based on β being (generically) identified 8 Let φ ϕ γ, the Ahn estimator is obtained from iteration on the following 8 he algorithm in Ahn (1997) is actually based on β I r β 0,whereβ 0 is an p r r matrix with the free parameters of the exactly identified β 6

system: { ˆφ i 1 ˆφ i 2 ] 1 } ln L ln L φ φ φ φ ˆφ i ˆφ i M ] ] ] 1 α Ω 1 α M 11 M M K r,p α Ω 1 M 1z G γ G γ ] Ω 1 α M z1 Kp,r M G γ M zz Ω 1] G γ M I r M 10 M 11 βα M 12 Γ Ω 1] ]} vec α G γ I p k 1 r Ω 1] vec M xz ΘM zz φ ˆφ i, where M z1 1/ t 1 z t Z 1t M 12 M 11 β and M 1z M z1 he iterations have converged when the gradient vector is sufficiently close to zero and as starting values we may, eg, choose the estimate of ϕ for unrestricted Θ, and the estimator of γ from the two-stage algorithm which is conditioned on the this initial estimate of ϕ Asymptotically, the two-stage and the approximate Newton-Raphson estimators of ϕ and γ are equivalent Furthermore, the two-stage estimator of ϕ and the maximum likelihood estimator of ϕ for unrestricted Θ are also asymptotically equivalent Hence, the limiting behavior of all these estimators of ϕ does not depend on the reduced rank structure of Θ Ahn (1997) notes, however, that for finite samples the efficiency in the sense of smaller mean squared error can be gained by exploring the structure among the parameters on stationary processes (24) 5 esting A natural approach to testing for s common cycles versus the alternative of s g (where g {1,p s}) within the current framework is to use a likelihood ratio test Estimation under the null can be achieved as discussed above, while estimation under the alternative is as discussed above for all g<p sand using the unrestricted estimate of Θ when g p s Let ˆΩ sj be the estimator of Ω subject to s j common cycles, with s j s under the null and s j s g under the alternative Asymptotically, it does not matter whether the two-stage or the one-stage algorithm has been used here he likelihood ratio statistic can be written as: ( ) det ˆΩ s LR ln (25) det ˆΩ s g Given the asymptotic normal behavior of ˆθ under the null hypothesis, the fact that all the common cycles restrictions are expressed as restrictions on parameters on stationary processes only, ie, the restrictions do not depend on β (see, eg, Sims, Stock and Watson, 1990), and thefactthattheestimatorofβ converges to its true value in probability at the rate (ie, it is super-consistent), it follows that LR is asymptotically χ 2 with q g pk r 2s g degrees of freedom In the event that g p s, the number of degrees of freedom q p s p k 1 r s An asymptotically equivalent form of the likelihood ratio test may also be derived he estimator of η, ie, the estimator obtained from the reduced rank regression before employing the normalization in (3), is obtained (just like in Johansen, 1996) by solving the eigenvalue problem det ( λm zz M zx Mxx 1 M xz) 0, for eigenvalues 1 > ˆλ 1 > > ˆλ p > ˆλ p 1 ˆλ p k 1 r 0 and eigenvectors ˆV ˆV 1 ˆV p k 1 r, normalized such that ˆV M zz ˆV I Here, M xx 1/ t 1 x t x t M 00,whileM zx M xz he estimator of η is given by ˆη ˆV 1 ˆV s under the null, and by ˆη ˆV 1 ˆV s g under the alternative Since the maximized likelihood function is given by L 2/ max s j det ( ) s j ( ) M xx 1 ˆλi, 7 i 1

for any choice of s j {r,,p}, the likelihood ratio test of s common cycles versus s g common cycles can then be formulated as: LR alt s g i s 1 ln ( 1 ˆλ i ) (26) he two test statistics in (25) and (26) are numerically equivalent if (and only if) the estimator of β is not updated under the two-stage approach In that case, ˆΩ s and ˆΩ s g in (25) are computed using the same β matrix he statistic in (26) always uses the same β matrix under the null and the alternative hypotheses he test statistic in (26) is similar to the statistic proposed by Vahid and Engle (1993) when g p s Infact,theonlydifference is that (26) uses the number of observations, while the statistic in Vahid and Engle employs the number of observations corrected for the number of lags References Ahn, S K (1997), Inference of Vector Autoregressive Models with Cointegration and Scalar Components, Journal of the American Statistical Association, 92, 350 356 Ahn, S K and Reinsel, G C (1988), Nested Reduced-Rank Autoregressive Models for Multiple ime Series, Journal of the American Statistical Association, 83, 849 856 Anderson, W (1951), Estimating Linear Restrictions on Regression Coefficients for Multivariate Normal Distributions, Annals of Mathematical Statistics, 22, 327 351 Hecq, A, Palm, F C and Urbain, J-P (2000), Permanent-ransitory Decomposition in VAR Models with Cointegration and Common Cycles, Oxford Bulletin of Economics and Statistics, 62, 511-532 Hecq, A, Palm, F C and Urbain, J-P (2001), esting for Common Cyclical Features in VAR Models with Cointegration, CESifo Working Paper No 451 Johansen, S (1995a) A Statistical Analysis of Cointegration for I(2) Variables, Econometric heory, 11, 25 59 Johansen, S (1995b), Identifying Restrictions of Linear Equations with Applications to Simultaneous Equations and Cointegration, Journal of Econometrics, 69, 111 132 Johansen, S (1996), Likelihood-Based Inference in Cointegrated Vector Autoregressive Models,2nd ed, Oxford: Oxford University Press Johansen, S (1997), Likelihood Analysis of the I(2) Model, Scandinavian Journal of Statistics, 24, 433 462 Lütkepohl, H (1991), Introduction to Multiple ime Series Analysis, Heidelberg: Springer-Verlag Magnus, J R and Neudecker, H (1988), Matrix Differential Calculus: With Applications in Statistics and Econometrics, New York: John Wiley & Sons Paruolo, P (2000), Asymptotic Efficiency of the wo Stage Estimator in I(2) Systems, Econometric heory, 16, 524 550 Pesaran, M H and Shin, Y (2002), Long-Run Structural Modelling, Econometric Reviews, 21, 49 87 Serfling, R J (1980), Approximation heorems of Mathematical Statistics, JohnWiley: New York Sims, C A, Stock, J H and Watson, M W (1990), Inference in Linear ime Series Models with Some Unit Roots, Econometrica, 58, 113 144 Vahid, F and Engle, R F (1993), Common rends and Common Cycles, Journal of Applied Econometrics, 8, 341 360 8