On the Correlations of Trend-Cycle Errors

Similar documents
Equivalence of several methods for decomposing time series into permananent and transitory components

Time Series Analysis. James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY

The Clark Model with Correlated Components

Trend-Cycle Decompositions

Business School Working Paper

The MIT Press Journals

Econ 424 Time Series Concepts

Time Series Analysis. James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY

Non-Stationary Time Series and Unit Root Testing

Univariate Nonstationary Time Series 1

BCT Lecture 3. Lukas Vacha.

Efficiency Tradeoffs in Estimating the Linear Trend Plus Noise Model. Abstract

Non-Stationary Time Series and Unit Root Testing

Non-Stationary Time Series and Unit Root Testing

CENTRE FOR APPLIED MACROECONOMIC ANALYSIS

Christopher Dougherty London School of Economics and Political Science

A TIME SERIES PARADOX: UNIT ROOT TESTS PERFORM POORLY WHEN DATA ARE COINTEGRATED

Trend and Cycles: A New Approach and Explanations of Some Old Puzzles

Intuitive and Reliable Estimates of the Output Gap from a Beveridge-Nelson Filter

SOME BASICS OF TIME-SERIES ANALYSIS

Class: Trend-Cycle Decomposition

Forecasting Levels of log Variables in Vector Autoregressions

A note on implementing the Durbin and Koopman simulation smoother

Working Paper Series. A note on implementing the Durbin and Koopman simulation smoother. No 1867 / November Marek Jarocinski

GMM-based inference in the AR(1) panel data model for parameter values where local identi cation fails

Lecture 5: Unit Roots, Cointegration and Error Correction Models The Spurious Regression Problem

A multivariate innovations state space Beveridge Nelson decomposition

Identifiability, Invertibility

B y t = γ 0 + Γ 1 y t + ε t B(L) y t = γ 0 + ε t ε t iid (0, D) D is diagonal

11/18/2008. So run regression in first differences to examine association. 18 November November November 2008

Cross validation of prediction models for seasonal time series by parametric bootstrapping

Inference in VARs with Conditional Heteroskedasticity of Unknown Form

A time series is called strictly stationary if the joint distribution of every collection (Y t

E 4101/5101 Lecture 9: Non-stationarity

Time Series Analysis -- An Introduction -- AMS 586

Trends and Unit Roots in Greek Real Money Supply, Real GDP and Nominal Interest Rate

IS THE NORTH ATLANTIC OSCILLATION A RANDOM WALK? A COMMENT WITH FURTHER RESULTS

UNSW Business School Working Paper

1 Teaching notes on structural VARs.

A Test of Cointegration Rank Based Title Component Analysis.

Research Article Least Squares Estimators for Unit Root Processes with Locally Stationary Disturbance

13. Time Series Analysis: Asymptotics Weakly Dependent and Random Walk Process. Strict Exogeneity

CHAPTER 21: TIME SERIES ECONOMETRICS: SOME BASIC CONCEPTS

Some Time-Series Models

E 4160 Autumn term Lecture 9: Deterministic trends vs integrated series; Spurious regression; Dickey-Fuller distribution and test

What are the Differences in Trend Cycle Decompositions by Beveridge and Nelson and by Unobserved Component Models?

Chapter 5: Models for Nonstationary Time Series

Cointegrated VARIMA models: specification and. simulation

Using all observations when forecasting under structural breaks

1 Teaching notes on structural VARs.

Lecture: Testing Stationarity: Structural Change Problem

Lecture 19 - Decomposing a Time Series into its Trend and Cyclical Components

Testing for non-stationarity

Covariance Stationary Time Series. Example: Independent White Noise (IWN(0,σ 2 )) Y t = ε t, ε t iid N(0,σ 2 )

TIMES SERIES INTRODUCTION INTRODUCTION. Page 1. A time series is a set of observations made sequentially through time

Estimating and Accounting for the Output Gap with Large Bayesian Vector Autoregressions

Ch 5. Models for Nonstationary Time Series. Time Series Analysis

Do Markov-Switching Models Capture Nonlinearities in the Data? Tests using Nonparametric Methods

Univariate linear models

Time Series Models and Inference. James L. Powell Department of Economics University of California, Berkeley

Cointegrated VAR s. Eduardo Rossi University of Pavia. November Rossi Cointegrated VAR s Financial Econometrics / 56

Gaussian processes. Basic Properties VAG002-

An algorithm for robust fitting of autoregressive models Dimitris N. Politis

Lesson 9: Autoregressive-Moving Average (ARMA) models

Empirical Market Microstructure Analysis (EMMA)

EC821: Time Series Econometrics, Spring 2003 Notes Section 9 Panel Unit Root Tests Avariety of procedures for the analysis of unit roots in a panel

Identifying SVARs with Sign Restrictions and Heteroskedasticity

Introduction to Modern Time Series Analysis

Estimating and Identifying Vector Autoregressions Under Diagonality and Block Exogeneity Restrictions

Inflation Revisited: New Evidence from Modified Unit Root Tests

G. S. Maddala Kajal Lahiri. WILEY A John Wiley and Sons, Ltd., Publication

Economics Division University of Southampton Southampton SO17 1BJ, UK. Title Overlapping Sub-sampling and invariance to initial conditions

Unit Roots and Structural Breaks in Panels: Does the Model Specification Matter?

TIME SERIES ANALYSIS. Forecasting and Control. Wiley. Fifth Edition GWILYM M. JENKINS GEORGE E. P. BOX GREGORY C. REINSEL GRETA M.

1 Linear Difference Equations

11. Further Issues in Using OLS with TS Data

Consider the trend-cycle decomposition of a time series y t

Problem Set 1 Solution Sketches Time Series Analysis Spring 2010

Ch. 15 Forecasting. 1.1 Forecasts Based on Conditional Expectations

Financial Time Series Analysis: Part II

Lecture 2: Univariate Time Series

Title. Description. var intro Introduction to vector autoregressive models

Macroeconomics Theory II

Covers Chapter 10-12, some of 16, some of 18 in Wooldridge. Regression Analysis with Time Series Data

Time Series Methods. Sanjaya Desilva

Multivariate Time Series

Cointegration and the joint con rmation hypothesis

Dynamic Regression Models (Lect 15)

Lecture 1: Fundamental concepts in Time Series Analysis (part 2)

Class 4: Non-stationary Time Series

at least 50 and preferably 100 observations should be available to build a proper model

This note discusses some central issues in the analysis of non-stationary time

Applied Time. Series Analysis. Wayne A. Woodward. Henry L. Gray. Alan C. Elliott. Dallas, Texas, USA

The Price Puzzle: Mixing the Temporary and Permanent Monetary Policy Shocks.

GLS detrending and unit root testing

7 Introduction to Time Series

GMM and SMM. 1. Hansen, L Large Sample Properties of Generalized Method of Moments Estimators, Econometrica, 50, p

10) Time series econometrics

TIME SERIES DATA ANALYSIS USING EVIEWS

On Consistency of Tests for Stationarity in Autoregressive and Moving Average Models of Different Orders

Transcription:

On the Correlations of Trend-Cycle Errors Tatsuma Wada Wayne State University This version: December 19, 11 Abstract This note provides explanations for an unexpected result, namely, the estimated parameter of the correlation coefficient of the trend shock and cycle shock in the state-space model is almost always (positive or negative) unity, even when the true variance of the trend shock is zero. It is shown that the set of the true parameter values lies on the restriction that requires the variancecovariance matrix of the errors to be nonsingular, therefore, almost always the likelihood function has its (constrained) global maximum on the boundary where the correlation coefficient implies perfect correlation. JEL Classification Number: C13, C. Keywords: Trend-Cycle Decomposition, Unit-root, Maximum likelihood. 1 Introduction When the trend-cycle decomposition of economic time series data is implemented through a statespace (or unobserved components, UC) model, the correlation of a shock (error) to the trend and a shock (error) to the cycle is often assumed to be zero. This is due to the fact that the correlation coefficient is generally unidentified, as Watson (1986) demonstrates. A recent influential paper by Morley et al. (3), however, shows that the correlation of errors can be identified only if identification conditions are satisfied; and that whether or not allowing such a correlation is a key to understanding the substantial differences between business cycles estimated by the Beveridge- Nelson decomposition and by the UC model. Among others, a study by Perron and Wada (9) finds that US GDP follows a stationary process, when the trend function is allowed to have sudden changes. Although there is no shocks to trend, 1 the maximum likelihood estimator implies perfect correlation. In addition, their simulation shows that the estimated correlation parameter is almost always (positive or negative) unity. This unexpected result is worth exploring. Since, in practice, researchers may estimate the correlation of the errors by simply assuming the data are non-stationary without applying a variety of unit root I am grateful to an anonymous referee for useful comments. All remaining errors are mine. Department of Economics, Wayne State University, 656 W. Kirby St., Detroit, MI, 48 (tatsuma.wada@wayne.edu). 1 Except the shock that causes the trend to change suddenly. 1

tests; hence, knowing properties of the estimator when the true parameter is not identified, serves the purpose of diagnosing model misspecification. In line with the idea of detecting misspecification via correlated UC models, Morley et al. (11) propose a likelihood ratio test of stationarity utilizing the fact that the correlation is observed only when trend errors exist. This note demonstrates that such a perfect correlation, when the true data generating process (DGP) is stationary, is artificially created due to the restriction requiring that the variancecovariance matrix of the errors be positive-semi-definite. Since the true parameters lie on the boundary of this restriction, where the correlation coefficient is positive or negative unity, or one of the variances is zero, the estimated correlation coefficient is almost always 1 or -1, as Table 1 displays, rather than the undefined correlation with zero variance. The rest of this note is organized as follows. Section explains our model. The likelihood function and its properties for a simple model under the restriction are analyzed in Section 3. An extension to the AR(p) error in the cyclical component is presented in Section 4. Section 5 concludes. Model Our model is given by: = + (1) = 1 + () = where is observable variable; and are the trend and cyclical components, respectively; () is the lag-polynomial; the shock to the trend andtheshocktothecycle are drawn from the bivariate Normal distribution: The correlation coefficient is defined as = ( ). 3 A Simple White Noise Case 3.1 The Likelihood Function Let the data DGP be a white noise process: = where is a zero-mean, normally distributed iid process with a variance of. state-space model: Consider the = + = 1 + As Table1 reveals, it is still possible to obtain an estimate that is neither 1 nor -1. The relative frequency of finding such estimates is, however, only about 1%. Note also that the results displayed in Table 1 are different from Perron and Wada s (9) simulation results, since their model includes a change in the slope of the deterministic trend, in addition to the stationary component.

Assuming that and are Normally distributed iid processes, the vector form of the model is (Tanaka 1996) = + + where =( 1 ) ; =( 1 ) ; = ; with the random walk generating matrix 1 1 1 =. 1 1........ 1 1 1 1 1 The log-likelihood function (without constant) is = 1 log + 1 + 1 In order to allow the correlation in errors and, let us assume s = + () where is a constant that represents the covariance of and ;shocks and are independent: 1 ; = By (), and are now correlated, 3 thereby allowing us to find the likelihood function (see Online Appendix 4 for details). Since the variance of, is non-zero, it is convenient to find the concentrated likelihood function (with respect to ): () = 1 log log 1 3 Alternatively, we can write the error processes as = 1 so that shocks and are contemporaneously correlated: ; = 1 4 http://www.clas.wayne.edu/multimedia/usercontent/file/economics/wada/tech_appendix_rho.pdf 3

where = + + ( + ). Using the concentrated likelihood function, we obtain the following three lemmas that clarify the properties of the likelihood function. Lemma 1 The first order conditions are given by: and () = 1 1 + ª + ( ) = 1 1 ª + 1 ( + ) 1 1 1 1 1 = = Lemma At = =, and for, we have ( ) = () 1 µ ( +1) Z 1 [ ()] 1 where represents weak convergence in distribution; and () is the standard Wiener process on C[ 1]. Lemma 3 At = =, and for, we have 3. The Restriction ( ) 1 ( ) = Z 1 µ ( + ) () () = 1 h i (1) 1 We impose the restriction that the variance-covariance matrix of the errors is positive-semi-definite. In our case, this restriction is. Figure 1 displays such a restriction on the parameter space: The horizontal axis represents, the covariance parameter, while the vertical axis represents, the ratio of the variance of the trend errors to the variance of the cycle error. As one can see, the boundary is a parabola: along the quadratic curve, the correlation coefficient is -1 (when corresponding is negative) or 1 (when corresponding is positive), except for the point where = =, at which the correlation is undefined and at which the set of the true parameter values is located. It is obvious from Figure 1 that the true parameter values are, in fact, close by the restriction that implies a perfect correlation. As is well known, the gradient at the true parameter values is not necessarily zero, since the gradient itself is composed of random variables. Because R h 1 [ ()] =1 and (1) i = 1, 5 the gradient becomes zero only as the expected value, asymptotically. Still, the fact that the probability of (1) being less than unity is.683, implies 1 ( ) is often negative at the 5 1 [ ()] = 1 [ ()] = 1 = 1 ; and [ (1)] = (1) =1. 4

true parameter values. Hence, the likelihood function has its global maximum at somewhere other than the true parameter values; many times it can be found on the boundary of the parameter space with negative. As an example, see Figure : The likelihood function does not have its global maximum at = =. It is important to keep in mind that increasing the sample size does not prevent one from finding the artificial perfect correlation. 4 An AR(p) Case Our framework can be extended to an AR(p) case. To do so, first we modify our model: = + Φ 1 where Φ 1 = 1 1 1...... 1 1 for an AR(1) case. The log-likelihood function (concentrated with respect to )isthen ( Φ) = 1 log log Φ 1 Φ (3) where = + Φ + Φ + Φ Φ. Similar to the previous section, our strategy here is that we concentrate the likelihood function with respect to Φ, so that (3) is a function of ( ): ( Φ ( )) = 1 log ( ) log Φ () ( ) 1 Φ ( ) (4) Note that Φ ( ) is obtained by solving ( Φ) =with respect to. Assuming DGP to be = Φ 1, Lemmas 1-3 are altered to: Lemma 4 The first order conditions are given by: ( Φ ( )) = 1 1 Φ + Φ ª + and (Φ ( )) = 1 1 Φ Φ ª + Lemma 5 At = =and for, we have ( Φ ( )) (1 ) Φ 1 ( Φ + Φ) 1 Φ Φ 1 Φ Φ 1 Φ Φ 1 Φ Φ 1 Φ Z 1 [ ()] 1 where represents weak convergence in distribution; and () is the standard Wiener process on C[ 1]. 5 = =

Lemma 6 At = =and for, we have 1 ( Φ ( )) (1 ) Z 1 () () = (1 ) n o [ (1)] 1 Clearly, = is the previous case. Also, it is not difficult to extend the Lemmas noted above to an AR() model, which is argued in Morley et al. (3). 6 For a finite sample, Figure 3 is computed as follows: first, b is obtained as a function of ( ), i.e., by solving ( Φ) = with respect to, using the MATLAB function fzero. Then, given, b the likelihood function is computed by (4). 5 Conclusion When data are generated by a stationary process, the correlation in the error of a stochastic trend and errors of cycles in the state-space model is undefined because there is no stochastic trend. If one allows for such a correlation and estimate the parameters, the correlation parameter will be unidentified, and unexpectedly, the estimated parameters will almost always be 1 or -1. It is shown that the following two facts explain such a result: (i) We impose the restriction that requires the variance covariance matrix of the errors to be positive semi-definite (in other words, either the variance of the trend error is zero or the correlation is perfect). The set of the true parameter values lies on the boundary in the parameter space and in the neighborhood of parameters that imply perfect correlation. (ii) The likelihood function has its (constrained) global maximum at the true parameters only on average. Almost always its global maximum is on the boundary. However, caution is necessary. Unlike Morley et al. (11), whose proposed likelihood ratio test of stationarity compares the likelihood value with the restriction of = =to the likelihood value without the restriction, our diagnosis principle does not fit the framework of a rigorous test of stationarity. This is because an estimated perfect correlation does not necessarily mean that the true DGP is stationary; it might as well be that the true DGP is non-stationary with a perfect correlation. References [1] Fuller, W.A. (1996) Introduction to Statistical Time Series John Wiley & Sons, Inc. [] Morley, J.C., C.R. Nelson and E. Zivot (3) Why are Beveridge-Nelson and Unobserved- Component Decompositions of GDP So Different? Review of Economics and Statistics 85, 35-43. [3] Morley, J.C., I. Panovska and T.M. Sinclair (11) A Likelihood Ratio Test of Stationarity Based on a Correlated Unobserved Components Model, RPF Working Paper 8-11, George Washington University. [4] Perron, P. and T. Wada (9) Let s Take a Break: Trends and Cycles in US Real GDP, Journal of Monetary Economics 56 (6) 749-765 6 In an AR() case, for example, the coefficients of the limit distributions in Lemmas 5 and 6 need to be replaced by (1 1 ) and (1 1 ), respectively. See Online Appendix. 6

[5] Tanaka, K. (1996) Time Series Analysis John Wiley & Sons, Inc. [6] Watson, M. (1986) Univariate Detrending Methods with Stochastic Trend, Journal of Monetary Economics 18, 49-75. 6 Appendix ProofofLemma Since ( ) = == = µ ( +1) µ ( +1) 1 X Z = =1 [ ()] where = 1 + with =; and denotes convergence in probability. We have then Proof of Lemma 3 Note further that ( ) ( ) 1 Z 1 [ ()] 1 == = 1 == + + = + ( + ) = + ( + ) 1 X 1 = 1 = 1 1 =1 n o [ (1)] 1 X =1 + 1 + 1 X =1 7

Hence, 1 ( ) 1+ 1 == = 1 n o [ (1)] 1 +1 n [ (1)] 1 o Proof of Lemma 4 (Sketch) The derivations are similar to those given in Lemma 1. Note that the Envelope theorem: ( Φ ( )) Φ ( ) =; Φ ( ) ( Φ ( )) Φ ( ) = Φ () is used to obtain the results. ProofofLemma5 At = =, = and =holds. Then, we can prove that the estimator for is a consistent estimator. To show this, noting that Φ =(1 ) + 1,thefirst order condition for, ³ ³ ( Φ) = = 1 µ 1 1 Φ 1 Φ 1 Φ Φ Φ 1 Φ is simplified at = =to = Φ Φ = 1 Φ and hence, 7 b = P = 1 P = 1 Since the least square estimator is consistent under the DGP process considered here (see Fuller 1996, for example), our estimator for the autoregressive parameters is consistent. 7 The first order condition is now 1 Φ = 1...... 1 1. 1 1...... 1 1. = ( 1) 1 = 8

Next,wenotethat Φ Φ = b + b ³ Φ Φ = b + b ³ 1 b 1 b + + + ³ 1 b ; ³ 1 b ( +1) Since b and the fact that DGP is = Φ 1, the same argument as Lemma leads to the limit distribution. ProofofLemma6 The limit distribution is obtained since ³ Φ + Φ = 1 b + +; b Φ + Φ = Table 1: Frequency Distributions of the Estimated Correlation Data Generating Process Estimated AR() AR(1) AR() 99 71 654 711 99 99 85 116 1 99 5 3 187 99 915 (915%) 884 (884%) 898 (898%) Notes: 1) The estimated frequencies are are computed from the sample size = 5 with 1, replications. ) For the AR(1) model, DGP is =9 1 +,where iid( 1) and =. 3) For the AR() model, DGP is =18 1 38 +,where iid( 1) and = 1 =. For more details, see Online Appendix. 9

1 Palameter Space.9.8 Positeve semi definite restriction.7.6 Delta.5.4.3..1 True parameter values 1.8.6.4...4.6.8 1 S Figure 1: Parameter space and the restriction Likelihood Surface: AR Likelihood Surface: AR1 Global Max 1 3 True Parameter Values 1 3 Global Max True Parameter Values 4 4 5 5 6 6 7 7 8 8 9 9 1 1.6 1.4 x 1 3 1. 1.8 Delta.6.4..4.3..1 S.1..3.4 1.4.35.3.5. Delta.15.1.5..15.1.5 S.5.1.15. Figure : Typical Likelihood Surface: AR() with T=5. Figure 3: Typical Likelihood Surface: AR(1) with T=5.