Bridging DSGE models and the raw data

Similar documents
Chapter 6. Maximum Likelihood Analysis of Dynamic Stochastic General Equilibrium (DSGE) Models

DSGE models: problems and some personal solutions. Fabio Canova EUI and CEPR. March 2014

Economics Discussion Paper Series EDP Measuring monetary policy deviations from the Taylor rule

Bridging cyclical DSGE models and the raw data

Time Series Models and Inference. James L. Powell Department of Economics University of California, Berkeley

Matching DSGE models,vars, and state space models. Fabio Canova EUI and CEPR September 2012

Choosing the variables to estimate singular DSGE models

Topics in Bayesian estimation of DSGE models. Fabio Canova EUI and CEPR February 2014

Chapter 1. GMM: Basic Concepts

Choosing the variables to estimate singular DSGE models

Are Policy Counterfactuals Based on Structural VARs Reliable?

NBER WORKING PAPER SERIES CORRELATED DISTURBANCES AND U.S. BUSINESS CYCLES. Vasco Cúrdia Ricardo Reis

RBC Model with Indivisible Labor. Advanced Macroeconomic Theory

Optimal Monetary Policy in a Data-Rich Environment

Equivalence of several methods for decomposing time series into permananent and transitory components

FEDERAL RESERVE BANK of ATLANTA

Paper. Series. Vasco Cúrdia Ricardo Reis. Discussion

4- Current Method of Explaining Business Cycles: DSGE Models. Basic Economic Models

A Composite likelihood approach for dynamic structural models

Identi cation and Frequency Domain QML Estimation of Linearized DSGE Models

Discussion of Juillard and Maih Estimating DSGE Models with Observed Real-Time Expectation Data

Combining Macroeconomic Models for Prediction

Correlated Disturbances and the Sources of U.S. Business Cycles

Dynamic Factor Models and Factor Augmented Vector Autoregressions. Lawrence J. Christiano

The Price Puzzle: Mixing the Temporary and Permanent Monetary Policy Shocks.

Monetary and Exchange Rate Policy Under Remittance Fluctuations. Technical Appendix and Additional Results

Estimation of moment-based models with latent variables

Notes on Time Series Modeling

On Econometric Analysis of Structural Systems with Permanent and Transitory Shocks and Exogenous Variables

Monetary Policy Regimes and Economic Performance: The Historical Record,

Identi cation and Frequency Domain QML Estimation of Linearized DSGE Models

Lecture 3, November 30: The Basic New Keynesian Model (Galí, Chapter 3)

Structural Macroeconometrics. Chapter 4. Summarizing Time Series Behavior

Graduate Macro Theory II: Notes on Quantitative Analysis in DSGE Models

Markov-Switching Models with Endogenous Explanatory Variables. Chang-Jin Kim 1

Adaptive Learning and Survey Expectations of In ation

Parametric Inference on Strong Dependence

Contents. University of York Department of Economics PhD Course 2006 VAR ANALYSIS IN MACROECONOMICS. Lecturer: Professor Mike Wickens.

Economics 701 Advanced Macroeconomics I Project 1 Professor Sanjay Chugh Fall 2011

Agnostic Structural Disturbances (ASDs): Detecting and Reducing Misspecification in Empirical Macroeconomic Models

GMM-based inference in the AR(1) panel data model for parameter values where local identi cation fails

Stochastic Trends & Economic Fluctuations

Estimating the Number of Common Factors in Serially Dependent Approximate Factor Models

Dynamic probabilities of restrictions in state space models: An application to the New Keynesian Phillips Curve

Department of Econometrics and Business Statistics

Introduction: structural econometrics. Jean-Marc Robin

Inference in VARs with Conditional Heteroskedasticity of Unknown Form

Assessing Structural VAR s

Labor-Supply Shifts and Economic Fluctuations. Technical Appendix

1 Regression with Time Series Variables

A Course on Advanced Econometrics

Chapter 1. Introduction. 1.1 Background

Advanced Economic Growth: Lecture 8, Technology Di usion, Trade and Interdependencies: Di usion of Technology

Y t = log (employment t )

ECONOMICS 7200 MODERN TIME SERIES ANALYSIS Econometric Theory and Applications

PANEL DISCUSSION: THE ROLE OF POTENTIAL OUTPUT IN POLICYMAKING

A Composite Likelihood Framework for Analyzing Singular DSGE Models

Identifying the Monetary Policy Shock Christiano et al. (1999)

in Trend University Gary Koop # 590

Problem set 1 - Solutions

Decomposing the effects of the updates in the framework for forecasting

Econometric Forecasting

Chapter 2. GMM: Estimating Rational Expectations Models

Fresh perspectives on unobservable variables: Data decomposition of the Kalman smoother

Macroeconomics Theory II

Estimation of Dynamic Nonlinear Random E ects Models with Unbalanced Panels.

An Extended Macro-Finance Model with Financial Factors: Technical Appendix

Monetary Policy Transmission and the Phillips Curve in a Global Context

COMMENT ON DEL NEGRO, SCHORFHEIDE, SMETS AND WOUTERS COMMENT ON DEL NEGRO, SCHORFHEIDE, SMETS AND WOUTERS

In Search of a Nominal Anchor: What Drives In ation Expectations?

When is it really justifiable to ignore explanatory variable endogeneity in a regression model?

Volume 29, Issue 4. Stability under learning: the neo-classical growth problem

Assessing Structural VAR s

Forecasting Levels of log Variables in Vector Autoregressions

Estimating Unobservable Inflation Expectations in the New Keynesian Phillips Curve

The Basic New Keynesian Model. Jordi Galí. June 2008

Introduction to Macroeconomics

Trend agnostic one-step estimation of DSGE models

MA Advanced Macroeconomics: 7. The Real Business Cycle Model

Estimating Output Gap, Core Inflation, and the NAIRU for Peru

Bayesian Inference for DSGE Models. Lawrence J. Christiano

1 A Non-technical Introduction to Regression

Research Division Federal Reserve Bank of St. Louis Working Paper Series

Do Long Run Restrictions Identify Supply and Demand Disturbances?

Perceived productivity and the natural rate of interest

Chapter 5. Structural Vector Autoregression

Assessing Structural VAR s

Identifying SVARs with Sign Restrictions and Heteroskedasticity

Discussion of "Noisy News in Business Cycle" by Forni, Gambetti, Lippi, and Sala

Federal Reserve Bank of New York Staff Reports

TAKEHOME FINAL EXAM e iω e 2iω e iω e 2iω

Indirect Inference- an introduction

Macroeconomics II Dynamic macroeconomics Class 1: Introduction and rst models

Panel Data. March 2, () Applied Economoetrics: Topic 6 March 2, / 43

Multivariate modelling of long memory processes with common components

Dynamic Factor Models Cointegration and Error Correction Mechanisms

Testing for a Trend with Persistent Errors

Employment, Wage, and Output Dynamics under External Habit Formation and E ciency Wages

What You Match Does Matter: The Effects of Data on DSGE Estimation

Measurement Errors and the Kalman Filter: A Unified Exposition

Transcription:

1 2 3 Bridging DSGE models and the raw data Fabio Canova EUI and CEPR March 13, 214 4 5 6 7 8 9 1 11 12 Abstract A method to estimate DSGE models using the raw data is proposed. The approach links the observables to the model counterparts via a exible speci cation which does not require the model-based component to be located solely at business cycle frequencies, allows the non model-based component to take various time series patterns, and permits certain types of model misspeci cation. Applying standard data transformations induce biases in structural estimates and distortions in the policy conclusions. The proposed approach recovers important model-based features in selected experimental designs. Two widely discussed issues are used to illustrate its practical use. 13 14 JEL classi cation: E3, C3. Keywords: DSGE models, Filters, Structural estimation, Business cycles. I would like to thank Urban Jermann and Bob King (the editors), two anonymous referees, Frank Schorfheide, Tim Cogley, Giorgio Primiceri, Tao Zha, Chris Sims, Harald Uhlig and the participants of seminars at the Bank of England, Bank of Italy, BIS, UCL, Yale, EUI, University of Modena, University of Sussex, Federal Reserve Bank of New York, University of Zurich, ESSIM, the Fed of Atlanta workshop on Methods of applications for DSGE models; the I t workshop in Time Series Econometrics, Zaragoza; and the conference Recent Development of Dynamic Analysis in Economics, Seoul; The Gent workshop in Macroeconomics for comments and suggestions. The nancial support of the Spanish Ministry of Economy and Competitiveness, through the grants ECO29-8556, ECO212-33247, of the Barcelona Graduate School of Economics and of the European University Institute is gratefully acknowledged. Department of Economics, EUI, Via della Piazzuola 43, 5133 Firenze, Italy, fabio.canova@eui.eu.

15 16 17 18 19 2 21 22 23 24 25 26 27 28 29 3 31 32 33 34 35 36 37 38 39 4 1 INTRODUCTION 1 1 Introduction There have been considerable developments in the speci cation of DSGE models in the last few years. Steps forward have also been made in the estimation of these models. Despite recent e orts, structural estimation of DSGE models is conceptually and practically di - cult. For example, classical estimation is asymptotically justi ed only when the model is the generating process (DGP) of the actual data, up to a set of serially uncorrelated measurement errors, and standard validation exercises are meaningless without such an assumption. Identi cation problems (see e.g. Canova and Sala, 29) and numerical di culties are widespread. Finally, while the majority of the models investigators use are intended to explain only the cyclical portion of observable uctuations, both permanent and transitory shocks may produce cyclical uctuations, and macroeconomic data contain many types of uctuations, some of which are hardly cyclical. The generic mismatch between what models want to explain and what the data contain creates headaches for applied investigators. A number of approaches, re ecting di erent identi cation assumptions, have been used: Fit a model driven by transitory shocks to the observables ltered with an arbitrary statistical device (see Smets and Wouters, 23, Ireland, 24a, Rubio and Rabanal, 25, among others). Such an approach is problematic for at least three reasons. First, since the majority of statistical lters can be represented as a symmetric, two-sided moving average of the raw data, the timing of the information is altered and dynamic responses hard to interpret. Second, while it is typical to lter each real variable separately and to demean nominal variables, there are consistency conditions that must hold - a resource constraint need not be satis ed if each variable is separately ltered - and situations when not all nominal uctuations are relevant. Thus, speci cation errors can be important. Finally, contamination errors could be present. For example, a Band Pass (BP) lter only roughly captures the power of the spectrum at the frequencies corresponding to cycles with 8-32

1 INTRODUCTION 2 41 42 43 44 45 46 47 48 49 5 51 52 53 54 55 56 57 58 59 6 61 62 63 64 65 66 quarters average periodicity in small samples and taking growth rates greatly ampli es the high frequency content of the data. Thus, rather than solving the problem, such an approach adds to the di culties faced by applied researchers. Fit a model driven by transitory shocks to transformations of the observables which, in theory, are void of non-cyclical uctuations, e.g. consider real great ratios (as in Cogley, 21, and McGrattan, 21) or nominal great ratios (as in Whelan, 25). As Figure 1 shows, such transformations may not solve the problem because many ratios still display low frequency movements. In addition, since the number and the nature of the shocks driving non-cyclical uctuations needs to be a-priori known, speci cation errors may be produced. Construct a model driven by transitory and permanent shocks; scale the model by the assumed permanent shocks; t the transformed model to the observables transformed in the same way (see e.g. Del Negro et al., 26, Fernandez and Rubio, 27, Justiniano et al., 21, among others). Such an approach puts stronger faith in the model than previous ones, explicitly imposes a consistency condition between the theory and the observables, but it is not free of problems. For example, since the choice of which shock is permanent is often driven by computational rather than economic considerations, speci cation errors could be present. In addition, structural parameter estimates may depend on nuisance features, such as the shock which is assumed to be permanent and its time series characteristics. As Cogley (21) and Gorodnichenko and Ng (21) have shown, misspeci cation of these nuisance features may lead to biased estimates of the structural parameters. Construct a model driven by transitory and/or permanent shocks; estimate the structural parameters by tting the transformed model to the transformed data over a particular frequency band (see e.g. Diebold et. al, 1998, Christiano and Vigfusson, 23). This approach is also problematic since it inherits the misspeci cation problems of the previous approach and the ltering problems of the rst approach. The paper shows rst that the approach one takes to match the model to the data matters 67 for structural parameter estimation and for economic inference. Thus, unless one has a

1 INTRODUCTION 3 68 69 7 71 72 73 74 75 76 77 78 79 8 81 82 83 84 85 86 87 88 89 9 91 92 93 94 strong view about what the model is supposed to capture and with what type of shocks, it is di cult to credibly select among various structural estimates (see Canova, 1998). In general, all preliminary data transformations should be avoided if the observed data is assumed to be generated by rational agents maximizing under constraints in a stochastic environment. Statistical ltering does not take into account that cross equation restrictions can rarely be separated by frequency, that the data generated by a DSGE model has power at all frequencies and that, if permanent and transitory shocks are present, both the permanent and the transitory component of the data will appear at business cycle frequencies. Model based transformations impose tight restrictions on the long run properties of the data. Thus, any deviations from the imposed structure must be captured by the shocks driving the transformed model, potentially inducing parameter distortions. As an alternative, one could estimate the structural parameters by creating a exible non-structural link between the DSGE model and the raw data that allows model-based and non model-based components to have power at all frequencies. Since the non model-based component is intended to capture aspects of the data in which the investigator is not interested but which may a ect inference, speci cation errors could be reduced. In addition, because the information present at all frequencies is used in the estimation, ltering distortions are eliminated and ine ciencies minimized. The methodology can be applied to models featuring transitory or transitory and permanent shocks and only requires that interesting features of the data are left out from the model - these could be low frequency movements of individual series, di erent long run dynamics of groups of series, etc.. The setup has two other advantages over competitors: structural estimates re ect the uncertainty present in the speci cation of non model-based features; what the model leaves out at interesting frequencies is quanti able with R-squared type measures. Thus, one can "test" the structure and to evaluate the explanatory power of additional shocks. The approach is related to earlier work of Altug (1989), McGrattan(1994) and Ireland (24b). As in these papers, a non-structural part is added to a structural model prior to

2 ESTIMATION WITH TRANSFORMED DATA 4 95 96 97 98 99 1 11 12 13 14 15 16 17 18 19 11 111 112 113 114 115 estimation, but here the non-structural part is not designed to eliminate singularity. More crucially, the approach does not substitute for theoretical e orts designed to strengthen the ability of DSGE models to account for all observable uctuations. But it can ll the gap between what is nowadays available and such a worthy long run aspiration, giving researchers a rigorous tool with which to address policy questions. Using a simple experimental design and two practically relevant cases, the paper documents the biases that standard transformations produce, interprets them using the tools developed in Hansen and Sargent (1993), and shows that crucial parameters are better estimated with the proposed procedure. To highlight how the approach can be used in practice, the paper examines nally two questions greatly discussed in macroeconomics: the time variations in the policy activism parameter and the sources of output and in ation uctuations. To focus attention on the issues of interest, two simplifying assumptions are made: (i) the estimated DSGE model features no missing variables or omitted shocks and (ii) the number of structural shocks equals the number of endogenous variables. While omitted variables and singularity issues are important, and the semi-structural methods suggested in Canova and Paustian (211) produce more robust inference when they are present, I sidestep them because the problems discussed here occur regardless of whether (i)-(ii) are present or not. The rest of the paper is organized as follows. The next section presents estimates of the structural parameters when a number of statistical and model based transformations are employed. Section 3 discusses the methodology. Section 4 compares approaches using a simple experimental design. Section 5 examines two economic questions. Section 6 concludes. 116 117 118 119 12 2 Estimation with transformed data The purpose of this section is to show that estimates of the structural parameters and inference about the e ect of certain shocks depend on the preliminary transformation employed to match a model to the data. Given the wide range of outcomes, we also argue that it is di cult to select a set of estimates for policy and interpretation purposes. We consider a textbook

2 ESTIMATION WITH TRANSFORMED DATA 5 121 122 123 124 125 126 127 128 129 13 131 132 133 134 135 136 137 138 139 14 141 142 143 144 145 146 147 small scale New-Keynesian model, where agents face a labour-leisure choice, production is stochastic and requires labour, there is external habit in consumption, an exogenous probability of price adjustments, and monetary policy is conducted with a conventional Taylor rule. Details on the structure are in the on-line appendix. The model features a technology disturbance z t, a preference disturbance t, a monetary policy disturbance t ; and a markup disturbance t. The latter two shocks are assumed to be iid. Depending on the speci cation z t ; t are either both transitory, with persistence z and respectively, or one of them is permanent. The structural parameters to be estimated are: c, the risk aversion coe cient, n, the inverse of the Frisch elasticity, h the coe cient of consumption habit, 1, the share of labour in production, r, the degree of interest rate smoothing, and y, the parameters of the monetary policy rule, 1- p, the probability of changing prices. The auxiliary parameters to be estimated are: ; z, the autoregressive parameters of transitory preference and technology shocks, and z ; ; r ; ; the standard deviations of the four structural shocks. The discount factor and the elasticity among varieties are not estimated since they are very weakly identi ed from the data. Depending on the properties of the technology and of the preference shocks, the optimality conditions will have a log-linear representation around the steady state or a growth path, driven either by the technology or by the preference shock, see table 1. Four observable variables are used in the estimation. When the model is assumed to be driven by transitory shocks, parameter estimates are obtained i) applying four statistical lters (linear detrending (LT), Hodrick and Prescott ltering (HP), growth rate ltering (FOD) and band pass ltering (BP)) to output, the real wage, the nominal interest rate and in ation or ii) using three data transformations. In the rst, the log of labour productivity, the log of real wages, the nominal rate and the in ation rate, all demeaned, are used as observables (Ratio 1). In the second the log ratio of output to the real wage, the log of hours worked, the nominal rate and the in ation rate, all demeaned, are used as observables (Ratio 2). In the third, the log of the labour share, the log ratio of real wages to output, the nominal interest rate and the

2 ESTIMATION WITH TRANSFORMED DATA 6 148 149 15 151 152 153 154 155 156 in ation rate all demeaned, are used as observables (Ratio 3). When the model features a trending TFP (TFP), the linear stochastic speci cation z t = bt + z t ; is used and, consistent with the theory, the observables for the transformed model are linearly detrended output, linearly detrended wages, demeaned in ation and demeaned interest rates. When the model features trending preferences shocks (Preferences), the unit root speci cation, t = t 1 + t, is employed and the observables for the transformed model are the demeaned growth rate of output, demeaned log of real wages, demeaned in ation and demeaned interest rates. Finally, when the model features a trending TFP, the likelihood function of the transformed model is approximated as in Hansen and Sargent (1993) and only the information present at business cycle frequencies ( ; 157 ) is used in the estimation (TFP FD). 32 8 The data used comes from the FRED quarterly database at the Federal Reserve Bank of 158 159 16 161 162 163 164 165 166 167 168 169 17 171 172 173 174 St. Louis and Bayesian estimation is employed. Since some of the statistical lters are two- sided, a recursive LT lter and a one-sided version of the HP lter have also been considered. The qualitative features of the results are unchanged by this re nement. Table 2 shows that the posterior distribution of several parameters depends on the pre- liminary transformation used (see e.g. the risk aversion coe cient c ; the Frisch elasticity n 1 ; the interest smoothing coe cient r ; persistence and the volatility of the shocks). Since posterior standard deviations are generally tight, di erences across columns are a-posteriori signi cant. Posterior di erences are also economically relevant. For example, the volatil- ity of markup shocks in the LT, the Ratio 1 and the Preference economies is considerably larger and, perhaps unsurprisingly, risk aversion stronger. Note that, even within classes of transformations, di erences are present. For example, comparing the Ratio 1 and Ratio 3 economies, it is clear that using the labour share and the ratio of real wages to output as ob- servables considerably reduces the persistence of the technology shocks - rendering the Ratio 3 transformation more appropriate as far as stationarity of the observables is concerned - at the cost of making the risk aversion and habit coe cient very low. Di erences in the location of the posterior of the parameters translate into important

2 ESTIMATION WITH TRANSFORMED DATA 7 175 176 177 178 179 18 181 182 183 184 185 186 187 188 189 19 191 192 193 194 195 196 197 198 199 2 21 di erences in the transmission of shocks. As shown in Figure 2, the magnitude of the impact coe cient and of the persistence of the responses to technology shocks vary with the preliminary transformation and, for the rst few horizons, di erences are statistically signi cant. Furthermore, the sign of output and interest rate responses is a ected. Why are parameter estimates so di erent? The rst four transformations only approximately isolate business cycle frequencies, leaving measurement errors in the transformed data. In addition, di erent approaches spread the measurement error across di erent frequencies: the LT transformation leaves both long and short cycles in the ltered data; the HP transformation leaves high frequencies variability unchanged; the FOD transformation emphasizes high frequency uctuations and reduces the importance of cycles with business cycle periodicity; and even a BP transformation induces signi cant small sample approximation errors (see e.g. Canova, 27). Since the magnitude of the measurement error and its frequency location is transformation dependent, di erences in parameter estimates emerge. An approach which can reduce the problematic part of the measurement error is in Canova and Ferroni (211). More importantly, ltering approaches neglect the fact that the spectral properties of a DSGE model are di erent from the output of a statistical lter. Data generated by a DSGE model driven by transitory shocks have power at all frequencies of the spectrum and if shocks are persistent most of the power will be in the low frequencies. Thus, concentrating on business cycles frequencies may lead to ine ciencies. When transitory and permanent shocks are present, the transitory and the permanent components of the model will jointly appear in any frequency band and it is not di cult to build examples where permanent shocks dominate the variability at business cycle frequencies (see Aguiar and Gopinath, 27). Hence, the association between the solution of the model and the ltered observables generally leads to biases. Implicit or explicit model-based transformations avoid these problems by specifying a permanent and a transitory component of the data with power at all frequencies of the spectrum. However, since speci cation problems are present (should we use a unit root process

3 THE ALTERNATIVE METHODOLOGY 8 22 23 24 25 26 27 28 29 21 211 212 or a trend stationary process? Should we allow trending preferences or trending technology shocks?), nuisance parameters problems could be important (the model estimated with a trending TFP has MA components which do not appear when the preferences are trending, see table 1), and tight cointegration relationships are imposed on the observables, any deviation from the assumed structure leads to biases. Finally, frequency domain estimation may require a model-based transformation (in which case the problems discussed in the previous paragraph apply) and is generally ine cient, since most of the variability the model produces is in the low frequencies. In general, while frequency domain estimation can help to tone down the importance of aspects of the model researchers do not trust, see Hansen and Sargent (1993), it cannot reduce the importance of what the model leaves unexplained at business cycle frequencies. 213 214 215 216 217 3 The alternative methodology Start from the assumption that the observable data has been generated by rational expectation agents, optimizing their objective functions under constraints in a stochastic environment. Suppose that the log of an N 1 demeaned vector of time series x d t can be decomposed in two mutually orthogonal parts x d t = z t + x t (1) 218 219 22 221 222 223 224 225 Assume that the econometrician is con dent about the process generating x t u t = x m t (), where is a vector of structural parameters, and u t a vector of iid measurement errors but he/she is unsure about the process generating z t = zt m (; ), where is another vector of structural parameters because she does not know the shocks which are driving z t ; because she does not feel con dent about their propagation properties; or because she does not know how to model the relationship between and. In the context of section 2, z t is the permanent component and x t the transitory component of the data, and the researcher is unsure about the modelling of z t because it could be deterministic or stochastic, it could

3 THE ALTERNATIVE METHODOLOGY 9 226 227 228 229 23 231 232 233 234 235 236 237 238 239 24 241 242 243 244 be driven by preference or technology shocks, and balance growth could hold or not. Still, she wants to employ x m t () for inference because z t may be tangential to the issues she is interested in. Thus, she is aware that the model is misspeci ed in at least two senses: there are shocks missing from the model; and there are cross equation restrictions that are ignored. An investigator interested in estimating and conducting structural inference does not necessarily have to construct an estimate of x m t (); ltering out from the data what the model is unsuited to explain; add ad-hoc structural features hoping that ~z t m D(`)(; )~e 1t is close to x d t x m t (), where, as in section 2, ~e 1t is a set of (permanent) shocks and D(`)(; ) a model propagating ~e 1t, or transform the observables so that z t becomes a vector of iid random variables, as is commonly done. Instead, she can use the raw data x d t ; the model x m t (); and build a non-structural link between the (misspeci ed) structural model and the raw data which is su ciently exible to capture what the model is unsuited to explain, and allows model-based and non model-based components to jointly appear at all frequencies of the spectrum. As a referee has pointed out, the assumption of orthogonality of z t and x t is crucial for the procedure outlined below to be e ective. When permanent drifts in the data occur because of drifting structure or drifting cyclical parameters rather than permanent shocks, alternative approaches need to be considered. Let the (log)-linearized stationary solution of a DSGE model be of the form: x 2t = A()x 1t 1 + B() t (2) x 1t = C()x 1t 1 + D() t (3) 245 246 247 248 249 where A(); B(); C(); D() depend on the structural parameters, x 1t (log ~x 1t log x 1t ) includes exogenous and endogenous states, x 2t = (log ~x 2t log x 2t ) all other endogenous variables, t the shocks and x 2t ; x 1t are the long run paths of ~x 2t and ~x 1t. Let x m t () = R[x 1t ; x 2t ] be an N 1 vector, where R is a selection matrix picking out of x 1t and x 2t variables which are observable and/or interesting from the point of view of the

3 THE ALTERNATIVE METHODOLOGY 1 25 251 analysis and let x m t () = R[x 1t ; x 2t ]. Let x d t = log ~x d t E(log ~x d t ) be the log demeaned N 1 vector of observable data. The speci cation for the raw data is: x d t = c t () + x nm t + x m t () + u t (4) 252 253 where c t () = log x m t () mutually orthogonal and x nm t E(log ~x d t ), u t is a iid (; u ) (proxy) noise, x nc t ; x m t and u t are is given by: x nm t = 1 x nm t 1 + w t 1 + v 1t v 1t iid (; 1 ) w t = 2 w t 1 + v 2t v 2t iid (; 2 ) (5) 254 255 256 257 258 259 26 261 262 263 264 265 266 267 268 269 27 271 where 1 = diag( 11 ; ::: 1N ); 2 = diag( 21 ; ::: 2N ); < 1i ; 2i 1; i = 1; :::N. To under- stand what (5) implies, notice that when 1 = 2 = I, and v 1t ; v 2t are uncorrelated x m t () is the locally linear trend speci cation used in state space models, see e.g. Gomez (1999). On the other hand, if 1 = 2 = I; 1 and 2 are diagonal, 1i =, and 2i > ; 8i, x nm t is a vector of I(1) processes while if 1i = 2i = ; 8i, x mn t is deterministic. When instead 1 = 2 = I, and 1i and 2i are functions of, (5) approximates the double exponential smoothing setup used in discounted least square estimation of state space models, see e.g. Delle Monache and Harvey (21). Thus, if x m t () = x m (); 8t, the observable x d t can display any of the typical structures that motivate the use of the statistical lters. Furthermore, as emphasized by Delle Monache and Harvey (21), (5) can capture several other types of structural model misspeci cation. For example, whenever 2 is di erent from zero, the growth rate of the endogenous variables may display persistent deviations from their mean, a feature that characterizes many real macroeconomic variables, see e.g. Ireland (212), even if the model is driven by transitory shocks. Finally, when x m t () is not constant, and 1i and 2i are complex conjugates for some i, the speci cation can account for residual low frequency variations with power at frequency!. To see this note that when N=1, (5) implies that (1 2 L)(1 1 L)x nm t = (1 2 L)v 1t + v 2t 1 (1 L) t. If the roots 1 1 ; 1 2 of the polynomial 1 ( 1 + 2 )z + 1 2 z 2 = are complex, they can be written as

272 3 THE ALTERNATIVE METHODOLOGY 11 1 1 = r(cos! + i sin!); 1 2 = r(cos! i sin!), where r = p 1 2 and! = cos 1 [ 1 + 2 2 p 1 2 ] and (5) is x nm t = P j r sin!(j+1) (1 L) sin! t, whose period of oscillation is p = 2 = 2 273! Thus, given r and p, there exists 1; 2 that produce x nm 274 with the required properties. 275 276 277 t. cos 1 [ 1 + 2 2 p ] 1 2 Given (2)-(5), identi cation of the structural parameters is achieved via the cross-equation restrictions that the model imposes on the data. Estimates of the non-structural parameters are implicitly obtained from the portion of the data the model cannot explain. 278 279 28 281 282 283 3.1 Two special cases Two special cases of the setup are of interest. Suppose that the model features only transitory shocks while the data may display common or idiosyncratic long run drifts, low frequency movements, and business cycle uctuations. Here x m t () = x m (); 8t, are the steady states of the model and, if the model is correctly speci ed on average, c t () =. Assume that no proxy errors are present. Then (4) is x d t = x nm t + x m t () (6) 284 285 286 287 288 289 29 291 292 293 294 295 and x nm t captures the features of x d t that the stationary model does not explain. Depend- ing on the speci cation of 1 and 2, these may include long run drifts, both of common and idiosyncratic type, and those idiosyncratic low and business cycle movements the model leaves unexplained. In this setup, x nm t has two interpretations: As in Altug (1989), McGrat- tan (1994) and Ireland (24b), it can be thought of as measurement error added to the structural model. However, rather than being iid or VAR(1), it has the richer representation (5) and it is present even when the number of structural shocks equals the number of en- dogenous variables. Alternatively, x nm t can be thought of as a reduced form representation for the components of the data the investigator decides not to model. Thus, as in Del Negro et al. (26), x nm t relaxes certain cross equations restrictions that the DGP imposes on x d t. Suppose, alternatively, that the model features transitory shocks and one or more per- manent shocks. In this case x m t () represents the (stationary) solution in deviation from a

3 THE ALTERNATIVE METHODOLOGY 12 296 297 growth path and x m t () is the model-based component generated by the permanent shocks. Suppose again that there are no proxy errors. Then (4) is x d t = c t () + x ;nm t + x m t () (7) 298 299 3 31 32 33 34 35 36 37 38 39 31 311 312 313 314 315 316 317 318 319 32 where x ;nm t captures the features of x d t which neither the transitory portion x m t () nor the permanent portion c t () of the model explains. These may include idiosyncratic long run patterns (such as diverging trends), idiosyncratic low frequency movements, or unaccounted cyclical uctuations. Comparing (6) and (7), one can see that x nm t = c t () + x ;nm t. Thus, the setup can be used to measure how much of the data the model leaves unexplained and to evaluate whether the introduction of certain structural shocks reduces the discrepancy. To illustrate, suppose as in the application discussed in section 5.1, one starts from a model featuring a few transitory shocks and nds that the relative importance of x nm t - measured, for example, by the variance decomposition at a particular set of frequencies - is large. Then, one could add a transitory shock or a permanent shock to the model and see how much the relative importance of x nm t has fallen. By comparing the relative size of x nm t the various cases, one can then assess whether adding a permanent or a transitory shock is more bene cial for understanding the dynamics of x d t. The same logic can be used to evaluate the model when, for example, the permanent shock takes the form of a stochastic linear trend, or of a unit root, or when all long run paths are left unmodelled. Hence, the approach provides a setup to judge the goodness of t of a model; a constructive criteria to increase its complexity; and a framework to examine the sensitivity of the estimation results to the speci cation of nuisance features. The speci cation has other advantages over existing approaches. As shown in Ferroni (211), the setup can be used to nd the most appropriate speci cation of the non model- based component, and to perform Bayesian averaging over di erent types of non model- based speci cations, both of which are not possible in standard setups. Finally, since joint estimation is performed, structural parameter estimates re ect the uncertainty present in in

3 THE ALTERNATIVE METHODOLOGY 13 321 322 323 324 the speci cation of the non model-based component. 3.2 Estimation Estimation of the parameters of the model can be carried out with both classical and Bayesian methods. (2)-(5) can be cast into the linear state space system: s t+1 = F s t + G! t+1! t (;! ) (8) x d t = c t () + Hs t (9) 325 326 327 328 329 33 331 332 333 334 335 336 337 338 339 34 341 where s t = F = B @ x nm t w t x m t () u t ;!t+1 = (v 1t+1 ; v 2t+1 ; u t+1 ; t+1 ), H = I I I 1 1 ; I C A ; G = B I C @ R[B D] A : Hence, the likelihood can I 1 I 2 R[A C] be computed with a modi ed Kalman lter (accounting for the possibility of di use initial observations) given # = (; 1 ; 2; 1 ; 2 ; u ) and maximized using standard tools. When a Bayesian approach is preferred, one can obtain the non-normalized posterior of #, using standard MCMC tools. For example, the estimates I present are obtained with a Metropolis algorithm where, given a # 1 and a prior g(#), candidate draws are obtained from # = # 1 + ; where t(; ; 5) and is a tuning parameter, and accepted if g(# jy) g(# 1 jy) exceeds the draw of a uniform random variable, where g(# i jy) = g(# i )L(yj# i ), i = ; 1, and L(yj# i ) is the likelihood of # i,. Iterated a large number of times, for appropriately chosen, limiting distribution of # is the target distribution (see e.g. Canova, 27). 3.3 The relationship with the existing literature Apart from the work of Altug (1989), McGrattan (1994), Ireland (24b), and Del Negro et al. (26) already mentioned, the procedure is related to a number of existing works. First, the state space setup (8)-(9) is similar to the one of Harvey and Jeager (1993), even though these authors consider only univariate processes and do not use a structural model to explain the observables. It also shares important similarities with the one employed by

3 THE ALTERNATIVE METHODOLOGY 14 342 343 344 345 346 347 348 349 35 351 352 353 354 355 356 357 358 359 36 361 362 363 364 365 Cayen et al. (29), who are interested in forecasting trends. Two are the most noticeable di erences. First, these authors use a two-step estimation approach, conditioning on ltered estimates of the parameters of the DSGE model; here a one-step approach is employed. Second, all the deviations from the model are bundled up in the non-model speci cation while here it is possible to split them into interpretable and non-interpretable parts. The contribution of the paper is also related to two distinct branches of the macroeconomic and macroeconometric literature. The rst attempts to robustify inference when the trend properties of data are misspeci ed (see Cogley, 21, and Gorodnichenko and Ng, 21). I share with the rst author the idea that economic theory may not have much to say about certain types of uctuations but rather than distinguishing between trend-stationary and di erence-stationary cycles, I design an estimation procedure which deals exibly with the mismatch between theoretical and empirical concepts of uctuations. The idea of jointly estimating structural and non-structural parameters without fully specifying the DGP is also present in Gorodnichenko and Ng. However, a likelihood based estimator, as opposed to a minimum distance estimator, is used here because it works regardless of the properties of the raw data. In addition, rather than assuming that the model is the DGP, the procedure assumes that the model is misspeci ed - a much more useful assumption in practice. The second branch points out that variations in trend growth are as important as cyclical uctuations in explaining the dynamics of macroeconomic variables (see Aguiar and Gopinath, 27, and Andrle, 28). While the rst paper characterizes di erences between emerging and developing economies, the latter is concerned with the misuse of models driven by transitory shocks in policy analyses for developing countries. The paper shows that the problems they highlight are generic and that policy analyses with misspeci ed models are possible without controversial assumptions on what the model is not designed to explain.

3 THE ALTERNATIVE METHODOLOGY 15 366 367 368 369 37 371 372 373 374 375 376 377 378 379 38 381 382 383 384 385 386 387 388 389 39 391 392 3.4 Setting the priors for 1 and 2 If the number of observables is small and the sample size large, one can estimate structural and non-structural parameters jointly from (8)-(9) in an unrestricted fashion. More realistically, when the number of observables and the sample size are moderate, unrestricted estimation is unfeasible - the model features 2N + 2N 2 non-structural parameters - and weak identi cation problems may be important - variations in the level from variations in the growth rates of the observables are hard to distinguish. Thus, it is worth imposing some structure to reduce the number of estimated non-structural parameters. For example, one may assume that 1 and 2 are diagonal (the non model-based component is series speci c) and of reduced rank (the non model-based component is common across [groups of] series). Alternatively, one may assume that these matrices have sparse non-zero elements on the diagonal (the non model-based component exists only in some observables) or that they are proportional to each other (shocks to the level and the growth rate are related). One may also want to make 1i and 2i common across certain variables. Restrictions of this type may be supported by plots or time series analysis of the observables. Additional restrictions may be needed to make estimation meaningful in small samples because, given a DSGE structure, the decomposition in model based and non model-based components is indexed by the relative intensity of the shocks driving the two components. Given that it is di cult to estimate this intensity parameter unrestrictedly in small samples, a sensible smoothness prior for 1 and 2 may avoid that estimates of non model-based components feature undesirable high frequency variability. Harvey and Jeager (1993) have indicated that in univariate state space models, estimation of the cycle depends on the as- sumptions about the trend - in particular whether it is deterministic or stochastic. The problem we highlight here is di erent: given assumptions about the trend, di erent decompositions of the observables in model and non-model based components are implied by di erent estimates of the relative variance of the permanent shocks. Thus, for example, if we assume that the trend is driven by permanent shocks, di erent decompositions may

4 THE PROCEDURE IN A CONTROLLED EXPERIMENT 16 393 394 395 396 397 398 399 4 41 42 43 44 45 46 47 48 49 41 411 412 413 414 415 416 417 418 be obtained if the relative magnitude of the shocks driving the two components is weak or strongly identi ed. The speci cation I found to work best in practice, and it is employed in the two ap- plications in section 5, involves making 1 and 2 a function of the structural shocks. As mentioned, it is possible to approximate the double exponential smoothing restrictions used in discounted least square estimation of state space models by selecting 1i = and q 2i = 2 ; where i indicates the non-zero elements of the matrices, (4) 2 t is one of structural shocks and a smoothing parameter. Thus, given a prior for t and, a prior for all non-zero elements of 1 and 2 is automatically generated. Since has the same interpretation as in the HP lter, an agnostic quarterly prior for could be uniform over [4,64], which allows for very smooth as well as relatively jagged non-model based components. It is worth noting that this speci cation is parsimonious and that selecting the signal to noise ratio is less controversial than assuming a particular format for the drifts the data displays or selecting a shock driving them. Since a structural shock needs to be selected, one could experiment and choose the disturbance with the largest or the smallest variance. For the applications in section 5, the way the prior is scaled is irrelevant. q 2 An alternative approach, suggested by one of the referees, would be to exploit the exi- bility of (5) to perform sensitivity analysis to alternative speci cations of 1 ; 2; 1 ; 2. Also in this case, restrictions to reduce the dimensionality of non-structural parameter space are generally needed to make estimation results sensible. 4 The procedure in a controlled experiment To examine the properties of the procedure and to compare them to those of standard transformations, I use the same setup employed in section 2 and simulate 15 data points 5 times, assuming rst that the preference shock has a transitory and a permanent component. Thus, t = 1t + 2t ; 1t = 1t 1 + T t and 2t = 2t 1 + P t where p = T is uniformly distributed [1.1, 1.9]. Because 2t is orthogonal to all transitory shocks, the design ts the

4 THE PROCEDURE IN A CONTROLLED EXPERIMENT 17 419 42 421 422 423 424 425 426 427 428 429 43 431 432 433 434 435 436 437 438 439 44 441 442 443 444 445 setup of section 3. The speci cation is chosen since Chang et al. (27) have indicated that a model with permanent preference shocks can capture well low frequency variations in hours worked. In this setup, the data displays stationary uctuations, driven by four transitory shocks (which are correctly captured with a model), and non-stationary uctuations, driven by the permanent preference shock (which will be either ltered out, eliminated with certain data transformations, or accounted for with a non-model based component). The estimated model is misspeci ed since the permanent preference shock is left out, but all the other features are correctly represented. Since the contribution of the permanent component is of the same order of magnitude as the contribution of the transitory component at almost all frequencies, standard transformations will feature both ltering and speci cation errors. When the proposed approach is used, the non model-based component is restricted to having a double exponential smoothing format and, consistently with the DGP, is allowed to enter only in output and the real wage (see on-line appendix). The second design features only transitory shocks, but measurement error is added to the data. The variability of the measurement error relative to the variability of the preference shock is uniformly distributed in the range [.8,.12]. Here the model captures the dynamics of the data correctly, but (a constant) noise is present at all frequencies. The question of interest is whether the suggested speci cation will be able to recognize that there is no non model-based component or whether the non model-based component will absorb part of the model dynamics. Note that, since the signal to noise ratio di ers in the two designs, we can also evaluate how our smoothness prior works in di erent situations. The structural parameters will be estimated in the most ideal situations one could consider - these include priors centered at the true parameter vector (the same prior distributions displayed in table 2 are used) and initial conditions equal to the true parameter vector which is listed in the rst column of table 3. The other columns report, for each of the six estimation procedures we consider, the mean square error (MSE) of each parameter separately, and two cumulative MSE measures, one for the structural parameters and one for all the

4 THE PROCEDURE IN A CONTROLLED EXPERIMENT 18 446 447 448 449 45 451 452 453 454 455 456 457 458 459 46 461 462 463 464 465 466 467 468 469 47 471 472 parameters. The MSE is calculated using the posterior mean estimate in each replication. In the rst design, estimation with HP and BP ltered data produce MSEs that are larger than with LT or FOD data, in particular, for the inverse of the Frisch elasticity and the share of labour in production. Moreover, all ltering procedures have a hard time to pin down the value of the Taylor rule coe cient on output. Perhaps unsurprisingly, all transformations fail to capture both the absolute and the relative variability of the shocks. The ratio transformation is also poor and the cumulative MSEs are the largest of all. In comparison, the exible approach does well in estimating structural parameters (the only exception is the consumption habit parameter) and captures the volatility and persistence of structural shocks much better than competitors. The pattern of the results with the second design is similar, even though several transformations induce larger distortions in the estimates of the Frisch elasticity. The performance of the exible approach is also good in this case. In particular, it does much better than other approaches in capturing the volatility and the persistence of the structural shocks. The implications of these results for standard dynamic analyses are clear. For example, variance decomposition exercises are likely to be distorted if parameter estimates are obtained with standard procedures. This is much less the case when the exible approach is employed. Furthermore, structural inference regarding, e.g. the sluggishness of the policy rate or its sensitivity to output gap uctuations, is less likely to be biased when the approach suggested in the paper is used. To highlight further the properties of the proposed approach, gure 3 compares the autocorrelation function and the spectral density of the true and estimated permanent and transitory components of output for rst design, where the latter is obtained using the me- dian estimates in one replication. The approach performs well: the rate of decay of the autocorrelation functions of the true and the estimated components is similar. As anticipated, the two estimated components have power at all frequencies, but at business cycle frequencies (indicated by the vertical bars in the last row of graphs) the permanent compo-

4 THE PROCEDURE IN A CONTROLLED EXPERIMENT 19 473 474 475 476 477 478 479 nent is more important than the transitory component. The conditional dynamics in response to transitory shocks with true and estimated parameters are in Figure 4. In general, the sign and the persistence of the responses are well matched. Magnitudes and shapes are occasionally imprecisely estimated (see e.g. the responses to technology shocks) but, overall, the approach does a reasonable job in reproducing the main qualitative features of the DGP. To understand the nature of the distortions produced by standard transformations, 48 note that the log-likelihood of the data can be represented as L(jy t ) = [A 1 () + A 2 () + P A 3 ()jy], see Hansen and Sargent (1993), where A 1 () = 1 481! j log det G (! j ), A 2 () = 1 P! j trace [G (! j ) 1 F (! j )], A 3 () = (E(y) ())G (! ) 1 (E(y) ()),! j = j 482 ; j = T 483 ; 1; : : : ; T 1. G (! j ) is the model-based spectral density matrix of y t, () the model-based 484 485 486 487 488 489 49 491 492 493 494 495 496 497 498 499 mean of y t, F (! j ) is the data-based spectral density and E(y) the unconditional mean of y t. A 2 () and A 3 () are penalty functions: A 2 () sums deviations of the model-based from the data-based spectral density over frequencies; A 3 () weights deviations of model-based from data-based means with the spectral density matrix of the model at frequency zero. 1 Suppose the data is transformed so that the zero frequency is eliminated and the low frequencies de-emphasized. Then, the log-likelihood consists of A 1 () and of A 2 () = P! j trace [G (! j )] 1 F (! j ), where F (! j ) = F (! j )I!j, and I!j is a function describ- ing the e ect of the lter at frequency! j. Suppose that I! = I [!1 ;! 2 ], i.e. an indicator function for the business cycle frequencies, as in an ideal BP lter. Then A 2 () matters only at business cycle frequencies. Since at these frequencies [G (! j )] < F (! j ), A 2 () and A 1 () enter additively L(jy t ), two types of biases will be present. Since estimates ^F (! j ) only approximately capture the features of F (! j ), ^A 2 () has smaller values at business cy- cle frequencies and a nonzero value at non-business cycle ones. Moreover, in order to reduce the contribution of the penalty function to the log-likelihood, parameters are adjusted so that [G (! j )] is close to ^F (! j ) at those frequencies where ^F (! j ) is not zero. This is done by allowing tting errors, (a larger A 1 ()), at frequencies where ^F (! j ) is zero - in particular,

4 THE PROCEDURE IN A CONTROLLED EXPERIMENT 2 5 51 52 53 54 55 56 57 58 59 51 511 512 513 514 515 516 517 518 519 52 521 522 523 524 525 526 the low frequencies. Hence, the volatility of the structural shocks will be overestimated (this makes G (! j ) close to ^F (! j ) at the relevant frequencies), in exchange for misspecifying their persistence. These distortions a ect agents decision rules: higher perceived volatility, for example, implies distortions in the Frisch elasticity. Inappropriate persistence estimates, on the other hand, imply that perceived substitution and income e ects are distorted with the latter typically underestimated. When I! is not the indicator function, the derivation of the size and the direction of the distortions is more complicated but the same logic applies. Clearly, di erent I! produce di erent ^F (! j ) and thus di erent distortions. Since estimates of F (! j ) are imprecise, even for large T, there are only two situations when estimation biases are small. First, the permanent component has low power at business cycle frequencies - in this case, the distortions induced by the penalty function are limited. This occurs when transitory volatility dominates. Second, when Bayesian estimation is performed, the prior is selected to limit the distortions induced by the penalty function. This is very unlikely, however, since priors are not elicited with such a scope in mind. If instead one ts a transformed version of the model to transformed data, as is done in P model- based approaches, the log-likelihood is composed of A 1 () = 1! j log jg (! j )I!j j and A 2 () - since the actual and model data are ltered in the same way, the lter does not a ect the penalty function. Suppose that I! = I [!1 ;! 2 ]. Then A 1 () matters only at business cycle frequencies while the penalty function is present at all frequencies. Therefore, parameter estimates are adjusted so as to reduce the misspeci cation at all frequencies. Since the penalty function is generally more important at low frequencies, parameters are selected to make [G (! j )] close to ^F (! j ) at those frequencies and large tting errors are permitted at medium and high frequencies. Consequently, the volatility of the shocks will be generally underestimated in exchange for overestimating their persistence - somewhat paradoxically, this procedure implies that the low frequency components of the data are those that matter most for estimation. Cross frequency distortions imply incorrect inference. For example since less noise is perceived, agents decision rules imply a higher degree of data predictability,

5 TWO APPLICATIONS 21 527 528 and higher perceived persistence implies that perceived substitution and income e ects are distorted with the latter overestimated. 529 53 531 532 533 534 535 536 537 538 5 Two applications This section shows how the proposed approach can be used to inform researchers about two questions which have received a lot of attention in the literature: the time variations in the policy activism parameter and the sources of output and in ation uctuations. The rst question is analyzed with the model presented in section 2. The second with a medium scale model, widely used in academic and policy circles. 5.1 The policy activism parameter What are the features of the monetary policy rule in place during the Great In ation of the 197s and the return to norm of the 198s and 199s? This question has been extensively studied in the literature, following Clarida et al. (2). One synthetic way to summarize y 1, 539 the information contained in the data is to compute the policy activism parameter 54 which gives a sense of the relative importance of the output and the in ation stabilization 541 542 543 544 545 546 547 548 549 55 551 objectives of the Central Bank. The conventional wisdom suggests that the absolute value of this parameter has declined over time, re ecting changes in the preferences of the monetary authorities, but most of the available evidence is obtained either with reduced form methods or, when structural methods are used, with ltered data. Are the results to be trusted? Is the characterization o ered by the approach of this paper di erent? Figure 5 plots the posterior density of the policy activism parameter when the data is linearly detrended (top left box) or HP ltered (top right box) and when the approach of this paper is employed (lower left box) for the samples 1964:1-1979:4 and 1984:1-27:4. The priors for the structural and auxiliary parameters are the same as in table 1. In the exible approach, e and v are assumed to be diagonal, a common non model-based component is assumed for all the variables, the signal-to-noise ratio in the four series is captured by a single parameter, a-priori uniformly