Discussion of Tests of Equal Predictive Ability with Real-Time Data by T. E. Clark and M.W. McCracken

Similar documents
Tests of Equal Predictive Ability with Real-Time Data

Research Division Federal Reserve Bank of St. Louis Working Paper Series

Tests of Equal Predictive Ability with Real-Time Data

EVALUATING DIRECT MULTI-STEP FORECASTS

Predicting bond returns using the output gap in expansions and recessions

Using all observations when forecasting under structural breaks

Reality Checks and Nested Forecast Model Comparisons

Bootstrapping Heteroskedasticity Consistent Covariance Matrix Estimator

Forecasting. Bernt Arne Ødegaard. 16 August 2018

Nested Forecast Model Comparisons: A New Approach to Testing Equal Accuracy

Are Forecast Updates Progressive?

The Comparative Performance of Alternative Out-ofsample Predictability Tests with Non-linear Models

Are Forecast Updates Progressive?

The Bootstrap: Theory and Applications. Biing-Shen Kuo National Chengchi University

Central Bank of Chile October 29-31, 2013 Bruce Hansen (University of Wisconsin) Structural Breaks October 29-31, / 91. Bruce E.

Econometrics 2, Class 1

Research Division Federal Reserve Bank of St. Louis Working Paper Series

Tests of Equal Forecast Accuracy for Overlapping Models

Multivariate Out-of-Sample Tests for Granger Causality

2.5 Forecasting and Impulse Response Functions

Approximating fixed-horizon forecasts using fixed-event forecasts

Forecasting Under Structural Break Uncertainty

General comments Linear vs Non-Linear Univariate vs Multivariate

Econ 423 Lecture Notes: Additional Topics in Time Series 1

TECHNICAL WORKING PAPER SERIES APPROXIMATELY NORMAL TEST FOR EQUAL PREDICTIVE ACCURACY IN NESTED MODELS. Todd E. Clark Kenneth D.

A Forecast Rationality Test that Allows for Loss Function Asymmetries

Testing for Regime Switching in Singaporean Business Cycles

Lecture Notes Prediction and Simulation Based Specification Testing and Model Selection

A Test for State-Dependent Predictive Ability based on a Markov-Switching Framework

Forecasting. A lecture on forecasting.

Working Paper Series. Reality Checks and Comparisons of Nested Predictive Models. Todd E. Clark and Michael W. McCracken. Working Paper A

VAR-based Granger-causality Test in the Presence of Instabilities

Warwick Business School Forecasting System. Summary. Ana Galvao, Anthony Garratt and James Mitchell November, 2014

Appendix of the paper: Are interest rate options important for the assessment of interest rate risk?

Supplemental Material for KERNEL-BASED INFERENCE IN TIME-VARYING COEFFICIENT COINTEGRATING REGRESSION. September 2017

Econometrics. Week 4. Fall Institute of Economic Studies Faculty of Social Sciences Charles University in Prague

Nonstationary Time Series:

VAR Models and Applications

Vector Autoregressive Model. Vector Autoregressions II. Estimation of Vector Autoregressions II. Estimation of Vector Autoregressions I.

Modified Variance Ratio Test for Autocorrelation in the Presence of Heteroskedasticity

ECON 4160, Spring term Lecture 12

DSGE-Models. Limited Information Estimation General Method of Moments and Indirect Inference

Econometrics. Week 11. Fall Institute of Economic Studies Faculty of Social Sciences Charles University in Prague

2. An Introduction to Moving Average Models and ARMA Models

Evaluating Conditional Forecasts from Vector Autoregressions

Estimation and Testing of Forecast Rationality under Flexible Loss

Inference in VARs with Conditional Heteroskedasticity of Unknown Form

Model Mis-specification

The Generalized Cochrane-Orcutt Transformation Estimation For Spurious and Fractional Spurious Regressions

Comparing Predictive Accuracy, Twenty Years Later: On The Use and Abuse of Diebold-Mariano Tests

Working Paper SerieS. A Predictability Test for a Small Number of Nested Models. NO 1580 / august 2013

Research Division Federal Reserve Bank of St. Louis Working Paper Series

Strict and Flexible Inflation Forecast Targets: An Empirical Investigation

Forecasting the unemployment rate when the forecast loss function is asymmetric. Jing Tian

Non-Stationary Time Series, Cointegration, and Spurious Regression

Vector Auto-Regressive Models

Volume 03, Issue 6. Comparison of Panel Cointegration Tests

ECON/FIN 250: Forecasting in Finance and Economics: Section 7: Unit Roots & Dickey-Fuller Tests

Forecasting Levels of log Variables in Vector Autoregressions

News Shocks: Different Effects in Boom and Recession?

Can a subset of forecasters beat the simple average in the SPF?

FORECAST-BASED MODEL SELECTION

A better way to bootstrap pairs

An estimate of the long-run covariance matrix, Ω, is necessary to calculate asymptotic

Economic modelling and forecasting

The Comparative Performance of Alternative Out-ofsample Predictability Tests with Non-linear Models

Identification of the New Keynesian Phillips Curve: Problems and Consequences

Mixed-Models. version 30 October 2011

Week 5 Quantitative Analysis of Financial Markets Characterizing Cycles

A time series is called strictly stationary if the joint distribution of every collection (Y t

DSGE Methods. Estimation of DSGE models: GMM and Indirect Inference. Willi Mutschler, M.Sc.

The Seasonal KPSS Test When Neglecting Seasonal Dummies: A Monte Carlo analysis. Ghassen El Montasser, Talel Boufateh, Fakhri Issaoui

The Empirical Behavior of Out-of-Sample Forecast Comparisons

GDP forecast errors Satish Ranchhod

FinQuiz Notes

A near optimal test for structural breaks when forecasting under square error loss

:Effects of Data Scaling We ve already looked at the effects of data scaling on the OLS statistics, 2, and R 2. What about test statistics?

Robust Backtesting Tests for Value-at-Risk Models

Finite-sample quantiles of the Jarque-Bera test

Evaluating Forecasts from Vector Autoregressions Conditional on Policy Paths

Introduction to Forecasting and Forecast Evaluation

Regression I: Mean Squared Error and Measuring Quality of Fit

Nowcasting GDP with Real-time Datasets: An ECM-MIDAS Approach

Non-nested model selection. in unstable environments

Evaluating density forecasts: forecast combinations, model mixtures, calibration and sharpness

Department of Economics, Vanderbilt University While it is known that pseudo-out-of-sample methods are not optimal for

E 4101/5101 Lecture 9: Non-stationarity

9) Time series econometrics

A primer on Structural VARs

Prof. Dr. Roland Füss Lecture Series in Applied Econometrics Summer Term Introduction to Time Series Analysis

Discussion of Bootstrap prediction intervals for linear, nonlinear, and nonparametric autoregressions, by Li Pan and Dimitris Politis

Error Statistics for the Survey of Professional Forecasters for Industrial Production Index

Principles of forecasting

Unit-Root Tests and the Burden of Proof

Error Statistics for the Survey of Professional Forecasters for Treasury Bond Rate (10 Year)

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A.

Research Division Federal Reserve Bank of St. Louis Working Paper Series

Forecasting the term structure interest rate of government bond yields

Identifying Aggregate Liquidity Shocks with Monetary Policy Shocks: An Application using UK Data

Mixed frequency models with MA components

Transcription:

Discussion of Tests of Equal Predictive Ability with Real-Time Data by T. E. Clark and M.W. McCracken Juri Marcucci Bank of Italy 5 th ECB Workshop on Forecasting Techniques Forecast Uncertainty in Macroeconomics and Finance Frankfurt, December 1, 2007

Forecasting with real-time data may be troublesome RTD affect forecasting in many dimensions: Data inputs Estimated parameters β Model (data revision may change model specification) Which Actual should we use for evaluation? Many choices: 1 st release (2 nd, 3 rd ) 4 quarters later Last benchmark (last snapshot before benchmark revision) Latest available

Overview Important question: What are the asymptotic and finite-sample properties of tests of Equal Predictive Ability (EPA) when forecasters use Real-Time Data (RTD)? Direct Multi-step predictions: Non-Nested models Nested models Main point of the paper: Suggest how to adjust tests of EPA to make correct out-of-sample inference in a framework with data revisions. Main results: ignoring data revision can lead to incorrect inference adjusted test with reasonable size and power properties application to real-time predictive content of GDP/GAP for inflation

Some (not easy) notation If observables are subject to a revision process their statistical properties might differ from those of final Want to predict y t+τ (t ) with x i,s (t) where t = t + τ + r and t = R,..., R + P τ + 1 vintage horizon is fixed (r R) Model i: y s+τ (t ) = x i,s (t)β i + u i,s+τ (t) Forecast errors û i,t+τ (t ) = y t+τ (t ) x i,t (t) β i,t Do you consider revisions only in y t? d = (P τ + 1) T t=r û2 1,t+τ (t ) û 2 2,t+τ (t )

Non-Nested Models MSE t(s dd ) = (P τ + 1) 1/2 d bsdd However, with parameter estimation uncertainty P 1/2 d d N(0, Ω) where Ω = S dd + 2(1 π 1 ln(1 + π))(fbs dh + FBS hh BF ) and F = ( 2Eu 1,t+τ x 1,t, 2Eu 2,t+τ x 2,t ) MSE t(s dd ) is asymptotically valid only when F = 0 (few special cases when forecast error is uncorrelated with predictors, same loss for estimation and evaluation) with data revisions population residuals [y s+τ x i,s β i ] and forecast errors [y t+τ (t ) x i,t (t) β i,t ] do not have the same covariance structure. Thus if E[y s+τ x i,s β i ]x i,s = 0 we cannot say anything about E[y t+τ (t ) x i,t (t) β i,t ]x i,t(t) = 0 Must use MSE t(ω)

Non-Nested Models (cont.) π = 0 Ω = 0 F = 0 when x is unrevised and y unrevised revisions of y uncorrelated with x final revised vintage of y used for evaluation vintages of y redefined so that data releases are used both for estimation and evaluation Revision process modeled as a combination of news and noise if data revision contain news, revisions are not forecastable if data revision reduce to noise, revisions are forecastable

Non-Nested Models (cont.) DGP for Final data: Initial estimates y t = βx 1,t 1 + βx 2,t 1 + e y,t + v y,t x t = e x,t + v x,t y t (t) = y t v y,t + w y,t x t (t) = x t v x,t + w x,t Noise component w creates a non-zero correlation between real-time forecast errors and predictors

Nested models MSE t and MSE F = (P τ + 1) d MSE 2. Clark and McCracken (2005) and McCracken (2006) show that with nested models these tests have non-standard distributions (CVs are suggested and bootstrap for multi-step ahead forecasts). With data revisions and nested models the asymptotic properties of these tests change dramatically (for example in the t-test, average loss differentials d is re-scaled by P instead of P) Thus, in contrast to previous finding, can construct a t-test that is asymptotically standard normal under H 0

Nested models (cont.) Pros and cons: Both MSE t and MSE F diverge under H 0 no assumptions of correct specification of the model can conduct asymptotically valid tests using neither the bootstrap nor non-standard tables Need estimate of Ω With data revisions the asymptotic distribution of MSE t can differ from Theorem 2 and can be highly non standard with complications due to nuisance parameters. (left for further research!) Strong belief that the approximations developed in the previous papers should reasonably approximate the true asymptotic distribution.

Monte Carlo and empirical application Monte Carlo for non-nested and nested case (3 DGPs) 1-step and 4-step ahead forecasts Assume a single revision published with 1 or 4 period delay. Revisions (news and noise) and DGPs calibrated and tailored to the inflation data used in empirical section (Aruoba, 2006) All long-run variances and covariances are estimated with Newey-West (h = 2τ) R = (40, 80) and P = (20, 40, 80, 160)

Some comments The non-nested case works. Promising results! In the nested case MSE t (Ω) should be preferred, but the motivations behind are not completely clear. I think that some further work should be done. DGP for the revision process: news-to-noise ratio for DGP1 seems quite high (large Ω) DGP1 σ2 v,y σ 2 w,y DGP2 σ2 v,y σ 2 w,y DGP3 σ2 v,y σ 2 w,y =.9.2 = 4.5 while σ2 v,x σ 2 w,x =.2.2 = 1 while σ2 v,x σ 2 w,x =.2.2 = 1 while σ2 v,x σ 2 w,x =.3 2 =.15 =.3.5 =.6 =.3.5 =.6

Monte Carlo results: Non-nested models Size: MSE t(s dd ) seems always oversized for 1-step ahead especially for low π. Well sized for larger π. For 4-step ahead highly oversized except when π 1. MSE t(ω) seems correctly sized in all cases and at both forecasting horizons. The only exception: for low π slightly oversized. Power: Here at 1-step ahead the two tests have similar power properties. Surprisingly, at 4-step horizons the MSE t(s dd ) seems almost always more powerful than MSE t(ω). Here I really don t see any gain in terms of power in using Ω instead of S dd! In sum, in the non-nested case, your tests seem to work fine.

Monte Carlo results: Nested models Size: MSE F [CM]: well sized in almost all cases. Slightly oversized for small π. MSE t(ω)[n] always oversized MSE t(s dd )[N]: oversized MSE t(s dd )[CM]: oversized but not for low π Power: MSE F [CM]: good power properties especially for high π MSE t(ω)[n]: best power π MSE t(s dd ): good power only in DGP1 with noise

Some comments on Monte Carlo results Why is MSE F working so well? (technically it s invalid with data revisions) Can you try also the DM-adjusted test (Clark and West, 2007)? Can t use ENC t or ENC F? Good properties in Clark and McCracken (2005) In nested case some size distortion remains even with Ω. Only exception: DGP1 (e.g. P=80). How do you explain it? Why increasing P the size distortion does not reduce so much? Is this related to the truncation lag?

Some comments on Monte Carlo results (cont.) If predictors x t are without noise there shouldn t be a problem and you could use unadjusted tests. What about qualitative regressors? (E.g. Consumer confidence or Business confidence indicators; not revised). Power worsens from 1-step to 4-step. This is in contrast with Clark and McCracken (2005) and some results I found (not RTD). With multi-step ahead power doesn t decrease so much. I would expect that if there are revisions 4-step ahead forecasts should be less affected by the noise due to the revision process. How do you explain this intuitively? Does this depend on the lag truncation parameter that you use to estimate the LR variance?

Empirical results with RT inflation forecasting Non-Nested models: Adjusted and unadjusted tests have same size. However MSE t(s dd ) is more powerful and rejects more often the null of EPA. MSE t(ω) always fails to reject. Nested models: MSE t(ω)[n] is more powerful in Monte Carlo than MSE t(s dd )[N] but empirically you fail to reject in 3 cases with sample 1970-2003 at 1-step (t + 2,t + 5 and t + 13) while MSE t(s dd )[N] rejects. Should it not be the reverse? Similar size properties but more powerful test fails to reject when the less powerful rejects. This is strange. AR(4) with GDP growth outperforms the naive AR(2). Interestingly MSE F shows good properties. If it rejects, either or both MSE t(s dd )[N] and MSE t(ω)[n] reject. However I might be wrong with the slanted font (not easy to see).

Features of the paper Strengths: First paper to deal with asymptotic properties of tests of EPA with real-time data Well done theoretical analysis with good insights for future empirical applications Weaknesses: Full asymptotic distribution of tests not developed with nested models and data revision Use of approximations of previous work to approximate a complicated unknown distribution in nested case There is some room for some improvements (especially for the nested case). Excellent paper!