Vanishing Predictability and Non-Stationary. Regressors

Similar documents
Lecture 5: Unit Roots, Cointegration and Error Correction Models The Spurious Regression Problem

Functional Coefficient Models for Nonstationary Time Series Data

New Methods for Inference in Long-Horizon Regressions

DEPARTMENT OF ECONOMICS AND FINANCE COLLEGE OF BUSINESS AND ECONOMICS UNIVERSITY OF CANTERBURY CHRISTCHURCH, NEW ZEALAND

Discussion of Bootstrap prediction intervals for linear, nonlinear, and nonparametric autoregressions, by Li Pan and Dimitris Politis

Inflation Revisited: New Evidence from Modified Unit Root Tests

Econ 423 Lecture Notes: Additional Topics in Time Series 1

BCT Lecture 3. Lukas Vacha.

Time Series Analysis. James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY

Testing for Unit Roots with Cointegrated Data

11. Further Issues in Using OLS with TS Data

Department of Economics, UCSB UC Santa Barbara

A Non-Parametric Approach of Heteroskedasticity Robust Estimation of Vector-Autoregressive (VAR) Models

Prof. Dr. Roland Füss Lecture Series in Applied Econometrics Summer Term Introduction to Time Series Analysis

Predicting bond returns using the output gap in expansions and recessions

Econometrics of Panel Data

Predictive Regression and Robust Hypothesis Testing: Predictability Hidden by Anomalous Observations

Time Series Analysis. James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY

ECON3327: Financial Econometrics, Spring 2016

Nonsense Regressions due to Neglected Time-varying Means

Christopher Dougherty London School of Economics and Political Science

Short T Panels - Review

Lecture 8a: Spurious Regression

Financial Econometrics

Using all observations when forecasting under structural breaks

Lecture 8a: Spurious Regression

Problem set 1 - Solutions

G. S. Maddala Kajal Lahiri. WILEY A John Wiley and Sons, Ltd., Publication

Response surface models for the Elliott, Rothenberg, Stock DF-GLS unit-root test

Econ 424 Time Series Concepts

7 Introduction to Time Series

A TIME SERIES PARADOX: UNIT ROOT TESTS PERFORM POORLY WHEN DATA ARE COINTEGRATED

A Robust Approach to Estimating Production Functions: Replication of the ACF procedure

The regression model with one stochastic regressor (part II)

Empirical Market Microstructure Analysis (EMMA)

Volatility. Gerald P. Dwyer. February Clemson University

Darmstadt Discussion Papers in Economics

Nonstationary Time Series:

7 Introduction to Time Series Time Series vs. Cross-Sectional Data Detrending Time Series... 15

This chapter reviews properties of regression estimators and test statistics based on

E 4160 Autumn term Lecture 9: Deterministic trends vs integrated series; Spurious regression; Dickey-Fuller distribution and test

Response surface models for the Elliott, Rothenberg, Stock DF-GLS unit-root test

Økonomisk Kandidateksamen 2004 (I) Econometrics 2. Rettevejledning

LECTURE 10: MORE ON RANDOM PROCESSES

Covers Chapter 10-12, some of 16, some of 18 in Wooldridge. Regression Analysis with Time Series Data

Predictive Regressions: A Reduced-Bias. Estimation Method

CHAPTER 21: TIME SERIES ECONOMETRICS: SOME BASIC CONCEPTS

A NOTE ON THE TIME SERIES MEASURE OF CONSERVATISM. Sebastian Brauer and Frank Westermann. Working Paper 81 February 2010

9) Time series econometrics

A nonparametric test for seasonal unit roots

Predictive Regressions: A Reduced-Bias. Estimation Method

Finite Sample Performance of A Minimum Distance Estimator Under Weak Instruments

Steven Cook University of Wales Swansea. Abstract

Lecture 2: Univariate Time Series

Understanding Regressions with Observations Collected at High Frequency over Long Span

A New Solution to Spurious Regressions *

Cointegration and Tests of Purchasing Parity Anthony Mac Guinness- Senior Sophister

On Perron s Unit Root Tests in the Presence. of an Innovation Variance Break

Predictive Systems: Living with Imperfect Predictors

Missing dependent variables in panel data models

Testing for non-stationarity

A Simple Nonlinear Predictive Model for Stock Returns

Econometrics. Week 4. Fall Institute of Economic Studies Faculty of Social Sciences Charles University in Prague

FinQuiz Notes

E 4101/5101 Lecture 9: Non-stationarity

Output correlation and EMU: evidence from European countries

Regression with time series

Lecture 7: Dynamic panel models 2

Online Appendix: Uncertainty and Economic Activity: Evidence from Business Survey Data" Rüdiger Bachmann, Steffen Elstner, and Eric R.

Modified Variance Ratio Test for Autocorrelation in the Presence of Heteroskedasticity

Time Series Methods. Sanjaya Desilva

Predictive Systems: Living With Imperfect Predictors. December

The Slow Convergence of OLS Estimators of α, β and Portfolio. β and Portfolio Weights under Long Memory Stochastic Volatility

Vector autoregressions, VAR

7. Integrated Processes

MA Advanced Econometrics: Applying Least Squares to Time Series

Unit Roots and Structural Breaks in Panels: Does the Model Specification Matter?

Advanced Econometrics

Time series: Cointegration

University of Pretoria Department of Economics Working Paper Series

Applied Microeconometrics (L5): Panel Data-Basics

Predictive Regressions with Imperfect Predictors

TAKEHOME FINAL EXAM e iω e 2iω e iω e 2iω

1 Estimation of Persistent Dynamic Panel Data. Motivation

Predictive Systems: Living with Imperfect Predictors

This note introduces some key concepts in time series econometrics. First, we

Lecture 6: Univariate Volatility Modelling: ARCH and GARCH Models

7. Integrated Processes

Efficiency Tradeoffs in Estimating the Linear Trend Plus Noise Model. Abstract

Non-Stationary Time Series and Unit Root Testing

Title. Description. Quick start. Menu. stata.com. xtcointtest Panel-data cointegration tests

Robust Bond Risk Premia

Comparing Nested Predictive Regression Models with Persistent Predictors

Bias-Correction in Vector Autoregressive Models: A Simulation Study

13. Time Series Analysis: Asymptotics Weakly Dependent and Random Walk Process. Strict Exogeneity

An Econometric Modeling for India s Imports and exports during

Introduction to Modern Time Series Analysis

Chapter 2: Unit Roots

1 Regression with Time Series Variables

Hypothesis Testing in Predictive Regressions

Transcription:

Vanishing Predictability and Non-Stationary Regressors Tamás Kiss June 30, 2017 For helpful suggestions I thank Ádám Faragó, Erik Hjalmarsson, Ron Kaniel, Riccardo Sabbatucci (discussant), arcin Zamojski, and seminar participants at the KWC-CFF workshop, Linköping University, Swedish House of Finance, University of Gothenburg. Centre for Finance, Department of Economics, University of Gothenburg; e-mail: tamas.kiss@cff.gu.se 1

Abstract This paper provides a potential explanation for why predictive regressions may have lost power in recent samples. In a noisy predictor framework, where expected returns are stationary and a non-stationary component masks the information in the regressor, the predictive power of the regression is shown to vanish as the sample size increases. An estimation method, subsample fixed effects, is introduced, for which the problem of decreasing predictive power is mitigated. Empirically, important predictors of the stock returns exhibit vanishing predictability but applying subsample fixed effects indicates that the underlying predictive relationship between these predictors and returns remain significant. Keywords: Return Predictability, Non-stationary Regressor, Subsample Fixed Effects JEL classification: C22, C58, G17 Surveying the recent literature on return predictability, Koijen and Van Nieuwerburgh (2011) report two disconcerting statistical features of in-sample forecasting. First, the high persistence of the predictors makes standard inference incorrect. Second, the relationships exhibit significant instability over time. In this paper, I propose that these two phenomena are interconnected. Specifically, I show that if a forecasting variable is masked by highly persistent noise, while returns are assumed to be stationary, the predictive relationship is bound to lack power in large samples. Since several prominent predictors of excess stock market returns have serial correlation close to unity, the literature on inference and estimation based on such variables is abundant. In particular, the persistent regressor bias has been extensively discussed (Goetzmann and Jorion (1993); Nelson and Kim (1993); Stambaugh (1999); Lewellen (2004); Ang and Bekaert (2007); Cochrane (2008); Boudoukh et al. (2008)). This bias is also apparent in the earlier work on cointegrating predictive regressions with endogenous regressors (Cavanagh et al. (1995); Valkanov (2003); Torous et al. (2004); Campbell and Yogo (2006); Ang and Bekaert (2007)). In the majority of these papers the presence of predictability implies that expected returns and the predictor have common time series 2

characteristics. In particular, if the predictor has high serial correlation, then expected returns must share this property. 1 To reconcile this feature with the stylized fact that realized returns are nearly serially uncorrelated, the usual approach is to assume that the persistent expected return component is small, and the unexpected returns dominate (c.f. oon et al., 2005; oon and Velasco, 2014). In contrast, I assume that expected returns are stationary even if they are potentially predictable by a non-stationary variable. Analogously to the work on noisy predictors (Torous et al., 2004; Pástor and Stambaugh, 2009), the information in the predictor, captured by a stationary process, is confounded by an uninformative component, which is here assumed to be non-stationary. If the variability in the informative part of the predictor is large enough, then it is possible to detect the predictive relationship in small samples. However, as the number of observations increases, the non-stationary component inevitably becomes dominant in the regressor, and therefore the estimated slope coefficient of the predictive regression converges to zero. This implies that the power of the predictive regression decreases over time, even if returns are indeed predictable. A similar model is considered in Osterrieder et al. (2015) who discuss the properties of the predictive equation with fractionally integrated regressors. However, their focus is to develop an instrumental variable method tailored for their specific set-up, and the method does not cover the cases with unit root predictors. Parameter instability in predictive regressions is often evident in two related forms. First, the evidence of predictability is usually stronger in sub-samples than in centurylong datasets (Ang and Bekaert, 2007; Lettau and Van Nieuwerburgh, 2008; Koijen and Van Nieuwerburgh, 2011). Second, predictors appear to lose power, as the evidence of predictability weakens over time (Ferson et al., 2003; Goyal and Welch, 2008; Deng, 2013). In the context of the current paper these forms of parameter instability are not surprising. In fact, based on the model proposed in the paper, if highly persistent (non-stationary) predictors are used, then (i) predictability is stronger in smaller samples, and (ii) the loss in predictive power is due to the increase in the sample size. Asymptotic results for the Ordinary Least Squares (OLS) estimator are derived to 3

show that predictability disappears under a general set of assumptions on the dynamics of the time-varying expected returns. Finite sample properties of the model based on onte Carlo simulations indicate that the high persistence of the regressor results in a substantially biased slope coefficient for sample sizes relevant for predictive regressions (approximately 100 observations for annual data, and 800 for monthly data). The bias is especially large if the persistence of the regressor is so high that the variable cannot be distinguished from a non-stationary process. To mitigate the bias caused by the decreasing predictive power, a simple and flexible estimation framework, subsample fixed effects (SFE), is proposed. It builds on the idea that the bias increases with the sample size because the non-stationary component becomes dominant in larger samples. The problem can therefore be reduced by dividing up the entire sample and pooling the information from different subsamples via a fixed effects estimator. By limiting the subsample size, one effectively puts a bound on the variance accumulation within the regressor. Therefore, the extent of the bias in the estimator is reduced and the estimated slope coefficient no longer vanishes asymptotically. The exact relationship between the bias and the subsample size is derived under the assumption of independent, identically distributed innovations and error terms. Simulations show that the proposed subsampling estimator substantially reduces the bias caused by the non-stationary component also under a more general set of assumptions. Since the non-stationary noise component is more dominant in larger samples, the bias in the subsampling estimator is positively related to the size of a given subsample. Therefore choosing a smaller subsample is more favourable for bias reduction. However, including more fixed effects results in a loss of estimation precision. This translates into an efficiency-bias trade-off for the choice of subsample size in the SFE estimator. Simulation results suggest that the optimal choice of subsample size depends on the (potentially unobservable) parameters of the data generating process. Therefore robustness can be empirically achieved by considering several subsample sizes simultaneously. To test the proposed model empirically, I investigate predictors of the excess returns on the S&P 500 stock market index. I focus on highly persistent variables, since in their 4

case the model with a non-stationary noise is a potentially good approximation. Looking at how regression estimates change over time, predictors of the excess returns (including the dividend-price ratio, the treasury bill rate and the book-to-market value) appear to exhibit vanishing predictability. Their slope coefficients approach zero as time progresses and thus the sample size grows. Applying subsample fixed effects shows an overall increase in the significance of these predictors. The estimated slope coefficients are the smallest in magnitude in case of OLS (no subsampling), and they grow as one introduces subsampling and moves towards smaller subsamples. All these empirical observations support the predictions of the proposed model. The rest of the paper is organised as follows. Section 1 describes the modelling framework and presents the main theoretical results, along with the proposed estimator, subsample fixed effects. Section 2 presents onte Carlo simulations analysing the performance of the model. Empirical results based on several important predictors of excess returns are collected in Section 3, and Section 4 concludes. Technical derivations are presented in the Appendix. 1 The model I define and characterize a model with stationary and possibly predictable returns, but where the information in the predictive variable is hidden by a non-stationary component. Furthermore, I show that the predictive power of the regressor decreases as the sample size increases and I also describe how it is related to the presence of a non-stationary component. A robust solution to this problem, subsample fixed effects, is also proposed, and it is shown to reduce the asymptotic bias in this setting. 1.1 Econometric background for vanishing predictability Consider returns that are stationary and potentially predictable. However, the explanatory variable has two components: the informative component (which is also stationary) and a unit root (non-stationary) component unrelated to the dependent variable. 5

The data generating process can then be written in the following form: y t = α 0 + β 0 η t 1 + u t (1) x t = η t + ξ t (2) ξ t = ξ t 1 + ε t (3) The stationary parts of the dependent variable, η t and u t, and the innovation of the unit root process, ε t, are assumed to be linear processes with zero unconditional mean. Specifically, define w t = (η t, u t, ε t ). Then w t = j=0 C jζ t j, where {C j } j=1 is a sequence of matrices and ζ t j is a martingale difference sequence with E(ζ t ) = 0, E(ζ t ζ t) = Σ ζ R 3 3 +. Furthermore, E(w t w t) = Σ < with diag(σ) = (σ 2 η, σ 2 u, σ 2 ε). The data generating process is fairly standard for the return predictability literature with noisy predictors (Pástor and Stambaugh, 2009). The non-standard element is the fact, that the noise component (the component that is not related to the predictive signal) is a unit-root process. To put the model assumptions in a more specific context, consider the setting of Lettau and Van Nieuwerburgh (2008), where changes in the long term mean of the dividend-price ratio cause the instability of the predictive relationship. The authors argue that these jumps in the mean reflect long term changes in the structure of the economy, going from one steady-state to another. They also relate these structural changes to the persistence of the dividend-price ratio. Infrequent but persistent shifts of the long term mean in an otherwise stationary process can generate unit-root type behaviour (Engle and Smith, 1999). Therefore one interpretation of the present model is that it is an alternative way of capturing the long term structural changes in the economy. The key parameter of interest when assessing predictability is the slope coefficient in equation (1), β 0. In most empirical work OLS estimator is applied to calculate the parameter estimates of the fitted regression, y t = ˆα + ˆβx t 1 + e t. (4) 6

Under the assumptions described by (1) (3), the predictive power of x t is masked by the non-stationary component. Since in this case the left hand side of equation (4) is stationary, while the right hand side is unit root non-stationary, the slope coefficient and the predictive power disappear in the limit (converges to zero) by construction. This result is stated in the following proposition: Proposition 1. Let the data generating process be described by equation (1) (3). Furthermore let w t = (η t, u t, ε t ) = j=0 C jζ t j, where {C j } j=1 is a sequence of matrices and ζ t j is a martingale difference sequence with E(ζ t ) = 0, E(ζ t ζ t) = Σ ζ R 3 3 +, and E(w t w t) = Σ <. Then ˆβ OLS p 0, where ˆβOLS is the OLS estimate of the slope coefficient of the fitted regression (4). If α 0 = 0 is also imposed, then ˆα OLS p 0, otherwise it converges to a random variable. Proof. In the Appendix. Intuitively the result is straightforward. If a stationary variable is regressed on a nonstationary predictor, then as the sample size grows, the variation in x t becomes arbitrarily large, and the only way to reconcile it with the finite variation in the stationary dependent variable is that the slope coefficient converges to zero. Since no specific assumptions are made about the autocovariance-structure of the y t and η t process, this argument is widely applicable. From a practical point of view, the advantage of this generality is that it allows for specifying η t as a weakly dependent series, for example a stationary ARA process. Proposition 1 also covers the case of endogenous regressors, as the error term of the dependent variable and either component of the regressor are allowed to be correlated. If predictors are tested for a unit root in a given sample, test statistics often suggest a narrow rejection of non-stationarity. This can happen if the unit root component is small relative to the informative part. In this case, statistical testing concludes that there is no unit root in the series. Therefore, running the predictive regression (4) seems to be appropriate and it may produce a good estimate for the slope coefficient and the predictive 7

power (c.f. oon et al. (2005)). However, Proposition 1 suggests that increasing the sample size leads to more biased point estimates, converging to zero eventually. This is in sharp contrast with the general notion that a growing sample size results in a better estimate of the slope coefficient. 1.2 Local demeaning and subsample fixed effects Removing the local mean from the regressor can mitigate the effect of the nonstationary component. The intuition comes from the observation that a demeaned random walk does not depend on its initial value. Therefore, if time series are split into subperiods, and they are demeaned in each of these subsamples, they become independent of their initial values in each subsample. Restarting a process in such a way limits the effects of non-stationarity. The process can only accumulate the variance within a given subsample. To formalize the idea, fix a subsample size, denoted by N, and assume that the entire sample size can be written as T = K, where K N. The processes x t and y t can then be written (with a slight abuse of notation in the time indexing) as sequences of processes {x k,t } K k=1 and {y k,t} K k=1, where x k,t = x (k 1)+t and y k,t = y (k 1)+t. Define the locally demeaned regressor as x k,t = x k,t 1 x k,m (5) for all k and t. If {x k,t } K k=1 is generated according to equations (2)-(3), where independence and identical distribution is imposed on the error terms, one can calculate the variance of the demeaned regressor. Lemma 1. Let x t be generated by equations (2)-(3), and ε t and η t be independent, identically distributed random variables with variances σ 2 ε, σ 2 η <, respectively. Then V ar ( x k,t ) = A()σ 2 η + B(, t)σ 2 ε, where the expressions for A() and B(, t) are given in the proof. Furthermore, A() = 8

O(1), and B(, t) = O(), therefore V ar ( x k,t ) = O(). Proof. In the Appendix. The key observation of Lemma 1 is that the variance of the demeaned regressor grows linearly in the subsample size. Based on this result, it is possible to characterise the properties of the least squares estimator using the locally demeaned explanatory variable. To obtain the exact asymptotic results the regression error u t is assumed to be independent of the other error variables. Then the asymptotic bias using the locally demeaned regressor is characterized by the following proposition. Proposition 2. Let the data generating process described by equations (1-3) with {ε t, η t, u t } t=0 i.i.d. sequences with unconditional variances σ 2 ε, σ 2 η, σ 2 u <, respectively. oreover fix the subsample size N such that = T, where T is the total number of observations K and K is the number of subsamples. Define ˆα SF E and ˆβ SF E as the OLS estimate of the regression y t = ˆα + ˆβ x t 1 + e t, (6) where x t is the locally demeaned regressor given by equation (5) using x (k 1)+t = x k,t. Then as T (and therefore K as is fixed) ˆβ SF E p β0 σ 2 η σ 2 η + +1 6 σ 2 ε ˆα SF E p E(yt ) = α 0 Proof. In the Appendix. SF E stands for subsample fixed effects in the above result. The notation comes from the observation that the estimator can be computed in a simple way. One simply needs to include a fixed effect for each subsample, and use OLS estimation. 2 This formulation is especially useful since fixed effects estimation is straightforward to carry out. Therefore, in the following discussion I use the terms subsample fixed effects and local demeaning estimator interchangeably, referring primarily to ˆβ SF E defined in Proposition 2. 9

The result for the slope coefficient is similar to the classical measurement error attenuation bias formula, as the bias enters the estimate as a multiplicative factor. However, in case of the subsampling fixed effects estimator the extent of the bias depends on the number of observations in each subsample,. Since their relationship is positive, a larger implies more biased estimation. 3 On the other hand, the variance of the estimator within a single subsample is decreasing as gets larger, since the number of observations within a given subsample gets larger. Therefore choosing the number of subsamples involves a bias-variance trade-off. In order to obtain the asymptotic results in Proposition 2, fairly strong assumptions need to be imposed. In particular, only strictly exogenous regressors are allowed in Proposition 2. Arguably, this restriction is the most relevant one, since endogeneity has been a prevalent concern in the return predictability literature. Although no asymptotic results are derived, I show in a simulation exercise in the following section, that the method appears robust also with endogenous regressors. Specifically, the endogeneity bias enters the estimates additively, but it does not affect the mechanism of vanishing predictability, i.e. shrinking point estimates as the sample size increases. 2 Simulations The results of Proposition 1 and 2 show the asymptotic bias of the OLS estimator in the model described by equations (1)-(4), and how subsample fixed effects mitigate the problem. This section carries out a onte Carlo experiment to complete the analysis with finite sample properties of the model and the proposed estimator. Simulations suggest that the presence of a non-stationary component in the regressor substantially biases OLS estimates even for moderate sample sizes. This bias is limited if subsample fixed effects are used. However, results of the subsampling estimator are sensitive to the choice of subsample size. The analysis involves drawing a sample of {x t, y t } T based on equations (1)-(3), and then estimating (4) by OLS, obtaining the estimated slope coefficient ˆβ. The common assumptions in these specifications is that the information part of the predictor, η t 1, 10

is independent and identically distributed, and σ η = σ u = 1 for normalization. The underlying slope coefficients of equation (1) is set to β 0 = 0.2 and no intercept is used. These parameter values are chosen to align with the empirical application in the following section, where standardized variables are used. For the remaining parameters of the model a set of difference scenarios are used. In particular, three values of the signal-to-noise ratio λ = ση σ ε are considered, λ = {1, 3, 10}, representing different levels of persistence caused by the non-stationary component. These values are chosen so that the autocorrelation of the regressors in the simulated samples is high, medium and low, respectively. Although it is difficult to formalize the exact relationship between λ and the sample autocorrelation for a given sample size, the simulation results indicate that the autocorrelations corresponding to these choices of the signal-tonoise ratio λ are approximately 0.99, 0.95 and 0.7. For the correlation between u t and ε t, two values, ρ u,ε = 0 and ρ u,ε = 0.8, are used in the simulations. They represent whether the predictor is strictly exogenous, or there is a negative correlation between the error terms that causes endogeneity. Results using 1000 repetitions are presented in Table 1 for sample sizes (T = {100, 300, 800}). These sample sizes are realistic in terms of representing the number of available observations in yearly, quarterly or monthly datasets generally used to assess return predictability. [Insert Table 1 here] The first three columns in Table 1 represent the case of an exogenous regressor, ρ u,ε = 0. The first observation to make is that the extent of bias for different sample sizes strongly depends on the persistence of the predictor. If the informative component is small (λ = 1), then the sample autocorrelation is almost completely driven by the non-stationary component (the sample autocorrelation is around 0.99, basically not distinguishable from the unit root specification), and there is a serious bias already for a small sample (T = 100). One can also observe that the higher the signal-to-noise ratio λ is, which corresponds to a less strong non-stationary component and thus a lower persistence, the less biased the estimation becomes. The convergence of the slope coefficient towards zero is apparent in each case, and it happens quickly. The coefficients are substantially closer to zero for 11

a sample size of T = 800. Columns (5)-(7) in Table 1 show results based on simulations where the strict exogeneity assumption of the regressor is violated. ρ u,ε = 0.8 means a strong negative relationship between the innovations u t and ε t. As argued by Stambaugh (1999), this correlation creates a positive bias in the estimation, which one can observe in columns (5)-(7). This results in an over-rejection of the null hypothesis of no predictability in the absence of non-stationarity. However, if the autocorrelation of the regressor is substantial, the parameter estimates are considerably biased towards zero even though the estimated slope coefficients are larger in absolute value. In fact, the endogeneity bias enters the estimates nearly additively and its extent is essentially not affected by the sample size. Thus, as the sample size increases, the effect of the non-stationary noise component becomes stronger than the endogeneity bias, and therefore the slope coefficients eventually converge to zero (the vanishing predictability phenomenon dominates). The findings of the simulations have important practical implications for predictive regressions. In particular, even if statistically significant predictive power is found in a given sample, it is not certain that the estimation precision of the relationship improves by using more observations. In fact, the contrary holds in the present case. The larger the sample, the more biased the least squares estimation becomes. To handle this problem the subsample fixed effects estimator is proposed, which estimates the coefficients using a locally demeaned regressor. Even though its asymptotic bias is smaller than that of the standard least squares estimator, it does not completely eliminate the problem caused by the persistence of the regressor. Therefore it is important to understand its finite sample characteristics. The simulations use the same data generating process as above and apply the SFE estimator with subsample sizes = {10, 50, 100}. Given and the sample size T ( ), the subsampling partition {T 1, T 2,..., T K } is uniquely determined. Panel I in Table 2 presents the point estimates of the coefficients. Corresponding standard errors (presented in Panel II) are calculated as the standard deviation of the simulated empirical distribution of the slope coefficients based on 1000 repetitions. 12

[Insert Table 2 here] As seen in Table 2, the SFE is robust to the size of the entire sample. The point estimates for T = 100 and T = 800 are almost identical. This is unsurprising given the theoretical results. Since fixed effects are included in each subsample (i.e. the local mean is removed from the regressor), the subsamples become independent of each other. Therefore, the accumulation of the variance of the non-stationary component is constrained to a given subsample and thus the overall bias is similar to the bias that appears in one subsample. Figure 1 provides an illustration of this fact by plotting the SFE estimates as a function of the sample size T, for a fixed subsample size = 50. [Insert Figure 1 here] Proposition 2 implies that the subsample size plays a key role in determining the bias in the subsample fixed effects estimator. This is also confirmed by the simulation results, showing that the subsample size does have a significant effect on the estimation results even in finite samples. Comparing results with = 10 to = 100 in Table 2, it is clear that the smaller subsample size results in a less biased, but more imprecise estimation. The point estimates are closer to the true value if the subsample size is small (Panel I), but their standard errors are larger (Panel II). This is independent of the fact whether or not there is correlation between the unexpected return and the innovation of the nonstationary component. In particular, the endogeneity bias enters as an additional factor to the point estimates. Both the theoretical and simulation results point to the importance of the efficiencybias trade-off that is present when one chooses the subsample size. In order to see how the bias and the variance of the estimator interact, I calculate the mean squared error (SE) of the subsampling estimator. The calculation is performed for the same subsample sizes as in Table 2, and it is applied to the data generating processes described in Table 1. [Insert Table 3 here] Table 3 shows that as the total number of observations increases, the SE tends to decrease, since the estimation becomes more precise. The only exception is if the non- 13

stationary component is large (λ = 1) and the β 0 coefficient is substantial, since in this case the bias dominates the estimation. The effect of the subsample size is ambiguous. Depending on the signal-to-noise ratio, choosing a larger subsample can cause either an increase or a reduction of the error. From a practical point of view the main implication of Table 3 is that the choice of optimal subsample size depends on the data generating process. Since a priori we do not have information about all the unobservable components of the model, a useful and transparent approach in empirical applications is to use several subsample sizes and describe results in terms of a set of estimates and the theoretical results. 3 Empirical results The aim of the section is to carry out an empirical analysis to illustrate the vanishing predictability phenomenon. In addition, the empirical performance of subsample fixed effects estimator is assessed. I use the monthly dataset compiled by Goyal and Welch (2008), who perform a comprehensive analysis on the predictors of the excess return on the S&P 500 stock market index. 4 The time window of the analysis is between January 1952 and December 2015. 5 Since the main focus is on prediction with highly persistent variables, I consider variables which have high serial correlation (their estimated first order autocorrelation is above 0.95). The analysis is further restricted to those variables that have monthly observations available for the full sample period. 6 As a result, there are seven time series whose predictive capacity is assessed, including the dividend-price ratio (dp), the earnings-price ratio (ep), the dividend payout ratio (de), the book-to-market value (bm), the three-month treasury bill rate (tbl), the term-spread on government bonds (tms), and the default yield spread (dfy). The summary statistics of the predictors and the excess returns are presented in Table 4. [Insert Table 4 here] Since the persistence of the explanatory variables is key to the analysis, the OLS 14

estimate of the largest autoregressive root and the p-value of an Augmented Dickey- Fuller test are shown in Table 4. The autoregressive roots are close to unity, and the existence of a unit-root cannot be rejected (except for the term spread that represents a borderline case: the null of a unit root is rejected at a five percent significance level, but not at the one percent level). The high persistence makes these variables good candidates to fulfil the assumptions of the model, namely that the stationary informative component is masked by a non-stationary noise. The theoretical results then imply that standard OLS becomes increasingly biased as the sample size grows and using subsampling fixed effects causes a reduction in the bias for these variables. 3.1 Vanishing predictability in the data Although the results of Proposition 1 are asymptotic, the implied bias is also substantial in finite samples. Empirically this is testable by looking at the changes in the estimated slope coefficient for different sample sizes. To obtain a set of slope coefficients that correspond to different sample sizes I perform an extending window analysis. First, I standardize the variables to get comparable results across different predictors. Then I consider the first 100 observations of the dataset and estimate equation (4) with Ordinary Least Squares. Then the sample is extended by adding one more observation, and equation (4) is re-estimated. This procedure is iterated until no new data points are available. I carry out this analysis for each of the predictors discussed above. According to the theoretical results, the variation in the non-stationary noise increases with a growing sample size, which deteriorates the signal to noise ratio. Therefore it becomes increasingly hard to detect the predictive relationship in larger samples. This in turn implies that the sequence of parameter estimates obtained by the extending window should approach zero under the assumptions of the model. Even though the overall evidence for predictability is weak with barely significant coefficients in most cases, one can observe the gradually decreasing predictability. This is true in particular in case of the dividend-price ratio, the book-to-market value and the interest related variables, the treasury-bill rate and the term spread. The sequences of the estimated slope coeffi- 15

cients for these variables are presented in Figure 2, where the corresponding 95 percent confidence intervals based on OLS standard errors are also shown. [Insert Figure 2 here] In contrast, the earnings-price ratio, the dividend payout ratio and the default yield spread appear to be essentially non-predictors. Their slope coefficients cannot be statistically distinguished from zero at any sample size, even with the relatively small OLS standard errors. It suggests that the information component in these variables is negligible, and does in effect not predict excess returns. This implies that even if based on their persistence these variables could be subject to vanishing predictability (decreasing predictive power on larger samples), it does not materialize. In particular, estimates of the slope coefficients are statistically indistinguishable from zero for all sample sizes. 7 The empirical findings so far are conditional on the selected starting date and the specific sample. Proposition 1 does not assume or require a specific initial condition. In fact, results are independent on the starting date and value of the processes, which means that vanishing predictability does not depend on the starting date of the sample according to the model. This can be tested by using subsets of observations. First, the time series between January 1952 and December 1994 are considered, omitting the last twenty years of observation. The sequences of slope coefficients are obtained by the extending window analysis described above (this gives the same sequences as before truncated in 1994). Then the starting and ending points of the sample are shifted ten years forward in time and the same exercise is carried out, giving a new set of sequences of slope coefficients for the shifted sample. This procedure is repeated two times in total, resulting in three sets of results presented in Figure 3. Although the coefficient series vary significantly over time, the tendency of decreasing predictive power also prevails in the shifted samples. This overall confirms the intuition obtained from the baseline results: non-stationary predictors that potentially have predictive power for the excess returns tend to lose power over time. [Insert Figure 3 here] 16

3.2 Applying subsample fixed effects Since the predictive power decreases with the sample size for highly persistent predictors, the subsample fixed effects estimator is applicable. To perform the analysis I estimated the slope coefficients of the predictive regression using subsampling fixed effects based on equation (5)-(6). The variables considered are those that are subject to vanishing predictability: the dividend-price ratio, the book-to-market value, the treasury bill rate, and the term spread. Since results are sensitive to the choice of subsample size, the models are estimated with subsample sizes 25, 50, 100 and 200 (resulting in 30, 15, 7 and 3 subsamples). Coefficient estimates and standard errors are reported in Table 5 (OLS results are presented as a benchmark in the last column). [Insert Table 5 here] The findings in Table 5 are in line with the theoretical results. First, while the OLS results on the entire sample are barely significant, the subsampling results are remarkably stronger. In particular, most of the subsampling coefficients are significant at a one percent level. Second, the point estimates become larger in absolute value as the subsample size decreases. This is completely in line with the theoretical prediction of Proposition 2, which suggests that for smaller subsamples the bias caused by the non-stationary component decreases. This, in turn, makes the relationship between the informative component of the predictor and the excess return easier to reveal. The results therefore suggest that the underlying predictive power of the variables is stronger than what the results based on standard least squares estimation show. Although the theoretical results are based on subsamples of equal size, subsample fixed effects can also be applied on subsamples of varying size. In particular, one can estimate structural breaks in the time series of the explanatory variable and define subsamples as observations between two breaks. This is analogous to the approach Lettau and Van Nieuwerburgh (2008) follow when they look at the dividend- price ratio, and identify structural breaks following the method by (Bai and Perron, 1998). 8 To see how results change if one uses estimated cut-off points, first I estimate the model by (Bai and Perron, 1998) to identify breaks between subsamples. The number 17

of breaks in each series is specified in advance, and I let the method to determine their location. This facilitates the comparison with the subsample fixed effects estimator, since by specifying the number of cut-off points, the average subsample size is also defined. Figure 4 illustrates the two different subsampling methods using the dividend-price ratio and eight subsamples. Panel (a) presents the original series and the cut-off points for both the equal size subsamples (solid black lines) and the estimated breaks (dashed red lines) and Panel (b) shows the break-adjusted dividend-price ratio series, using both subsampling approaches. Despite the differences in the cut-off points in Panel (a), the adjusted series in Panel (b) look fairly similar. [Insert Figure 4 here] Next, I estimate the predictive regression (6), using the break-adjusted explanatory variables based on the estimated cut-off values. Results based on this approach are presented for the same set of regressors in Table 6. They are qualitatively similar to the ones using the subsample fixed effects approach. In particular, the results are generally strongly significant, and the coefficients seem to be even further away from zero for a given (average) subsample size in the estimated cut-off case. Overall, the evidence based on estimated cut-off is in line with the theoretical predictions of the model, suggesting that estimation using subsamples can provide stronger evidence for the presence of predictability. [Insert Table 6 here] 4 Conclusion Several predictors of stock market returns (such as financial ratios or interest-related variables) exhibit highly persistent behaviour. This is in contrast to excess returns, which are usually found to be weakly dependent, almost white noise processes. The traditional way to address this difference is to assume persistent expected returns under the presence of predictability. In contrast to this approach, I reconcile potential non-stationarity in the explanatory variables and a stationary expected return by assuming a noisy predictor. 18

The main result is that if stationary returns are regressed on the lagged values of a nonstationary explanatory variable, the slope coefficient and the predictive power approach zero as the sample size increases. This observation is in line with the empirical evidence on weakening predictive power of several regressors presented in Section 3. The key result of the model holds for a general set of assumptions about the innovations in the regressor and about the unexpected returns. Using onte Carlo simulations I also show that the convergence of the slope coefficients towards zero happens quickly, therefore estimates are biased even for moderate sample sizes. The proposed subsample fixed effects estimator puts a bound on the variance of the non-stationary component, and therefore it reduces the bias caused by the high persistence of the explanatory variable. Exact theoretical results about the extent of the bias for a fixed subsample size are derived for a restrictive set of assumptions. Simulations show that the estimator also works well for more general assumptions. Applying this estimator to the non-stationary predictors of the returns, the point estimates improve and tend to become significant. 19

References Ang, A. and Bekaert, G. (2007). Stock return predictability: Is it there? Review of Financial Studies, 20(3):651 707. Bai, J. and Perron, P. (1998). Estimating and testing linear models with multiple structural changes. Econometrica, pages 47 78. Billingsley, P. (1968). Convergence of probability easures. Wiley. Boudoukh, J., Richardson,., and Whitelaw, R. F. (2008). The myth of long-horizon predictability. Review of Financial Studies, 21(4):1577 1605. Campbell, J. Y. and Yogo,. (2006). Efficient tests of stock return predictability. Journal of Financial Economics, 81(1):27 60. Cavanagh, C. L., Elliott, G., and Stock, J. H. (1995). Inference in models with nearly integrated regressors. Econometric theory, 11(05):1131 1147. Cochrane, J. H. (2008). The dog that did not bark: A defense of return predictability. Review of Financial Studies, 21(4):1533 1575. Deng, A. (2013). Understanding Spurious Regression in Financial Economics. Journal of Financial Econometrics, 12(1):122 150. Engle, R. F. and Smith, A. D. (1999). Stochastic Permanent Breaks. The Review of Economics and Statistics, 81(4):553 574. Ferson, W. E., Sarkissian, S., and Simin, T. T. (2003). Spurious Regressions in Financial Economics? Journal of Finance, 58(4):1393 1414. Goetzmann, W. N. and Jorion, P. (1993). Testing the predictive power of dividend yields. The Journal of Finance, 48(2):663 679. Goyal, A. and Welch, I. (2008). A comprehensive look at the empirical performance of equity premium prediction. Review of Financial Studies, 21(4):1455 1508. 20

Koijen, R. S. and Van Nieuwerburgh, S. (2011). Predictability of returns and cash flows. Annual Review of Financial Economics, 3:467 491. Lettau,. and Van Nieuwerburgh, S. (2008). Reconciling the Return Predictability Evidence. Review of Financial Studies, 21(4):1607 1652. Lewellen, J. (2004). Predicting returns with financial ratios. Journal of Financial Economics, 74(2):209 235. aynard, A. and Shimotsu, K. (2009). Covariance-Based Orthogonality Tests For Regressors With Unknown Persistence. Econometric Theory, 25(01):63 116. oon, R., Rubia, A., and Valkanov, R. (2005). Long horizon regressions when the predictor is slowly varying. Working Paper. oon, S. and Velasco, C. (2014). On the properties of regression tests of stock return predictability using dividend-price ratios. Journal of Financial Econometrics, 12(1):151 173. Nelson, C. R. and Kim,. J. (1993). Predictable stock returns: The role of small sample bias. The Journal of Finance, 48(2):641 661. Osterrieder, D., Ventosa-Santaulria, D., and Vera-Valds, J. E. (2015). Unbalanced Regressions and the Predictive Equation. CREATES Research Papers 2015-09, Department of Economics and Business Economics, Aarhus University. Pástor, L. and Stambaugh, R. F. (2009). Predictive systems: Living with imperfect predictors. The Journal of Finance, 64(4):1583 1628. Phillips, P. C. B. (1986). Understanding spurious regressions in econometrics. Journal of Econometrics, 33(3):311 340. Phillips, P. C. B. (1987). Time Series Regression with a Unit Root. Econometrica, 55(2):277 301. 21

Stambaugh, R. F. (1999). Predictive regressions. Journal of Financial Economics, 54(3):375 421. Torous, W., Valkanov, R., and Yan, S. (2004). On predicting stock returns with nearly integrated explanatory variables. The Journal of Business, 77(4):937 966. Valkanov, R. (2003). Long-horizon regressions: theoretical results and applications. Journal of Financial Economics, 68(2):201 232. 22

Notes 1 An interesting deviation from this approach is by aynard and Shimotsu (2009), who consider a setting where returns are stationary and the innovations of the non-stationary predictor have predictive power. They develop a test based on a generalized covariance concept between a stationary and a unit root variable, and conclude that the evidence on predictability based on standard inference may be understated. 2 To see the equivalence note that if {y k,t } K k=1 is generated according to equation (5), then K k=1 ỹk,t x k,t 1 K k=1 x2 k,t 1 K k=1 = y k,t x k,t 1 K k=1 x2 k,t 1 K k=1 ỹ k,t x k,t 1 K k=1 x2 k,t 1 K k=1 = y k,t x k,t 1 K, k=1 x2 k,t 1 where the last equality follows from the fact that x k,t 1 = 0. 3, implies ˆβ SF E p 0, which is in line with Proposition 1. 4 Professor Goyal graciously made available their dataset (updated until 2015) on his webpage (http: //www.hec.unil.ch/agoyal/). 5 Since variables related to the short rate are included in the analysis, the sample starts in 1952, when independent monetary policy was made possible, making the interest rate variables informative. Campbell and Yogo (2006) and Pástor and Stambaugh (2009), among others, start their sample in the same year for this reason. 6 Although the results in this section are based on monthly frequency data, I also considered the prediction on a quarterly basis. The empirical findings remain unchanged, which suggests that aggregation to lower frequencies does not change the results. 7 All results based on this group of variables are shown in the Appendix. 8 For a given number of breaks, the method by (Bai and Perron, 1998) estimates a linear model with subsample dummies, where the cut-offs between subsamples are determined so that the resulting equation model has the smallest mean squared error. 23

Tables Table 1: Slope coefficients of the predictive regression ρ u,ε = 0 ρ u,ε = 0.8 λ = 1 λ = 3 λ = 10 λ = 1 λ = 3 λ = 10 T=100 0.0185 0.0836 0.1792 0.0527 0.1429 0.2028 T=300 0.0062 0.0432 0.1411 0.0201 0.0736 0.1712 T=800 0.0027 0.0191 0.1010 0.0077 0.0343 0.1183 Notes: This table presents OLS estimates of the slope coefficient β in the regression y t = α+βx t 1 +e t with different sample sizes (T ). (y t, x t ) are generated by equations (1)-(2) with β 0 = 0.2. The information component of the predictor, η t is independent, identically distributed with σ η (= σ u ) = 1. λ is the signal-to-noise ratio that determines the persistence of the explanatory variable (higher λ means more informative predictor with lower sample autocorrelation). Columns (2)-(4) represent exogenous cases, when ρ u,ε = Corr(u, ε) = 0, while in columns (5)-(7) ρ u,ε = Corr(u, ε) = 0.8, corresponding to endogenous regressors. 24

Table 2: Subsampling fixed effects estimator I: Coefficients ρ u,ε = 0 ρ u,ε = 0.8 λ = 1 λ = 3 λ = 10 λ = 1 λ = 3 λ = 10 = 25 T=100 0.0418 0.1342 0.1853 0.1246 0.2307 0.2316 T=300 0.0388 0.1357 0.1933 0.1180 0.2250 0.2290 T=800 0.0370 0.1364 0.1908 0.1149 0.2258 0.2305 = 50 T=100 0.0256 0.1149 0.1771 0.0790 0.1856 0.2188 T=300 0.0229 0.1069 0.1859 0.0695 0.1756 0.2178 T=800 0.0223 0.1052 0.1848 0.0652 0.1720 0.2229 = 100 T=100 0.0186 0.0850 0.1697 0.0554 0.1387 0.2061 T=300 0.0140 0.0778 0.1709 0.0420 0.1258 0.2063 T=800 0.0120 0.0738 0.1704 0.0367 0.1205 0.2079 II: Standard errors ρ u,ε = 0 ρ u,ε = 0.8 λ = 1 λ = 3 λ = 10 λ = 1 λ = 3 λ = 10 = 25 T=100 0.0503 0.0876 0.0993 0.0559 0.0835 0.0959 T=300 0.0274 0.0510 0.0571 0.0326 0.0499 0.0573 T=800 0.0167 0.0308 0.0348 0.0195 0.0300 0.0367 = 50 T=100 0.0393 0.0814 0.0957 0.0477 0.0865 0.0986 T=300 0.0211 0.0470 0.0547 0.0270 0.0467 0.0565 T=800 0.0122 0.0279 0.0345 0.0152 0.0284 0.0334 = 100 T=100 0.0344 0.0744 0.0954 0.0411 0.0807 0.0940 T=300 0.0167 0.0421 0.0532 0.0223 0.0454 0.0533 T=800 0.0095 0.0259 0.0355 0.0132 0.0294 0.0340 Notes: This table presents the subsample fixed effects estimates (Panel I) and standard errors (Panel II) in Equation (4) with subsample sizes = {25, 50, 100}. The data generation and further parameter specifications are given in the description of Table 1. Standard errors are calculated from the simulated empirical distribution of the parameter estimates (the simulation is repeated 1000 times). 25

Table 3: ean squared error of the subsampling estimator ρ u,ε = 0 ρ u,ε = 0.8 λ = 1 λ = 3 λ = 10 λ = 1 λ = 3 λ = 10 = 25 T=100 0.0276 0.0120 0.0101 0.0088 0.0079 0.0102 T=300 0.0268 0.0067 0.0033 0.0078 0.0031 0.0041 T=800 0.0268 0.0050 0.0013 0.0076 0.0016 0.0023 = 50 T=100 0.0320 0.0139 0.0097 0.0169 0.0077 0.0101 T=300 0.0318 0.0109 0.0032 0.0177 0.0028 0.0035 T=800 0.0317 0.0098 0.0014 0.0184 0.0016 0.0016 = 100 T=100 0.0341 0.0188 0.0100 0.0226 0.0103 0.0089 T=300 0.0349 0.0167 0.0037 0.0255 0.0076 0.0029 T=800 0.0354 0.0166 0.0021 0.0268 0.0072 0.0012 Notes: This table presents mean squared errors of the subsampling fixed effects estimator using subsample sizes = {25, 50, 100}. The data generation and further parameter specifications are given in the description of Table 1. 26

Table 4: Summary statistics of the predictors and the excess returns mean stdev φ p-value N dp -3.5278 1.4034 0.9930 0.8607 768 ep -2.7963 1.4515 0.9889 0.7279 768 de -0.7316 1.0363 0.9865 0.3744 768 tbl 0.0444 0.1064 0.9915 0.2516 768 bm 0.5212 0.8571 0.9939 0.2327 768 dfy -0.0097 0.0154 0.9705 0.2290 768 tms 0.0171 0.0490 0.9573 0.0126 768 SP500 0.0568 0.1460 0.0514 0.0000 768 Notes: stdev stands for the standard deviation of the variable and N is the number of observations. Column (3) presents first order autocorrelation of the variables, while column (4) shows the empirical significance level of the Augmented Dickey Fuller test (without deterministic trend and drift). The variables presented in the table are the dividend-price ratio (dp), the earning price ratio (ep), the dividend payout ratio (de), the book-to-market value (bm), the three-month treasury bill rate (tbl), the term-spread on the government bonds (tms), and the default yield spread (dfy). 27

Table 5: The regression results from the one period ahead forecasts using fixed subsample size Subsample size () 25 50 100 200 OLS dp 0.0921 0.0537 0.0281 0.0120 0.0061 (0.0151)*** (0.0130)*** (0.0096)*** (0.0075) (0.0039) bm 0.1113 0.0913 0.0533 0.0173 0.0034 (0.0252)*** (0.0221)*** (0.0161)*** (0.0114) (0.0063) tbl -0.5525-0.4065-0.2236-0.1574-0.1114 (0.1396)*** (0.1131)*** (0.0964)** (0.0725)** (0.0586)* tms 0.5338 0.3922 0.1914 0.1780 0.2191 (0.1856)*** (0.1643)** (0.1350) (0.1479) (0.1121)* Notes: The table presents slope coefficients and standard errors of the univariate predictive regression. The subsample fixed effects estimator is used. Standard errors are calculated using residual block bootstrapping, where the length of the blocks is O(T 1/3 ). The column header specifies the size of the subsamples. The variables presented in the table are the dividend-price ratio (dp), the book-to-market value (bm), the three-month treasury bill rate (tbl), the term-spread on the government bonds (tms), and the excess returns (SP500).,, and represent statistical significance at the 1%, 5%, and 10% level, respectively. 28

Table 6: The regression results from the one period ahead forecasts, using estimated subsample cut-offs 8 4 2 dp 0.0472 0.0316 0.0217 (0.0128)*** (0.0080)*** (0.0067)*** bm 0.0922 0.0518 0.0120 (0.0238)*** (0.0163)*** (0.0100) tbl -0.4490-0.3265-0.1618 (0.1341)*** (0.0848)*** (0.0587)*** tms 0.4441 0.2913 0.2342 (0.1784)** (0.1416)** (0.1290)* Av. size 96 192 384 Notes: The table presents slope coefficients and standard errors of the univariate predictive regression. Subsample fixed effects are used together with the approach in Bai and Perron (1998) to estimate cut-off values for the subsamples. Standard errors are from the classical OLS formula. The column header specifies the number of subsamples (which also determine the average sample size, shown in the last row). The variables presented in the table are the dividend-price ratio (dp), the book-to-market value (bm), the three-month treasury bill rate (tbl), and the term-spread on the government bonds (tms).,, and represent statistical significance at the 1%, 5%, and 10% level, respectively. 29