Robust Standard Errors in Transformed Likelihood Estimation of Dynamic Panel Data Models

Similar documents
Robust Standard Errors in Transformed Likelihood Estimation of Dynamic Panel Data Models with Cross-Sectional Heteroskedasticity

Robust Standard Errors in Transformed Likelihood Estimation of Dynamic Panel Data Models with Cross-Sectional Heteroskedasticity

Improving GMM efficiency in dynamic models for panel data with mean stationarity

Linear dynamic panel data models

Econometrics of Panel Data

GMM based inference for panel data models

Efficient Estimation of Dynamic Panel Data Models: Alternative Assumptions and Simplified Estimation

Iterative Bias Correction Procedures Revisited: A Small Scale Monte Carlo Study

Transformed Maximum Likelihood Estimation of Short Dynamic Panel Data Models with Interactive Effects

1 Introduction The time series properties of economic series (orders of integration and cointegration) are often of considerable interest. In micro pa

Exogenous Treatment and Endogenous Factors: Vanishing of Omitted Variable Bias on the Interaction Term

What s New in Econometrics. Lecture 15

1 Estimation of Persistent Dynamic Panel Data. Motivation

Estimation of Dynamic Panel Data Models with Sample Selection

A Bias-Corrected Method of Moments Approach to Estimation of Dynamic Short-T Panels *

Subset-Continuous-Updating GMM Estimators for Dynamic Panel Data Models

Chapter 6. Panel Data. Joan Llull. Quantitative Statistical Methods II Barcelona GSE

Consistent OLS Estimation of AR(1) Dynamic Panel Data Models with Short Time Series

Asymptotic Properties of Empirical Likelihood Estimator in Dynamic Panel Data Models

The Time Series and Cross-Section Asymptotics of Empirical Likelihood Estimator in Dynamic Panel Data Models

Bias Correction Methods for Dynamic Panel Data Models with Fixed Effects

A Note on Adapting Propensity Score Matching and Selection Models to Choice Based Samples

GMM ESTIMATION OF SHORT DYNAMIC PANEL DATA MODELS WITH INTERACTIVE FIXED EFFECTS

An Exponential Class of Dynamic Binary Choice Panel Data Models with Fixed Effects

Simplified Implementation of the Heckman Estimator of the Dynamic Probit Model and a Comparison with Alternative Estimators

GMM-based inference in the AR(1) panel data model for parameter values where local identi cation fails

Dynamic Panels. Chapter Introduction Autoregressive Model

Asymptotic distributions of the quadratic GMM estimator in linear dynamic panel data models

Advanced Econometrics

Missing dependent variables in panel data models

Inference about Clustering and Parametric. Assumptions in Covariance Matrix Estimation

Panel Threshold Regression Models with Endogenous Threshold Variables

Short T Panels - Review

17/003. Alternative moment conditions and an efficient GMM estimator for dynamic panel data models. January, 2017

Appendix A: The time series behavior of employment growth

Econometric Methods for Panel Data

GMM estimation of spatial panels

Linear Panel Data Models

Specification Test for Instrumental Variables Regression with Many Instruments

GMM Estimation of Empirical Growth Models

Nearly Unbiased Estimation in Dynamic Panel Data Models

A Course in Applied Econometrics Lecture 4: Linear Panel Data Models, II. Jeff Wooldridge IRP Lectures, UW Madison, August 2008

A Transformed System GMM Estimator for Dynamic Panel Data Models. February 26, 2014

Chapter 2. Dynamic panel data models

GMM ESTIMATION FOR DYNAMIC PANELS WITH FIXED EFFECTS AND STRONG INSTRUMENTS AT UNITY. Chirok Han and Peter C. B. Phillips.

Testing Overidentifying Restrictions with Many Instruments and Heteroskedasticity

Panel Data Models. James L. Powell Department of Economics University of California, Berkeley

On GMM Estimation and Inference with Bootstrap Bias-Correction in Linear Panel Data Models

Lecture 8 Panel Data

Lecture 7: Dynamic panel models 2

Small Sample Bias Properties of the System GMM Estimator in Dynamic Panel Data Models

Final Exam. Economics 835: Econometrics. Fall 2010

Bootstrap Based Bias Correction for Homogeneous Dynamic Panels*

Econometrics. Week 4. Fall Institute of Economic Studies Faculty of Social Sciences Charles University in Prague

Dynamic panel data methods

Repeated observations on the same cross-section of individual units. Important advantages relative to pure cross-section data

xtdpdml for Estimating Dynamic Panel Models

Efficient Estimation of Non-Linear Dynamic Panel Data Models with Application to Smooth Transition Models

Using all observations when forecasting under structural breaks

What s New in Econometrics. Lecture 13

Econometric Analysis of Cross Section and Panel Data

Modified Variance Ratio Test for Autocorrelation in the Presence of Heteroskedasticity

New GMM Estimators for Dynamic Panel Data Models

On IV estimation of the dynamic binary panel data model with fixed effects

Consistent estimation of dynamic panel data models with time-varying individual effects

Department of Economics

Non-linear panel data modeling

Casuality and Programme Evaluation

The Role of "Leads" in the Dynamic Title of Cointegrating Regression Models. Author(s) Hayakawa, Kazuhiko; Kurozumi, Eiji

Single Equation Linear GMM with Serially Correlated Moment Conditions

Regression with time series

Estimation of Time-invariant Effects in Static Panel Data Models

EC327: Advanced Econometrics, Spring 2007

Efficiency of repeated-cross-section estimators in fixed-effects models

Lecture 6: Dynamic panel models 1

Panel Data Models. Chapter 5. Financial Econometrics. Michael Hauser WS17/18 1 / 63

Applied Microeconometrics (L5): Panel Data-Basics

Alternative Over-identifying Restriction Test in GMM with Grouped Moment Conditions

Single Equation Linear GMM

Motivation Non-linear Rational Expectations The Permanent Income Hypothesis The Log of Gravity Non-linear IV Estimation Summary.

1. GENERAL DESCRIPTION

Panel Data Seminar. Discrete Response Models. Crest-Insee. 11 April 2008

Least Squares Estimation of a Panel Data Model with Multifactor Error Structure and Endogenous Covariates

The weak instrument problem of the system GMM estimator in dynamic panel data models

DSGE Methods. Estimation of DSGE models: GMM and Indirect Inference. Willi Mutschler, M.Sc.

Spring 2017 Econ 574 Roger Koenker. Lecture 14 GEE-GMM

Wooldridge, Introductory Econometrics, 4th ed. Chapter 15: Instrumental variables and two stage least squares

Switching Regime Estimation

A Test of Cointegration Rank Based Title Component Analysis.

Strength and weakness of instruments in IV and GMM estimation of dynamic panel data models

Flexible Estimation of Treatment Effect Parameters

Finite Sample Performance of A Minimum Distance Estimator Under Weak Instruments

When Should We Use Linear Fixed Effects Regression Models for Causal Inference with Longitudinal Data?

LM threshold unit root tests

When Should We Use Linear Fixed Effects Regression Models for Causal Inference with Longitudinal Data?

Econometrics I KS. Module 2: Multivariate Linear Regression. Alexander Ahammer. This version: April 16, 2018

Reliability of inference (1 of 2 lectures)

A Course in Applied Econometrics Lecture 14: Control Functions and Related Methods. Jeff Wooldridge IRP Lectures, UW Madison, August 2008

The exact bias of S 2 in linear panel regressions with spatial autocorrelation SFB 823. Discussion Paper. Christoph Hanck, Walter Krämer

Tests of the Co-integration Rank in VAR Models in the Presence of a Possible Break in Trend at an Unknown Point

Transcription:

D I S C U S S I O N P A P E R S E R I E S IZA DP No. 6583 Robust Standard Errors in Transformed Likelihood Estimation of Dynamic Panel Data Models Kazuhiko Hayakawa M. Hashem Pesaran May 0 Forschungsinstitut zur Zukunft der Arbeit Institute for the Study of Labor

Robust Standard Errors in Transformed Likelihood Estimation of Dynamic Panel Data Models Kazuhiko Hayakawa Hiroshima University M. Hashem Pesaran University of Cambridge, USC and IZA Discussion Paper No. 6583 May 0 IZA P.O. Box 740 5307 Bonn Germany Phone: +49-8-3894-0 Fax: +49-8-3894-80 E-mail: iza@iza.org Any opinions expressed here are those of the author(s) and not those of IZA. Research published in this series may include views on policy, but the institute itself takes no institutional policy positions. The Institute for the Study of Labor (IZA) in Bonn is a local and virtual international research center and a place of communication between science, politics and business. IZA is an independent nonprofit organization supported by Deutsche Post Foundation. The center is associated with the University of Bonn and offers a stimulating research environment through its international network, workshops and conferences, data service, project support, research visits and doctoral program. IZA engages in (i) original and internationally competitive research in all fields of labor economics, (ii) development of policy concepts, and (iii) dissemination of research results and concepts to the interested public. IZA Discussion Papers often represent preliminary work and are circulated to encourage discussion. Citation of such a paper should account for its provisional character. A revised version may be available directly from the author.

IZA Discussion Paper No. 6583 May 0 ABSTRACT Robust Standard Errors in Transformed Likelihood Estimation of Dynamic Panel Data Models * This paper extends the transformed maximum likelihood approach for estimation of dynamic panel data models by Hsiao, Pesaran, and Tahmiscioglu (00) to the case where the errors are crosssectionally heteroskedastic. This extension is not trivial due to the incidental parameters problem that arises, and its implications for estimation and inference. We approach the problem by working with a mis-specified homoskedastic model. It is shown that the transformed maximum likelihood estimator continues to be consistent even in the presence of cross-sectional heteroskedasticity. We also obtain standard errors that are robust to cross-sectional heteroskedasticity of unknown form. By means of Monte Carlo simulation, we investigate the finite sample behavior of the transformed maximum likelihood estimator and compare it with various GMM estimators proposed in the literature. Simulation results reveal that, in terms of median absolute errors and accuracy of inference, the transformed likelihood estimator outperforms the GMM estimators in almost all cases. JEL Classification: C, C3, C3 Keywords: dynamic panels, cross-sectional heteroskedasticity, Monte Carlo simulation, GMM estimation Corresponding author: M. Hashem Pesaran Trinity College University of Cambridge CB TQ Cambridge United Kingdom E-mail: mhp@cam.ac.uk * This paper was written whilst Hayakawa was visiting the University of Cambridge as a JSPS Postdoctoral Fellow for Research Abroad. He acknowledges the financial support from the JSPS Fellowship and the Grant-in-Aid for Scientific Research (KAKENHI 73078) provided by the JSPS. Pesaran acknowledges financial support from the ESRC Grant No. ES/0366/. Elisa Tosetti contributed to a preliminary version of this paper. Her assistance in coding of the transformed ML estimator and some of the derivations is gratefully acknowledged.

Introduction In dynamic panel data models where the time dimension (T ) is short, the presence of lagged dependent variables among the regressors makes standard panel estimators inconsistent, and complicates statistical inference on the model parameters considerably. Over the last few decades, a sizable literature has been developed on the estimation of dynamic panel data models. Early work includes the Instrumental Variables (IV) approach by Anderson and Hsiao (98, 98). More recently, a large number of studies have been focusing on the generalized method of moments (GMM), see, among others, Holtz-Eakin, Newey, and Rosen (988), Arellano and Bond (99), Arellano and Bover (995), Ahn and Schmidt (995) and Blundell and Bond (998). One important reason for the popularity of GMM in applied economic research is that it provides asymptotically valid inference under a minimal set of statistical assumptions. Arellano and Bond (99) suggested to transform the dynamic model into first differences to eliminate the individual-specific effects, and then use a set of moment conditions where lagged variables in levels are used as instruments. Blundell and Bond (998) showed that the performance of this estimator deteriorates when the parameter associated with the lagged dependent variable is close to one and/or the variance ratio of the individual effects to the idiosyncratic errors is large since in these cases the instruments are only weakly related to the lagged dependent variables. Among others, the poor finite sample properties of GMM has been documented in Monte Carlo studies by Kiviet (007). To deal with this problem, Arellano and Bover (995) and Blundell and Bond (998) proposed the use of extra moment conditions arising from the model in levels, available when the initial observations satisfy certain conditions. The resulting GMM estimator, known as system GMM, combines moment conditions for the model in first differences with moment conditions for the model in levels. We refer to Blundell, Bond, and Windmeijer (000) for an extension to the multivariate case, and for a Monte Carlo study of the properties of GMM estimators using moment conditions from either the first differenced and/or levels models. Bun and Windmeijer (00) proved that the equation in levels suffers from a weak instrument problem when the variance ratio is large. Hayakawa (007) also shows that the finite sample bias of the system GMM estimator becomes large when the variance ratio is large. The GMM estimators discussed so far have been widely adopted in the empirical literature, to investigate problems in areas such as labour economics, development economics, health economics, macroeconomics and finance. Theoretical and applied research on dynamic panels has mostly focused on the GMM, and has by and large neglected the maximum likelihood (ML) approach. Indeed, the incidental parameters issue and the initial values problem lead to a violation of the standard regularity conditions for the ML estimators of the structural parameters to be consistent. Hsiao et al. (00) developed a transformed likelihood approach to overcome the incidental parameters problem. Binder et al. (005) have extended this approach for estimating panel VAR (PVAR) models. Alvarez and Arellano (004) have studies ML estimation of autoregressive panels in the presence of time-specific See also the discussion in Binder, Hsiao, and Pesaran (005), who proved that the asymptotic variance of the Arellano and Bond (99) GMM estimator depends on the variance of the individual effects.

heteroskedasticity (see also Bhargava and Sargan (983)). Kruiniger (008) considers ML estimation of a stationary/unit root AR() panel data models. In this paper, we extend the analysis of Hsiao et al. (00) to allow for cross-sectional heteroskedasticity. This extension is not trivial due to the incidental parameters problem that arises, and its implications for estimation and inference. To deal with the problem, we follow the GMM literature and ignore the error variance heterogeneity and work with a mis-specified homoskedastic model, and show that the transformed maximum likelihood estimator by Hsiao et al. (00) continues to be consistent. We then derive a covariance matrix estimator which is robust to cross-sectional heteroskedasticity. Using Monte Carlo simulations, we investigate the finite sample performance of the transformed likelihood estimator and compare it with a range of GMM estimators. Simulation results reveal that, in terms of median absolute errors and accuracy of inference, the transformed likelihood estimator outperforms the GMM estimators in almost all cases when the model contains an exogenous regressor, and in many cases if we consider pure autoregressive panels. The rest of the paper is organized as follows. Section describes the model and its underlying assumptions. Section 3 proposes the transformed likelihood estimator for cross-sectionally heteroskedastic errors. Section 4 reviews the GMM approach as applied to dynamic panels. Section 5 describes the Monte Carlo design and comments on the small sample properties of the transformed likelihood and GMM estimators. Finally, Section 6 ends with some concluding remarks. The dynamic panel data model Consider the panel data model y it = α i + γy i,t + βx it + u it, () for i =,,..., N. It is supposed that these dynamic processes have started at time t = m, (m being a finite positive constant) but we only observe the observations (y it, x it ) over the period t = 0,,,..., T. We assume that x it is a scalar to simplify the notation. Extension to the case of multiple regressors is straightforward at the expense of notational complexity. We further assume that x it is generated either by x it = µ i + ϕt + a j ε i,t j, a j < () or x it = ϕ + d j ε i,t j, d j < (3) where µ i can either be fixed constants, differing across i, or randomly distributed with a common mean, and ε it are independently distributed over i and t with E(ε it ) = 0, and var(ε it ) = σεi, with 0 < σεi < K <. We shall also consider the following assumptions: 3

Assumption (Initialization) Depending on whether the y it process has reached stationarity, one of the following two assumptions holds: (i) γ <, and the process has been going on for a long time, namely m ; (ii) The process has started from a finite period in the past not too far back from the 0th period, namely for given values of y i, m+ with m finite, such that E( y i, m+ x i, x i,..., x it ) = b, for all i, where b is a finite constant. Assumption (shocks to equations) Disturbances u it are serially and cross-sectionally independently distributed, with E (u it ) = 0, E ( u it) = σ i, and E ( u 4 ) it /σ4 i = κ, such that 0 < σ i < K <, and 0 < κ < K <, for i =,,..., N and t =,,..., T. Assumption 3 (shocks to regressors) ε it in x it are independently distributed over all i and t, with E (ε it ) = 0, and E ( ε it) = σ εi, and independent of u is for all s and t. Assumption 4 (constant variance ratio) σ εi /σ i = c, for i =,,..., N, with 0 < c < K <. Remark Assumption.(ii) constrains the expected changes in the initial values to be the same across all individuals, but does not necessarily require that γ <. Assumptions, 3, and 4 allow for heteroskedastic disturbances in the equations for y it and x it, but to avoid the incidental parameter problem require their ratio to be constant over i. Also Assumption 3 requires x it to be strictly exogenous. These restrictions can be relaxed by considering a panel vector autoregressive specification of the type considered in Binder et al. (005). However, these further developments are beyond the scope of the present paper. See also the remarks in Section 6. 3 Transformed likelihood estimation Take the first differences of () to eliminate the individual effects: y it = γ y i,t + β x it + u it, (4) which is well defined for t =, 3,..., T, but not for t =, because the observations y i,, i =,,..., N, are not available. However, starting from y i, m+, and by continuous substitution, we obtain m m y i = γ m y i, m+ + β γ j x i, j + γ j u i, j. 4

Note that the mean of y i conditional on y i, m+, x i, x i0,..., given by m η i = E ( y i y i, m+, x i, x i0,...) = γ m y i, m+ + β γ j x i, j, (5) is unknown, since the observations x i, j, for j =,,..., m, i =,,..., N are unavailable. To solve this problem, we need to express the expected value of η i, conditional on the observables, in a way that it only depends on a finite number of parameters. The following theorem provides the conditions under which it is possible to derive a marginal model for y i, which is a function of a finite number of unknown parameters. Theorem Consider model (), where x it follows either () or (3). Suppose that Assumptions,, 3, and 4 hold. Then y i can be expressed as: y i = b + π x i + v i, (6) where b is a constant, π is a T -dimensional vector of constants, x i = ( x i, x i,..., x it ), and v i is independently distributed across i, such that E(v i ) = 0, and E(vi )/σ i = ω, with 0 < ω < K <. Note that Assumption 4 is used to show that E(v i )/σ i does not vary with i. It is now possible to derive the likelihood function of the transformed model given by equations (6) and (4) for t =, 3,..., T. Let y i = ( y i, y i,..., y it ), x i 0 0 0 0 y W i = i x i T (T +3)...., 0 0 y i,t x it and note that the transformed model can be rewritten as y i = W i φ + r i, (7) with φ = (b, π, γ, β). The covariance matrix of r i = (v i, u i,..., u it ) has the form: E(r i r i) = Ω i = σi where ω is a free parameter defined in Theorem. ω 0......... 0 = σi Ω, (8) 5

The log-likelihood function of the transformed model (7) is given by l (ψ N ) = NT ln (π) T σ i ln σi N ln + T (ω )] ( y i W i φ) Ω ( y i W i φ), where ψ N = ( φ, ω, σ, σ,...σ N). Unfortunately, the maximum likelihood estimation based on l(ψn ) encounters the incidental parameter problem of Neyman and Scott (948) since the number of parameters grows with the sample size, N. Following the mis-specification literature in econometrics, (White, 98; Kent, 98), we examine the asymptotic properties of the ML estimators of the parameters of interest, φ and ω, using a mis-specified model where the heteroskedastic nature of the errors is ignored. Accordingly, suppose that it is incorrectly assumed that the regression errors u it are homoskedastic, i.e., σi = σ, i =,,..., N. Then under this mis-specification the pseudo log-likelihood function of the transformed model (7), is given by where θ = ( φ, ω, σ ) l p (θ) = NT σ NT ln (π) ln ( σ ) N ln + T (ω )] ( y i W i φ) Ω ( y i W i φ), (9) is the vector of unknown parameters. maximizing the pseudo log-likelihood in (9), and consider the pseudo-score vector l p (θ) θ = σ N W i Ω ( y i W i φ) NT g + σ g N r i Φr i NT σ + σ 4 N r i Ω r i Let θ be the estimator obtained by, where g = Ω = + T (ω ), (see (40)), and Φ = T T (T ) T (T )... T T (T ) (T ) (T )(T )... (T )....... T (T ) (T ).... (0) 6

Under heteroskedastic errors, the pseudo-true value of θ denoted by θ = (φ, ω, σ ), is the solution of E l p (θ ) / θ] = 0, namely E W iω ( y i W i φ ) ] = 0, () NT g + σ g NT σ + σ 4 E ( r ) iφr i = 0, () E ( r iω r i ) = 0, (3) where expectations are taken with respect to the true probability measure, and g = + T (ω ). Focusing first on () and (3), we have E ( r ) N iφr i = σi tr (ΦΩ) = N σ NT g, E ( r iω ) r i = T N σ N tr(ω Ω)/T, where σ N = N N σ i and (4) is used. Hence, using the above results in () and (3), we have NT g + NT σ + σ 4 σ g N σ Ntr (ΦΩ) = NT + g σ g N σ NT g = 0, E ( r iω r i ) = NT σ + σ 4 T N σ Ntr(Ω Ω)/T = 0. From the first equation, we have σ / σ N = g/g = + T (ω )] / + T (ω )]. From the second equation, we have σ / σ N = tr(ω Ω)/T. Using these two, we have To solve this equation for ω, we first note that note that + T (ω ) + T (ω ) = T tr(ω Ω). (4) tr(ω Ω)/T = + g (ω ω ). This result follows since all elements of = Ω Ω are zero, except for the first element of which is given by ω ω. Substituting this into (4), and after some algebra we have (T )( ω ω) = 0, which yields ω = ω for all T >. It also follows that σ = lim N σ N. Using the former result in (), we have φ = φ. These results are stated formally in the following theorem. Theorem Suppose that Assumptions,, 3, and 4 hold, and let θ = ( φ, ω, σ ) be the pseudo 7

true values of the ML estimator obtained by maximizing the pseudo log-likelihood function in (9). Then, we have φ = φ, ω = ω, σ = lim N N This is one of the key results of this paper. of θ are identical to the first (T + 4) entries of ψ N. σi. This theorem shows that the first (T + 4) entries This indicates that the ML estimator of φ and ω obtained under mis-specified homoskedastic models will continue to be consistent, namely, the transformed ML estimator by Hsiao et al. (00) is consistent even if cross-sectional heteroskedasticity is present. The following theorem establishes the asymptotic distribution of the ML estimator of the transformed model. Theorem 3 Suppose that Assumptions,, 3 and 4 hold and let θ = ( φ, ω, σ ) be the ML estimator obtained by maximizing the pseudo log-likelihood function in (9). Then as N tends to infinity, θ is asymptotically normal with N ( θ θ ) d N ( 0, A B A ) (5) where θ = (φ, ω, σ ), A = lim N N ] l p (θ ) θ θ, and B = lim E N N l p (θ ) θ ] l p (θ ) θ. To obtain consistent estimators of A and B, robust to unknown heteroskedasticity, let r i = y i W i φ. Further, let σ NT = (T N) N r i Ω r i, be an estimator of N N σ i. Then a consistent estimator of A, denoted as Â, is given by  = N σ NT g N σ NT N W i Ω W i g N σ NT N r iφ W i T g 0 T g σ NT N W i Φ r i 0 T g σ NT T ( σ NT ) To obtain a consistent estimator of B, denoted by B, we also need to assume that the fourth moment of (v i u i )/σ i is homogeneous across i. In particular, Assumption 5 (kurtosis condition) Assume that E(η 4 i ) = κ = γ + 3 for i =,,..., N, where η i = (v i u i )/σ i (ω ) / ], and γ is the Pearson s measure of kurtosis. 8.

This assumption is used in combination with Assumption to consistently estimate N N σ4 i by σ 4 NT defined in the Appendix by (66). Then the elements of B are given by: 4 GMM approach B N = N ( σ ) NT W Ω r i i r Ω i W i, { B T N = 4ĝ ( σ ) 4 NT N ( r i Φ r i T ( B T 33 = 4 ( σ ) NT 4 N r Ω r ) i i σ 4 T ) ĝ σ 4 NT NT },, B N ( = Nĝ ( σ ) NT r Ω ) i W i ( r ) i Φ r i, B N 3 = N ( σ ) NT 3 B T 3 = 4ĝ ( σ ) NT 3 N ( r i Ω W i ) ( r i Ω r i ), r iφ r i r Ω r i i ĝ σ NT 4 T T In this section, we review the GMM approach as a basis for the simulation studies in the next section. In the GMM approach, it is assumed that α i and u it have an error components structure, in which E (α i ) = 0, E (u it ) = 0, E(α i u it ) = 0, for i =,.., N; and t =,,..., T, (6) and the errors are uncorrelated with the initial values E (y i0 u it ) = 0, for i =,,.., N, and t =,,..., T. (7) As with the transformed likelihood approach, it is also assumed that the errors, u it, are serially and cross-sectionally independent: E (u it u is ) = 0, for i =,,.., N, and t s =,,..., T. (8) ]. Note that no restrictions are placed on E(α i u it ) under the transformed likelihood approach 9

4. Estimation 4.. The first-difference GMM estimator Under (6)-(8), and focusing on the equation in first differences, (4), Arellano and Bond (99) suggest the following T (T )/ moment conditions: E y is u it ] = 0, (s = 0,,..., t, t =, 3,..., T ). (9) If regressors, x it, are strictly exogenous, i.e., if E (x is u it ) = 0, for all t and s, then the following additional moments can also be used E x is u it ] = 0, (s, t =,..., T ). (0) The moment conditions (9) and (0) can be written compactly as: ] E Ż i u i = 0, where u i = q i Ẇiδ, δ = (γ, β) = (δ, δ ) and y i0, x i,..., x it 0... 0 0 y Ż i = i0, y i, x i,..., x it... 0....., q i = y i. y it 0 0... y i0,..., y i,t, x i,..., x it, Ẇ i = y i. y i,t x i. x it. The one and two-step first-difference GMM estimators based on the above moment conditions are given by δ dif GMM = δ dif GMM = ( ) ) ) Ṡ ZW (Ḋstep ṠZW Ṡ ZW (Ḋstep ṠZq, () ( ) ) ) Ṡ ZW (Ḋstep ṠZW Ṡ ZW (Ḋstep ṠZq, () where ṠZW = N N Ż iẇi, ṠZq = N N Ż i q i, Ḋstep = N N Ż i HŻi, Ḋstep = N N Ż i u i u iżi, dif u i = q i Ẇi δ GMM, and H is a matrix with s on the main diagonal, - s on the first sub-diagonal and 0 s otherwise. 0

4.. System GMM estimator Although consistency of the first-difference GMM estimator is obtained under a mild assumption of no serial correlation, Blundell and Bond (998) demonstrated that it suffers from the so called weak instruments problem when γ is close to one and/or the variance ratio var(α i )/var(u it ) is large. As a solution, these authors propose the system GMM estimator due to Arellano and Bover (995) and show that it works well even if γ is close to unity. But as shown recently by Bun and Windmeijer (00), the system GMM estimator continues to suffer from the weak instruments problem when the variance ratio var(α i )/var(u it ) is large. To introduce the moment conditions for the system GMM estimator, the following additional homogeneity assumptions are required: E(y is α i ) = E(y it α i ), for all s and t, E(x is α i ) = E(x it α i ), for all s and t. Under these assumptions, we have the following moment conditions: E y is (α i + u it )] = 0, (s =,..., t, t =, 3,..., T ), (3) E x is (α i + u it )] = 0, (s, t =, 3,..., T ). (4) For the construction of the moment conditions for the system GMM estimator, given the moment conditions for the first-difference GMM estimator, some moment conditions in (3) and (4) are redundant. Hence, to implement the system GMM estimation, in addition to (9) and (0), we use the following moment conditions: E y i,t (α i + u it )] = 0, (t =, 3,..., T ), (5) E x it (α i + u it )] = 0, (t =, 3,..., T ). (6) The moment conditions (9), (0), (5) and (6) can be written compactly as ] E Z i ü i = 0,

where ü i = q i Ẅiδ, Z i = diag (Żi, Z ) i, Zi = y i, x i 0... 0 0 y i, x i3 0....., q i = ( qi q i ), q i = y i. y it 0 0... y i,t, x it ( ) y i x i, Ẅ Ẇi i =, Wi = W.. i y i,t x it. The one and two-step system GMM estimators based on the above conditions are given by δ sys GMM = δ sys GMM = ( ( ) ) ( ) S ZW Dstep SZW S ZW Dstep SZq, (7) ( ( ) ) ( ) S ZW Dstep SZW S ZW Dstep SZq, (8) where S ZW = N N Z iẅi, S Zq = N N Z i q i and D step = diag ( N N Ż i HŻi, N N Z i Z i ). The two-step system GMM estimator is obtained by replacing D step with D step = N Z N i ü i ü Z i i where ü sys i = q i Ẅi δ GMM. 4..3 Continuous-updating GMM estimator Since the two-step GMM estimators have undesirable finite sample bias property, (Newey and Smith, 004), alternative estimation methods have been proposed in the literature. These include the empirical likelihood estimator, (Qin and Lawless, 994), the exponential tilting estimator (Kitamura and Stutzer, 997; Imbens, Spady, and Johnson, 998) and the continuous updating (CU-) GMM estimator (Hansen, Heaton, and Yaron, 996), where these are members of the generalized empirical likelihood estimator (Newey and Smith, 004). Amongst these estimators, we mainly focus on the CU-GMM estimator as an alternative to the two-step GMM estimator. To define the CU-GMM estimator, we need some additional notation. Let Ži denote Żi or Z i, and ǔ i denote u i or ü i. Also, let m be the number of columns of Ži, i.e., the number of instruments, and set g i (δ) = Ž iǔi, ĝ(δ) = N Then, the CU-GMM estimator is defined as g i (δ), Ω(δ) = N g i (δ) ĝ(δ)] g i (δ) ĝ(δ)]. δ GMM CU = arg min δ Q(δ), (9) Q(δ) = ĝ(δ) Ω(δ) ĝ(δ)/. (30)

Newey and Smith (004) demonstrate that the CU-GMM estimator has a smaller finite sample bias than the two-step GMM estimator. 4. Inference 4.. Alternative standard errors In the case of GMM estimators the choice of the covariance matrix is often as important as the choice of the estimator itself for inference. Although, it is clearly important that the estimator of the covariance matrix should be consistent, in practice it might not have favorable finite sample properties and result in inaccurate inference. To address this problem, some modified standard errors have been proposed. For the two-step GMM estimators, Windmeijer (005) proposes corrected standard errors for linear static panel data models which are applied to dynamic panel models by Bond and Windmeijer (005). For the CU-GMM, while it is asymptotically equivalent to the two-step GMM estimator, it is more dispersed than the two-step GMM estimator in finite samples and inference based on conventional standard errors formula results in a large size distortion. To overcome this problem, Newey and Windmeijer (009) propose an alternative estimator for the covariance matrix of CU-GMM estimators under many-weak moments asymptotics and demonstrate by simulation that the use of the modified standard errors improve the size property of the tests based on the CU-GMM estimators. 3 4.. Weak instruments robust inference As noted above, the first-difference and system GMM estimators could be subject to the weak instruments problem. In the presence of weak instruments, the estimators are biased and inference becomes inaccurate. As a remedy for this, some tests that have correct size regardless of the strength of instruments have been proposed in the literature. These include Stock and Wright (000) and Kleibergen (005). Stock and Wright (000) propose a GMM version of the Anderson and Rubin(AR) test (Anderson and Rubin, 949). Kleibergen (005) proposes a Lagrange Multiplier (LM) test. This author also extends the conditional likelihood ratio (CLR) test of Moreira (003) to the GMM case since the CLR test performs better than other tests in linear homoskedastic regression models. We now introduce these tests. The GMM version of the AR statistic proposed by Stock and Wright (000) is defined as AR(δ) = N Q(δ). (3) Under the null hypothesis H 0 : δ = δ 0, this statistic is asymptotically (as N ) distributed as χ m, regardless of the strength of the instruments, where m is the dimension of δ. 3 For the precise definition of many weak moments, see Newey and Windmeijer (009). 3

The LM statistic proposed by Kleibergen (005) is LM(δ) = N Q(δ) δ where D(δ) = ( d (δ), d ) (δ) with D(δ) Ω(δ) D(δ) ] Q(δ) δ, (3) d j (δ) = N g i (δ) δ j ( N ) g i (δ) g δ i (δ) Ω(δ) ĝ(δ), for j = and. j Under the null hypothesis H 0 : δ = δ 0, this statistic follows χ, asymptotically The GMM version of the CLR statistic proposed by Kleibergen (005) is given by CLR(δ) = ] ( AR(δ) R(δ) + AR(δ) R(δ) ) + 4LM(δ) R(δ) (33) where R(δ) is a statistic which is large when instruments are strong and small when the instruments are weak, and is random only through D(δ) asymptotically. In the simulation, following Newey and Windmeijer (009), we use R(δ) ( ) = N λ min D(δ) Ω(δ) D(δ) where λ min (A) denotes the smallest eigenvalue of A. Under the null hypothesis H 0 : δ = δ 0, this statistic asymptotically follows a nonstandard distribution which can be obtained by simulation 4. These tests are derived under the standard asymptotic where the number of moment conditions is fixed. Recently, Newey and Windmeijer (009) show that these results are valid even under many weak moments asymptotics. 5 Monte Carlo simulations In this section, we conduct Monte Carlo simulations to investigate the finite sample properties of the transformed log-likelihood approach and compare them to those of the various GMM estimators proposed in the literature and discussed in the previous section. 5. ARX() model We first consider a distributed lag model with one exogenous regressor (ARX()), which is likely to be more relevant in practice than the pure AR() model which will be considered later. 4 For the details of computation, see Kleibergen (005) or Newey and Windmeijer (009). 4

5.. Monte Carlo design For each i, the time series processes {y it } are generated as y it = α i + γy i,t + βx it + u it, for t = m +, m +,.., 0,,..., T, (34) with the initial value y i, m = α i + βx i, m + u i, m, and u it N (0, σi ), with σ i U0.5,.5], so that E(σi ) =. We discard the first m observations, and use the observations t = 0 through T for estimation and inference. 5 The regressor, x it, is generated as where x it = µ i + gt + ζ it, for t = m, m +,.., 0,,..., T, (35) ζ it = ϕζ i,t + ε it, for t = 49 m, 48 m,..., 0,,..., T, (36) ε it N (0, σ εi), ξ i, m 50 = 0. (37) with ϕ <. We discard the first 50 observations of ζ it and use the remaining T + + m observations for generating x it and y it. In the simulations, we try the values γ = 0.4, 0.9, β = 0.5, ϕ = 0.5, and g = 0.0. The error variances σεi are set so that to ensure a reasonable fit, namely6 σεi = σ i R y ( + ϕ)( ϕγ) ), β ( R y with R y = 0.4. The sample sizes considered are N = 50, 50, 500 and T = 5, 0, 5. For the ( ) individual effects, we set α i = τ qi, where q i χ. For the value of τ, which is the variance ratio, τ = var(α i )/var(u it ), we consider the values of τ = often used in the literature, and the high value of τ = 5. Further, we assume that both y it and x it depend linearly on the same individual effects, by taking µ i = ηα i where the value of η is computed by (69) in the Appendix A.5 with R y = 0.4. 7 In computing the transformed ML estimators we use the minimum distance estimator of Hsiao et al. (00) as starting values for the nonlinear optimization where ω is estimated by the one-step 5 Hence, T + is the actual length of the estimation sample. 6 For the derivation of R y, see Appendix A.5. 7 Since (69) is a quadratic equation, we have two solutions. In the simulations, we used the positive solution. 5

first-difference GMM estimator () in which Żi is replaced with y i0 x i 0 0 y Ż i = i x i y i0 x i..... y i,t x i,t y i,t 3 x i,t For the GMM estimators, although there are many moment conditions for the first-difference GMM estimator as in (9) and (0), we consider two sets of moment conditions which only exploit a subset of instruments. The first set of moment conditions, denoted as DIF, consists of E(y is u it ) = 0 for s = 0,..., t ; t =,..., T and E(x is u it ) = 0 for s =,..., t; t =,..., T. In this case, the number of moment conditions are 4, 99, 4 for T = 5, 0, 5, respectively. The second set of moment conditions, denoted as DIF, consist of E(y i,t l u it ) = 0 with l = 0 for t =, l = 0, for t = 3,..., T and E(x i,t l u it ) = 0 with l = 0, for t =, l = 0,, for t = 3,..., T. In this case, the number of moment conditions are 8, 43, 68 for T = 5, 0, 5, respectively. Similarly, for the system GMM estimator, we add moment conditions (5) and (6) in addition to DIF and DIF, which are denoted as SYS and SYS, respectively. For SYS we have 3, 7, 5 moment conditions for T = 5, 0, 5, respectively, while for SYS we have 6, 6, 96 moment conditions for T = 5, 0, 5, respectively. In a number of cases where N is not sufficiently large relative to the number of moment conditions (for example, when T = 5 and N = 50) the inverse of the weighting matrix can not be computed. Such cases are denoted in the summary result tables. For inference, we use the robust standard errors formula given in Theorem for the transformed likelihood estimator. For the GMM estimators, in addition to the conventional standard errors, we also compute Windmeijer (005) s standard errors with finite sample correction for the two-step GMM estimators and Newey and Windmeijer (009) s alternative standard errors formula for the CU-GMM estimators. In addition to the MC results for γ and β, we also report simulation results for the long-run coefficient defined by δ = β/( γ). We report median biases, median absolute errors (MAE), size and power for γ, β and δ. The power is computed at γ 0., β 0. and (β 0.)/( (γ 0.)), for selected null values of γ and β. All tests are carried out at the 5% significance level, and all experiments are replicated, 000 times. 5.. Results for the ARX() model To save space, we report the results of the GMM estimators which exploit moment conditions DIF and SYS only. The reason for selecting these moment conditions is that, in practice, these moment conditions are often used to mitigate the finite sample bias caused by using too many instruments. A complete set of results giving the remaining GMM estimators that make use of additional instruments 6

are provided in a supplement available from the authors on request. The small sample results for γ are summarized in Tables to 4. Table provides the results for the case of γ = 0.4, and shows that the transformed likelihood estimator has a smaller bias than the GMM estimators in all cases with the exception of the CU-GMM estimator (the last panel of Table ). In terms of MAE the transformed likelihood estimator outperforms the GMM estimators in all cases. As for the effect of increasing the variance ratio, τ, on the various estimators, we first recall that the transformed likelihood estimator must be invariant to the choice of τ, although the estimates reported in Table do show some effects, albeit small. The observed impact of changes in τ on the performance of the transformed likelihood estimator is solely due to computational issues and reflects the dependence of the choice of initial values on τ in computation of the transformed ML estimators. One would expect that these initial value effects to disappear as N is increased, and this is seen to be the case from the results summarized in Table. In contrast, the performance of the GMM estimators deteriorates (in some case substantially) as τ is increased from to 5. This tendency is especially evident in the case of the system GMM estimators, and is in sharp contrast to the performance of the transformed likelihood estimator which are robust to changes in τ. These observations also hold if we consider the experiments with γ = 0.9 (Table ). Although the GMM estimators have smaller biases than the transformed likelihood estimator in a few cases, in terms of MAE, the transformed likelihood estimator performs best in all cases. We next consider size and power of the various tests, summarized in Tables 3 and 4. Table 3 shows that the empirical size of the transformed likelihood estimator is close to the nominal size of 5% for all values of T, N and τ. For the GMM estimators, we find that the test sizes vary considerably depending on T, N, τ, the estimation method (step, step, CU), and whether corrections are applied to the standard errors. In the case of the GMM results without standard error corrections, most of the GMM methods are subject to substantial size distortions when N is small. For instance, when N = 50, T = 5, and τ =, the size of the test based on DIF(step) estimator is 30.4%. But the size distortion gets smaller as N increases. Increasing N to 500, reduces the size of this test to 6.6%. However, even with N = 500, the size distortion gets larger for two-step and CU-GMM estimators as T increases. As to the effects of changes in τ on the estimators, we find that the system GMM estimators are significantly affected when τ is increased. When τ = 5, all the system GMM estimators have large size distortions even when T = 5 and N = 500, where conventional asymptotics are expected to work well. This may be due to large finite sample biases caused by a large τ. Amongst the tests based on corrected GMM standard errors, Windmeijer (005) s correction seems to be quite useful, and in many cases it leads to accurate inference, although the corrections do not seem able to mitigate the size problem of the system GMM estimator when τ is large. The standard errors of Newey and Windmeijer (009) are not always helpful, and although they improve the size property in some cases, they have either little effects or tend to worsen the test sizes in other cases. 7

Comparing power of the tests, we observe that the transformed likelihood estimator is in general more powerful than the GMM estimators. For example when N = 50, the transformed likelihood estimators have higher power than SYS(step W ) which is the most efficient amongst the reported GMM estimators. The above conclusions hold generally when we consider experiments with γ = 0.9 (Table 4), except that the system GMM estimators now perform rather poorly even for a relatively large N. For example, when γ = 0.9, T = 5, N = 500 and τ =, size distortions of the system GMM estimators are substantial, as compared to the case where γ = 0.4. Although it is known that the system GMM estimators break down when τ is large 8, the simulation results in Table 4 reveal that they perform poorly even when τ is not so large (τ = ). The small sample results for β (Tables 5 to 8), are similar to the results reported for γ. The transformed likelihood estimator tends to have smaller biases and MAEs than the GMM estimators in many cases, and there are almost no size distortions for all values of T, N and τ. The performance of the GMM estimators crucially depends on the values of T, N and τ. Unless N is large, the GMM estimators perform poorly and the system GMM estimators are subject to substantial size distortions when τ is large even for N = 500, although the magnitude of size distortions are somewhat smaller than those reported for γ. The results for the long-run coefficient, δ = β/( γ), are reported in a supplementary appendix, and are very similar to those of γ and β. Although the GMM estimators outperform the transformed likelihood estimator in some cases, in terms of MAE, the transformed likelihood estimator performs best in almost all cases. As for inference, the transformed likelihood estimator has correct sizes for all values of T, N and τ when γ = 0.4. However, it shows some size distortions when γ = 0.9 and the sample size is small, say, when T = 5 and N = 50. However, size improves as T and/or N increase(s). When T = 5 and N = 500, there is essentially no size distortions. For the GMM estimators, it is observed that although the sizes are correct in some cases, say, the case with T = 5 and N = 500 when γ = 0.4, it is not the case when γ = 0.9; even for the case of T = 5 and N = 500, there are size distortions and a large τ aggravates the size distortions. Finally, we consider weak instruments robust tests, which are reported in Tables 9 and 0. We find that test sizes are close to the nominal value only when T = 5 and N = 500. In other cases, especially when N is small and/or T is large, there are substantial size distortions. Although Newey and Windmeijer (009) prove the validity of these tests under many weak moments asymptotics, they are essentially imposing m /N 0 or a stronger restriction where m is the number of moment conditions, which is unlikely to hold when N is small and/or T is large. Therefore, the weak instruments robust tests are less appealing, considering the very satisfactory size properties of the transformed likelihood estimator, the difficulty of carrying out inference on subset of the parameters using the weak instruments robust tests, and large size distortions observed for these tests when N is small. In summary, for estimation of ARX panel data models the transformed likelihood estimator has 8 See Hayakawa (007) and Bun and Windmeijer (00). 8

several favorable properties over the GMM estimators in that the transformed likelihood estimator generally performs better than the GMM estimators in terms of biases, MAEs, size and power, and unlike GMM estimators, it is not affected by the variance ratio of individual effects to disturbances. 5. AR() model 5.. Monte Carlo design The data generating process is the same as that in the previous section with β = 0. More specifically, y it are generated as y it = α i + γy i,t + u it, for t =,..., T and i =,..., N, (38) α i y i0 = γ + u i0 γ, (39) where u it N (0, σ i ) with σ i U0.5,.5]. Note that y it are covariance stationary. Individual effects are generated as α i = τ(q i )/ where q i χ. For parameters and sample sizes, we consider γ = 0.4, 0.9, T = 5, 0, 5, 0 N = 50, 50, 500, and τ =, 5. Some comments on the computations are in order. For the starting value in the nonlinear optimization routine used to compute the transformed log-likelihood estimator, we use ( b, γ, ω, σ ) where b = N N y i, γ is the one-step first-difference GMM estimator () where Ẇ i and Żi are replaced with 9 y i0 0 0 y i y i y i0 0 Ẇ i =., Ż i = y i y i y i0, y i,t... y i,t y i,t 3 y i,t 4 ω = (N ) σ u] ( N y i b ) and σ u = N(T )] N ( y it γ y i,t ). For the first-difference GMM estimators, we consider two sets of moment conditions. The first set of moment conditions, denoted as DIF, consists of E(y is u it ) = 0 for s = 0,..., t ; t =,..., T. In this case, the number of moment conditions are 0, 45, 05 for T = 5, 0, 5, respectively. The second set of moment conditions, denoted by DIF, consist of E(y i,t l u it ) = 0 with l = 0 for t =, and l = 0, for t = 3,..., T. In this case, the number of moment conditions are 7, 7, 7 for T = 5, 0, 5, respectively. Similarly, for the system GMM estimator, we add moment conditions E y i,t (α i + u it )] = 0 for t =,..., T in addition to DIF and DIF, which are denoted as SYS and SYS, respectively. 9 This type of estimator is considered in Bun and Kiviet (006). Since the number of moment conditions are three, this estimator is always computable for any values of N and T considered in this paper. Also, since there are two more moments, we can expect that the first and second moments of the estimator to exist. 9

For the moment conditions SYS, we have 4, 54, 9 moment conditions for T = 5, 0, 5, respectively, while for the moment conditions SYS, we have, 6, 4 moment conditions for T = 5, 0, 5, respectively. With regard to the inference, we use the robust standard errors formula given in Theorem for the transformed log-likelihood estimator. For the GMM estimators, in addition to the conventional standard errors, we also compute Windmeijer (005) s standard errors for the two-step GMM estimators and Newey and Windmeijer (009) s standard errors for the CU-GMM estimators. We report the median biases, median absolute errors (MAE), sizes (γ = 0.4 and 0.9) and powers (resp. γ = 0.3 and 0.8) with the nominal size set to 5%. As before, the number of replications is set to, 000. 5.. Results As in the case of ARX() experiments, to save space, we report the results of the transformed likelihood estimator and the GMM estimators exploiting moment conditions DIF and SYS. Complete set of results are provided in a supplement, which is available upon request. The biases and MAEs of the various estimators for the case of γ = 0.4 are summarized in Table. As can be seen from this table, the transformed likelihood estimator performs best (in terms of MAE) in almost all cases, the exceptions being the CU-GMM estimators that show smaller biases in some experiments. As to be expected, the one- and two-step GMM estimators deteriorate as the variance ratio, τ, is increased from to 5, and this tendency is especially evident for the system GMM estimator. For the case of γ = 0.9 (Table ), we find that the system GMM estimators have smaller biases and MAEs than the transformed likelihood estimator in some cases. However, when τ = 5, the transformed likelihood estimator outperforms the GMM estimators in all cases, both in terms of bias and MAE. Consider now the size and power properties of the alternative procedures. The results for γ = 0.4 are summarized in Table 3. We first note that the transformed likelihood procedure shows almost correct sizes for all experiments. For the GMM estimators, although there are substantial size distortions when N = 50, the empirical sizes become close to the nominal value as N is increased. When T = 5, 0 and N = 500 and τ =, the size distortions of the GMM estimators are small. However, when τ = 5, there are severe size distortions for the system GMM estimator even when N = 500. For the effects of corrected standard errors, similar results to the ARX() case are obtained. Namely, Windmeijer (005) s correction is quite useful, and in many cases it leads to accurate inference although the corrections do result in severely under-sized tests in some cases. Also, this correction does not seem that helpful in mitigating the size problem of the system GMM estimator when τ is large. The standard errors of Newey and Windmeijer (009) used for the CU-GMM estimators are not always helpful: although they improve the size property in some cases, they have almost no effects in some cases or worsen the test sizes in other cases. Size and power of the tests in the case of experiments with γ = 0.9 are summarized in Table 4, and show significant size distortions in many cases. The size distortion of the transformed likelihood 0

gets reduced for relatively large sample sizes and its size declines to 7.7% when τ =, N > 50 and T > 5. As to be expected, increasing the variance ratio, τ, to 5, does not change this result. A similar pattern can also be seen in the case of GMM-DIF estimators if we consider τ =. But the size results are much less encouraging if we consider the system GMM estimators. Also, as to be expected, size distortions of GMM type estimators become much more pronounced when the variance ratio is increased to τ = 5. Finally, we consider the small sample performance of the weak instruments robust tests which are provided in a supplement, to save space. These results show that size distortions are reduced only when N is large (N = 500). In general, size distortions of these tests get worse as T, or the number of moment conditions, increases. In terms of power, although LM(SYS) and CLR(SYS) tests have almost the same power as the transformed likelihood estimator when γ = 0.4, T = 5, N = 500 and τ =, their powers decline when τ = 5, unlike the transformed likelihood estimator which is invariant to changes in τ. For the case of γ = 0.9, the results are very similar to the case of γ = 0.4. Size distortions are small only when N is large. When N is small, there are substantial size distortions. 6 Concluding remarks In this paper, we proposed the transformed likelihood approach to estimation and inference in dynamic panel data models with cross-sectionally heteroskedastic errors. It is shown that the transformed likelihood estimator by Hsiao et al. (00) continues to be consistent and asymptotically normally distributed, but the covariance matrix of the transformed likelihood estimators must be adjusted to allow for the cross-sectional heteroskedasticity. By means of Monte Carlo simulations, we investigated the finite sample performance of the transformed likelihood estimator and compared it with a range of GMM estimators. Simulation results revealed that the transformed likelihood estimator for an ARX() model with a single exogenous regressor has very small bias and accurate size property, and in most cases outperformed GMM estimators, whose small sample properties vary considerably across parameter values (γ and β), the choice of moment conditions, and the value of the variance ratio, τ. In this paper, x it is assumed to be strictly exogenous. However, in practice, the regressors may be endogenous or weakly exogenous (c.f. Keane and Runkle, 99). To allow for endogenous and weakly exogenous variables, one could consider extending the panel VAR approach advanced in Binder et al. (005) to allow for cross-sectional heteroskedasticity. More specifically, consider the following bivariate model: y it = α yi + γy i,t + βx it + u it x it = α xi + ϕy i,t + ρx i,t + v it where cov(u it, v it ) = θ. In this model, x it is strictly exogenous if ϕ = 0 and θ = 0, weakly exogenous

if θ = 0, and endogenous if θ 0. This model can be written as a PVAR() model as follows ( y it x it ) ( ) ( α yi + βα xi γ + βϕ = + ϕ α xi ) ( βρ ρ y i,t x i,t ) ( ) u it + βv it +, for i =,,..., N. Let A = {a ij }(i, j =, ) be the coefficient matrix of (y i,t, x i,t ) in the above VAR model. Then, we have β = a /a, γ = a a a /a, ρ = a and ϕ = a. Thus, if we estimate a PVAR model in (y it, x it ), allowing for fixed effects and cross-sectional heteroskedasticity, we can recover the parameters of interest, γ and β, from the estimated coefficients of such a PVAR model. However, detailed analysis of such a model is beyond the scope of the present paper and is left to future research. v it A Proofs A. Preliminary results In this appendix we provide some definitions and results useful for the derivations in the paper. Lemma A Let Ω be given by (8). Then the determinant and inverse of Ω are: Ω = g = + T (ω ), (40) T T... T (T ) ω... ω ω Ω = g T. ω (T ) ω (T 3)] (T ) ω (T 3) ω... (T ) ω (T 3) (T ) ω (T ) The generic (t, s)th element of the (T ) (T ) lower block of Ω, denoted by Ω, can be calculated using the following formulas, for t, s =,,..., T : { Ω} Proof. See Hsiao et al. (00). Lemma A Let Φ be defined in (0). We have ts = { s (T t) ω (s ) (T t), (s t) t (T s) ω (t ) (T s), (s > t). (4) Φ = ϑϑ,

where ϑ = (T, T,...,, ) and tr (ΦΩ) = tr ( ϑϑ Ω ) = ϑ Ωϑ = T g, (4) where g is given by (40). Proof. See Hsiao et al. (00). Lemma A3 Let {x i, i =,,..., N} and {z i, i =,,..., N} be two sequences of independently distributed random variables, such that x i z i are independently distributed across i, although x i and z i need not be independently distributed of each other. Then ( N ) ( N )] N ] N ] E x i z i = Cov(x i, z i ) + E (x i ) E (z i ). Lemma A4 Consider the transformed model (7). Under Assumptions and we have E ( W iω ) r i = 0, (i =,,..., N), (43) where Ω is given in (8). Further, E ( r iφ W i ) = ( 0 0 ϖ 0 ), (i =,,..., N), (44) where Φ is given by (0), and ϖ 0. Proof. Let ỹ i, = )(0, y i,..., y i,t ) and note that, for (43) to hold, it is only needed to prove that E ( ỹ i, Ω r i = 0. To show this, let p i = Ω r i = (p i,..., p it ) where by (4) T p i = T v i + (T s + ) u is, s= p it = (T t + )v i + p it = v i + and T h T s u is s= t T h ts u is + k ts u is, (t =,..., T ) s= s=t+ h ts = (T t + ) (s )ω (s )], (45) k ts = (T s + ) (t )ω (t )]. 3

Then, we have E ỹ i, Ω r i ] = = = T T E p it y i,t ] = E p it y i,t ] + E p it y i,t ] t= T t= E (T t + )v i y i,t + ] t T h ts u is y i,t + k s u is y i,t t= s= s=t+ +E (p it y i,t ) T T t (T t + )E (v i y i,t ) + h ts E ( u is y i,t ) t= = A + A. t= s= where we used E( u is y it ) = 0 for t < s. To derive A and A, we use the followings 0 : σ i E(v i y it ) = σ i E( u is y it ) = { ω t = γ t (γω ) t =,..., T t = s ( γ) s = t ( γ) γ t s s < t (46) (47) Using (46) and (47), we have A = σ i (T )ω + (γω ) ] T (T t + )γ t 3, (48) t=3 0 These results are obtained by noting that y it can be written as follows y i = b + π x i + v i, y it = γ t y i + β ( t ) t γ j x i,t j + γ j u i,t j = γ t ( b + π x i ) + γ t v i + β ( t ) t γ j x i,t j + γ j u i,t j, (t =,..., T ). 4