Economics 536 Lecture 7. Introduction to Specification Testing in Dynamic Econometric Models

Similar documents
Econometrics. Week 4. Fall Institute of Economic Studies Faculty of Social Sciences Charles University in Prague

F9 F10: Autocorrelation

Econometrics Summary Algebraic and Statistical Preliminaries

Empirical Market Microstructure Analysis (EMMA)

Lecture 6: Dynamic Models

AUTOCORRELATION. Phung Thanh Binh

Spring 2017 Econ 574 Roger Koenker. Lecture 14 GEE-GMM

Økonomisk Kandidateksamen 2004 (I) Econometrics 2. Rettevejledning

Econometrics II - EXAM Answer each question in separate sheets in three hours

Lecture 4: Heteroskedasticity

Auto correlation 2. Note: In general we can have AR(p) errors which implies p lagged terms in the error structure, i.e.,

Bootstrap Testing in Econometrics

Formulary Applied Econometrics

Diagnostics of Linear Regression

Ch.10 Autocorrelated Disturbances (June 15, 2016)

Short Questions (Do two out of three) 15 points each

Econometrics - 30C00200

Heteroskedasticity and Autocorrelation

Applied Econometrics (MSc.) Lecture 3 Instrumental Variables

Econometrics. Week 8. Fall Institute of Economic Studies Faculty of Social Sciences Charles University in Prague

Reliability of inference (1 of 2 lectures)

Empirical Economic Research, Part II

LECTURE 10: MORE ON RANDOM PROCESSES

1 Introduction to Generalized Least Squares

Econometrics. 9) Heteroscedasticity and autocorrelation

G. S. Maddala Kajal Lahiri. WILEY A John Wiley and Sons, Ltd., Publication

Dynamic Regression Models (Lect 15)

Outline. Nature of the Problem. Nature of the Problem. Basic Econometrics in Transportation. Autocorrelation

Econometrics I. Professor William Greene Stern School of Business Department of Economics 25-1/25. Part 25: Time Series

Econometrics of Panel Data

The regression model with one stochastic regressor (part II)

Panel Data Models. Chapter 5. Financial Econometrics. Michael Hauser WS17/18 1 / 63

Linear Regression with Time Series Data

Lecture 10: Panel Data

Economics 582 Random Effects Estimation

Econometrics. Week 11. Fall Institute of Economic Studies Faculty of Social Sciences Charles University in Prague

Freeing up the Classical Assumptions. () Introductory Econometrics: Topic 5 1 / 94

Finite Sample Performance of A Minimum Distance Estimator Under Weak Instruments

Linear Regression. Junhui Qian. October 27, 2014

Bootstrapping Heteroskedasticity Consistent Covariance Matrix Estimator

ORTHOGONALITY TESTS IN LINEAR MODELS SEUNG CHAN AHN ARIZONA STATE UNIVERSITY ABSTRACT

Generalized Method of Moments (GMM) Estimation

ECON3327: Financial Econometrics, Spring 2016

ECON 4160, Spring term Lecture 12

Testing Restrictions and Comparing Models

E 4160 Autumn term Lecture 9: Deterministic trends vs integrated series; Spurious regression; Dickey-Fuller distribution and test

Lecture: Simultaneous Equation Model (Wooldridge s Book Chapter 16)

10. Time series regression and forecasting

GARCH Models Estimation and Inference

Lecture 2: Univariate Time Series

LECTURE 13: TIME SERIES I

Non-Stationary Time Series and Unit Root Testing

Cointegration Lecture I: Introduction

Economics 308: Econometrics Professor Moody

Regression with time series

Section 6: Heteroskedasticity and Serial Correlation

Stationary and nonstationary variables

Wooldridge, Introductory Econometrics, 4th ed. Chapter 15: Instrumental variables and two stage least squares

VAR Models and Applications

Reading Assignment. Serial Correlation and Heteroskedasticity. Chapters 12 and 11. Kennedy: Chapter 8. AREC-ECON 535 Lec F1 1

Econ 510 B. Brown Spring 2014 Final Exam Answers

Linear Regression with Time Series Data

Intermediate Econometrics

Introductory Econometrics

Recent Advances in the Field of Trade Theory and Policy Analysis Using Micro-Level Data

ECON 312 FINAL PROJECT

Economics 620, Lecture 13: Time Series I

Statistics 910, #5 1. Regression Methods

Section 2 NABE ASTEF 65

Prof. Dr. Roland Füss Lecture Series in Applied Econometrics Summer Term Introduction to Time Series Analysis

Advanced Econometrics

Vector Auto-Regressive Models

Review of Classical Least Squares. James L. Powell Department of Economics University of California, Berkeley

Answers to Problem Set #4

Short T Panels - Review

Linear Regression with Time Series Data

(c) i) In ation (INFL) is regressed on the unemployment rate (UNR):

Non-Stationary Time Series and Unit Root Testing

Darmstadt Discussion Papers in Economics

Likely causes: The Problem. E u t 0. E u s u p 0

The outline for Unit 3

Efficient Estimation of Dynamic Panel Data Models: Alternative Assumptions and Simplified Estimation

LECTURE 2 LINEAR REGRESSION MODEL AND OLS

MS&E 226: Small Data. Lecture 11: Maximum likelihood (v2) Ramesh Johari

Econ 423 Lecture Notes: Additional Topics in Time Series 1

Answer all questions from part I. Answer two question from part II.a, and one question from part II.b.

Iris Wang.

Title. Description. var intro Introduction to vector autoregressive models

Non-Stationary Time Series and Unit Root Testing

Økonomisk Kandidateksamen 2004 (II) Econometrics 2 June 14, 2004

11.1 Gujarati(2003): Chapter 12

Economic modelling and forecasting

MEI Exam Review. June 7, 2002

A time series is called strictly stationary if the joint distribution of every collection (Y t

Peter Hoff Linear and multilinear models April 3, GLS for multivariate regression 5. 3 Covariance estimation for the GLM 8

Introduction to Eco n o m et rics

Lecture 3: Multiple Regression

1 Outline. 1. Motivation. 2. SUR model. 3. Simultaneous equations. 4. Estimation

Missing dependent variables in panel data models

Econometrics II - EXAM Outline Solutions All questions have 25pts Answer each question in separate sheets

Transcription:

University of Illinois Fall 2016 Department of Economics Roger Koenker Economics 536 Lecture 7 Introduction to Specification Testing in Dynamic Econometric Models In this lecture I want to briefly describe some techniques for evaluating dynamic econometric models like the models for gasoline demand you have been estimating. Until now, we have implicitly assumed that these models satisfied the classical assumptions of the Gaussian linear model. In particular, we have assumed that the error sequences {u t } were iid and approximately Gaussian, thus justifying the application of elementary least squares methods of estimation. Testing for Autocorrelation We might begin by recalling some basic facts about autocorrelation. In classical regression with fixed regressors, y i = x iβ + u i we know that if the vector, u = (u i ) N (0, σ 2 I) then ˆβ = (X X) 1 X y N (β, σ 2 (X X) 1 ) but when the errors are autocorrelated, for example, u N (0, Ω), then ˆβ N (β, (X X) 1 X ΩX(X X) 1 ) and therefore the conventional estimates of standard errors from ordinary least squares regression may badly misrepresent the true precision of ˆβ. Of course, in this case it is preferable to use ˇβ = (X Ω 1 X) 1 X Ω 1 y N (0, (X Ω 1 X) 1 ) which, in effect, restores the model to the original iid structure and thereby achieves optimality. DiNardo and Johnston give an example, a standard 1

one, showing that the precision of ˇβ can be considerably greater than that of ˆβ, even for modest amounts of autocorrelation. One can estimate deirectly the OLS covariance matrix using a variant of the proposal of Newey and West (1987) where ˆΩ = ˆΓ n 0 + (1 j/(p + 1))(ˆΓ j + ˆΓ j) j=1 ˆΓ j = n 1 n t=j+1 u t u t j x t x t j When dynamic models are specified with lagged endogenous explanatory variables, autocorrelation raises more serious problems. Even in the simplest examples, there are bias as well as efficiency costs to ignoring the autocorrelation. To see this consider the model y t = β 0 y t 1 + β 1 x t + u t where u t = ρu t 1 + ε t. with {ε t } assumed to be iid N (0, σ 2 ). Consistent estimation of β requires orthogonality of u t and the explanatory variables (y t 1, x t ), but note that Ey t 1 u t = E(β 0 y t 2 + β 1 x t 1 + u t 1 )(ρu t 1 + ε t ) = ρσ 2 + β 0 ρe(y t 2 u t 1 ) But, by stationarity, a concept introduced in the next lecture, Ey t 1 u t = Ey t 2 u t 1 so Ey t 1 u t = ρσ 2 u/(1 β 0 ρ) Not surprisingly, given the serious consequences of this bias effect, there is a large literature on testing for autocorrelation in this context. The most straight forward approach is that of Breusch and Godfrey which derives from work of Durbin. While the details are usually reserved for Ec575, the Breusch and Godfrey test is easily described. Let û t = y t x t ˆβ, t = 1,, n denote the residuals from a least squares regression in which the vector x t may include lagged endogenous variables. Suppose that we wish to test the hypothesis H o : ρ 1 = ρ 2 = = ρ s = 0 2

in the potential autocorrelation model u t = ρ 1 u t 1 + + ρ s u t s. Consider the auxiliary regression equation and the associated test statistic û t = ρ 1 û t 1 + + ρ s û t s + x tγ + v t T n = nr 2 based on the conventional R 2 of the auxiliary regression. Under H 0, T n χ 2 s, providing a rather general testing strategy for this important class of models. Note that it is crucial that the variables x t appear in the auxiliary regression, even though in some superficially similar circumstances this is found to be unnecessary. Digression on R 2 asymptotics The connection between R 2 and F is an important aspect of trying to interpret nr 2 as a reasonable test statistic. For a general linear hypothesis, say Rβ = r, in the regression setting, let S ω and S Ω denote the restricted and unrestricted sums of squared residuals, and F = (S ω S Ω )/q S Ω /(n p) and we can define an R 2 of ω relative to Ω as, thus n p q R 2 = 1 S Ω S ω R 2 1 R 2 = (S ω S Ω )/q S Ω /(n p) Under the null hypothesis we have suggested that nr 2 can be used as a asymptotically valid test statistic which has χ 2 q behavior under the null, where q = rank (R). What is the connection of this to the foregoing algebra which relates the finite sample value of R 2 to the appropriate F statistic for testing H 0? Note first that under H 0 S Ω /(n p) σ 2 and S ω /(n p) σ 2 so for n large, nr 2 = S ω S Ω S ω /n 3 S ω S Ω S Ω /n

and therefore nr 2 is approximately equal to the numerator χ 2 q of the F statistic. Now to make the connection between χ 2 q and F we need only note that χ 2 q/q F q,. Alternatively, we may observe that under H 0, R 2 0 so n p q R 2 1 R 2 n q R2 The next obvious question is: what do we do if the Breusch-Godfrey test rejects H 0? There are two general approaches which I will describe briefly. Nonlinear Models Dynamic models of the type we have been discussing can always be written in the form y t = D(L)x t + u t where D(L) = B(L)/A(L) is called a rational lag polynomial. As an example, take the simple model ( ) y t = αy t 1 + βx t + u t with u t = ρu t + ε t and {ε t } iid N (0, σ 2 ). Subtracting ρy t 1 = ραy t 2 + ρβx t 1 + ρu t from (*) we have y t = (ρ + α)y t 1 ραy t 2 + βx t βρx t 1 + ε t. Since this version of the model has a nice (iid) error structure, it can be consistently estimated by ordinary least squares. Note, however, that we now have 4 parameters not 3, as in the original formulation of the model. Since these 4 parameters are simple functions of the original 3 parameters, we can impose the implied constraints and estimate the model by nonlinear least squares. A related approach, which we may also illustrate with this simple model, is to write the original model in the form y t = δ j x t j + v t. j=0 This form may appear impractical since we have an infinite number of δ j s, but as we have seen in the second lecture these δ s may be expressed in terms of a finite number of α s and β s. Harvey (1989) discusses several versions of this in some simple models and the resulting nonlinear least squares estimation strategy. 4

Instrumental Variable Estimation An alternative strategy for estimating models of this type relies on instrumental variables. Recall that our fundamental problem, the bias resulting from autocorrelation in dynamic models, was attributed to the lack of orthogonality between errors and lagged endogenous variables. An obvious strategy for dealing with this problem is to identify suitable instrumental variables (IV s) which have the properties: orthogonality with u, i.e., EZ u = 0 Mutual association with X, i.e., E ˆX X positive definite, where ˆX = Z(Z Z) 1 Z X is the projection of X onto the column space of Z. The immediate question is where would such variables, Z, come from? In the case of dynamic models of the type we have been discussing it is quite straightforward to answer this question. They may be chosen to be lagged exogenous variables. But this question changes a worry that there might be too few IV s into a new worry that there might be too many. This question does not have a good analytical answer and remains a question of active research concern. Nevertheless, empirical research has suggested some rules of thumb which can be used in applications. Generally, we would suggest that the number of IV s be kept to some modest number above the number of existing explanatory variables, say, p q 2p where p is the parametric dimension of the original model and q is the number of IV s. We will return to this question later in the course. ARCH in Brief Frequently, we observe (particularly in financial data) time-varying heteroscedasticity. In the early 80 s Engel coined the term autoregressive conditional heteroscedasticity ARCH to refer to model in which V (u t past ) = h t = α 0 + α 1 u 2 t 1 + + α p u 2 t p In subsequent work this has been generalized in several directions notably to GARCH (1,1), V (u t past ) = h t = α 0 + α 1 u 2 t 1 + γ 1 h t 1 These are simple examples of a broad class of nonlinear time series models. In the ARCH model we have unconditional expectations, 1.) Eu t = Eh t ε t = 0 2.) V u t = α 0 /A(1) where A(L) = (1 + α 1 L + + α p L p ) 5

As long as the lag polynomial A(L) is stable, i.e., if the roots of A(z) = 0 lie outside the unit circle, then we get a gradual oscillation of h t around the unconditional variance. In integrated ARCH, i.e., roots on the unit circle, we get long swings away from the initial h t. Testing for ARCH Naive LM tests can be implemented just like LM-AR tests. Regress û 2 t on {û 2 t 1,, û 2 t q and x t } compute nr 2 compare to χ 2 q or nr 2 /q compare to F q,n p. Note that in this form the test may be regarded as a joint test for ARCH and heteroscedasticity of the form usually tested by the Breusch- Pagan, and related tests. One could obviously consider refining the hypothesis under consideration in light of the results obtained for this expanded version of the test. The LM Principle Let l( θ) denote log likelihood evaluated at mle under H 0 : θ Θ 0 Θ 1 we are interested in testing H 0 vs H 1 : θ Θ 1. One way to do this is to ask how does l change as we move the restricted estimator θ Θ 0 toward the unrestricted ˆθ Θ 1. To explore this we need to say something about nonlinear optimization, but this would take us too far away from the main topic. Sufficient to say that the LM-test is based on the magnitude of the gradient of l at ˆθ Θ 0 in the direction of ˆθ 1 Θ 1. Again, there are questions to be answered about what to do when ARCH effects are found to be present in the model. Joint estimation of ARCH and regression shift effects is the preferred solution when it is computationally feasible. Various iterative solutions are obviously available as alternatives. 6