Generalized Method of Moments (GMM) Estimation

Similar documents
GMM estimation is an alternative to the likelihood principle and it has been

GMM estimation is an alternative to the likelihood principle and it has been

1 Outline. 1. Motivation. 2. SUR model. 3. Simultaneous equations. 4. Estimation

Linear Regression with Time Series Data

Economic modelling and forecasting

Econometrics of Panel Data

Maximum Likelihood (ML) Estimation

Generalized Method of Moments: I. Chapter 9, R. Davidson and J.G. MacKinnon, Econometric Theory and Methods, 2004, Oxford.

The generalized method of moments

GMM, HAC estimators, & Standard Errors for Business Cycle Statistics

GMM and SMM. 1. Hansen, L Large Sample Properties of Generalized Method of Moments Estimators, Econometrica, 50, p

ORTHOGONALITY TESTS IN LINEAR MODELS SEUNG CHAN AHN ARIZONA STATE UNIVERSITY ABSTRACT

An estimate of the long-run covariance matrix, Ω, is necessary to calculate asymptotic

Econometrics II - EXAM Outline Solutions All questions have 25pts Answer each question in separate sheets

Economics 536 Lecture 7. Introduction to Specification Testing in Dynamic Econometric Models

Nonlinear GMM. Eric Zivot. Winter, 2013

Chapter 1. GMM: Basic Concepts

Final Exam. Economics 835: Econometrics. Fall 2010

Introduction to Estimation Methods for Time Series models. Lecture 1

Instrumental Variables

Spring 2017 Econ 574 Roger Koenker. Lecture 14 GEE-GMM

GMM Estimation and Testing

Generalized Method of Moment

Chapter 11 GMM: General Formulas and Application

Estimation of Dynamic Regression Models

Lecture 6: Dynamic panel models 1

Chapter 2. GMM: Estimating Rational Expectations Models

Introduction to Estimation Methods for Time Series models Lecture 2

Linear Regression with Time Series Data

Testing in GMM Models Without Truncation

Chapter 6: Endogeneity and Instrumental Variables (IV) estimator

Chapter 6. Panel Data. Joan Llull. Quantitative Statistical Methods II Barcelona GSE

Binary Models with Endogenous Explanatory Variables

Estimating Deep Parameters: GMM and SMM

Linear models. Linear models are computationally convenient and remain widely used in. applied econometric research

MLE and GMM. Li Zhao, SJTU. Spring, Li Zhao MLE and GMM 1 / 22

Chapter 2. Dynamic panel data models

Time Series Analysis

Single Equation Linear GMM with Serially Correlated Moment Conditions

Lecture 4: Heteroskedasticity

Applied Econometrics (MSc.) Lecture 3 Instrumental Variables

Single Equation Linear GMM with Serially Correlated Moment Conditions

ECON 4160, Spring term 2015 Lecture 7

1 Estimation of Persistent Dynamic Panel Data. Motivation

1 Motivation for Instrumental Variable (IV) Regression

Business Economics BUSINESS ECONOMICS. PAPER No. : 8, FUNDAMENTALS OF ECONOMETRICS MODULE No. : 3, GAUSS MARKOV THEOREM

Monte Carlo Simulations and the PcNaive Software

ECONOMETRICS II (ECO 2401S) University of Toronto. Department of Economics. Spring 2013 Instructor: Victor Aguirregabiria

DSGE Methods. Estimation of DSGE models: GMM and Indirect Inference. Willi Mutschler, M.Sc.

Instrumental Variables and GMM: Estimation and Testing. Steven Stillman, New Zealand Department of Labour

Asymptotic distribution of GMM Estimator

Econometrics. Week 8. Fall Institute of Economic Studies Faculty of Social Sciences Charles University in Prague

GMM - Generalized method of moments

Univariate Time Series Analysis; ARIMA Models

Linear Regression with Time Series Data

Time Series Analysis Fall 2008

The BLP Method of Demand Curve Estimation in Industrial Organization

Spatial Regression. 11. Spatial Two Stage Least Squares. Luc Anselin. Copyright 2017 by Luc Anselin, All Rights Reserved

Panel Data Exercises Manuel Arellano. Using panel data, a researcher considers the estimation of the following system:

Linear Regression. Junhui Qian. October 27, 2014

Econometrics I. Ricardo Mora

Generalized Method of Moments Estimation

Economics 241B Estimation with Instruments

Slide Set 14 Inference Basded on the GMM Estimator. Econometrics Master in Economics and Finance (MEF) Università degli Studi di Napoli Federico II

Cointegration Lecture I: Introduction

Multivariate Regression: Part I

Wooldridge, Introductory Econometrics, 4th ed. Chapter 15: Instrumental variables and two stage least squares

Eksamen på Økonomistudiet 2006-II Econometrics 2 June 9, 2006

MIT Spring 2015

Econometrics II - EXAM Answer each question in separate sheets in three hours

A Course on Advanced Econometrics

Instrumental Variables Estimation in Stata

ECON 4160: Econometrics-Modelling and Systems Estimation Lecture 9: Multiple equation models II

Panel Data Models. Chapter 5. Financial Econometrics. Michael Hauser WS17/18 1 / 63

LECTURE ON HAC COVARIANCE MATRIX ESTIMATION AND THE KVB APPROACH

Motivation Non-linear Rational Expectations The Permanent Income Hypothesis The Log of Gravity Non-linear IV Estimation Summary.

Instrumental Variables and the Problem of Endogeneity

GARCH Models Estimation and Inference

Non-Stationary Time Series and Unit Root Testing

Quick Review on Linear Multiple Regression

Økonomisk Kandidateksamen 2004 (I) Econometrics 2. Rettevejledning

Heteroscedasticity and Autocorrelation

Economics 620, Lecture 14: Lags and Dynamics

Instrumental Variables and Two-Stage Least Squares

Heteroskedasticity and Autocorrelation Consistent Standard Errors

Econometrics - 30C00200

ASSET PRICING MODELS

Birkbeck Working Papers in Economics & Finance

IV and IV-GMM. Christopher F Baum. EC 823: Applied Econometrics. Boston College, Spring 2014

4 Instrumental Variables Single endogenous variable One continuous instrument. 2

Non-Stationary Time Series and Unit Root Testing

Single Equation Linear GMM

Greene, Econometric Analysis (5th ed, 2003)

Dealing With Endogeneity

DSGE-Models. Limited Information Estimation General Method of Moments and Indirect Inference

We begin by thinking about population relationships.

Discrete Dependent Variable Models

Advanced Econometrics

Short Questions (Do two out of three) 15 points each

Economics 620, Lecture 20: Generalized Method of Moment (GMM)

Transcription:

Econometrics 2 Fall 2004 Generalized Method of Moments (GMM) Estimation Heino Bohn Nielsen of29 Outline of the Lecture () Introduction. (2) Moment conditions and methods of moments (MM) estimation. Ordinary least squares (OLS) estimation. Instrumental variables (IVE) estimation. (3) GMM defined in the general case. (4) Specification test. (5) Linear GMM. Generalized instrumental variables (GIVE or 2SLS) estimation. 2of29

Idea of GMM Estimation under weak assumptions; based on so-called moment conditions. Moment conditions are statements involving the data and the parameters. Arise naturally in many contexts. For example: (A) In a regression model, y t = x 0 tβ + t, we might think that E[y t x t ]=x 0 tβ. his implies the moment condition (B) Consider the economic relation E[x t t ]=E[x t (y t x 0 tβ)] = 0. y t = β E[x t+ I t ]+ t = β x t+ +(β (E[x t+ I t ] x t+ )+ t ) {z } u t Under rational expectation, the expectation error, E[x t+ I t ] x t+, should be orthogonal to the information set, I t,andforz t I t we have the moment condition E[z t u t ]=0. 3of29 GMM is a large sample estimator. Desirable properties as. Properties of GMM Consistent under weak assumptions. No distributional assumptions like in maximum likelihood (ML) estimation. Asymptotically efficient in the class of models that uses the same amount of information. Many estimators are special cases of GMM. Unifying framework for comparing estimators. GMM is a nonlinear procedure. We do not need a regression setup E[y t ]=h(x t ; β). We can have E[f(y t,x t ; β)] = 0. 4of29

Moment Conditions and MM Estimation Consider a variable y t with some (possibly unknown) distribution. Assume that the mean µ = E[y t ] exists. We want to estimate µ. We could state the population moment condition: E[y t µ] =0, or E[f(y t,µ)] = 0, where f(y t,µ)=y t µ. he parameter µ is identified by the condition if there is a unique solution, in the sense E[f(y t,µ)] = 0 only if µ = µ 0. 5of29 We cannot calculate E[f(y t,µ)] from an observed sample, y,y 2,..., y t,..., y. Define the sample moment condition as g (µ) = f(y t,µ)= (y t µ) =0. ( ) By Law of Large Numbers, sample moments converge to population moments, g (µ) E[f(y t,µ)] for. ( ) he method of moments estimator, bµ MM,isthesolutionto( ), i.e. bµ MM = y t. he sample average can be seen as a MM estimator. MM estimator is consistent. Under weak regularity conditions ( ) implies bµ MM µ 0. 6of29

OLSasaMMEstimator Consider the regression model with K explanatory variables y t = x 0 tβ + t. Assume no-contemporaneous-correlation (minimum for consistency of OLS): E[x t t ]=E[x t (y t x 0 tβ)] = 0. (K ) K moment conditions for the K parameters in β. Define the sample moments g (β) = (x t (y t x 0 tβ)) = X0 (Y Xβ)= 0. (K ) he MM estimator is given by the solution Ã! bβ MM = x t x 0 t x t y t =(X 0 X) X 0 Y = β b OLS. 7of29 Instrumental Variables as a MM Estimator Consider the regression model y t = x 0 tβ + x 2t β 2 + t, where the K variables in x t are predetermined and x 2t is endogenous: E[x t t ] = 0 (K ) E[x 2t t ] 6= 0. (#) OLS is inconsistent! Assume there exists a variable, z 2t, such that corr(x 2t,z 2t ) 6= 0 E[z 2t t ] = 0 ( ) (##) he new moment condition (##) can replace (#). 8of29

Define µ xt x t = K x 2t and µ xt z t =. K z 2t z t are called instruments. z 2t is the new instrument; the predetermined variables, x t, are instruments for themselves. he K population moment conditions are E[z t t ]=E[z t (y t x 0 tβ)] = 0 (K ). he K corresponding sample moment conditions are g (β) = (z t (y t x 0 tβ)) = Z0 (Y Xβ)= 0. (K ) he MM estimator is given by the unique solution Ã! bβ MM = z t x 0 t z t y t =(Z 0 X) Z {z } 0 Y = β b IV. (K K) 9of29 (Where do Instruments Come From?) Consider the two simple equations c t = β 0 + β y t + β 2 w t + t y t = β 20 + β 2 c t + + β 22 w t + β 23 r t + β 24 τ t + 2t Say that we are only interested in the first equation. Assume that w t is predetermined. If β 2 6=0,theny t is endogenous and E[y t t ] 6= 0. In this setup r t and τ t are possible instruments. We need β 23 and β 24 different from zero and E[(r t,τ t ) t ]=0. In dynamic models we can often used lagged values as instruments. Note,thatinthiscasewehavemorepotentialinstrumentsthanwehaveendogenous variables. his is adressed in GMM. 0 of 29

he GMM Problem Defined Let w t =(y t,x 0 t) 0 be a vector of model variables and let z t be instruments. Consider the R moment conditions E[f(w t,z t,θ)] = 0. Here θ is a K vector and f( ) is a R dimensional vector function. Consider the corresponding sample moment conditions g (θ) = f(w t,z t,θ)=0. When can the R sample moments be used to estimate the K parameters in θ? of 29 R<K No unique solution to g (θ) =0. he parameters are not identified. Order Condition R = K Unique solution to g (θ) =0. Exact identification. his is the MM estimator (OLS, IV). Note, that g (θ) =0is potentially a non-linear problem numerical solution. R>K More equations than parameters. Over-identified case. No solution in general (Z 0 X is a R K matrix). Not optimal to drop moments! Instead, choose θ to make g (θ) as close as possible to zero. 2 of 29

GMM Estimation (R >K) We want to make the R moments g (θ) asclosetozeroaspossible...how? Assume we have a R R symmetric and positive definite weight matrix W. hen we could define the quadratic form Q (θ) =g (θ) 0 W g (θ) ( ). he GMM estimator is defined as the vector that minimizes Q (θ), i.e. n b θgmm (W )=arg min g ( b θ) 0 W g ( b o θ). bθ he matrix W tells how much weight to put on each moment condition. Different W give different estimators, b θgmm (W ). GMM is consistent for any weight matrix, W. What is the optimal choice of W? 3 of 29 Optimal GMM Estimation he R sample moments g (θ) are estimators of E[f( )]; and random variables. he law of large numbers implies: g (θ) E[f( )] for. A central limit theorem implies: g (θ) N(0,S), where S istheasymptoticvarianceofthemoments, g (θ). Intuitively, moments with little variance should have large weights. he optimal weight matrix for GMM is a matrix W opt such that plim W opt = W opt = S. 4 of 29

Without autocorrelation, a natural estimator S b of S is h bs = V g (θ)i = V [g (θ)] " # = V f(w t,z t,θ) = f(w t,z t,θ)f(w t,z t,θ) 0. his implies that W opt = b S = Ã! f(w t,z t,θ)f(w t,z t,θ) 0. Note, that W opt depends on θ in general. 5 of 29 wo-step (efficient) GMM. Estimation in Practice () Choose some initial weight matrix W []. E.g. W [] = I or W [] =(Z 0 Z). Find a (consistent) estimate b θ[] =argmin θ Estimate the optimal weights, W opt. (2) Find the optimal GMM estimate g (θ) 0 W [] g (θ). b θgmm =argmin θ g (θ) 0 W opt g (θ). Iterated GMM. Start with some initial weight matrix W []. () Find an estimate b θ []. (2) Find a new weight matrix, W opt [2]. Iterate between b θ [ ] and W opt [ ] until convergence. 6 of 29

Properties of Optimal GMM he GMM estimator, b θ GMM (W opt ), is asymptotically efficient. Lowest variance in a class of models that uses same information. he GMM estimator is asymptotically normal, i.e. ³ bθgmm θ N (0,V), where V = D 0 W opt D = D 0 S D D = plim g (θ) θ 0 (R K). S measures the variance of the moments. he larger S the larger V. D measures the sensitivity of the moments wrt. changes in θ. If this is large the parameter can be estimated precisely. Little is known in finite samples. 7 of 29 Specification est If R>K, we have more moments than parameters. All moments have expectation zero. In a sense K moments are zero by estimating the parameters. additional R K moments are close to zero. If not, some orthogonality condition is violated. est if the Remember, that g (θ) N(0,S). his implies that if the weights are optimal, W opt S,then ξ = g ( b θ GMM ) 0 µ S g ( b θ GMM ) = g ( b θ GMM ) 0 W opt g ( b θ GMM ) χ 2 (R K). Hansen test for overidentifying restrictions. (J-test, Sargan test). A test for R K overidentifying conditions. 8 of 29

Famous Example: Hansen and Singleton (982) Consider an optimizing agent with a power utility function on consumption, U(C t )= C γ t γ.hefirst order condition for maximizing the discounted utility of future concumption is given by " µ # γ Ct+ E δ ( + r t+) I t =0, where I t is the conditioning information set a time t. C t Assume rational expectations. Now if z t I t, then it must be orthogonal to the expectation error, i.e. "Ã µ # γ Ct+ f(c t+,c t,r t+ ; z t ; δ, γ) =E δ ( + r t+)! z t =0. his is a moment condition. We need at least R =2instruments in z t. Note: Specification is theory driven, nonlinear, and not in regression format. C t 9 of 29 Linear GMM and GIVE Consider the linear regression model with K explanatory variables y t = x 0 tβ + x 0 2tβ 2 + t = x 0 tβ + t, where E[x t t ]=0, but the variables in x 2t are endogenous E[x 2t t ] 6= 0. Assume there exist R>K instruments z t =(x 0,z 0 2), suchthat E[z t t ]=E[z t (y t x 0 tβ)] = 0 (R ). Identification requires a non-zero correlation of z 2t and x 2t. Rank condition. he sample moments are g (β) = (z t (y t x 0 tβ)) = Z0 (Y Xβ). Note, that we cannot solve g (β) =0directly. Z 0 X is a R K matrix (of rank K) and cannot be inverted. 20 of 29

Instead we want to minimize the quadratic form Q (β) = g (β) 0 W g (β) µ 0 µ = Z0 (Y Xβ) W Z0 (Y Xβ) = 2 Y 0 Z β 0 X 0 Z W (Z 0 Y Z 0 Xβ) = 2 Y 0 ZW Z 0 Y 2β 0 X 0 ZW Z 0 Y + β 0 X 0 ZW Z 0 Xβ. o minimize Q (β) wetakethederivativeandsolvethek equations Q (β) β = 0 (K ). 2 of 29 he GMM estimator solves the K equations i.e. Q (β) β = 2 Y 0 ZW Z 0 Y 2β 0 X 0 ZW Z 0 Y + β 0 X 0 ZW Z 0 Xβ β = 2 2 X 0 ZW Z 0 Y +2 2 X 0 ZW Z 0 Xβ = 0 bβ GMM (W )=(X 0 ZW Z 0 X) X 0 ZW Z 0 Y. he optimal weight matrix is the inverse variance of the moments, i.e. where S = V W opt = S, h g (θ)i = V [Z0 ]= E [Z0 0 Z]= Z0 ΩZ, where we define E[ 0 ]=Ω. 22 of 29

Case : Homoscedastic Errors If E[ 0 ]=Ω = σ 2 I, the natural estimator of S is bs = Z0 b ΩZ = bσ2 Z 0 Z, where bσ 2 is a consistent estimator for σ 2. hen the GMM estimator becomes bβ GMM = ³ X 0 ZS b Z 0 X X 0 ZS b Z 0 Y = Ã µ! µ X 0 Z bσ2 Z 0 Z Z 0 X X 0 Z bσ2 Z 0 Z Z 0 Y ³ = X 0 Z (Z 0 Z) Z 0 X X 0 Z (Z 0 Z) Z 0 Y = b β GIV E = b β 2SLS Under homoscedasticity the optimal GMM estimator is GIVE (and 2SLS). 23 of 29 Recall, that bβ GMM N µ β, D 0 W opt D. he derivative is given by D = g (β) (R K) β 0 = Z0 (Y Xβ) β 0 = Z0 X = z t x 0 t. hevariancecanbeestimatedby h i V bβgmm = D 0 W opt D Ã µ = 0 µ µ Z0 X bσ2 Z 0 Z! Z0 X known from 2SLS. = bσ 2 (X 0 Z (Z 0 Z) Z 0 X), 24 of 29

he specification test is given by ξ = g ( b θ GMM ) 0 S b g ( b θ GMM ) Ã! 0 Ã! Ã bσ 2! = b t z t z t zt 0 b t z t Ã! 0 Ã! Ã! = b t z t bσ 2 z t zt 0 b t z t χ 2 (R K). In the linear case it is denoted the Sargan test. A simple way to calculate ξ is to consider the regression b t = z 0 tγ + residual, and calculate the test statistic as ξ = R 2. 25 of 29 Case 2: Heteroscedasticity, No Autocorrelation Inthecaseofheteroscedasticity but no autocorrelation, σ 2 0 0 E[ 0 ]=Ω = 0 σ 2 2..... 0, 0 0 σ 2 we can use the estimator bs = Z0 b ΩZ = b 2 tz t zt. 0 We only need an estimate of the K K matrix Z 0 ΩZ and not the matrix Ω. Weget β GMM ( S b ³ )= X 0 ZS b Z 0 X X 0 ZS b Z 0 Y. Note, that a constant in the weight is not important for the estimation. 26 of 29

hevarianceoftheestimatorbecomes h i V bβgmm = D 0 W opt D Ã!Ã! Ã = x t zt 0 b 2 tz t zt 0 Ã! Ã = x t zt 0 b 2 tz t zt 0 z t xt! 0, z t xt! 0 which is the heteroscedasticity consistent variance estimator of White. 27 of 29 Case 3: Autocorrelation he weight-matrix is W opt = S b,where " h # bs = V g (θ)i = V (z t t ). With autocorrelation, we need to take into account the covariances. his is done by the heteroscedasticity and autocorrelation consistent (HAC) estimator. Let Γ j = cov (z t t,z t j t j )=E[(z t t )(z t j t j ) 0 ] (R R) beacovariancematrixforlagj. hen S = {V (z t t )+Cov(z t t,z t t )+Cov(z t t,z t 2 t 2 )+... +Cov(z t t,z t+ t+ )+Cov(z t t,z t+2 t+2 )+...} = X Γ j. j= 28 of 29

If we can argue that Γ j =0for j larger than some lag, q, we can use the estimator qx bs = bγ j, where we estimate the covariances by j= q bγ j = (z t b t )(z t j b t j ) 0. t=j+ he obtained b S is not necessarily positive definite. Instead the covariances can be given decreasing weights, Newey-West estimator. Finite sample properties are unknown. he HAC covariance estimator can also be used for OLS. 29 of 29