Instrumental Variables and Two-Stage Least Squares

Similar documents
PANEL DATA RANDOM AND FIXED EFFECTS MODEL. Professor Menelaos Karanasos. December Panel Data (Institute) PANEL DATA December / 1

Chapter 6: Endogeneity and Instrumental Variables (IV) estimator

Chapter 2. Dynamic panel data models

1. The Multivariate Classical Linear Regression Model

Repeated observations on the same cross-section of individual units. Important advantages relative to pure cross-section data

Economics 241B Estimation with Instruments

Linear Model Under General Variance

Economics 620, Lecture 7: Still More, But Last, on the K-Varable Linear Model

A Course on Advanced Econometrics

ECONOMET RICS P RELIM EXAM August 24, 2010 Department of Economics, Michigan State University

Introductory Econometrics

Introduction to Econometrics. Heteroskedasticity

2014 Preliminary Examination

Quantitative Techniques - Lecture 8: Estimation

the error term could vary over the observations, in ways that are related

Intermediate Econometrics

The regression model with one stochastic regressor (part II)

Lecture: Simultaneous Equation Model (Wooldridge s Book Chapter 16)

We begin by thinking about population relationships.

ECONOMETRICS II (ECO 2401S) University of Toronto. Department of Economics. Spring 2013 Instructor: Victor Aguirregabiria

11. Bootstrap Methods

When is it really justifiable to ignore explanatory variable endogeneity in a regression model?

1 Correlation between an independent variable and the error

Econ 510 B. Brown Spring 2014 Final Exam Answers

Föreläsning /31

Notes on Generalized Method of Moments Estimation

1 Introduction to Generalized Least Squares

Linear Regression. Junhui Qian. October 27, 2014

Econometrics - 30C00200

ECONOMETRICS FIELD EXAM Michigan State University May 9, 2008

Econometrics Summary Algebraic and Statistical Preliminaries

Suggested Solution for PS #5

x i = 1 yi 2 = 55 with N = 30. Use the above sample information to answer all the following questions. Show explicitly all formulas and calculations.

Linear Regression. y» F; Ey = + x Vary = ¾ 2. ) y = + x + u. Eu = 0 Varu = ¾ 2 Exu = 0:

Heteroskedasticity ECONOMETRICS (ECON 360) BEN VAN KAMMEN, PHD

Econometrics. 7) Endogeneity

GLS and FGLS. Econ 671. Purdue University. Justin L. Tobias (Purdue) GLS and FGLS 1 / 22

Strength and weakness of instruments in IV and GMM estimation of dynamic panel data models

Environmental Econometrics

Generalized Method of Moments (GMM) Estimation

The Simple Regression Model. Simple Regression Model 1

Simple Linear Regression: The Model

Regression and Statistical Inference

Lecture Notes on Measurement Error

Statistics 910, #5 1. Regression Methods

1. You have data on years of work experience, EXPER, its square, EXPER2, years of education, EDUC, and the log of hourly wages, LWAGE

Heteroscedasticity and Autocorrelation

Panel Data. March 2, () Applied Economoetrics: Topic 6 March 2, / 43

plim W 0 " 1 N = 0 Prof. N. M. Kiefer, Econ 620, Cornell University, Lecture 16.

Chapter 6. Panel Data. Joan Llull. Quantitative Statistical Methods II Barcelona GSE

Instrumental Variables and GMM: Estimation and Testing. Steven Stillman, New Zealand Department of Labour

Applied Econometrics (MSc.) Lecture 3 Instrumental Variables

Heteroskedasticity. y i = β 0 + β 1 x 1i + β 2 x 2i β k x ki + e i. where E(e i. ) σ 2, non-constant variance.

Spatial Regression. 3. Review - OLS and 2SLS. Luc Anselin. Copyright 2017 by Luc Anselin, All Rights Reserved

OLS, MLE and related topics. Primer.

1 Estimation of Persistent Dynamic Panel Data. Motivation

Heteroskedasticity. We now consider the implications of relaxing the assumption that the conditional

Multivariate Regression: Part I

Microeconometrics: Clustering. Ethan Kaplan

Economics 620, Lecture 20: Generalized Method of Moment (GMM)

Sensitivity of GLS estimators in random effects models

Econ 836 Final Exam. 2 w N 2 u N 2. 2 v N

Topic 10: Panel Data Analysis

Topic 7: Heteroskedasticity

Econometrics Multiple Regression Analysis: Heteroskedasticity

A time series plot: a variable Y t on the vertical axis is plotted against time on the horizontal axis

Dealing With Endogeneity

1 The Multiple Regression Model: Freeing Up the Classical Assumptions

REVIEW (MULTIVARIATE LINEAR REGRESSION) Explain/Obtain the LS estimator () of the vector of coe cients (b)

Finnancial Development and Growth

4.8 Instrumental Variables

Econometrics Midterm Examination Answers

Advanced Econometrics

Iris Wang.

Wooldridge, Introductory Econometrics, 4th ed. Chapter 15: Instrumental variables and two stage least squares

Panel Data Models. Chapter 5. Financial Econometrics. Michael Hauser WS17/18 1 / 63

LECTURE 11: GENERALIZED LEAST SQUARES (GLS) In this lecture, we will consider the model y = Xβ + ε retaining the assumption Ey = Xβ.

Economics 620, Lecture 14: Lags and Dynamics

Multivariate Regression Analysis

Econometric Analysis of Cross Section and Panel Data

Chapter 14. Simultaneous Equations Models Introduction

6.435, System Identification

ECON 4230 Intermediate Econometric Theory Exam

Econometrics. Week 4. Fall Institute of Economic Studies Faculty of Social Sciences Charles University in Prague

IV Estimation and its Limitations: Weak Instruments and Weakly Endogeneous Regressors

Least Squares Estimation-Finite-Sample Properties

Business Economics BUSINESS ECONOMICS. PAPER No. : 8, FUNDAMENTALS OF ECONOMETRICS MODULE No. : 3, GAUSS MARKOV THEOREM

Introductory Econometrics

Simultaneous Equation Models

CLUSTER EFFECTS AND SIMULTANEITY IN MULTILEVEL MODELS

The BLP Method of Demand Curve Estimation in Industrial Organization

Simultaneous Equation Models (Book Chapter 5)

13 Endogeneity and Nonparametric IV

Linear Regression with Time Series Data

Instrumental Variables

LECTURE 2 LINEAR REGRESSION MODEL AND OLS

Heteroskedasticity. Part VII. Heteroskedasticity

Linear models. Linear models are computationally convenient and remain widely used in. applied econometric research

MS&E 226: Small Data. Lecture 11: Maximum likelihood (v2) Ramesh Johari

Econometrics I. Professor William Greene Stern School of Business Department of Economics 25-1/25. Part 25: Time Series

Transcription:

Instrumental Variables and Two-Stage Least Squares Generalised Least Squares Professor Menelaos Karanasos December 2011

Generalised Least Squares: Assume that the postulated model is y = Xb + e, (1) where e N(0, σ 2 Ω), where Ω is a positive de nite matrix-this implies that Ω 1 is also a positive de nite matrix. Thus it is possible to nd a nonsingular matrix P such that Ω 1 = P 0 P. or Ω = (P 0 P) 1 = P 1 (P 0 ) 1.

Premultiply the linear model in equation (1) by P, to obtain Py = PXb + Pe (2) Denote Pe by u. Then Var(u) = E (uu 0 ) = E (Pee 0 P 0 ) = PE (ee 0 ) P 0 = σ 2 P {z } {z} Ω P 0 σ 2 Ω P 1 (P 0 ) 1 = σ 2 PP 1 (P 0 ) 1 P 0 = σ 2 I. Thus the transformed variables in equation (2) satisfy the conditions under which OLS is BLUE.

The coe cient estimated vector from OLS regression of Py on PX is the generalized least squares (GLS) estimator: From the OLS theory it follows that β G = [(PX ) 0 PX ] 1 (PX ) 0 Py = (X 0 P 0 P {z} Ω 1 X ) 1 X 0 P 0 Py = (X 0 Ω 1 X ) 1 X 0 Ω 1 y. Var(β G ) = σ 2 [(PX ) 0 PX ] 1 = σ 2 (X 0 P 0 P {z} Ω 1 X ) 1 = σ 2 (X 0 Ω 1 X ) 1.

An unbiased estimator of the unknown σ 2 is readily obtained from the application of OLS to the transformed model. It is s 2 = bu 0 bu/n k where β G = (X 0 Ω 1 X ) 1 X 0 Ω 1 y. = (Py PX β G ) 0 (Py PX β G )/N k = (y X β G ) 0 P 0 P {z} Ω 1 (y X β G )/N k = (y X β G ) 0 Ω 1 (y X β G )/N k,

Note that the procedures outlined so far imply knowledge of Ω. In practice Ω is unknown, and it is important to develop feasible generalized least squares (FGLS). Finally, note that if σ 2 Ω = V, where V is a positive de nite variance-covariance matrix. Then, it follows directly that and β G = (X 0 {z} Ω 1 X ) 1 X 0 {z} Ω 1 y = σ 2 V 1 σ 2 V 1 = (X 0 V 1 X ) 1 X 0 V 1 y, Var(β G ) = σ 2 (X 0 {z} Ω 1 X ) 1 σ 2 V 1 = σ 2 (X 0 V 1 X ) 1.

Feasible GLS procedure: Let V = Further, we hypothesize that 2 6 4 σ 2 1 0..... 0 σ 2 N 3 7 5. σ 2 i = a 0 + a 1 z a 2 i, i = 1,..., N, where z i is a single variable, possibly one of the regressors, thought to determine the heteroscedasticity.

Because the OLS residuals be = y can run the nonlinear regression X β are consistent estimates of e one be 2 i = a 0 + a 1 z a 2 i + v i. Estimates of the disturbance variances are then bσ 2 i = α 0 + α 1 z α 2 i, i = 1,..., N. These estimates give the bv matrix and a feasible GLS procedure.

Instrumental variables (IV) estimators: Consider the relation y i = bx i + e i, (3) where for simplicity the constant term has been dropped. Suppose however, that the observed value x i can be represented as the sum of the true value ex i and a random measurement error v i, that is x i = ex i + v i. In this case the apropriate relation may be y i = bex i + e i (4)

If we assume that equation (4) is the maintained speci cation but that observations are only available on x i and not on ex i, what happens if we use OLS? The OLS slope is β = y x {z} bex i +e i x(bex + e) x 2 = x 2 = b xex x 2 + xe x 2 ) E (β) = b xex 6= b. x 2 Thus OLS is biased. This is an example of speci cation error.

In this case one should make use of instrumental variables which are also commonly referred to as instruments. Suppose that it is possible to nd a data matrix Z of order N l (l k), which posseses two vital properties: 1. The variables in Z are correlated with those of X 2. The variables in Z are (in the limit) uncorrelated with the disturbance term e. Premultiplying the general relation by Z 0 gives with Z 0 y = {z} {z} Z 0 X b + {z} Z 0 e, y X e Var(e ) = (Z 0 e) = E [(Z 0 e)(z 0 e) 0 ] = E (Z 0 ee 0 Z ) = Z 0 E (ee 0 )Z = σ 2 Z 0 Z.

This suggests the use of the GLS. The resultant estimator is β GLS = β IV = [X {z} 0 Z (Z 0 Z ) 1 {z} Z 0 X ] 1 {z} X 0 Z (Z 0 Z ) 1 Z 0 y (5) {z} (X ) 0 X (X ) 0 y = (X 0 P z X ) 1 X 0 P z y, where P z = Z (Z 0 Z ) 1 Z 0. The variance-covariance matrix is Var(β IV ) = σ 2 [X 0 Z (Z 0 Z ) 1 Z 0 X ] 1 = σ 2 (X 0 P z X ) 1 and the distrurbance variance maybe estimated consistenly from bσ 2 = (y X β IV ) 0 (y X β IV )/N. Note that the use of N or N k or N l in the divisor does not matter asymptotically.

Special case: When l = k, that is, when Z contains the same number of columns as X, we have a special case of the foregoing results. Now X 0 Z is k k and nonsingular. This implies that [(X {z} 0 Z (Z 0 Z ) 1 (Z {z} 0 X )] 1 = (X ) 0 X (Z 0 X ) 1 (Z 0 Z )(X 0 Z ) 1. Thus the estimator in equation (5) reduces to β IV = (Z 0 X ) 1 (Z 0 Z )(X 0 Z ) 1 (X 0 Z )(Z 0 Z ) 1 Z 0 y = (Z 0 X ) 1 Z 0 y Moreover, Var(β IV ) simpli es to Var(β IV ) = σ 2 [X 0 Z (Z 0 Z ) 1 Z 0 X ] 1 = σ 2 (Z 0 X ) 1 (Z 0 Z )(X 0 Z ) 1.

Two-stage least square (2SLS) The IV estimator may also be seen as the result of a double application of least squares: Stage (i): Regress each variable in the X matrix on Z (X = Zd + u) to obtain a matrix of tted values X bx = Z {z} δ (Z 0 Z ) 1 Z 0 X = Z (Z 0 Z ) 1 Z 0 {z } P Z X = P Z X.

Stage (ii): Regress y on bx to obtain the 2SLS estimated β vector β 2SLS = ( bx 0 bx ) 1 bx 0 y = (X 0 PZ 0 P {z } Z X ) {z} 1 X 0 PZ 0 y {z } bx 0 bx 0 bx = (X 0 P Z X )X 0 P Z y = β GLS, since P 0 Z P Z = [Z (Z 0 Z ) 1 Z 0 ] 0 = Z (Z 0 Z ) 1 Z 0 = P Z and P 0 Z P Z = P 2 Z = Z (Z 0 Z ) 1 Z 0 Z (Z 0 Z ) 1 Z 0 = Z (Z 0 Z ) 1 Z 0 = P Z.

Thus the IV estimator can be obtained by a two-stage least-squares procedure. The variance-covariance matrix and the estimated disturbance term are given by Var(β IV ) = σ 2 [X 0 Z (Z 0 Z ) 1 Z 0 X ] 1 = σ 2 (X 0 P z X ) 1 and bσ 2 = (y X β IV ) 0 (y X β IV )/N.

Choice of instruments: The crucial question is, where do we nd the instruments? Some of them are often variables from X matrix itself. Any variables that are thought to be exogenous and indepedent of the disturbance are retained to serve in the Z matrix. When some of the X variables are used as instruments, we may particion X and Z as X = [X 1 X 2 ], Z = [X 1 Z 1 ], where X 1 is of order N r (r < k), X 2 is N (k r), and Z 1 is N (l r).

It can be shown that bx, the matrix of regressors in the second-stage regression, is then bx = [X 1 bx 2 ], and bx 2 = Z (Z 0 Z ) 1 Z 0 X 2, {z } P Z that is bx 2 are the tted values of X 2 obtained from the regression of X 2 on the full set of instruments: X 2 = Z γ + v and bx 2 = Z bγ, bγ = (Z 0 Z ) 1 Z 0 X 2.

There still remains the question of how many instruments to use. The minimum number is k. The asymptotic e ciency increases with the number of isntruments. However, the small sample bias also increases with the number of instruments. If, in fact, we select N instruments, it is simple to show that P Z = I in which case the IV estimator is simply the OLS which is biased and inconsistent. If on the other hand, we use the minimum or close to the minimum, number of instruments, the results may also be poor. It has been shown that the mth moment of the 2SLS estimator exists if and only if m < l k + 1. Thus, if there are just as many instruments as explanatory variables, the 2SLS estimator will not have a mean. With one more instrument there will be a mean but not variance, and so forth.