Economics 620, Lecture 7: Still More, But Last, on the K-Varable Linear Model

Similar documents
Economics 620, Lecture 4: The K-Varable Linear Model I

Economics 620, Lecture 4: The K-Variable Linear Model I. y 1 = + x 1 + " 1 y 2 = + x 2 + " 2 :::::::: :::::::: y N = + x N + " N

Economics 620, Lecture 14: Lags and Dynamics

Models, Testing, and Correction of Heteroskedasticity. James L. Powell Department of Economics University of California, Berkeley

1 The Multiple Regression Model: Freeing Up the Classical Assumptions

Economics 620, Lecture 9: Asymptotics III: Maximum Likelihood Estimation

Semester 2, 2015/2016

Quantitative Techniques - Lecture 8: Estimation

Economics 620, Lecture 15: Introduction to the Simultaneous Equations Model

Economics 620, Lecture 20: Generalized Method of Moment (GMM)

Chapter 8 Heteroskedasticity

Multiple Regression Analysis

Chapter 6: Endogeneity and Instrumental Variables (IV) estimator

the error term could vary over the observations, in ways that are related

Economics 620, Lecture 13: Time Series I

Economics 620, Lecture 5: exp

Economics 620, Lecture 18: Nonlinear Models

Economics 620, Lecture 2: Regression Mechanics (Simple Regression)

Föreläsning /31

Economics 620, Lecture 19: Introduction to Nonparametric and Semiparametric Estimation

Uses of Asymptotic Distributions: In order to get distribution theory, we need to norm the random variable; we usually look at n 1=2 ( X n ).

Instrumental Variables and Two-Stage Least Squares

Multiple Regression Analysis: Heteroskedasticity

Finansiell Statistik, GN, 15 hp, VT2008 Lecture 15: Multiple Linear Regression & Correlation

Graduate Econometrics Lecture 4: Heteroskedasticity

Making sense of Econometrics: Basics

ECONOMET RICS P RELIM EXAM August 24, 2010 Department of Economics, Michigan State University

Econometrics Multiple Regression Analysis: Heteroskedasticity

Lecture 4: Heteroskedasticity

Econometrics Summary Algebraic and Statistical Preliminaries

Heteroskedasticity. We now consider the implications of relaxing the assumption that the conditional

Econometrics. Week 8. Fall Institute of Economic Studies Faculty of Social Sciences Charles University in Prague

Intermediate Econometrics

LECTURE 11: GENERALIZED LEAST SQUARES (GLS) In this lecture, we will consider the model y = Xβ + ε retaining the assumption Ey = Xβ.

ECON 497: Lecture Notes 10 Page 1 of 1

Contest Quiz 3. Question Sheet. In this quiz we will review concepts of linear regression covered in lecture 2.

Heteroskedasticity and Autocorrelation

Lecture Notes on Measurement Error

Econometrics - 30C00200

1 Correlation between an independent variable and the error

A time series plot: a variable Y t on the vertical axis is plotted against time on the horizontal axis

Economics 241B Estimation with Instruments

Heteroskedasticity. y i = β 0 + β 1 x 1i + β 2 x 2i β k x ki + e i. where E(e i. ) σ 2, non-constant variance.

We begin by thinking about population relationships.

Linear Model Under General Variance

Week 11 Heteroskedasticity and Autocorrelation

Applied Econometrics (MSc.) Lecture 3 Instrumental Variables

1 A Non-technical Introduction to Regression

Economics Introduction to Econometrics - Fall 2007 Final Exam - Answers

Introductory Econometrics

Panel Data. March 2, () Applied Economoetrics: Topic 6 March 2, / 43

11. Bootstrap Methods

Linear Regression. y» F; Ey = + x Vary = ¾ 2. ) y = + x + u. Eu = 0 Varu = ¾ 2 Exu = 0:

Spatial Regression. 3. Review - OLS and 2SLS. Luc Anselin. Copyright 2017 by Luc Anselin, All Rights Reserved

13 Endogeneity and Nonparametric IV

Outline. Possible Reasons. Nature of Heteroscedasticity. Basic Econometrics in Transportation. Heteroscedasticity

An Introduction to Parameter Estimation

Testing Linear Restrictions: cont.

x i = 1 yi 2 = 55 with N = 30. Use the above sample information to answer all the following questions. Show explicitly all formulas and calculations.

ECONOMETRICS II (ECO 2401S) University of Toronto. Department of Economics. Spring 2013 Instructor: Victor Aguirregabiria

Topic 7: Heteroskedasticity

ECONOMETRICS FIELD EXAM Michigan State University May 9, 2008

Chapter 2. Dynamic panel data models

Suggested Solution for PS #5

Economics 113. Simple Regression Assumptions. Simple Regression Derivation. Changing Units of Measurement. Nonlinear effects

Inference in Regression Analysis

1 Motivation for Instrumental Variable (IV) Regression

A Course on Advanced Econometrics

Simple Linear Regression

LECTURE 13: TIME SERIES I

Introduction to the Simultaneous Equations Model

Wooldridge, Introductory Econometrics, 2d ed. Chapter 8: Heteroskedasticity In laying out the standard regression model, we made the assumption of

1. The Multivariate Classical Linear Regression Model

Ordinary Least Squares Regression

Max. Likelihood Estimation. Outline. Econometrics II. Ricardo Mora. Notes. Notes

PANEL DATA RANDOM AND FIXED EFFECTS MODEL. Professor Menelaos Karanasos. December Panel Data (Institute) PANEL DATA December / 1

Review of Classical Least Squares. James L. Powell Department of Economics University of California, Berkeley

Variable Selection and Model Building

Simple and Multiple Linear Regression

1. You have data on years of work experience, EXPER, its square, EXPER2, years of education, EDUC, and the log of hourly wages, LWAGE

LECTURE 10. Introduction to Econometrics. Multicollinearity & Heteroskedasticity

Christopher Dougherty London School of Economics and Political Science

Math 423/533: The Main Theoretical Topics

OMISSION OF RELEVANT EXPLANATORY VARIABLES IN SUR MODELS: OBTAINING THE BIAS USING A TRANSFORMATION

Lecture 4: Linear panel models

Introduction: structural econometrics. Jean-Marc Robin

Quantitative Analysis of Financial Markets. Summary of Part II. Key Concepts & Formulas. Christopher Ting. November 11, 2017

Introduction to Econometrics. Heteroskedasticity

The regression model with one fixed regressor cont d

Freeing up the Classical Assumptions. () Introductory Econometrics: Topic 5 1 / 94

Notes on Generalized Method of Moments Estimation

Econometrics I. Professor William Greene Stern School of Business Department of Economics 3-1/29. Part 3: Least Squares Algebra

REVIEW (MULTIVARIATE LINEAR REGRESSION) Explain/Obtain the LS estimator () of the vector of coe cients (b)

7 Semiparametric Methods and Partially Linear Regression

Chapter 2 The Simple Linear Regression Model: Specification and Estimation

Lecture 5: Omitted Variables, Dummy Variables and Multicollinearity

A Course in Applied Econometrics Lecture 7: Cluster Sampling. Jeff Wooldridge IRP Lectures, UW Madison, August 2008

Regression Estimation Least Squares and Maximum Likelihood

Economics 620, Lecture 8: Asymptotics I

Least Absolute Value vs. Least Squares Estimation and Inference Procedures in Regression Models with Asymmetric Error Distributions

Transcription:

Economics 620, Lecture 7: Still More, But Last, on the K-Varable Linear Model Nicholas M. Kiefer Cornell University Professor N. M. Kiefer (Cornell University) Lecture 7: the K-Varable Linear Model IV 1 / 17

Speci cation Error: Suppose the model generating the data is y = X + " However, the model tted is y = X + ", with the LS estimator b = (X 0 X ) 1 X 0 y = (X 0 X ) 1 X 0 X + (X 0 X ) 1 X 0 ". Then Eb = (X 0 X ) 1 X 0 X and V (b ) = 2 (X 0 X ) 1 Professor N. M. Kiefer (Cornell University) Lecture 7: the K-Varable Linear Model IV 2 / 17

Application 1: Excluded variables Let X = [X 1 X 2 ] and X = X 1. That is, the model that generates the data is Consider b as an estimator of 1 : y = X 1 1 + X 2 2 + ": Proposition: b is biased. Proof : b = (X 0 X ) 1 X 0 y = (X1X 0 1 ) 1 X1(X 0 1 1 + X 2 2 + ") = 1 + (X1X 0 1 ) 1 X1X 0 2 2 + (X1X 0 1 ) 1 X1" 0 Eb = 1 + (X1X 0 1 ) 1 X1X 0 2 2 The second expression on the right hand side is the bias. Professor N. M. Kiefer (Cornell University) Lecture 7: the K-Varable Linear Model IV 3 / 17

A classic example: Suppose that the model generating the data is y i = 0 + 1 S i + a i + " i y: natural logarithm of earnings S: schooling a: ability a is unobserved and omitted, but it is positively correlated with S. Then P 1 P Eb 0 N S = + P P P a S S 2 as 1 supposing a is measured so that its coe cient is 1. If we suppose that P a = 0, then the bias in the coe cient of schooling is positive. Professor N. M. Kiefer (Cornell University) Lecture 7: the K-Varable Linear Model IV 4 / 17

A classic example (cont d) Generally, we cannot sign the bias, it depends not only on 2 but also on (X 0 1 X 1) 1 X 0 1 X 2, which of course can be positive or negative. Note that Vb = 2 (X1 0X 1) 1. So if 2 = 0, there is an e ciency gain from imposing the restriction and leaving out X 2. This con rms our earlier results. Professor N. M. Kiefer (Cornell University) Lecture 7: the K-Varable Linear Model IV 5 / 17

Estimation of Variance: e = M 1 y = M 1 (X 1 1 + X 2 2 + ") = M 1 X 2 2 + M 1 ) e 0 e = 0 2X 0 2 M 1X 2 2 + " 0 M 1 " + 2 0 2X 0 2 M 1 Note the expected value of the last term is 0. Clearly, we cannot estimate 2 by usual methods even if X 0 1 X 2 = 0 (no bias) since still M 1 X 2 6= 0. There is hope of detecting misspeci cation from the residuals since Ee e 0 = 2 M 1 under correct speci cation and Ee e 0 = 2 M 1 + M 1 X 2 2 0 2X 0 2 M 1 under misspeci cation. Professor N. M. Kiefer (Cornell University) Lecture 7: the K-Varable Linear Model IV 6 / 17

Application 2: Inclusion of unnecessary variables. Let X = X 1 and X = [X 1 X 2 ] X 1 is N K 1 and X 2 is N K 2. That is, the true model is y = X 1 + ". Proposition: b is unbiased. Proof : Eb = (X 0 X ) 1 X 0 X X 0 = 1 X 1 X1 0X 2 X2 0X 1 X2 0X 2 1 X 0 1 X 1 X 0 2 X 1 Professor N. M. Kiefer (Cornell University) Lecture 7: the K-Varable Linear Model IV 7 / 17

The partitioned inversion formula gives (X1 0X 1) 1 X1 0X 2D DX2 0X 1(X1 0X 1) 1 D for (X 0 X ) 1 where D = (X 0 2 M 1X 2 ) 1 and = (X 0 1X 1 ) 1 + (X 0 1X 1 ) 1 X 0 1X 2 DX 0 2X 1 (X 0 1X 1 ) 1 : This is a symmetric matrix. Multiplying this out veri es that Eb =. 0 Note that the variance of b is V (b ) = 2 (X 0 X ) 1. Professor N. M. Kiefer (Cornell University) Lecture 7: the K-Varable Linear Model IV 8 / 17

Proposition: The variance of the coe cients of X 1 in the unrestricted (where the matrix of explanatory variables is X ) is greater than the variance of the coe cients in the restricted model (where the matrix of explanatory variables is X 1 ). Proof : Using partitioned inversion, the variance of the rst K 1 elements (coe cients on X 1 ) is 2 (X1 0M 2X 1 ) 1 2 (X1 0X 1) 1 = variance of the restricted estimator. (why?) Estimation of 2 : e = M y = M " Under normality, (e 0 e = 2 ) = (" 0 M "= 2 ) 2 (N K 1 K 2 ) ) s 2 has higher variance than in the restricted model. (why?) Professor N. M. Kiefer (Cornell University) Lecture 7: the K-Varable Linear Model IV 9 / 17

Note on the interpretation of bias: Ey = X de nes and LS gives unbiased estimates of that. of bias really require a model. Questions Further statistical assumptions like Vy = 2 I allow some sorting out of speci cations, but is this assumption really attractive? Professor N. M. Kiefer (Cornell University) Lecture 7: the K-Varable Linear Model IV 10 / 17

Cross products matrix: In LS, everything comes from the cross products matrix. De nition: y 0 y X 0 y The cross products matrix is 2 P P P y 0 y 2 i yi yi x 2i P P P y i x Ki X X 0 = 6 1 x2i P P x Ki X 4 x 2 2i P x 2i x Ki P xki 2 with a column of ones in X. It is a symmetric matrix. Note that x j refers to the jth column of X. 3 7 5 Professor N. M. Kiefer (Cornell University) Lecture 7: the K-Varable Linear Model IV 11 / 17

Heteroskedasticity V (Y ) = V 6= 2 I Is the LS estimator unbiased? Is it BLUE? Proposition: Under the assumption of heteroskedasticity, V (^) = (X 0 X ) 1 X 0 VX (X 0 X ) 1 : Proof : V (^) = E(X 0 X ) 1 X 0 "" 0 X (X 0 X ) 1 = (X 0 X ) 1 X 0 VX (X 0 X ) 1. Suppose P = V (") is a diagonal matrix. Then X 0 P X = E P X i " 2 i X 0 i : Note that the cross products have expectation 0. This suggests using P X i ei 2X i 0. So we can estimate standard errors under the assumption that V (y) is diagonal. Professor N. M. Kiefer (Cornell University) Lecture 7: the K-Varable Linear Model IV 12 / 17

Testing for heteroskedasticity: 1. Goldfeld-Quandt test: Suppose we suspect that 2 i varies with x i. Then reorder the observations in the order of x i. Suppose N is even. If " was observed, then " 2 1 + "2 2 + ::: + "2 N=2 " 2 [(n=2)+1] + "2 [(N=2)+2] + ::: + "2 N F (N=2; N=2) could be used. We are tempted to use e i, but we can t because the rst N=2 e i s are not independent of the last. Here comes the Goldfeld-Quandt trick: Estimate e separately for each half of the sample with K parameters. The statistic is F ((N=2) K; (N=2) K): It turns out that this works better if you delete the middle N=3 observations. Professor N. M. Kiefer (Cornell University) Lecture 7: the K-Varable Linear Model IV 13 / 17

Testing for heteroskedasticity (cont d): 2. Breusch-Pagan test: The disturbances " i are assumed to be normally and independently distributed with variance 2 i = h(zi 0 ) where h denotes a function, and z0 i is a 1 P vector of variables in uencing heteroskedasticity. Let Z be an N P matrix with row vectors z 0 Z could be the same as the variables in X. i. Some of the variables in Regress e 2 = 2 ML on Z, including an intercept term. Note that (sum of squares due to Z)=2 2 (P 1) approximately. The factor 1=2 appears here since under normality the variance of " 2 = 2 is 2(E" 4 = 3 4 ). Professor N. M. Kiefer (Cornell University) Lecture 7: the K-Varable Linear Model IV 14 / 17

Testing for heteroskedasticity (cont d): An alternative approach (Koenker) drops normality and estimates the variance of e 2 i directly by N 1 P (e 2 i ^ 2 ) 2. The resulting statistic can be obtained by regressing e 2 on z and looking at NR 2 from this regression. Other tests are available for time series. Professor N. M. Kiefer (Cornell University) Lecture 7: the K-Varable Linear Model IV 15 / 17

Testing Normality The moment generating function of a random variable x is m(t) = E(exp(tx)); note m 0 (0) = Ex; m 00 (0) = Ex 2 ; etc. The MGF of the normal distribution n(; 2 ) is m(t) = exp(t + t 2 2 =2): Proof : let c = (2) 1=2 Z m(t) = c Z = c exp(tx) exp( 1=2(x ) 2 = 2 )dx exp( 1=2(x 2 t) 2 = 2 + t + 2 t 2 =2)dx = exp(t + 2 t 2 =2): Professor N. M. Kiefer (Cornell University) Lecture 7: the K-Varable Linear Model IV 16 / 17

Testing Normality (cont d) Thus for the regression errors " we have E" = 0; E" 2 = 2 ; E" 3 = 0; E" 4 = 3 4 ; E" 5 = 0; etc. It is easier to test the 3rd and 4th moment conditions than normality directly. If we knew the ", it would be easy to come up with a 2 test. In fact a test can be formed using the residuals e instead (and relying on asymptotic distibution theory). The test statistic is Which is 2 with 2 df : n[((e=s) 3 ) 2 =6 + ((e=s) 4 3) 2 =24]: This is the Kiefer/Salmon test (also called Jarque/Bera). Professor N. M. Kiefer (Cornell University) Lecture 7: the K-Varable Linear Model IV 17 / 17