ECONOMETRICS II (ECO 2401S) University of Toronto. Department of Economics. Spring 2013 Instructor: Victor Aguirregabiria

Similar documents
ECONOMETRICS II (ECO 2401S) University of Toronto. Department of Economics. Winter 2016 Instructor: Victor Aguirregabiria

ECONOMETRICS II (ECO 2401S) University of Toronto. Department of Economics. Winter 2014 Instructor: Victor Aguirregabiria

ECONOMETRICS FIELD EXAM Michigan State University May 9, 2008

i=1 y i 1fd i = dg= P N i=1 1fd i = dg.

ECONOMETRICS II (ECO 2401) Victor Aguirregabiria. Spring 2018 TOPIC 4: INTRODUCTION TO THE EVALUATION OF TREATMENT EFFECTS

ECONOMET RICS P RELIM EXAM August 24, 2010 Department of Economics, Michigan State University

1. The Multivariate Classical Linear Regression Model

We begin by thinking about population relationships.

Testing for Regime Switching: A Comment

Notes on Generalized Method of Moments Estimation

Econometrics Midterm Examination Answers

A Course on Advanced Econometrics

Economics 620, Lecture 19: Introduction to Nonparametric and Semiparametric Estimation

Exercises Chapter 4 Statistical Hypothesis Testing

Simple Estimators for Semiparametric Multinomial Choice Models

Chapter 1. GMM: Basic Concepts

Föreläsning /31

Chapter 2. Dynamic panel data models

Introduction to Nonparametric and Semiparametric Estimation. Good when there are lots of data and very little prior information on functional form.

ECON 594: Lecture #6

Chapter 6: Endogeneity and Instrumental Variables (IV) estimator

Economics 241B Estimation with Instruments

Collection of Formulae and Statistical Tables for the B2-Econometrics and B3-Time Series Analysis courses and exams

Economics 620, Lecture 13: Time Series I

Models, Testing, and Correction of Heteroskedasticity. James L. Powell Department of Economics University of California, Berkeley

GMM-based inference in the AR(1) panel data model for parameter values where local identi cation fails

LECTURE 13: TIME SERIES I

ECONOMETRICS II (ECO 2401) Victor Aguirregabiria. Winter 2017 TOPIC 3: MULTINOMIAL CHOICE MODELS

Panel Data. March 2, () Applied Economoetrics: Topic 6 March 2, / 43

Quantitative Techniques - Lecture 8: Estimation

HEC Lausanne - Advanced Econometrics

Chapter 6. Panel Data. Joan Llull. Quantitative Statistical Methods II Barcelona GSE

Lecture Notes on Measurement Error

13 Endogeneity and Nonparametric IV

Discrete Dependent Variable Models

(c) i) In ation (INFL) is regressed on the unemployment rate (UNR):

Testing Linear Restrictions: cont.

Economics 620, Lecture 9: Asymptotics III: Maximum Likelihood Estimation

Econometric Methods and Applications II Chapter 2: Simultaneous equations. Econometric Methods and Applications II, Chapter 2, Slide 1

Dynamic Panels. Chapter Introduction Autoregressive Model

Simple Estimators for Monotone Index Models

11. Bootstrap Methods

Econometrics of Panel Data

ERSA Training Workshop Lecture 5: Estimation of Binary Choice Models with Panel Data

WISE International Masters

Economics 620, Lecture 20: Generalized Method of Moment (GMM)

Robust Con dence Intervals in Nonlinear Regression under Weak Identi cation

Introduction: structural econometrics. Jean-Marc Robin

A New Approach to Robust Inference in Cointegration

Economics Introduction to Econometrics - Fall 2007 Final Exam - Answers

Econometric Analysis of Cross Section and Panel Data

1 Outline. Introduction: Econometricians assume data is from a simple random survey. This is never the case.

Pseudo panels and repeated cross-sections

Environmental Econometrics

Multivariate Regression: Part I

WISE International Masters

A Course in Applied Econometrics Lecture 14: Control Functions and Related Methods. Jeff Wooldridge IRP Lectures, UW Madison, August 2008

Nonparametric Identi cation and Estimation of Truncated Regression Models with Heteroskedasticity

Non-linear panel data modeling

Econometrics Honor s Exam Review Session. Spring 2012 Eunice Han

Empirical Industrial Organization (ECO 310) University of Toronto. Department of Economics Fall Instructor: Victor Aguirregabiria

Problem set 1 - Solutions

Finite Sample Performance of A Minimum Distance Estimator Under Weak Instruments

Econ 510 B. Brown Spring 2014 Final Exam Answers

Economics 620, Lecture 18: Nonlinear Models

Applied Health Economics (for B.Sc.)

ECONOMETRICS HONOR S EXAM REVIEW SESSION

MULTIVARIATE POPULATIONS

Instrumental Variables. Ethan Kaplan

A Note on the Closed-form Identi cation of Regression Models with a Mismeasured Binary Regressor

MC3: Econometric Theory and Methods. Course Notes 4

ECON Introductory Econometrics. Lecture 11: Binary dependent variables

Binary Models with Endogenous Explanatory Variables

Notes on Time Series Modeling

Extensions to the Basic Framework II

ECONOMETRICS II (ECO 2401) Victor Aguirregabiria (March 2017) TOPIC 1: LINEAR PANEL DATA MODELS

Linear Regression. Junhui Qian. October 27, 2014

On GMM Estimation and Inference with Bootstrap Bias-Correction in Linear Panel Data Models

Lecture 6: Discrete Choice: Qualitative Response

Day 3B Nonparametrics and Bootstrap

Some Recent Developments in Spatial Panel Data Models

Suggested Solution for PS #5

Industrial Organization II (ECO 2901) Winter Victor Aguirregabiria. Problem Set #1 Due of Friday, March 22, 2013

Economic modelling and forecasting

When is it really justifiable to ignore explanatory variable endogeneity in a regression model?

Lecture 3: Multiple Regression

Missing dependent variables in panel data models

MAXIMUM LIKELIHOOD ESTIMATION AND UNIFORM INFERENCE WITH SPORADIC IDENTIFICATION FAILURE. Donald W. K. Andrews and Xu Cheng.

Finansiell Statistik, GN, 15 hp, VT2008 Lecture 15: Multiple Linear Regression & Correlation

Testing Weak Convergence Based on HAR Covariance Matrix Estimators

WISE MA/PhD Programs Econometrics Instructor: Brett Graham Spring Semester, Academic Year Exam Version: A

GENERIC RESULTS FOR ESTABLISHING THE ASYMPTOTIC SIZE OF CONFIDENCE SETS AND TESTS. Donald W.K. Andrews, Xu Cheng and Patrik Guggenberger.

Econ 583 Homework 7 Suggested Solutions: Wald, LM and LR based on GMM and MLE

Tests forspatial Lag Dependence Based onmethodof Moments Estimation

Time Series Models and Inference. James L. Powell Department of Economics University of California, Berkeley

A Course in Applied Econometrics Lecture 18: Missing Data. Jeff Wooldridge IRP Lectures, UW Madison, August Linear model with IVs: y i x i u i,

1 The Multiple Regression Model: Freeing Up the Classical Assumptions

Generalized Linear Models 1

GMM based inference for panel data models

Control Functions in Nonseparable Simultaneous Equations Models 1

Transcription:

ECONOMETRICS II (ECO 2401S) University of Toronto. Department of Economics. Spring 2013 Instructor: Victor Aguirregabiria SOLUTION TO FINAL EXAM Friday, April 12, 2013. From 9:00-12:00 (3 hours) INSTRUCTIONS: The exam consists of two sets of questions. Try to answer all the questions. No study aids, including calculators, are allowed. TOTAL MARKS = 100 PROBLEM 1 (60 points). Consider the Binary Choice Logit model, Y = 1fX 0 " > 0g, where 1f:g is the indicator function, and " is independent of X with Logistic distribution with mean zero and unit variance, such that Pr(" c) = (c) expfcg=[1 + expfcg]. Let fy i ; x i : i = 1; 2; :::; ng be a random sample of the variables (Y; X). (a) [5 points] Write the expression for the log-likelihood function for this model and data. ANSWER. Given that we have a random sample, Pr(y 1 ; y 2 ; :::; y n j x 1 ; x 2 ; :::; x n ) = Q n Pr(y i; j x i ). The model implies that, Pr(Y = 1j X = x i ) = Pr(" i < x 0 i ) = (x0 i ), and Pr(Y = 0 j X = x i) = Pr(" i x 0 i ) = 1 (x 0 i). Therefore, the likelihood function is: And the log-likelihood function is: L() = Q i:y (x 0 i) Q i:y i=0 [1 (x 0 i)] P `() = n y i ln (x 0 i) + (1 y i ) ln[1 (x 0 i)] (b) [5 points] Derive the expression for the likelihood-equations that implicitly de ne the Maximum Likelihood estimator, and show that this system of equations is: P n x i[y i (x 0 i )] = 0. ANSWER. The likelihood equation is the system of rst order conditions of optimality for the MLE: @`()=@ = 0. Using the previous expression for the log-likelihood function, we have: @`() @ = n P y i 0 (x 0 i ) (x 0 i ) x i (1 y i ) 0 (x 0 i ) 1 (x 0 i )x i = 0 where 0 (:) is the derivative of the CDF (:), i.e., the density function of the logistic distribution. Operating in this equation we get: 0 (x 0 y i 1 y i i) x i (x 0 i ) 1 (x 0 i ) = 0 1

and 0 (x 0 i ) (x 0 i ) [1 (x0 i )] x i [y i (x 0 i)] = 0 Finally, by de nition of the function (c) expfcg=[1 + expfcg], we have that 0 (c) = expfcg=[1 + expfcg] 2 such that 0 (c) = (c) [1 (c)]. Therefore, the system of likelihood equations is: x i [y i (x 0 i)] = 0 (c) [5 points] Explain why this MLE can be interpreted as a Methods of Moments estimator. What are the moment conditions? ANSWER. The MLE is the value of that solves the system of equations n 1 P n x i [y i (x 0 i)] = 0. The vector n 1 P n x i [y i (x 0 i )] is the sample mean of x i [y i (x 0 i)]. Therefore, these conditions are the sample counterpart of the population moment conditions E( X [Y (X 0 )] ) = 0. The estimator that solves the sample counterpart of a set of population moment conditions is a MM estimator. Therefore, the MLE can be interpreted as a MM estimator. (d) [10 points] Obtain the expression for Hessian matrix of the log-likelihood function. Show that the log-likelihood function is globally concave (i.e., for any value of ). ANSWER. By de nition, the Hessian matrix is H @ 2`()=@ @ 0 such that H = @`()=@ @ 0 = @ @ 0 x i [y i (x 0 i)] = x i x 0 i 0 (x 0 i) = (x i x 0 i ) where x i x i 0 (x 0 i). Provided that there is not perfect collinearity between the vector of explanatory variables in X, it is clear that P n ) is positive de nite matrix, such as the Hessian H = P x 0 i (x i x0 i ) is negative de nite. This implies that the log-likelihood function is globally concave in. n (x i (e) [10 points] Describe the BHHH iterative algorithm (or alternatively, the Newton-Raphson algorithm) for the computation of the MLE. Write the formula describing an iteration in this algorithm. Explain how to implement the algorithm. ANSWER. The BHHH is an iterative algorithm to compute the maximum of the log-likelihood function. It is based on the following formula: at iteration k 1, we obtain the vector b k as: b k = b h Pn k 1 + s i( b k 1 ) s i ( b k 1 ) 0i 1 h Pn s i( b i k 1 ) where s i () is the vector of scores, i.e., the contribution of observation i to the likelihood equations. In this model we have that s i () = x i [y i b k = b k 1 + Pn x i x 0 i (x 0 i)], and therefore, y i (x 0 b 2 1 hpn i k 1 ) x i y i (x 0 b i i k 1 ) 2

The algorithm is implemented in the following way. We start with an arbitrary initial value b 0. Then, we apply recursively the previous formula that describes at iteration, for k = 1; 2; :::. At every iteration k, we calculate the Euclidean norm jj b b k k 1 jj. If this distance is lower than a pre-speci ed small constant (e.g., say 10 6 ), we stop the algorithm and use b k as the MLE. Otherwise we continue with iteration k + 1. (f) [5 points] Let b be the ML estimate of. Based on this estimate, describe a test of the null hypothesis "" is independent of X with Logistic distribution." ANSWER. According to the model, E(Y j X = x i ) = (x 0 i). Therefore, if the speci cation of the distribution is correct we have that y i = (x 0 i ) + u i, where E(u i jx i ) = 0. We can construct a test for this prediction of the model. The null hypothesis is that E(u i jx i ) = 0. Given b, we can construct the residuals: bu i = y i (x 0 b i). Note that by construction of b (by the likelihood equations) we have that P n x i bu i = 0. Therefore, we cannot test the null hypothesis E(x i u i ) = 0 because by construction the residuals are not correlated with X. This is exactly the same result as in a linear regression model. However, we can test for E(h(x i ) u i ) = 0 where h(x i ) is a vector of nonlinear functions of x i, i.e., power functions, cross-products, etc. We run an OLS regression of bu i on the vector h(x i ). Let b be n R 2, where R 2 is the R-square coe cient in this regression. Under the null hypothesis b is asymptotically distributed as a Chi-square random variable with degrees of freedom equal to the dimension of the vector h(x i ). We reject the null hypothesis with a con dence level of % if b is greater than the critical value, where is 1 distribution. quantile of Chi-square(q) (g) [10 points] Suppose that you reject the null hypothesis in the previous test. Describe how to use Klein-Spady semiparametric method to obtain a consistent estimator of (; F ), where F (:) is a nonparametrically speci ed distribution of ". ANSWER. If the vector x i contains at least one continuous explanatory variable, we can apply Klein-Spady estimator. We can implement this estimator using the following iterative procedure. At every iteration t 1, we obtain the estimates ff b t, b t g using the following formulas. Given b t 1, we update our estimate of the distribution F at any possible value z using the following kernel estimator: P n bf y z bz t 1 i i K b t (z) = P n K z bz t 1 i b where, for every observation i, bz t 1 i is the index x 0 i b t 1, K(:) is a kernel function and b is a pre-speci ed bandwidth parameter. Then, given b F t (:), we update our estimate of by obtaining b t as the value of that maximizes the log-likelihood function where we use b F t (:) as the distribution of ". At every iteration t, we calculate the Euclidean norm jj b t b t 1 jj. If this distance is lower than a pre-speci ed small constant (e.g., say 10 6 ), we stop the algorithm and use f b F t, b t g as our semiparametric MLE. Otherwise we continue with iteration t + 1. Note that, by construction, convergence in f b t g implies convergence in f b F t g and viceversa. (h) [10 points] Suppose that the assumption "" is independent of X with Logistic distribution " cannot be rejected. However, the researcher considers that one of the regressors in X, say 3

W, is potentially endogenous such that W and " are not independently distributed. Explain how to construct a test of the null hypothesis "W and " are not independently distributed". [Note: Normality of " is not necessary to implement the test.] ANSWER. Let X = ( X; e W ), i.e., X e represents the exogenous regressors in X. Suppose that there is an observable variable (or vector of variables) Z such that: (1) Z that does not enter in X e and it is not perfectly collinear with X; e (2) we are willing to assume that Z is independent of "; and (3) Z is correlated with W even after controlling for the correlation between W and X. e Under these conditions, we can test for the endogeneity of W in the binary choice model. The null hypothesis is that W and " are independent. The procedure to implement the test is the following. First, we ran an OLS regression for W on X e and Z. Let bw i be the tted values from this regression. Second, we estimate by ML a logit model for y i with explanatory variables ex i, w i, and bw i. Under the null hypothesis, the coe cient associated with bw i should be zero. This null hypothesis can be tested simply using a t-ratio test for that coe cient. Note that we have not made any assumption about the distribution of the error term in the regression model for W. The form of that distribution is relevant for the consistent estimation of the parameters in the second step, but it does not a ect the validity of this test (Rivers and Vuong, 1988). PROBLEM 2 (40 points). Consider the following Panel Data system of simultaneous equations: y it = 1 x it + 2 y it 1 + i + t + u it x it = 1 y it + 2 x it 1 + i + t + v it where u it and v it are not serially correlated, but they can be correlated with each other. The individual e ects i and i can be also correlated with each other. The panel dataset contains a large number N of individuals but a small number of time periods, say T = 4. (a) [15 points] Show whether each of the following estimators is consistent as N! 1 and xed T. (1) OLS in levels; (2) Within-Groups estimator; (3) OLS in rst di erences; (4) IV estimation of the equation for y it in levels using as instruments fy it 1,x it 1 g; (5) IV estimation of the equation for y it in levels using as instruments fx it 2,x it 3 g; (6) IV estimation of the equation y it in rst di erences using as instruments fx it 1,y it 2 g; and (7) IV estimation of the equation y it in rst di erences using as instruments fx it 2,y it 2 g. ANSWER. (1) OLS in levels. Inconsistent because E(f i + u it g y it 1 ) 6= 0 and E(f i + u it g x it ) 6= 0. (2) Within-Groups estimator. Inconsistent. In dynamic PD models with small T, the WG estimator is inconsistent. (3) OLS in rst di erences. Inconsistent. In dynamic PD models with small T, the OLS-FD estimator is inconsistent. 4

(4) IV estimation of the equation for y it in levels using as instruments fy it 1,x it 1 g. Inconsistent because E(f i + u it g y it 1 ) 6= 0 and E(f i + u it g x it 1 ) 6= 0. (5) IV estimation of the equation for y it in levels using as instruments fx it 2,x it 3 g. Inconsistent because E(f i + u it g x it 2 ) 6= 0 and E(f i + u it g x it 3 ) 6= 0 unless 1 = 0 and i and i are not correlated. (6) IV estimation of the equation y it in rst di erences using as instruments fx it 1,y it 2 g. Inconsistent because E(u it x it 1 ) 6= 0 because E(u it 1 x it 1 ) 6= 0 unless 1 = 0 and u it 1 and v it 1 are not correlated. (7) IV estimation of the equation y it in rst di erences using as instruments fx it 2,y it 2 g. Consistent because E(u it x it 2 ) 6= 0 and E(u it y it 2 ) 6= 0 as long as u it and v it are not serially correlated. We also need that 2 6= 0 and 2 6= 0 for these instruments to have predictive power. (b) [15 points] Consider the estimation of 1 and 2 in the equation for y it using Arellano- Bond GMM estimator. Write the expressions of all the Arellano-Bond moment conditions in this model and dataset with T = 4 time periods. ANSWER. With T = 4 we have two equations (two time periods) in the model in rst di erences. y i3 = 1 x i3 + 2 y i2 + 3 + u i3 y i4 = 1 x i4 + 2 y i3 + 4 + u i4 For the rst equation we have the following moment conditions: E(u i3 ) = 0, E(y i1 u i3 ) = 0, and E(x i1 u i3 ) = 0. For the second equation we have the following moment conditions: E(u i4 ) = 0, E(y i1 u i4 ) = 0, E(y i2 u i4 ) = 0, E(x i1 u i4 ) = 0, and E(x i1 u i4 ) = 0. In total we have 8 moment conditions to estimate the vector of parameters = ( 1, 2, 3, 4 ) 0. Let m N () be the vector with the 8 sample moment conditions such that m N () = N 1 P N 2 6 4 We can represent this moment conditions as N 1 P N Z i (y i3 1 x i3 2 y i2 3 ) y i1 (y i3 1 x i3 2 y i2 3 ) x i1 (y i3 1 x i3 2 y i2 3 ) (y i4 1 x i4 2 y i3 4 ) y i1 (y i4 1 x i4 2 y i3 4 ) x i1 (y i4 1 x i4 2 y i3 4 ) y i2 (y i4 1 x i4 2 y i3 4 ) x i2 (y i4 1 x i4 2 y i3 4 ) yi3 1 x i3 2 y i2 3 y i4 1 x i4 2 y i3 4 3 7 5 where 2 Z i = 6 4 1 0 y i1 0 x i1 0 0 1 0 y i1 0 x i1 0 y i2 0 x i2 3 7 5 5

(c) [10 points] Explain how to obtain an Arellano-Bond 1-step-GMM estimator, and a 2-step optimal GMM estimator. The one-step AB estimator is obtained as the value of that minimizes the criterion function m N () 0 h W (1) m N (), where W (1) is the matrix N 1 P i 1 N 2 1 Z i H Zi 0 where H is the matrix. The 1 2 matrix H represents, up to a scale parameter, the variance-covariance matrix of (u i3 ; u i4 ) under the assumptions of homoscedasticity and no serial correlation in u it. It is the optimal weighting matrix under these assumptions. We can relax the assumption of time-homoscedasticity in the second step. Given the residuals in the rst step, say ([u i3 ; [u i4 ), we can obtain: W (2) = N 1 P N Z i " ([u i3 ; [u i4 ) [u i3 [u i4 Then, the two-step AB estimator is obtained as the value of that minimizes the criterion function m N () 0 W (2) m N (). This estimator is asymptotically e cient.!# Z 0 i! 1 6