Econ 583 Homework 7 Suggested Solutions: Wald, LM and LR based on GMM and MLE

Similar documents
Asymptotics for Nonlinear GMM

Introduction to Maximum Likelihood Estimation

LECTURE 10: NEYMAN-PEARSON LEMMA AND ASYMPTOTIC TESTING. The last equality is provided so this can look like a more familiar parametric test.

Exercises Chapter 4 Statistical Hypothesis Testing

Introduction Large Sample Testing Composite Hypotheses. Hypothesis Testing. Daniel Schmierer Econ 312. March 30, 2007

Economics 583: Econometric Theory I A Primer on Asymptotics: Hypothesis Testing

Maximum Likelihood (ML) Estimation

6. MAXIMUM LIKELIHOOD ESTIMATION

Economics 582 Random Effects Estimation

Economics 583: Econometric Theory I A Primer on Asymptotics

Final Exam. 1. (6 points) True/False. Please read the statements carefully, as no partial credit will be given.

Nonlinear GMM. Eric Zivot. Winter, 2013

Maximum Likelihood Tests and Quasi-Maximum-Likelihood

Testing and Model Selection

Institute of Actuaries of India

Greene, Econometric Analysis (6th ed, 2008)

HEC Lausanne - Advanced Econometrics

Statistics and econometrics

Composite Hypotheses and Generalized Likelihood Ratio Tests

A Very Brief Summary of Statistical Inference, and Examples

1/24/2008. Review of Statistical Inference. C.1 A Sample of Data. C.2 An Econometric Model. C.4 Estimating the Population Variance and Other Moments

Chapter 6 - Sections 5 and 6, Morris H. DeGroot and Mark J. Schervish, Probability and

An Introduction to Generalized Method of Moments. Chen,Rong aronge.net

McGill University. Faculty of Science. Department of Mathematics and Statistics. Part A Examination. Statistics: Theory Paper

1.5 Testing and Model Selection

Lecture 17: Likelihood ratio and asymptotic tests

Mathematics Ph.D. Qualifying Examination Stat Probability, January 2018

Hypothesis Testing. 1 Definitions of test statistics. CB: chapter 8; section 10.3

Econ 583 Final Exam Fall 2008

simple if it completely specifies the density of x

Maximum Likelihood. F θ, θ Θ. X 1,..., X n. L(θ) = f(x i ; θ) l(θ) = ln L(θ) = i.i.d. i=1. n ln f(x i ; θ) Sometimes

Chapter 4: Constrained estimators and tests in the multiple linear regression model (Part III)

(θ θ ), θ θ = 2 L(θ ) θ θ θ θ θ (θ )= H θθ (θ ) 1 d θ (θ )

Single Equation Linear GMM

Likelihood-based inference with missing data under missing-at-random

Psychology 282 Lecture #4 Outline Inferences in SLR

Practice Problems Section Problems

Poisson Regression. Ryan Godwin. ECON University of Manitoba

Statistical Inference

Chapter 11. Regression with a Binary Dependent Variable

Chapter 2. Review of basic Statistical methods 1 Distribution, conditional distribution and moments

ECONOMETRICS II (ECO 2401S) University of Toronto. Department of Economics. Spring 2013 Instructor: Victor Aguirregabiria

Hypothesis Test. The opposite of the null hypothesis, called an alternative hypothesis, becomes

Math 70 Homework 8 Michael Downs. and, given n iid observations, has log-likelihood function: k i n (2) = 1 n

ECONOMETRIC THEORY. MODULE VI Lecture 19 Regression Analysis Under Linear Restrictions

2014/2015 Smester II ST5224 Final Exam Solution

Chapter 4. Theory of Tests. 4.1 Introduction

Two hours. To be supplied by the Examinations Office: Mathematical Formula Tables THE UNIVERSITY OF MANCHESTER. 21 June :45 11:45

GARCH Models Estimation and Inference

Mathematics Qualifying Examination January 2015 STAT Mathematical Statistics

GARCH Models Estimation and Inference. Eduardo Rossi University of Pavia

Cherry Blossom run (1) The credit union Cherry Blossom Run is a 10 mile race that takes place every year in D.C. In 2009 there were participants

TUTORIAL 8 SOLUTIONS #

MLE and GMM. Li Zhao, SJTU. Spring, Li Zhao MLE and GMM 1 / 22

The Logit Model: Estimation, Testing and Interpretation

Module 6: Methods of Point Estimation Statistics (OA3102)

Maximum-Likelihood Estimation: Basic Ideas

Lecture 6: Hypothesis Testing

δ -method and M-estimation

Chapter 1. GMM: Basic Concepts

13.2 Example: W, LM and LR Tests

A Primer on Asymptotics

Multiple Equation GMM with Common Coefficients: Panel Data

GARCH Models Estimation and Inference

Review: General Approach to Hypothesis Testing. 1. Define the research question and formulate the appropriate null and alternative hypotheses.

Confirmatory Factor Analysis: Model comparison, respecification, and more. Psychology 588: Covariance structure and factor models

7.2 One-Sample Correlation ( = a) Introduction. Correlation analysis measures the strength and direction of association between

Stat 5101 Lecture Notes

10. Composite Hypothesis Testing. ECE 830, Spring 2014

Introduction to Estimation Methods for Time Series models Lecture 2

A CONDITIONAL LIKELIHOOD RATIO TEST FOR STRUCTURAL MODELS. By Marcelo J. Moreira 1

ECON 4160, Autumn term Lecture 1

Math 494: Mathematical Statistics

Chapter 4 HOMEWORK ASSIGNMENTS. 4.1 Homework #1

ST495: Survival Analysis: Hypothesis testing and confidence intervals

Review Quiz. 1. Prove that in a one-dimensional canonical exponential family, the complete and sufficient statistic achieves the

8. Hypothesis Testing

Assumptions of classical multiple regression model

Summary: the confidence interval for the mean (σ 2 known) with gaussian assumption

Wald s theorem and the Asimov data set

Central Limit Theorem ( 5.3)

Generalized, Linear, and Mixed Models

CIVL /8904 T R A F F I C F L O W T H E O R Y L E C T U R E - 8

Testing Hypothesis. Maura Mezzetti. Department of Economics and Finance Università Tor Vergata

VAR Models and Cointegration 1

Ch. 5 Hypothesis Testing

Testing Statistical Hypotheses

Some General Types of Tests

ML Testing (Likelihood Ratio Testing) for non-gaussian models

Let us first identify some classes of hypotheses. simple versus simple. H 0 : θ = θ 0 versus H 1 : θ = θ 1. (1) one-sided

Econometrics I, Estimation

Logistic Regression Analysis

Analogy Principle. Asymptotic Theory Part II. James J. Heckman University of Chicago. Econ 312 This draft, April 5, 2006

Least Absolute Value vs. Least Squares Estimation and Inference Procedures in Regression Models with Asymmetric Error Distributions

Density Estimation: ML, MAP, Bayesian estimation

Discrete Dependent Variable Models

Bivariate normal population distribution, 694 BLUE. See Best linear unbiased estimator (BLUE) Bonferroni joint tests,

Comparing two independent samples

2. We care about proportion for categorical variable, but average for numerical one.

Universität Potsdam Institut für Informatik Lehrstuhl Maschinelles Lernen. Hypothesis testing. Anna Wegloop Niels Landwehr/Tobias Scheffer

Transcription:

Econ 583 Homework 7 Suggested Solutions: Wald, LM and LR based on GMM and MLE Eric Zivot Winter 013 1 Wald, LR and LM statistics based on generalized method of moments estimation Let 1 be an iid sample of Bernoulli random variables; that is, each has density (; 0 )= 0(1 0 ) 1 Instead of using ML estimation, consider instead estimation of 0 using the generalized method of moments (GMM). Here 1. Consider GMM estimation of using the single moment condition ( )= Show that [( )] = 0 for = 0,and[( )] 6= 0for 6= 0 Additionally, show that = [( 0 ) ]= 0 (1 0 ) [( )] = [ ] = 0 Clearly, [( )] = 0 for = 0,and[( )] 6= 0for 6= 0 Next, = [( 0 ) ]=[( 0 ) ]=( )= 0 (1 0 ) since has a Bernoulli distribution with parameter 0 Note that = ( 0 ) 1. Using the sample moment () = P 1 ( ), show that the GMM estimator of is ˆ = P 1 What is the value of the GMM objective function ( ˆ 1 ˆ )= (ˆ ) 0 ˆ 1 (ˆ )= (ˆ ) ˆ 1? How would you consistently estimate? 1

Equating the sample moment with the population moment gives the estimating equation (ˆ ) = 1 ( ˆ )= 1 ˆ =0 ˆ = 1 = Hence, the GMM estimator is the same as the ML estimator. The GMM objective function evaluated at ˆ is! ( ˆ 1 ˆ )= (ˆ ) ˆ 1 = 1 ˆ ˆ 1 =0 regardless of the value of ˆ since ˆ = P 1 This is to be expected since with one moment condition and one parameter to estimate the model is exactly identified. Since = 0 (1 0 ) a nature estimate is ˆ = ˆ (1 ˆ ) Alternatively, one could use the the usual non-parametric estimator of : ˆ = 1 ( ˆ ) = 1 ( ˆ ) 3. Determine the asymptotic normal distribution for ˆ Is this distribution the same as the asymptotic normal distribution for ˆ? Using the asymptotic normality result for GMM estimators we have Now, so that and ˆ ( 0 1 ( 0 1 ) 1 ) ( 0 ) = = [( 0 ) ] 0 ( 0 ) 0 = 0 ( 0 )= 1 = 0 (1 0 ) ( 0 1 ) 1 = = 0 (1 0 ) ˆ ( 0 1 0 (1 0 )) which is equivalent to the asymptotic distribution for the MLE. Next, consider testing the hypotheses 0 : = 0 vs. 1 : 6= 0 using a 5 significance level.

4. and 5. Derive the Wald, LR and LM statistics for testing the above hypothesis. Because is a scalar, these statistics have the form = (ˆ 0 ) ( (ˆ ) ˆ 1 ) = ( ( 0 ) ˆ 1 ( 0 )) ( ( 0 ) ˆ 1 ) 1 = ( ˆ 1 0 ) ( ˆ 1 ˆ ) where () = () For the Wald and LR statistic it is useful to estimate using ˆ For the LM statistic, it is useful to an estimate of imposing = 0 Are the GMM Wald, LM and LR statistics numerically equivalent? Compare the GMM Wald, LR and LM statistics with the ML equivalents. Do you see any similarities or differences? The Wald statistic has the form = (ˆ 0 ) ( (ˆ ) ˆ 1 ) = (ˆ 0 ) ˆ (1 ˆ ) which is equivalent to the ML Wald statistic. For the LM statistic, first note that ( 0 ) = 1 ( 0 ) = 1 0 = ˆ 0 ˆ = 0 (1 0 ) Then, the LM statistic has the form = ( ( 0 ) ˆ 1 ( 0 )) ( ( 0 ) ˆ 1 ) 1 = (ˆ 0 ) 0 (1 ˆ 0 ) which is identical to the ML LM statistic. For the LR statistic, recall that (ˆ ˆ 1 )=0since the model is just identified. Therefore, the LR statistic has the form = ( ˆ 1 0 ) ( ˆ 1 ˆ ) = ( ˆ 1 0 )= ( 0 ) 0 ˆ 1 ( 0 )= ( 0 ) ˆ 1 = (ˆ 0 ) ˆ (1 ˆ ) 3

which is not equivalent to the ML LR statistic. However, the GMM LR statistic is equivalent to the GMM Wald statistic. If you use ˆ = 0 (1 0 ) then the GMM LR statistics is equivalent to the the GMM LM statistic. Note: for GMM, = = if the same ˆ is used for all statistics. If ˆ = P 1 ( ˆ ) = P 1 ( ˆ ) is used then the GMM statistics willnotbeequaltothemlstatisticsinfinitesamples,butwillbeasthesamplesize gets large. Wald, LR and LM statistics based on maximum likelihood estimation Let 1 be an iid sample of Bernoulli random variables; that is, each has density (; 0 )= 0(1 0 ) 1 First, consider estimation of 0 using maximum likelihood (ML). Let =( 1 ) 1. Derive the log-likelihood function, ln ( x) score function, ( x) = ln ( x) Hessian function, ( x) = ln ( x) and information matrix, ( 0 x) = [( 0 x)] = (( 0 x)) Verify that [( 0 x)] = 0 and show that the ML estimate for is ˆ = P 1 The joint density/likelihood function for the sample is given by Y (x; ) =( x) = (1 ) 1 = (1 ) The log-likelihood function is ln ( x) = ln (1 )! = ln()+ ln(1 ) The score function for the Bernoulli log-likelihood is ln ( x) ( x) = = 1 1 1 P = ( ) (1 ) Using the chain rule it can be shown that ( x) = (1 ) ( x) = (1 ) µ + ( x) ( x) = (1 ) 4!

where = 1 P Therefore, the information matrix is ( 0 x) = [( 0 x)] = + [( 0 x)] 0 [( 0 x)] 0 (1 0 ) = 0 (1 0 ) since [( 0 x)] = P ([ ] 0 ) 0 (1 0 ) =0 since [ ]= 0 The MLE solves (ˆ x) = P ( ˆ ) ˆ (1 ˆ =0 ) ˆ = 1 =. Determine the asymptotic normal distribution for ˆ Using the asymptotic properties of the MLE for a regular model, ˆ ( 0 ( 0 x) 1 ) ( 0 x) 1 = 0(1 0 ) Since 0 is unknown, a practically useful result replaces ( 0 x) 1 with (ˆ x) 1 Next, consider testing the hypotheses 0 : = 0 vs. 1 : 6= 0 usinga5significance level. 3. Derive the Wald, LR and LM statistics for testing the above hypothesis. These statistics have the form = (ˆ 0 ) (ˆ x) = ( 0 x) ( 0 x) 1 = [ln ( 0 x) ln (ˆ x)] Are these statistics numerically equivalent? For a 5% test, what is the decision rule to reject 0 for each statistic? 5

The Wald statistic has the form For the LM statistic, note that = (ˆ 0 ) (ˆ x) = (ˆ 0 ) ˆ 1 ˆ = (ˆ 0 ) ˆ 1 ˆ so that P ( 0 x) = ( ) (1 ) ( 0 x) 1 = 0(1 0 ) = (ˆ 0 ) 0 (1 0 ) = ( 0 x) ( 0 x) 1 (ˆ 0 ) 0 (1 0 ) = 0 (1 0 ) = (ˆ 0 ) 0 (1 0 ) Notice that the LM statistic has the same form as the Wald statistic. The difference between the two statistics is based on how ( x) is estimated. With the Wald statistic, ( x) is estimated using (ˆ x) andwiththelmstatistic,( x) is estimated using ( 0 x) For the statistic, note that! ln ( x) = ln(ˆ )+ ln(1 ˆ ) Therefore ln ( 0 x) = ln( 0 )+! ln(1 0 ) = [ln ( 0 x) ln (ˆ x)] "! # = (ln 0 ln )+ ln(1 ) ln(1 ˆ ) µ µ 0 = ˆ ln +(1 ˆ 1 0 )ln ˆ 1 ˆ 6

4. ProvethattheWaldandLMstatisticshaveasymptotic (1) distributions. First, consider the Wald statistic: = (ˆ 0 ) ˆ 1 ˆ = (ˆ 0 ) rˆ 1 ˆ From the asymptotic normality of the MLE, we have that (ˆ 0 ) (0 0 (1 0 )) and from the consistency of the MLE we have that ˆ 0 Slutsky s theorem then gives rˆ 1 ˆ p 0 (1 0 ) Finally, by a combination of Slutsky s theorem and the continuous mapping theorem (CMT) we have (ˆ 0 ) (0 0 (1 0 )) rˆ 1 ˆ p 0 (1 0 ) = (0 1) = (1) where (1) denotes a chi-square random variable with 1 degree of freedom. A similar proof works for the LM statistic: = (ˆ 0 ) 0 (1 0 ) (ˆ 0 ) = p 0 (1 0 ) (0 0 (1 0 )) p = (1) 0 (1 0 ) 7