Statistics and econometrics

Similar documents
Introduction Large Sample Testing Composite Hypotheses. Hypothesis Testing. Daniel Schmierer Econ 312. March 30, 2007

Chapter 4: Constrained estimators and tests in the multiple linear regression model (Part III)

(θ θ ), θ θ = 2 L(θ ) θ θ θ θ θ (θ )= H θθ (θ ) 1 d θ (θ )

Econometrics of Panel Data

The outline for Unit 3

8. Hypothesis Testing

LECTURE 10: NEYMAN-PEARSON LEMMA AND ASYMPTOTIC TESTING. The last equality is provided so this can look like a more familiar parametric test.

Maximum Likelihood Tests and Quasi-Maximum-Likelihood

Testing Hypothesis. Maura Mezzetti. Department of Economics and Finance Università Tor Vergata

Practical Econometrics. for. Finance and Economics. (Econometrics 2)

Exercises Chapter 4 Statistical Hypothesis Testing

Some General Types of Tests

Ch. 5 Hypothesis Testing

Introduction to Estimation Methods for Time Series models Lecture 2

Greene, Econometric Analysis (6th ed, 2008)

Let us first identify some classes of hypotheses. simple versus simple. H 0 : θ = θ 0 versus H 1 : θ = θ 1. (1) one-sided

Final Exam. 1. (6 points) True/False. Please read the statements carefully, as no partial credit will be given.

Maximum Likelihood (ML) Estimation

Hypothesis Testing. 1 Definitions of test statistics. CB: chapter 8; section 10.3

Advanced Quantitative Methods: maximum likelihood

MS&E 226: Small Data

Econ 583 Homework 7 Suggested Solutions: Wald, LM and LR based on GMM and MLE

Econ 583 Final Exam Fall 2008

A Very Brief Summary of Statistical Inference, and Examples

Cointegration Lecture I: Introduction

Economics 582 Random Effects Estimation

GARCH Models Estimation and Inference

Chapter 4. Theory of Tests. 4.1 Introduction

Functional Form. Econometrics. ADEi.

Advanced Quantitative Methods: maximum likelihood

Introductory Econometrics

Testing and Model Selection

Quick Review on Linear Multiple Regression

1.5 Testing and Model Selection

Likelihood-Based Methods

Econometrics I KS. Module 2: Multivariate Linear Regression. Alexander Ahammer. This version: April 16, 2018

MATH5745 Multivariate Methods Lecture 07

Economics 536 Lecture 7. Introduction to Specification Testing in Dynamic Econometric Models

Spring 2017 Econ 574 Roger Koenker. Lecture 14 GEE-GMM

Recall that in order to prove Theorem 8.8, we argued that under certain regularity conditions, the following facts are true under H 0 : 1 n

Assumptions of classical multiple regression model

Chapter 7. Hypothesis Testing

So far our focus has been on estimation of the parameter vector β in the. y = Xβ + u

Advanced Econometrics

Appendix A: The time series behavior of employment growth

Testing an Autoregressive Structure in Binary Time Series Models

DA Freedman Notes on the MLE Fall 2003

Theory of Maximum Likelihood Estimation. Konstantin Kashin

Graduate Econometrics I: Maximum Likelihood II

Econometrics II - EXAM Answer each question in separate sheets in three hours

F9 F10: Autocorrelation

Multiple Linear Regression

Lecture 4: Testing Stuff

Spatial Regression. 9. Specification Tests (1) Luc Anselin. Copyright 2017 by Luc Anselin, All Rights Reserved

Introductory Econometrics

Econometrics of Panel Data

1/24/2008. Review of Statistical Inference. C.1 A Sample of Data. C.2 An Econometric Model. C.4 Estimating the Population Variance and Other Moments

Statistical Tests. Matthieu de Lapparent

Asymptotics for Nonlinear GMM

Interpreting Regression Results

Topic 19 Extensions on the Likelihood Ratio

ECON 4160, Autumn term Lecture 1

Hypothesis testing: theory and methods

ECON 5350 Class Notes Nonlinear Regression Models

Problem Set #6: OLS. Economics 835: Econometrics. Fall 2012

Lecture 10: Generalized likelihood ratio test

6. MAXIMUM LIKELIHOOD ESTIMATION

Inference about the Indirect Effect: a Likelihood Approach

Hypothesis Testing: The Generalized Likelihood Ratio Test

GARCH Models Estimation and Inference

Outline. 1. Define likelihood 2. Interpretations of likelihoods 3. Likelihood plots 4. Maximum likelihood 5. Likelihood ratio benchmarks

Test of hypotheses with panel data

MFE Financial Econometrics 2018 Final Exam Model Solutions

Econometrics II - EXAM Outline Solutions All questions have 25pts Answer each question in separate sheets

Public Sector Management I

Master s Written Examination

MISCELLANEOUS TOPICS RELATED TO LIKELIHOOD. Copyright c 2012 (Iowa State University) Statistics / 30

Graduate Econometrics I: Maximum Likelihood I

Testing Hypothesis after Probit Estimation

Economics 583: Econometric Theory I A Primer on Asymptotics

GARCH Models Estimation and Inference. Eduardo Rossi University of Pavia

Lecture 6: Hypothesis Testing

ML Testing (Likelihood Ratio Testing) for non-gaussian models

Multiple Regression Analysis: Inference ECONOMETRICS (ECON 360) BEN VAN KAMMEN, PHD

parameter space Θ, depending only on X, such that Note: it is not θ that is random, but the set C(X).

[y i α βx i ] 2 (2) Q = i=1

LECTURE 5. Introduction to Econometrics. Hypothesis testing

Multiple Regression Analysis: Heteroskedasticity

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A.

2.1 Linear regression with matrices

Empirical Economic Research, Part II

Generalized Linear Models

Master s Written Examination - Solution

simple if it completely specifies the density of x

Review of Classical Least Squares. James L. Powell Department of Economics University of California, Berkeley

13.2 Example: W, LM and LR Tests

The regression model with one fixed regressor cont d

Outline of GLMs. Definitions

Central Limit Theorem ( 5.3)

Intermediate Econometrics

Transcription:

1 / 36 Slides for the course Statistics and econometrics Part 10: Asymptotic hypothesis testing European University Institute Andrea Ichino September 8, 2014

2 / 36 Outline Why do we need large sample hypothesis testing? The trinity of large sample testing strategies The Wald test The Lagrance Multiplier (or Score) test The Likelihood Ratio test Large sample testing and regression Wald test in the context of regression The Lagrange multiplier test in the context of linear regression The Likelihood Ratio Test in the context of regression Final comment on the trinity of tests

3 / 36 Section 1 Why do we need large sample hypothesis testing?

4 / 36 No need of distributional assumptions More than precision, the advantage of a large sample for testing is that no distributional assumption on the outcome Y is needed. In particular we do not have to assume MLR 6: normality of U X. This is particularly important from a methodological point of view because it allows us to use all the machinery of regression analysis also in cases where normality is clearly a wrong assumption: Discrete dependent variables Limited dependent variables Conditional on positive" models Duration analysis Count data analysis Thanks to large samples, econometrics becomes considerably simpler!

5 / 36 Section 2 The trinity of large sample testing strategies

Justification of the three tests in a nutshell H 0 : θ = θ 0 (1) 1. Wald test Under the null, the normalised distance between ˆθ and θ 0 should be small. 2. Lagrange multiplier (or Score) test Under the null, the score lθ (X, θ 0 ) evaluated at θ 0 should be close to zero. 3. Likelihood ratio test The ratio between the likelihood L(X, ˆθ) evaluated at at the estimate ˆθ and the likelihood L(X, θ 0 ) evaluated at θ 0 should be close to 1. Figure 3.1 in Engle s Handbook chapter is a useful way to think at the three tests 6 / 36

7 / 36 Subsection 1 The Wald test

8 / 36 The logic of the Wald test in detail We have established that n(ˆθn θ) d Normal(0, Ω) (2) where (see previous slides on asymptotics): Ω = 1 I 1 (θ) in the case of ML; Therefore, under the null, W = n(ˆθ n θ) Ω d Normal(0, 1) (3) or, taking squares W 2 = n(ˆθ n θ) 2 Ω d χ 2 1 (4)

9 / 36 The logic of the Wald test in detail (comments) Given a random sample, estimates of Ω must be used. : ˆΩ = 1 Î n (ˆθ ML ) = n n i=1 l θθ(x, ˆθ ML ) (5) The null H 0 : θ = θ 0 is rejected at a significance s when: Ŵ 2 > χ 2 1,s (6) where χ 2 1,s is the critical value c of the χ2 1 such that Pr(W 2 > c H 0 ) = s with W 2 χ 2 1 (7) The p-value is p = Pr(χ 2 1 > Ŵ 2 )

10 / 36 Subsection 2 The Lagrance Multiplier (or Score) test

11 / 36 The logic of the LM test in detail It can be shown that under the null: n(lθ (X, θ)) d Normal(0, 1 I 1 (θ) ) (8) Intuitively, l θ (X, θ) is the average of iid random variables (the ln of the derivatives of the pdf of each observation) with mean zero and variance 1 I 1 (θ)

12 / 36 The logic of the LM test in detail (cont.) Therefore, under the null, n(lθ (X, θ 0 ) LM = 1 I 1 (θ 0 ) or, taking squares d Normal(0, 1) (9) LM 2 = n(l θ(x, θ 0 ) 2 1 I 1 (θ 0 ) d χ 2 1 (10) Note that LM test statistics can be computed without having to find the ML estimate. The name of the test comes from the fact that implicitly we are maximizing the likelihood subject to the constraint θ = θ 0 max L(X, θ) + λ(θ θ 0 ) λ = L θ (11)

13 / 36 The logic of the LM test in detail (comments) Estimates of the Information at θ 0 must be used: 1 Î n (θ 0 ) = n n i=1 l θθ(x, θ 0 ) (12) The null H 0 : θ = θ 0 is rejected at a significance s when: ˆLM 2 > χ 2 1,s (13) where χ 2 1,s is the critical value c of the χ2 1 such that Pr(LM 2 > c H 0 ) = s with LM 2 χ 2 1 (14) The p-value is p = Pr(χ 2 1 > ˆLM 2 )

14 / 36 Subsection 3 The Likelihood Ratio test

15 / 36 The logic of the LR test in detail Under H 0 : θ = θ 0, expand the log likelihood l(θ 0 ) around ˆθ ML l(θ 0 ) = l(ˆθ ML ) + l θ (ˆθ ML )(θ 0 ˆθ ML ) 1 2 (ˆθ ML θ 0 ) 2 l θθ ( θ) (15) where θ is some intermediate point between ˆθ ML and θ 0, and 1 N l θθ( θ) p I 1 (θ) (16) Note that l θ (ˆθ ML ) = 0 and therefore we can write: LR = 2(l(θ 0 ) l(ˆθ ML )) = ( N(ˆθ ML θ 0 )) 2 (I 1 (θ)) 1 χ 2 1 (17) where the LHS is the Likelihood Ratio Test

16 / 36 The logic of the LR test in detail (comments) The null H 0 : θ = θ 0 is rejected at a significance s when: LR 2 > χ 2 1,s (18) where χ 2 1,s is the critical value c of the χ2 1 such that Pr(LR 2 > c H 0 ) = s with LR 2 χ 2 1 (19) The p-value is p = Pr(χ 2 1 > LR 2 )

17 / 36 Section 3 Large sample testing and regression

18 / 36 A general formulation of an hypothesis concerning β Given the PRF Y = Xβ + U (20) let s now consider the most general formulation of an hypothesis concerning β: H 0 : r(β) = q against H 1 : r(β) q (21) where r(.) is any function of the parameters and r(β) q is a ρ 1 vector, if ρ is the number of restrictions. So H 0 and H 1 are systems of ρ equations if there are ρ restrictions.

19 / 36 An example Example : H 0 : r(β) = Rβ = q against H 1 : r(β) = Rβ q (22) where R is a ρ k + 1 matrix which charaterize the ρ restrictions on the parameters that we would like to test.

20 / 36 Exercise on the specification of a set of restrictions Suppose that you are estimating the log of a Cobb Douglas production function in which output depends on labor and capital and you want to test: constant returns to scale; the return to one unit of labor is twice the return to one unit of capital; there exist neutral technological progress/regress. What is R for these restrictions?

21 / 36 Subsection 1 Wald test in the context of regression

22 / 36 The logic of the Wald test If the restrictions are valid the quantity r( ˆβ) q should be close to 0 while otherwise it should be far away from 0. The Wald form of the test statistic that captures this logic is W = [r( ˆβ) q] [Var(r( ˆβ) q)] 1 [r( ˆβ) q] (23) In other words we want to evaluate how far away from 0 is r( ˆβ) q after normalizing it by its average variability. Note that W is a scalar.

23 / 36 Implementation of the test If r( ˆβ) q is normally distributed, under H 0 W χ 2 ρ (24) where the number of degrees of freedom ρ is the number of restrictions to be tested. The difficulty in computing the test statistics is how to determine the variance at the denominator.

24 / 36 The Variance of the Wald Test Statistic Using the Delta Method in a setting in which h( ˆβ) = r( ˆβ) q [ ] [ ] Var[r( ˆβ) r( ˆβ) q] = [Var( ˆβ ˆβ)] r( ˆβ) (25) ˆβ [ ] where note that r( ˆβ) ˆβ is a ρ k + 1 matrix and therefore Var[r( ˆβ) q] is a ρ ρ matrix.

25 / 36 Back to the example Going back to the example in which r( ˆβ) q = R ˆβ q Var[R ˆβ q] = R[Var( ˆβ)]R (26) and the Wald test is W = [R ˆβ q] [RVar( ˆβ)R ] 1 [R ˆβ q] (27) and Var( ˆβ) is in practice estimated by substituting the sample counterparts of the asymptotic variance-covariance matrice

26 / 36 Exercise: Wald test and simple restrictions Consider again the unrestricted regression in matrix form Y = X 1 β 1 + X 2 β 2 + U ur (28) where X 1 is a n 2 matrix including the constant; β 1 is dimension 2 vector of parameters; X 2 is a n 1 matrix; β 2 is dimension 1 vector of parameters; and suppose that we want to test the following joint hypothesis on the β 2 parameters: H 0 : β 2 = 0 against H 1 : β 2 0 (29) What is R in this case?

27 / 36 Exercise: Wald test and simple restriction (cont.) It is easy to verify that in this case the Wald test is W = [R ˆβ q] [RVar( ˆβ)R ] 1 [R ˆβ q] (30) = ˆβ 2 2 Var( ˆβ 2 ) which is the square of a standard t-test, and is distributed as a χ 2 distribution

28 / 36 A drawback of the Wald Test The Wald test is a general form of a large sample test that requires the estimation of the unrestricted model. There are cases in which this may be difficult or even impossible. An alternative large sample testing procedure is the Lagrange Multiplier test, which requires instead only the estimation of the restricted model. A third alternative is the Likelihood Ratio test, which requires the estimation of both the restricted and the unrestricted models, on an equal basis.

29 / 36 Subsection 2 The Lagrange multiplier test in the context of linear regression

30 / 36 The logic of the test In the simple context of linear regression we can define a LM test for multiple exclusion restrictions. Consider again the unrestricted regression in matrix form Y = X 1 β 1 + X 2 β 2 + U ur (31) X 1 is a n k 1 + 1 matrix including the constant; β 1 is dimension k 1 + 1 vector of parameters; X 2 is a n k 2 matrix; β 2 is dimension k 2 vector of parameters; Suppose that we want to test the following joint hypothesis on the β 2 : H 0 : β 2 = 0 against H 1 : β 2 0 (32)

31 / 36 The logic of the test (cont.) Suppose that you estimate the restricted PRF Y = X 1 β r1 + U r (33) where the subscript r indicates that the population parameters and unobservables of this restricted equation may differ from the corresponding one of the unrestricted PRF. It is intuitive to hypothesize that in the auxiliary regression Û r = X 1 γ 1 + X 2 γ 2 + V (34) if the restrictions in the primary PRF are valid then H 0 : β 2 = 0 γ 2 = 0 (35)

32 / 36 The logic of the test (cont.) Let the R-squared of the auxiliary regression 34 be RU 2 and consider the statistics LM = nru 2 (36) If the restrictions are satisfied, this statistics should be close to zero because; X 1 is by construction orthogonal to U R and therefore γ 1 = 0; and γ 2 = 0 if the restrictions are satisfied. Since, given k 2 exclusion restrictions: LM = nru 2 χ 2 k 2 (37) we can use the Classical testing procedure to test H 0.

33 / 36 Subsection 3 The Likelihood Ratio Test in the context of regression

34 / 36 Derivation of the LR test Continuing with the regression framework used to discuss the LM test, the LR test is derived as follows: Get the ML estimate of the unrestricted regression and retrieve the unrestricted normalized log likelihood l U Get the ML estimate of the restricted regression and retrieve the restricted normalized log likelihood l R Then given k 2 the test statisic is LR = 2(l R l U ) χ 2 k 2 (38)

35 / 36 Subsection 4 Final comment on the trinity of tests

36 / 36 Relationship between the three tests From the Handbook Chapter 13 by Robert Engle These three general principles have a certain symmetry which has revolutionized the teaching of hypothesis tests and the development of new procedures Essentially, the Lagrance Multiplier approach starts at the null and asks whether movement toward the alternative would be an improvement...... while the Wald approace starts at the alternative and considers movement toward the null The Likelihood ratio method compares the two hypothesis directly on an equal basis. Figure 3.1 in Engle s chapter is a useful way to think at the three tests.