Economics 620, Lecture 5: exp

Similar documents
Economics 620, Lecture 4: The K-Variable Linear Model I. y 1 = + x 1 + " 1 y 2 = + x 2 + " 2 :::::::: :::::::: y N = + x N + " N

Economics 620, Lecture 4: The K-Varable Linear Model I

Economics 620, Lecture 9: Asymptotics III: Maximum Likelihood Estimation

Lecture 15. Hypothesis testing in the linear model

Economics 620, Lecture 7: Still More, But Last, on the K-Varable Linear Model

Economics 620, Lecture 2: Regression Mechanics (Simple Regression)

Multivariate Linear Models

Economics 573 Problem Set 5 Fall 2002 Due: 4 October b. The sample mean converges in probability to the population mean.

plim W 0 " 1 N = 0 Prof. N. M. Kiefer, Econ 620, Cornell University, Lecture 16.

Computational functional genomics

Peter Hoff Linear and multilinear models April 3, GLS for multivariate regression 5. 3 Covariance estimation for the GLM 8

DS-GA 1002 Lecture notes 10 November 23, Linear models

REVIEW (MULTIVARIATE LINEAR REGRESSION) Explain/Obtain the LS estimator () of the vector of coe cients (b)

MLES & Multivariate Normal Theory

LECTURE 13: TIME SERIES I

STAT 100C: Linear models

' Liberty and Umou Ono and Inseparablo "

Maximum Likelihood Estimation

Suggested Solution for PS #5

MA 575 Linear Models: Cedric E. Ginestet, Boston University Midterm Review Week 7

2 Statistical Estimation: Basic Concepts

Linear Regression. y» F; Ey = + x Vary = ¾ 2. ) y = + x + u. Eu = 0 Varu = ¾ 2 Exu = 0:

Ma 3/103: Lecture 24 Linear Regression I: Estimation

Large Sample Properties of Estimators in the Classical Linear Regression Model

Properties of the least squares estimates

TUTORIAL 8 SOLUTIONS #

LECTURE 11: GENERALIZED LEAST SQUARES (GLS) In this lecture, we will consider the model y = Xβ + ε retaining the assumption Ey = Xβ.

1. The Multivariate Classical Linear Regression Model

Gaussian random variables inr n

Lecture Notes. Introduction

Inverse of a Square Matrix. For an N N square matrix A, the inverse of A, 1

AGEC 621 Lecture 16 David Bessler

Moment Generating Function. STAT/MTHE 353: 5 Moment Generating Functions and Multivariate Normal Distribution

Multivariate Regression

Lecture Notes 15 Prediction Chapters 13, 22, 20.4.

Economics 620, Lecture 13: Time Series I

ECONOMETRICS II (ECO 2401S) University of Toronto. Department of Economics. Spring 2013 Instructor: Victor Aguirregabiria

CS229 Lecture notes. Andrew Ng

Linear Models Review

Part IB Statistics. Theorems with proof. Based on lectures by D. Spiegelhalter Notes taken by Dexter Chua. Lent 2015

Ma/CS 6b Class 23: Eigenvalues in Regular Graphs

We begin by thinking about population relationships.

THE UNIVERSITY OF CHICAGO Booth School of Business Business 41912, Spring Quarter 2016, Mr. Ruey S. Tsay

PCA and admixture models

Regression: Lecture 2

R = µ + Bf Arbitrage Pricing Model, APM

Notes on Time Series Modeling

The Multivariate Gaussian Distribution

Economics 620, Lecture 14: Lags and Dynamics

The purpose of this section is to derive the asymptotic distribution of the Pearson chi-square statistic. k (n j np j ) 2. np j.

CAS MA575 Linear Models

Restricted Maximum Likelihood in Linear Regression and Linear Mixed-Effects Model

The outline for Unit 3

Exercises Chapter 4 Statistical Hypothesis Testing

Educjatipnal. L a d ie s * COBNWALILI.S H IG H SCHOOL. I F O R G IR L S A n B k i n d e r g a r t e n.

Variations. ECE 6540, Lecture 10 Maximum Likelihood Estimation

For GLM y = Xβ + e (1) where X is a N k design matrix and p(e) = N(0, σ 2 I N ), we can estimate the coefficients from the normal equations

Linear models. Linear models are computationally convenient and remain widely used in. applied econometric research

Multilevel Models in Matrix Form. Lecture 7 July 27, 2011 Advanced Multivariate Statistical Methods ICPSR Summer Session #2

[y i α βx i ] 2 (2) Q = i=1

THE UNIVERSITY OF CHICAGO Graduate School of Business Business 41912, Spring Quarter 2008, Mr. Ruey S. Tsay. Solutions to Final Exam

Introduction to Machine Learning

MA 575 Linear Models: Cedric E. Ginestet, Boston University Mixed Effects Estimation, Residuals Diagnostics Week 11, Lecture 1

Lecture 13: Data Modelling and Distributions. Intelligent Data Analysis and Probabilistic Inference Lecture 13 Slide No 1

ECONOMETRICS II (ECO 2401S) University of Toronto. Department of Economics. Winter 2014 Instructor: Victor Aguirregabiria

Ross (1976) introduced the Arbitrage Pricing Theory (APT) as an alternative to the CAPM.

Lecture 34: Properties of the LSE

Linear Regression. In this problem sheet, we consider the problem of linear regression with p predictors and one intercept,

x i = 1 yi 2 = 55 with N = 30. Use the above sample information to answer all the following questions. Show explicitly all formulas and calculations.

Next is material on matrix rank. Please see the handout

High-resolution Parametric Subspace Methods

Supplemental Material 1 for On Optimal Inference in the Linear IV Model

MC3: Econometric Theory and Methods. Course Notes 4

Lecture 07 Hypothesis Testing with Multivariate Regression

STAT 100C: Linear models

Econometrics I. Professor William Greene Stern School of Business Department of Economics 5-1/36. Part 5: Regression Algebra and Fit

E2 212: Matrix Theory (Fall 2010) Solutions to Test - 1

Chapter 4: Factor Analysis

Probability Background

Likelihood Methods. 1 Likelihood Functions. The multivariate normal distribution likelihood function is

Probabilistic Graphical Models

Brief Suggested Solutions

Introduction to Estimation Methods for Time Series models. Lecture 1

Random vectors X 1 X 2. Recall that a random vector X = is made up of, say, k. X k. random variables.

Fiedler s Theorems on Nodal Domains

Multivariate Regression: Part I

Asymptotics for Nonlinear GMM

This model of the conditional expectation is linear in the parameters. A more practical and relaxed attitude towards linear regression is to say that

3. For a given dataset and linear model, what do you think is true about least squares estimates? Is Ŷ always unique? Yes. Is ˆβ always unique? No.

Computer Vision Group Prof. Daniel Cremers. 2. Regression (cont.)

Mathematical foundations - linear algebra

Mathematical foundations - linear algebra

Review: General Approach to Hypothesis Testing. 1. Define the research question and formulate the appropriate null and alternative hypotheses.

Partitioned Covariance Matrices and Partial Correlations. Proposition 1 Let the (p + q) (p + q) covariance matrix C > 0 be partitioned as C = C11 C 12

Switching Regime Estimation

SOLUTIONS TO PROBLEMS

Regression #5: Confidence Intervals and Hypothesis Testing (Part 1)

Lecture 3: Multiple Regression

Cointegrated VAR s. Eduardo Rossi University of Pavia. November Rossi Cointegrated VAR s Financial Econometrics / 56

Copula Regression RAHUL A. PARSA DRAKE UNIVERSITY & STUART A. KLUGMAN SOCIETY OF ACTUARIES CASUALTY ACTUARIAL SOCIETY MAY 18,2011

Transcription:

1 Economics 620, Lecture 5: The K-Variable Linear Model II Third assumption (Normality): y; q(x; 2 I N ) 1 ) p(y) = (2 2 ) exp (N=2) 1 2 2(y X)0 (y X) where N is the sample size. The log likelihood function is `(; 2 ) = c Proposition: N 2 ln 2 1 2 2(y X)0 (y X): The LS estimator ^ is the ML estimator.

2 Proposition: The ML estimator for 2 is 2 ML = e0 e=n. Proof : FOC: To nd the ML estimator for 2, we solve the @` N @ 2 = 2 2 + 1 2 4(y X)0 (y X) = 0 ) 2 = (y X) 0 (y X)=N Plugging in the MLE for gives the MLE for 2 Proposition: The distribution of ^ given a value of 2 is q(; 2 (X 0 X) 1 ). Proof : Since ^ is a linear combination of jointly normal variables, it is normal.

3 Fact: If A is an N N idempotent matrix with rank r, then there exists an N N matrix C with C 0 C = I = CC 0 (orthogonal) C 0 AC =, where: = 2 6 4 1:::0:::0 0:::1:::0 :::::::::: 0::::::::0 3 7 5 = Ir 0 0 0 # : C is the matrix whose columns are orthornormal eigenvectors of A.

4 Lemma: Let z q(0; I N ) and A be an N N idempotent matrix with rank r. Then z 0 Az 2 (r): Proof : z 0 Az = z 0 CC 0 ACC 0 z = ~zc 0 AC ~z = ~z 0 ~z, where ~z 0 = z 0 C. But ~z is normal with mean zero and variance: E~z~z 0 = EC 0 zz 0 C = C 0 (Ezz 0 )C = C 0 C = I. So, z 0 Az = ~z 0 ~z is the sum of squares of r standard normal variables, i.e., z 0 Az 2 (r).

5 Proposition: N 2 ML 2 2 (N K) Proof : Note that 2 ML = e0 e=n = 0 M=N: ) N2 ML 2 = 0 M 2 2 (N K); using the previous lemma with z = =. Proposition: cov( 2 ML ; ^) = 0 Proof : Ee(^ ) 0 = EM((X 0 X) 1 X 0 ) 0 = EM 0 X(X 0 X) 1 = 2 MX(X 0 X) 1 = 0

6 ) e and ^ are independent. (This depends on normality: zero covariance ) independence) ) e 0 e and ^ are independent. So, we have the complete sampling distribution of ^ and 2 ML : Note on t-testing: We now that ^ k k q(0; 1) where 2, is the kth diagonal element of 2 (X 0 X) 1. Estimating 2 by s 2 gives a statistic which is t(n using the same argument as in simple regression. K),

7 Simultaneous Restrictions In multiple regression we can test several restrictions simultaneously. Why is this useful? Recall our expenditure system: ln z j = ln a j P a` + ln m ln p j or y = 0 + 1 ln m + 2 ln p j + We are interested in the hypothesis 1 = 1 and 2 = 1. A composite hypothesis like this cannot be tested with the tools we have developed so far.

8 Lemma: Let z q(0; I), and A and B be symmetric idempotent matrices such that AB = 0: Thus A and B are projections to orthogonal spaces. Then a = z 0 Az and b = z 0 Bz are independent. Proof : a = z 0 A 0 Az = sum of squares of Az b = z 0 B 0 Bz = sum of squares of Bz. Note that both Az and Bz are normal with mean zero. cov(az; Bz) = EAzz 0 B 0 = AEzz 0 B 0 = AB 0 = 0 We are done. (why?) Note: A similar argument shows that z 0 Az and Lz are independent if AL 0 = 0.

9 Testing De nition: Suppose v 2 (k) and u 2 (p) are independent. Then F = v=k u=p F (k; p): Lemma: Let M and M be idempotent with MM = M, e = M, e = M, q(0; 2 I): Then F = (e0 e e 0 e )=(trm trm ) e 0 e =trm F (trm trm, trm ).

10 Proof : 2 trm times the denominator is 2 (trm ) As for the numerator: e 0 e e 0 e = 0 M 0 M 0 M 0 M = 0 (M M ). Note that: (M M )(M M ) = M 2 M M MM + M 2 = M M (idempotent). So e 0 e e 0 e = 0 (M M ). Thus, the numerator upon multiplication by 2 tr(m M ) is distributed as 2 (tr(m M )). It only remains to show that the numerator and the denominator are independent. But (M M )M = 0, so we are done.

11 Interpretation: R[M ] R[M], i.e. e 0 e is a restricted sum of squares e 0 e is an unrestricted sum of squares. F looks at the normalized reduction in t caused by the restriction. What sort of restrictions meet the conditions of the lemma? Proposition: Let X be N H and X be N K where H < K. (R[X] R[X ]). Suppose X = X A (A is K H). Let M = I X(X 0 X) 1 X 0 and M = I X (X 0 X ) 1 X 0. Then M and M are idempotent and MM = M.

12 Example 1: Leaving out variables Consider y = X 1 1 + X 2 2 + where X 1 is N K 1 and X 2 is N K 2. Hypothesis: 2 = 0, i.e., X 2 is not in the model. Using the notation from the previous proposition, X = X 1 and X = [X 1 X 2 ] X = X A, A = I 0 Note that: trm = N K 1, trm = N K 1 K 2. Thus: F = (e0 e e 0 e )=K 2 e 0 e =(N K 1 K 2 ) : e is from the regression of y on X = X 1, and e is from the regression of y on X = [X 1 X 2 ]. The degrees of freedom in the numerator is the number of restrictions. #

13 Example 2: Testing the equality of regression coe cients in two samples. Consider y 1 = X 1 1 + 1 where y 1 is N 1 1 and X 1 is N 1 K, and y 2 = X 2 2 + 2 where y 2 is N 2 1 and X 2 is N 2 K. Hypothesis: 1 = 2 Combine the observations from the samples: y1 # X1 0 # 1 # y = y 2, X = 0 X 2, = 2

14 The unrestricted model is y = X + = X 1 1 + X 2 2 + : X = X I A, A = I # Note that trm = N 1 + N 2 2K and trm = N 1 + N 2 K: Run the restricted and unrestricted regressions, and calculate (e 0 e e 0 e )=K F = e 0 e =(N 1 + N 2 2K) :

15 Example 3: Testing the equality of a subset of coe cients Consider y 1 = X 1 1 + X 2 2 + 1 where X 1 is N 1 K 1 and X 2 is N 1 K 2 and y 2 = X 3 3 + X 4 4 + 2 where X 3 is N 2 K 1 and X 4 is N 2 K 4 Hypothesis: 1 = 3 The unrestricted regression is y = y1 y 2 # = X1 X 2 0 0 0 0 X 3 X 4 2 # 6 4 1 2 3 4 3 7 5 + = X +.

16 With the restriction, we have y = X1 X 2 0 X 3 0 X 4 = X ~ + : X = X A, A = 2 6 4 # 2 6 4 1 2 3 I 0 0 0 I 0 I 0 0 0 0 I 3 7 5 + 3 7 5 : Thus, the test statistics is: F = (e 0 e e 0 e )=K 1 e 0 e =(N 1 + N 2 2K 1 K 2 K 4 ) :

17 Another way to look at the condition of the lemma: Let be the unrestricted coe cient vector and be the restricted coe cient vector. The lemma requires that there exist a matrix A such that = A. What kinds or restrictions cannot be brought into this framework?? Consider Ey = X 1 1 versus Ey = X 2 2 : The combined model is not in consideration.