Economics 240A, Section 3: Short and Long Regression (Ch. 17) and the Multivariate Normal Distribution (Ch. 18)

Similar documents
In the bivariate regression model, the original parameterization is. Y i = β 1 + β 2 X2 + β 2 X2. + β 2 (X 2i X 2 ) + ε i (2)

Review of Classical Least Squares. James L. Powell Department of Economics University of California, Berkeley

Reference: Davidson and MacKinnon Ch 2. In particular page

Econ 2120: Section 2

Interpreting Regression Results

Regression #5: Confidence Intervals and Hypothesis Testing (Part 1)

[y i α βx i ] 2 (2) Q = i=1

Multilevel Models in Matrix Form. Lecture 7 July 27, 2011 Advanced Multivariate Statistical Methods ICPSR Summer Session #2

STAT 100C: Linear models

MA 575 Linear Models: Cedric E. Ginestet, Boston University Midterm Review Week 7

Part 6: Multivariate Normal and Linear Models

So far our focus has been on estimation of the parameter vector β in the. y = Xβ + u

Ordinary Least Squares Regression

Next is material on matrix rank. Please see the handout

Chapter 5. The multivariate normal distribution. Probability Theory. Linear transformations. The mean vector and the covariance matrix

Interpreting Regression Results -Part II

Multiple Regression Model: I

Introduction to Estimation Methods for Time Series models. Lecture 1

A Probability Review

Lecture 6: Geometry of OLS Estimation of Linear Regession

Problem Set #6: OLS. Economics 835: Econometrics. Fall 2012

Multivariate Regression (Chapter 10)

ECON 3150/4150, Spring term Lecture 6

Algebra of Random Variables: Optimal Average and Optimal Scaling Minimising

Math 423/533: The Main Theoretical Topics

Introduction to Econometrics. Heteroskedasticity

Lecture 3: Linear Models. Bruce Walsh lecture notes Uppsala EQG course version 28 Jan 2012

Linear Algebra Review

We begin by thinking about population relationships.

ECON The Simple Regression Model

ST 740: Linear Models and Multivariate Normal Inference

STAT5044: Regression and Anova. Inyoung Kim

The Simple Regression Model. Part II. The Simple Regression Model

Financial Econometrics

Copula Regression RAHUL A. PARSA DRAKE UNIVERSITY & STUART A. KLUGMAN SOCIETY OF ACTUARIES CASUALTY ACTUARIAL SOCIETY MAY 18,2011

Multivariate Regression Analysis

1 Appendix A: Matrix Algebra

Introduction to Normal Distribution

Brief Suggested Solutions

MS&E 226: Small Data

Ma 3/103: Lecture 24 Linear Regression I: Estimation

Multivariate Regression: Part I

Functions of Several Variables

Gaussian Processes. Le Song. Machine Learning II: Advanced Topics CSE 8803ML, Spring 2012

Lecture 4: Testing Stuff

MS&E 226: Small Data. Lecture 11: Maximum likelihood (v2) Ramesh Johari

Topic 7: Heteroskedasticity

Peter Hoff Linear and multilinear models April 3, GLS for multivariate regression 5. 3 Covariance estimation for the GLM 8

PANEL DATA RANDOM AND FIXED EFFECTS MODEL. Professor Menelaos Karanasos. December Panel Data (Institute) PANEL DATA December / 1

Simultaneous Equation Models Learning Objectives Introduction Introduction (2) Introduction (3) Solving the Model structural equations

Weighted Least Squares

The Linear Regression Model

The outline for Unit 3

Economics 583: Econometric Theory I A Primer on Asymptotics

ECONOMICS 210C / ECONOMICS 236A MONETARY HISTORY

Linear Regression 9/23/17. Simple linear regression. Advertising sales: Variance changes based on # of TVs. Advertising sales: Normal error?

Restricted Maximum Likelihood in Linear Regression and Linear Mixed-Effects Model

Essential of Simple regression

Randomized Complete Block Designs

Topic 3: Inference and Prediction

STA 2201/442 Assignment 2

Multivariate Random Variable

Lecture 19. Condence Interval

Lecture 9 SLR in Matrix Form

The Multivariate Normal Distribution

Answers to Problem Set #4

Elliptically Contoured Distributions

The Multivariate Gaussian Distribution

Bayesian linear regression

STAT 100C: Linear models

Lecture 15. Hypothesis testing in the linear model

CAS MA575 Linear Models

Regression Analysis Tutorial 34 LECTURE / DISCUSSION. Statistical Properties of OLS

3. For a given dataset and linear model, what do you think is true about least squares estimates? Is Ŷ always unique? Yes. Is ˆβ always unique? No.

Probability. Table of contents

Topic 3: Inference and Prediction

The Matrix Algebra of Sample Statistics

The linear model is the most fundamental of all serious statistical models encompassing:

CHAPTER 6: SPECIFICATION VARIABLES

Econ 510 B. Brown Spring 2014 Final Exam Answers

FENG CHIA UNIVERSITY ECONOMETRICS I: HOMEWORK 4. Prof. Mei-Yuan Chen Spring 2008

Bayesian Linear Regression [DRAFT - In Progress]

Lecture 11. Multivariate Normal theory

Projection. Ping Yu. School of Economics and Finance The University of Hong Kong. Ping Yu (HKU) Projection 1 / 42

Heteroskedasticity ECONOMETRICS (ECON 360) BEN VAN KAMMEN, PHD

Multiple Random Variables

01 Probability Theory and Statistics Review

Econometrics I KS. Module 2: Multivariate Linear Regression. Alexander Ahammer. This version: April 16, 2018

MBF1923 Econometrics Prepared by Dr Khairul Anuar

Problem Set - Instrumental Variables

Categorical Predictor Variables

Inverse of a Square Matrix. For an N N square matrix A, the inverse of A, 1

The returns to schooling, ability bias, and regression

Probability Background

Statistics Homework #4

5.1 Consistency of least squares estimates. We begin with a few consistency results that stand on their own and do not depend on normality.

Lecture 13: Simple Linear Regression in Matrix Format. 1 Expectations and Variances with Vectors and Matrices

Lecture 4: Heteroskedasticity

Motivation for multiple regression

Transcription:

Economics 240A, Section 3: Short and Long Regression (Ch. 17) and the Multivariate Normal Distribution (Ch. 18) MichaelR.Roberts Department of Economics and Department of Statistics University of California at Berkeley March 26, 2000 1 Introduction This handout reviews some of the key points regarding regression algebra and the multivariate normal distribution. It follows closely Goldberger Ch. s 17 and Ch. 18. 2 Short and Long Regressions The basic set-up is y = Xβ + ε = X 1 β 1 + X 2 β 2 + ε (1) wherewehavepartitionedthen k matrix X into two submatrices X 1 R n k 1 and X 2 R n k 2. We can think of two regressions: 1. a short one y = X 1 β 1 + ε (2) and, 2. a long one y = X 1 β 1 + X 2 β 2 + ε (3) I ll use the same notation as Goldberger so let b i be a vector of OLS parameter estimates for the subvector β i in a long regression and b i be the OLS estimates of β i in a short regression. And, let e be the residuals from the long regression and e be the residuals from the short regression. 1

2 Exercise 1: Let b 1 be the OLS estimates of β 1 from regression 2 and b 1 be the OLS estimates of β 1 from regression 3. Show b 1 = b 1 + ( X 1X ) 1 1 X 1 X 2 b 2 (4) Exercise 2: Letting e be the residuals from 2 show that e = M 1 X 2 b 2 + e In words, what is M 1 X 2? Exercise 3: Show that e e = b 2X 2M 1 X 2 b 2 + e e and interpret this result. What implication does this have for the fit of the long regression relative to the short regression? Result 1 Some exceptions 1. If b 2 =0then b 1 = b 1 and e = e. 2. If X 1X 2 =0then b 1 = b 1 but e e. 3 Frisch-Waugh-Lovell Problem 2 proves the Frisch-Waugh-Lovell theorem which can be thought of as an alternative way of getting at the OLS estimator of β 2. 1. Regress each column of X 2 on X 1 and save the corresponding set of residuals in a matrix, X2. 2. Regress y on X 1 and save its residual as y.(in fact, this step is unnecessary and Goldberger refers to this as a double residual regression. Exercise: 4 Prove that this step is in fact unnecessary) 3. Regress y on X2 and the resulting coefficient vector is the same as the OLS coefficients from the original regression in 1.

3 To see this consider regressing y on M 1 X 2 (= X 2 in Goldberger). The coefficient vector is c 2 = ( X 2M ) 1 1 X 2 X 2 M 1 y = ( X 2M ) 1 1 X 2 X 2 M 1 (X 1 b 1 + X 2 b 2 + e) = ( X 2 M ) 1 1X 2 X 2 M 1 (X 2 b 2 + e) (M 1 X 1 =0) = b 2 (cancelling and noting M 1 e = e and X 2e =0) For some applications see section 17.4. 4 The CR Model Recall the set-up E (y) = Xβ = X 1 β 1 + X 2 β 2 V (y) = σ 2 I X : full rank and nonstochastic 4.1 The Parameters Exercise 5: (Omitted Variables Bias): Show that the estimated coefficients from the short regression (2) b 1 are biased. Exercise 6: What is the variance of the short regression coefficents b 1 and what is its relation relative to the variance of the long regression coefficients b 1? 4.2 The Residuals Exercise 7: Find the expectation and variance of the short regression residual vector e. Exercise 8: Find the expectation of the sum of squared residuals, e e.

4 5 The Normal Distribution You better become REAL familiar with this. There are just a zillion different properties that the normal (univariate and multivariate) distribution has. Here s a short list of some things that might be worth knowing: 5.1 Univariate Normal Distribution 1. X N ( µ, σ 2) means X has a univariate normal distribution with mean parameter µ and σ 2. The density is of course { 1 f X (x) = exp 1 (x µ) 2 } 2πσ 2 2 σ 2 which is often denoted by φ (x) and there is no closed form for the corresponding distribution, Φ(x) 2. The distribution is symmetric implying Φ( x) =1 Φ(x) This is easily seen by thinking of the area under the normal density. 3. Closed under affine transformations. If x N ( µ, σ 2) then y = α+βx is distributed N ( α + βµ, β 2 σ 2). 4. Is uniquely determined by it s first two moments. 5. φ (x) =xφ (x) 6. If Z is standard normal than all odd moments are equal to 0 and E (Z 2k) = (2k)! 2 k, k =1, 2, 3,... k! (This can be shown using integration by parts and induction) 5.2 Multivariate Normal Distribution 1. The vector x R n is distributed multivariate normal with mean vector µ and variance-covariance matrix Σ and has the corresponding density { f X (x) =(2π) n/2 Σ 1/2 exp 1 } 2 (x µ) Σ 1 (x µ) where means determinant.

5 2. Two normal random variables are independent if and only if they are uncorrelated. 3. Affine transformations of a vector of normal random variables are again normal. So, if x MVN (µ, Σ) then y = Hx + b is distributed multivariate normal with mean Hµ + b and variance HΣH 4. Important!!! Consider a pair of random vectors x and y each multivariate normal such that (x, y ) has mean and covariance matrix given by µ = µ x and Σ = Σ xx Σ xy µ y Σ yx Σ yy respectively. Then the distribution of x conditional on y is also multivariate normal with mean ) µ x y = µ x + Σ xy Σ 1 yy (y µ y and covariance matrix Σ x y = Σ xx Σ xy Σ 1 yy Σ yx Note that the conditional covariance matrix does not depend on y and that while Σ and Σ yy are assumed to be nonsingular, Σ 1 yy can be replaced by a pseudo inverse. 5.3 Functions of Normal Random Variables 1. Let x be a k dimensional vector of standard normal random variables. Then x x is distributed χ 2 with k degrees of freedom. 2. Extending the above result, if x R n is distributed MVN (µ, Σ) then (x µ) Σ (x µ) is distributed χ 2 n 3. If x R n is distributed MVN (0, I)andM is any nonrandom idempotent matrix with rank r n then u Mu is distributed χ 2 r. 4. Let x R n be distributed MVN (0, I). Let M be any nonrandom idempotent matrix with rank r n and let L be a nonrandom matrix such that LM =0. Then a = Mu and b = Lu are independent random vectors. 5. Let v χ 2 n and w χ 2 d be two independent chi-square random variables. Then z = v/n w/d is distributed Snedecor-F : F (n, m)

6 6. Let z N (0, 1) and w χ 2 n independent of z. Then t = z w/n has a Student s t-distribution with n degrees of freedom (t n ) 7. If u t n then u 2 F (1,n).