Multiple Linear Regression

Similar documents
coefficients n 2 are the residuals obtained when we estimate the regression on y equals the (simple regression) estimated effect of the part of x 1

Econometrics I KS. Module 2: Multivariate Linear Regression. Alexander Ahammer. This version: April 16, 2018

The Simple Regression Model. Simple Regression Model 1

Multiple Regression Analysis: Heteroskedasticity

Testing Linear Restrictions: cont.

Introductory Econometrics

Econometrics Multiple Regression Analysis: Heteroskedasticity

Multiple Linear Regression CIVL 7012/8012

Introductory Econometrics

Homoskedasticity. Var (u X) = σ 2. (23)

Lecture 4: Multivariate Regression, Part 2

Econometrics. Week 8. Fall Institute of Economic Studies Faculty of Social Sciences Charles University in Prague

1 Motivation for Instrumental Variable (IV) Regression

Multiple Regression Analysis

Review of Econometrics

Econometrics. Week 4. Fall Institute of Economic Studies Faculty of Social Sciences Charles University in Prague

Heteroskedasticity and Autocorrelation

Econometrics - 30C00200

LECTURE 2 LINEAR REGRESSION MODEL AND OLS

Intermediate Econometrics

Multiple Regression Analysis

Multivariate Regression Analysis

Business Economics BUSINESS ECONOMICS. PAPER No. : 8, FUNDAMENTALS OF ECONOMETRICS MODULE No. : 3, GAUSS MARKOV THEOREM

Reliability of inference (1 of 2 lectures)

Lecture 4: Multivariate Regression, Part 2

ECNS 561 Multiple Regression Analysis

ECON3150/4150 Spring 2016

Introduction to Econometrics. Heteroskedasticity

Econometrics Summary Algebraic and Statistical Preliminaries

Statistical Inference with Regression Analysis

1. You have data on years of work experience, EXPER, its square, EXPER2, years of education, EDUC, and the log of hourly wages, LWAGE

LECTURE 10. Introduction to Econometrics. Multicollinearity & Heteroskedasticity

Introductory Econometrics

ECON3150/4150 Spring 2015

Applied Econometrics (QEM)

ECON Introductory Econometrics. Lecture 6: OLS with Multiple Regressors

Heteroskedasticity. Part VII. Heteroskedasticity

Model Mis-specification

THE MULTIVARIATE LINEAR REGRESSION MODEL

Lecture 3: Multiple Regression

Wooldridge, Introductory Econometrics, 4th ed. Chapter 15: Instrumental variables and two stage least squares

statistical sense, from the distributions of the xs. The model may now be generalized to the case of k regressors:

Graduate Econometrics Lecture 4: Heteroskedasticity

Heteroskedasticity ECONOMETRICS (ECON 360) BEN VAN KAMMEN, PHD

INTRODUCTION TO BASIC LINEAR REGRESSION MODEL

Least Squares Estimation-Finite-Sample Properties

Multiple Regression Analysis. Part III. Multiple Regression Analysis

Problem Set #6: OLS. Economics 835: Econometrics. Fall 2012

Steps in Regression Analysis

Business Statistics. Lecture 10: Correlation and Linear Regression

Econometrics Honor s Exam Review Session. Spring 2012 Eunice Han

13. Time Series Analysis: Asymptotics Weakly Dependent and Random Walk Process. Strict Exogeneity

The Statistical Property of Ordinary Least Squares

The general linear regression with k explanatory variables is just an extension of the simple regression as follows

ECON 3150/4150, Spring term Lecture 7

Introduction to Econometrics Chapter 6

1 The Multiple Regression Model: Freeing Up the Classical Assumptions

OSU Economics 444: Elementary Econometrics. Ch.10 Heteroskedasticity

Ma 3/103: Lecture 24 Linear Regression I: Estimation

Multiple Regression Analysis

Applied Statistics and Econometrics

Simple Linear Regression Model & Introduction to. OLS Estimation

Econometrics II - EXAM Answer each question in separate sheets in three hours

Intermediate Econometrics

Regression with a Single Regressor: Hypothesis Tests and Confidence Intervals

Spatial Regression. 3. Review - OLS and 2SLS. Luc Anselin. Copyright 2017 by Luc Anselin, All Rights Reserved

Maximum Likelihood (ML) Estimation

Heteroscedasticity and Autocorrelation

Ability Bias, Errors in Variables and Sibling Methods. James J. Heckman University of Chicago Econ 312 This draft, May 26, 2006

Applied Quantitative Methods II

Heteroskedasticity. Occurs when the Gauss Markov assumption that the residual variance is constant across all observations in the data set

Heteroskedasticity (Section )

Simple Linear Regression

Outline. 11. Time Series Analysis. Basic Regression. Differences between Time Series and Cross Section

Linear Regression. Junhui Qian. October 27, 2014

1 Appendix A: Matrix Algebra

Regression Analysis. y t = β 1 x t1 + β 2 x t2 + β k x tk + ϵ t, t = 1,..., T,

Heteroskedasticity. (In practice this means the spread of observations around any given value of X will not now be constant)

LECTURE 5. Introduction to Econometrics. Hypothesis testing

ECON The Simple Regression Model

Econometrics. 7) Endogeneity

Chapter 8 Heteroskedasticity

Lecture 4: Heteroskedasticity

Basic econometrics. Tutorial 3. Dipl.Kfm. Johannes Metzler

Iris Wang.

x i = 1 yi 2 = 55 with N = 30. Use the above sample information to answer all the following questions. Show explicitly all formulas and calculations.

Regression #3: Properties of OLS Estimator

Basic Econometrics - rewiev

Inference in Regression Analysis

Ordinary Least Squares Regression

Warwick Economics Summer School Topics in Microeconometrics Instrumental Variables Estimation

INTRODUCTORY ECONOMETRICS

Wooldridge, Introductory Econometrics, 2d ed. Chapter 8: Heteroskedasticity In laying out the standard regression model, we made the assumption of

Linear models. Linear models are computationally convenient and remain widely used in. applied econometric research

Multicollinearity. Filippo Ferroni 1. Course in Econometrics and Data Analysis Ieseg September 22, Banque de France.

Econometrics. Final Exam. 27thofJune,2008. Timeforcompletion: 2h30min

Econometrics I Lecture 3: The Simple Linear Regression Model

Applied Econometrics (QEM)

Lecture: Simultaneous Equation Model (Wooldridge s Book Chapter 16)

Chapter 2 The Simple Linear Regression Model: Specification and Estimation

Transcription:

Multiple Linear Regression Asymptotics Asymptotics

Multiple Linear Regression: Assumptions Assumption MLR. (Linearity in parameters) Assumption MLR. (Random Sampling from the population) We have a random sample: satisfying the equation above Assumption MLR.3 (No perfect Collinearity) In the sample, none of the independent variables is a linear combination of the others. Assumption MLR.4 (Zero Conditional Mean) this implies Assumption MLR.5 (Homoskedasticity) Var( u x σ, x,..., x k ) These are the Gauss-Markov assumptions. Under MLR. to MLR.5 OLS is BLUE Asymptotics

Consistency Under the Gauss-Markov assumptions OLS is BLUE (Best Linear Unbiased Estimator) With only MLR. to MLR.4 we have shown that OLS is unbiased Under weaker assumptions, we cannot prove unbiasedness In those cases, we may be happy with estimators that are consistent, meaning that as n, the distribution of the estimator collapses to the parameter value (or: the probability that the estimator is really close to the true parameter converges to ) Asymptotics 3

Sampling Distributions as n n 3 n < n < n 3 n n Asymptotics 4

Consistency of OLS Under MLR. to MLR.4, the OLS estimator is consistent besides being unbiased For unbiasedness, assumption MLR.4 was crucial: E(u x, x,,x k ) 0 Consistency can be proved with the weaker assumption MLR.4 : E(u) 0 and Cov(x,u) 0, for,,, k Will need to take probability limits (plim) to establish consistency Asymptotics 5

Rough proof of Consistency For the simple linear regression, using assumptions MLR. to MLR.4 : ˆ plimˆ ( ) i ( ) u n x x ( ( x x ) y ) ( x x ) + i because Cov ( n ( x x ) ) ( ) + ( x, u) 0 i Cov i ( x, u) Var( x ) i E(u) 0 and Cov(x,u) 0, for,,, k is a crucial assumption Without this assumption, OLS will be inconsistent Asymptotics 6 i

Inconsistency with omitted variables Just as we analysed the omitted variable bias earlier, now we can think about inconsistency, or asymptotic bias True model: You think : u x whereδ + v y y and, Cov ( x, x ) Var( x ) 0 0 + + x ~ plim x + + x + v + u, so that Asymptotic bias depends on the population variance and covariance, while bias uses the sample counterparts δ Asymptotics 7

Large Sample Inference With assumption MLR.6 (which makes MLR.4 and MLR.5 redundant), the sampling distributions of the estimators are normal, so we could derive t and F distributions for testing Assumption MLR.6 (Normality) The distribution of the population error u is independent of x, x,,x k and u is normally distributed with mean 0 and variance σ We write: u ~ Normal (0,σ ) What if normality is an unreasonable assumption? (remember wage regression) It turns out that only with MLR. through MLR.5, the distribution of the estimators previously derived is still valid asymptotically Asymptotics 8

Asymptotic Normality Under assumptions MLR. through MLR.5: (i) where a x n ( ) a ˆ ~ Normal( 0, σ a ) plim ( n rˆ ) and rˆ on all the other regressors; σ i i, are the residuals from regressing a is called the asymptotic variance (ii) σˆ is a consistent estimator of σ (iii) ( ) ( ) a ˆ se ˆ ~ Normal( 0,) Asymptotics 9

Asymptotic Standard Errors With assumptions MLR. through MLR.5, we have thus asymptotic normality of the OLS estimators If we cannot guarantee that normality holds (MLR.6) we refer to the standard errors as asymptotic standard errors Formulas for standard errors are exactly the same as before, but the derived distribution of the estimators is not exact, it is only asymptotic se ( ˆ ) σˆ SST ( ), R Asymptotics 0

Asymptotic Standard Errors se se and ( ˆ ) ( ˆ ) SST c σˆ SST ( R ) nvar( x ), n, since R converges to some constant Standard errors are expected to decrease at a rate equal to the inverse of the square root of the sample size Can perform hypothesis tests on individual coefficients ust as before, using the Normal distribution (although t-distribution could also be used). Can also perform F tests since the F statistics will also have approximate F distributions in large samples Asymptotics

Lagrange Multiplier statistic With only assumptions MLR. through MLR.5 we rely on large sample distributions for inference. But we can use more than t and F statistics The Lagrange multiplier or LM statistic is an alternative for testing multiple exclusion restrictions Asymptotics

LM Statistic (cont) Suppose we have a standard model, y 0 + x + x +... k x k + u and our null hypothesis is H 0 : k-q+ 0,..., k 0; the number of restrictions is q First, we ust run the restricted model ~ ~ ~ y ~ 0+ x+... + k qxk q+ u Now take the residuals, u~, and regress u~ on x, x,..., x (i.e. all the variables) LM nr u, k where R u is the R - squared from this regression Asymptotics 3

LM Statistic (cont) LM a ~ χ q value, c, from a under the null, so can choose a critical level or calculate the p - value χ q distribution, given a significance With a large sample, the results from an F test and from an LM test should be similar Asymptotics 4

LM Statistic (Example) Dependent variable: Log of monthly wages, n064 Test wether Labor market experience (and its square) have an effect on wages, α0.05 q (number of restrictions) LM obs 064 0.4338.996>5.99 so we reect the null Asymptotics 5

Asymptotic Efficiency Other estimators besides OLS will be consistent However, under the Gauss-Markov assumptions, the OLS estimators will have the smallest asymptotic variances (check theorem 5.3 in the textbook) We say that OLS is asymptotically efficient Asymptotics 6