LINEAR REGRESSION ANALYSIS. MODULE IX Lecture Multicollinearity

Similar documents
LINEAR REGRESSION ANALYSIS. MODULE IX Lecture Multicollinearity

Chapter 7 Generalized and Weighted Least Squares Estimation. In this method, the deviation between the observed and expected values of

Polynomial Regression Models

Dr. Shalabh Department of Mathematics and Statistics Indian Institute of Technology Kanpur

Interval Estimation in the Classical Normal Linear Regression Model. 1. Introduction

Chapter 5. Solution of System of Linear Equations. Module No. 6. Solution of Inconsistent and Ill Conditioned Systems

a. (All your answers should be in the letter!

Dr. Shalabh Department of Mathematics and Statistics Indian Institute of Technology Kanpur

Chapter 11: Simple Linear Regression and Correlation

Chapter 8 Indicator Variables

Dr. Shalabh Department of Mathematics and Statistics Indian Institute of Technology Kanpur

Chapter 13: Multiple Regression

Economics 130. Lecture 4 Simple Linear Regression Continued

Chapter 2 - The Simple Linear Regression Model S =0. e i is a random error. S β2 β. This is a minimization problem. Solution is a calculus exercise.

e i is a random error

Chapter 3. Two-Variable Regression Model: The Problem of Estimation

LINEAR REGRESSION ANALYSIS. MODULE VIII Lecture Indicator Variables

Department of Quantitative Methods & Information Systems. Time Series and Their Components QMIS 320. Chapter 6

Sampling Theory MODULE V LECTURE - 17 RATIO AND PRODUCT METHODS OF ESTIMATION

Psychology 282 Lecture #24 Outline Regression Diagnostics: Outliers

x i1 =1 for all i (the constant ).

Statistics for Economics & Business

The Multiple Classical Linear Regression Model (CLRM): Specification and Assumptions. 1. Introduction

Econ107 Applied Econometrics Topic 3: Classical Model (Studenmund, Chapter 4)

Systems of Equations (SUR, GMM, and 3SLS)

A Comparative Study for Estimation Parameters in Panel Data Model

Econ107 Applied Econometrics Topic 9: Heteroskedasticity (Studenmund, Chapter 10)

where I = (n x n) diagonal identity matrix with diagonal elements = 1 and off-diagonal elements = 0; and σ 2 e = variance of (Y X).

Dr. Shalabh Department of Mathematics and Statistics Indian Institute of Technology Kanpur

ECONOMICS 351*-A Mid-Term Exam -- Fall Term 2000 Page 1 of 13 pages. QUEEN'S UNIVERSITY AT KINGSTON Department of Economics

Comparison of Regression Lines

Statistics for Business and Economics

1. Inference on Regression Parameters a. Finding Mean, s.d and covariance amongst estimates. 2. Confidence Intervals and Working Hotelling Bands

Lecture 4 Hypothesis Testing

2016 Wiley. Study Session 2: Ethical and Professional Standards Application

[ ] λ λ λ. Multicollinearity. multicollinearity Ragnar Frisch (1934) perfect exact. collinearity. multicollinearity. exact

is the calculated value of the dependent variable at point i. The best parameters have values that minimize the squares of the errors

Fall 2012 Analysis of Experimental Measurements B. Eisenstein/rev. S. Errede

STAT 511 FINAL EXAM NAME Spring 2001

STAT 3340 Assignment 1 solutions. 1. Find the equation of the line which passes through the points (1,1) and (4,5).

4 Analysis of Variance (ANOVA) 5 ANOVA. 5.1 Introduction. 5.2 Fixed Effects ANOVA

Lecture 9: Linear regression: centering, hypothesis testing, multiple covariates, and confounding

Resource Allocation and Decision Analysis (ECON 8010) Spring 2014 Foundations of Regression Analysis

Lecture 9: Linear regression: centering, hypothesis testing, multiple covariates, and confounding

β0 + β1xi and want to estimate the unknown

Statistics MINITAB - Lab 2

SIO 224. m(r) =(ρ(r),k s (r),µ(r))

Chapter 12 Analysis of Covariance

Negative Binomial Regression

x = , so that calculated

Statistics for Managers Using Microsoft Excel/SPSS Chapter 13 The Simple Linear Regression Model and Correlation

Durban Watson for Testing the Lack-of-Fit of Polynomial Regression Models without Replications

The Ordinary Least Squares (OLS) Estimator

ISQS 6348 Final Open notes, no books. Points out of 100 in parentheses. Y 1 ε 2

Week3, Chapter 4. Position and Displacement. Motion in Two Dimensions. Instantaneous Velocity. Average Velocity

Chapter 14 Simple Linear Regression

Now we relax this assumption and allow that the error variance depends on the independent variables, i.e., heteroskedasticity

Chapter 15 - Multiple Regression

Module 3 LOSSY IMAGE COMPRESSION SYSTEMS. Version 2 ECE IIT, Kharagpur

DO NOT OPEN THE QUESTION PAPER UNTIL INSTRUCTED TO DO SO BY THE CHIEF INVIGILATOR. Introductory Econometrics 1 hour 30 minutes

Linear Approximation with Regularization and Moving Least Squares

18. SIMPLE LINEAR REGRESSION III

Lecture 16 Statistical Analysis in Biomaterials Research (Part II)

28. SIMPLE LINEAR REGRESSION III

[The following data appear in Wooldridge Q2.3.] The table below contains the ACT score and college GPA for eight college students.

3.1 Expectation of Functions of Several Random Variables. )' be a k-dimensional discrete or continuous random vector, with joint PMF p (, E X E X1 E X

Basic Business Statistics, 10/e

Tests of Single Linear Coefficient Restrictions: t-tests and F-tests. 1. Basic Rules. 2. Testing Single Linear Coefficient Restrictions

Lecture 6: Introduction to Linear Regression

A Bound for the Relative Bias of the Design Effect

Econ Statistical Properties of the OLS estimator. Sanjaya DeSilva

Originated from experimental optimization where measurements are very noisy Approximation can be actually more accurate than

CHAPTER 5 NUMERICAL EVALUATION OF DYNAMIC RESPONSE

Lecture 2: Prelude to the big shrink

STAT 3008 Applied Regression Analysis

Outline. Zero Conditional mean. I. Motivation. 3. Multiple Regression Analysis: Estimation. Read Wooldridge (2013), Chapter 3.

Analysis of Variance and Design of Experiments-II

January Examinations 2015

Global Sensitivity. Tuesday 20 th February, 2018

Statistics Chapter 4

This column is a continuation of our previous column

Inner Product. Euclidean Space. Orthonormal Basis. Orthogonal

LECTURE 9 CANONICAL CORRELATION ANALYSIS

Lecture Notes for STATISTICAL METHODS FOR BUSINESS II BMGT 212. Chapters 14, 15 & 16. Professor Ahmadi, Ph.D. Department of Management

Correlation and Regression

Predictive Analytics : QM901.1x Prof U Dinesh Kumar, IIMB. All Rights Reserved, Indian Institute of Management Bangalore

Lecture 10 Support Vector Machines II

x yi In chapter 14, we want to perform inference (i.e. calculate confidence intervals and perform tests of significance) in this setting.

STATISTICS QUESTIONS. Step by Step Solutions.

Statistics II Final Exam 26/6/18

Linear Feature Engineering 11

β0 + β1xi. You are interested in estimating the unknown parameters β

Y = β 0 + β 1 X 1 + β 2 X β k X k + ε

SIMPLE LINEAR REGRESSION

Professor Chris Murray. Midterm Exam

CHAPTER 8. Exercise Solutions

β0 + β1xi. You are interested in estimating the unknown parameters β

Biostatistics 360 F&t Tests and Intervals in Regression 1

Chapter 9: Statistical Inference and the Relationship between Two Variables

Liu-type Negative Binomial Regression: A Comparison of Recent Estimators and Applications

Transcription:

LINEAR REGRESSION ANALYSIS MODULE IX Lecture - 31 Multcollnearty Dr. Shalabh Department of Mathematcs and Statstcs Indan Insttute of Technology Kanpur

6. Rdge regresson The OLSE s the best lnear unbased estmator of regresson coeffcent n the sense that t has mnmum varance n the class of lnear and unbased estmators. However f the condton of unbasedness can be relaxed then t s possble to fnd a based estmator of regresson coeffcent say error (MSE) of ˆ s ˆ that has smaller varance them the unbased OLSE b. The mean squared. MSE( ˆ ) = E( ˆ ) { ˆ ( ˆ ) } { ( ˆ ) } = E E + E = Var( ˆ ) + E( ˆ ) = Var( ˆ ) + Bas( ˆ ). ˆ Thus MSE( ) can be made smaller than Var( ˆ ) by ntroducng small bas n. One of the approach to do so s the regresson. The regresson estmator s obtaned by solvng the normal equatons of least squares estmaton. The normal equatons are modfed as ˆ ˆ ( ' ) X X + I ˆ = X ' y ( ) 1 ˆ = X ' X + I X ' y. s the regresson estmator of and 0 s any characterzng scalar termed as basng parameter.

3 As 0, ˆ b ( OLSE) and as, ˆ 0. So larger the value of, larger shrnkage towards zero. Note that the OLSE s napproprate to use n the sense that t has very hgh varance when multcollnearty s present n the data. On the other hand, a very small value of accept the null hypothess H : = 0 0 parameter controls the amount of shrnkage n the estmates. may tend to ndcatng that the correspondng varables are not relevant. The value of basng ˆ Bas of regresson estmator ˆ The bas of s Bas( ˆ ) = E( ˆ ) ( ' ) ' ( ) = X X + I X E y = ( X ' X + I) X ' X I = ( X ' X + I) X ' X X ' X I = ( X ' X + I). [ ] Thus the regresson estmator s a based estmator of.

4 Covarance matrx The covarance matrx of ˆ s defned as { }{ } ' V( ˆ ) ˆ ( ˆ ) ˆ ( ˆ = E E E ). Snce ˆ E( ˆ ) = ( X ' X + I) X ' y ( X ' X + I) X ' X = + ( X ' X I) X '( y X) = + ( X ' X I) X ' ε, so V( ˆ ) = ( X ' X + I) X ' V( ε) X( X ' X + I) = σ ( X ' X + I) X ' X( X ' X + I).

Mean squared error 5 ' =Λ= (,,..., ), X X dag λ1 λ λ k ˆ Wrtng the mean squared error of s MSE( ˆ ) ( ˆ ) ( ˆ = Var + bas ) = tr V ( ˆ ) ( ˆ + bas ) = ( ' + ) ' ( ' + ) + '( ' + ) σ tr X X I X X X X I X X I λ = σ + k k j j j= 1 ( λj + ) j= 1 ( λj + ) where λ1, λ,..., λk are the egenvalues of X ' X. Thus as ncreases, the bas n ncreases but ts varance decreases. The trade off between bas and varance hnges upon the value of. It can be shown that there exsts a value of such that provded ' Choce of ' s bounded. ˆ MSE( ˆ ) < Var( b) The estmaton of regresson estmator depends upon the value of. Varous approaches have been suggested n the lterature to determne the value of. The value of can be chosen on the bass of crtera lke - stablty of estmators wth respect to. - reasonable sgns. - magntude of resdual sum of squares etc. We consder here the determnaton of by the nspecton of trace.

6 Rdge trace Rdge trace s the graphcal dsplay of regresson estmator versus. If multcollnearty s present and s severe, then the nstablty of regresson coeffcents s reflected n the trace. As ncreases, some of the estmates vary dramatcally and they stablzes at some value of. The objectve n trace s to nspect the trace (curve) and fnd the reasonably small value of at whch the regresson estmators are stable. The regresson estmator wth such a choce of wll have smaller MSE than the varance of OLSE. An example of trace for a model wth 6 parameters s as follows. In ths trace, the ˆ s evaluated for varous ˆ j ( ) choces of and the correspondng values of all regresson coeffcents s, j = 1,,,6 are plotted versus. These values are denoted by dfferent symbols and are joned by a smooth curve. Ths produces a trace for respectve parameter. Now choose the value of where all the curves stablze and become nearly parallel. For example, the curves n followng fgure become nearly parallel startng from or so. Thus one possble choce of s and parameters can be estmated as ˆ ' '. ( ) 1 = X X + 4I X y = 4 = 4

7 The fgure drastcally exposes the presence of multcollnearty n the data. The behavour of at s very dfferent than at other values of. For small values of, the estmates change rapdly. The estmates stablze gradually as ncreases. The value of at whch all the estmates stablze gves the desred value of because movng away from such wll not brng any apprecable reducton n the resdual sum of squares. If multcollnearty s present, then the varaton n regresson estmators s rapd around The optmal s chosen such that after that value of ˆ ( ) 0 0 0 0., almost all traces stablze.

8 Lmtatons 1. The choce of s data dependent and therefore s a random varable. Usng t as a random varable volates the assumpton that of. s a constant. Ths wll dsturb the optmal propertes derved under the assumpton of constancy (0, ). The value of les n the nterval. So large number of values are requred for exploraton. Ths results s wastng of tme. However, ths s not a bg ssue when workng wth software. 3. The choce of from graphcal dsplay may not be unque. Dfferent people may choose dfferent and consequently the values of regresson estmators wll be changng. However estmators of all the coeffcents stablze. Hence small varaton n choosng the value of change n estmators of the coeffcents. Another choce of s s chosen so that all the may not produce much = k ˆ σ bb ' where b and are obtaned from the least squares estmaton. 4. The stablty of numercal estmates of s a rough way to determne. Dfferent estmates may exhbt stablty for dfferent σˆ estmators are used. ˆ ' s and t may often be hard to strke a compromse. In such stuaton, generalzed regresson 5. There s no gudance avalable regardng the testng of hypothess and for confdence nterval estmaton.

9 Idea behnd regresson estmator The problem of multcollnearty arses because some of the egenvalues roots of X X are close to zero or are zero. So f λ are the characterstc roots, and f 1, λ,..., λp then X X =Λ= dag λ1 λ λ k ' (,,..., ) ˆ ( ) ( ) 1 = Λ+ I X y = I + Λ b where b s the OLSE of gven by b= X X X y = Λ X y 1 1 ( ' ) ' '. Thus a partcular element n ˆ wll be of the form 1 ' λ xy = b. λ + λ + λ λ = 0, So a small quantty s added to so that f even then remans meanngful. 1 λ +

10 Another nterpretaton of regresson estmator In the model obtan the least squares estmator of when = C, where C s some constant. So mnmze y = X + ε, ( ) = ( y X)'( y X) + ( ' C) S( ) where s the Lagrangan multpler. Dfferentatng wth respect to, the normal equatons are obtaned as k = 1 1 S( ) = 0 X ' y+ X ' X + = 0 ˆ ( X ' X I) X ' y. = + Note that f s very small, t may ndcate that most of the regresson coeffcents are close to zero and f s large, then t may ndcate that the regresson coeffcents are away from zero. So coeffcents to enable ts estmaton. puts a sort of penalty on the regresson