LINEAR REGRESSION ANALYSIS. MODULE IX Lecture Multicollinearity

Similar documents
LINEAR REGRESSION ANALYSIS. MODULE IX Lecture Multicollinearity

Polynomial Regression Models

Psychology 282 Lecture #24 Outline Regression Diagnostics: Outliers

[ ] λ λ λ. Multicollinearity. multicollinearity Ragnar Frisch (1934) perfect exact. collinearity. multicollinearity. exact

Dr. Shalabh Department of Mathematics and Statistics Indian Institute of Technology Kanpur

Chapter 7 Generalized and Weighted Least Squares Estimation. In this method, the deviation between the observed and expected values of

Dr. Shalabh Department of Mathematics and Statistics Indian Institute of Technology Kanpur

Econ107 Applied Econometrics Topic 3: Classical Model (Studenmund, Chapter 4)

LECTURE 9 CANONICAL CORRELATION ANALYSIS

Chapter 13: Multiple Regression

LINEAR REGRESSION ANALYSIS. MODULE VIII Lecture Indicator Variables

The Multiple Classical Linear Regression Model (CLRM): Specification and Assumptions. 1. Introduction

Inner Product. Euclidean Space. Orthonormal Basis. Orthogonal

Dr. Shalabh Department of Mathematics and Statistics Indian Institute of Technology Kanpur

Chapter 3. Two-Variable Regression Model: The Problem of Estimation

Chapter 8 Indicator Variables

Module 9. Lecture 6. Duality in Assignment Problems

Sampling Theory MODULE V LECTURE - 17 RATIO AND PRODUCT METHODS OF ESTIMATION

However, since P is a symmetric idempotent matrix, of P are either 0 or 1 [Eigen-values

Chapter 12 Analysis of Covariance

Chapter 5. Solution of System of Linear Equations. Module No. 6. Solution of Inconsistent and Ill Conditioned Systems

Composite Hypotheses testing

x = , so that calculated

NUMERICAL DIFFERENTIATION

1. Inference on Regression Parameters a. Finding Mean, s.d and covariance amongst estimates. 2. Confidence Intervals and Working Hotelling Bands

Negative Binomial Regression

ISSN: ISO 9001:2008 Certified International Journal of Engineering and Innovative Technology (IJEIT) Volume 3, Issue 1, July 2013

where I = (n x n) diagonal identity matrix with diagonal elements = 1 and off-diagonal elements = 0; and σ 2 e = variance of (Y X).

Comparison of Regression Lines

Economics 130. Lecture 4 Simple Linear Regression Continued

See Book Chapter 11 2 nd Edition (Chapter 10 1 st Edition)

3.1 Expectation of Functions of Several Random Variables. )' be a k-dimensional discrete or continuous random vector, with joint PMF p (, E X E X1 E X

x i1 =1 for all i (the constant ).

Econ107 Applied Econometrics Topic 9: Heteroskedasticity (Studenmund, Chapter 10)

Number of cases Number of factors Number of covariates Number of levels of factor i. Value of the dependent variable for case k

REGRESSION ANALYSIS II- MULTICOLLINEARITY

Department of Quantitative Methods & Information Systems. Time Series and Their Components QMIS 320. Chapter 6

Dr. Shalabh Department of Mathematics and Statistics Indian Institute of Technology Kanpur

Factor models with many assets: strong factors, weak factors, and the two-pass procedure

Statistics for Business and Economics

Topic 23 - Randomized Complete Block Designs (RCBD)

Salmon: Lectures on partial differential equations. Consider the general linear, second-order PDE in the form. ,x 2

U.C. Berkeley CS294: Spectral Methods and Expanders Handout 8 Luca Trevisan February 17, 2016

Lecture 9: Linear regression: centering, hypothesis testing, multiple covariates, and confounding

Lecture 9: Linear regression: centering, hypothesis testing, multiple covariates, and confounding

Chapter 11: Simple Linear Regression and Correlation

Linear Approximation with Regularization and Moving Least Squares

Now we relax this assumption and allow that the error variance depends on the independent variables, i.e., heteroskedasticity

SIMPLE LINEAR REGRESSION

The Ordinary Least Squares (OLS) Estimator

Statistics II Final Exam 26/6/18

CSci 6974 and ECSE 6966 Math. Tech. for Vision, Graphics and Robotics Lecture 21, April 17, 2006 Estimating A Plane Homography

Lectures - Week 4 Matrix norms, Conditioning, Vector Spaces, Linear Independence, Spanning sets and Basis, Null space and Range of a Matrix

Limited Dependent Variables

Chapter 15 Student Lecture Notes 15-1

ECONOMICS 351*-A Mid-Term Exam -- Fall Term 2000 Page 1 of 13 pages. QUEEN'S UNIVERSITY AT KINGSTON Department of Economics

4 Analysis of Variance (ANOVA) 5 ANOVA. 5.1 Introduction. 5.2 Fixed Effects ANOVA

Chapter 15 - Multiple Regression

e - c o m p a n i o n

Analysis of Variance and Design of Experiments-II

Chapter 2 - The Simple Linear Regression Model S =0. e i is a random error. S β2 β. This is a minimization problem. Solution is a calculus exercise.

CHAPTER 5 NUMERICAL EVALUATION OF DYNAMIC RESPONSE

Econ Statistical Properties of the OLS estimator. Sanjaya DeSilva

Week3, Chapter 4. Position and Displacement. Motion in Two Dimensions. Instantaneous Velocity. Average Velocity

9. Binary Dependent Variables

Statistics for Economics & Business

Primer on High-Order Moment Estimators

[The following data appear in Wooldridge Q2.3.] The table below contains the ACT score and college GPA for eight college students.

Chapter 7 Partially Balanced Incomplete Block Design (PBIBD)

Kernel Methods and SVMs Extension

Discussion of Extensions of the Gauss-Markov Theorem to the Case of Stochastic Regression Coefficients Ed Stanek

First Year Examination Department of Statistics, University of Florida

Lecture 6: Introduction to Linear Regression

January Examinations 2015

Fall 2012 Analysis of Experimental Measurements B. Eisenstein/rev. S. Errede

A Comparative Study for Estimation Parameters in Panel Data Model

7. Products and matrix elements

Issues To Consider when Estimating Health Care Costs with Generalized Linear Models (GLMs): To Gamma/Log Or Not To Gamma/Log? That Is The New Question

FREQUENCY DISTRIBUTIONS Page 1 of The idea of a frequency distribution for sets of observations will be introduced,

Estimation: Part 2. Chapter GREG estimation

# c i. INFERENCE FOR CONTRASTS (Chapter 4) It's unbiased: Recall: A contrast is a linear combination of effects with coefficients summing to zero:

Global Sensitivity. Tuesday 20 th February, 2018

Chapter 3 Describing Data Using Numerical Measures

Errors for Linear Systems

Introduction to Vapor/Liquid Equilibrium, part 2. Raoult s Law:

Problem Set 9 Solutions

The Prncpal Component Transform The Prncpal Component Transform s also called Karhunen-Loeve Transform (KLT, Hotellng Transform, oregenvector Transfor

Formulas for the Determinant

Perron Vectors of an Irreducible Nonnegative Interval Matrix

Foundations of Arithmetic

Linear Regression Analysis: Terminology and Notation

Outline. Zero Conditional mean. I. Motivation. 3. Multiple Regression Analysis: Estimation. Read Wooldridge (2013), Chapter 3.

Report on Image warping

2.3 Nilpotent endomorphisms

Computation of Higher Order Moments from Two Multinomial Overdispersion Likelihood Models

Random Partitions of Samples

8.4 COMPLEX VECTOR SPACES AND INNER PRODUCTS

Basically, if you have a dummy dependent variable you will be estimating a probability.

Tests of Single Linear Coefficient Restrictions: t-tests and F-tests. 1. Basic Rules. 2. Testing Single Linear Coefficient Restrictions

10-701/ Machine Learning, Fall 2005 Homework 3

Transcription:

LINEAR REGRESSION ANALYSIS MODULE IX Lecture - 30 Multcollnearty Dr. Shalabh Department of Mathematcs and Statstcs Indan Insttute of Technology Kanpur

2 Remedes for multcollnearty Varous technques have been proposed to deal wth the problems resultng from the presence of multcollnearty n the data.. Obtan more data The harmful multcollnearty arses essentally because ran of X X falls below and then X X = 0 whch clearly suggests the presence of lnear dependences n the columns of X. It s the case when X X s close to zero whch needs attenton. Addtonal data may help n reducng the samplng varance of the estmates. The data need to be collected such that t helps n breang up the multcollnearty n the data. It s always not possble to collect addtonal data to varous reasons as follows. The experment and process have fnshed and no longer avalable. The economc constrans may also not allow to collect the addtonal data. The addtonal data may not match wth the earler collected data and may be unusual. If the data s n tme seres, then longer tme seres may force to tae data that s too far n the past. If multcollnearty s due to any dentty or exact relatonshp, then ncreasng the sample sze wll not help. Sometmes, t s not advsable to use the data even f t s avalable. For example, f the data on consumpton pattern s avalable for the years 950-200, then one may not le to use t as the consumpton pattern usually does not remans same for such a long perod.

2. Drop some varables that are collnear 3 If possble, dentfy the varables whch seems to causng multcollnearty. These collnear varables can be dropped so as to match the condton of fall ran of X-matrx. The process of omttng the varables may be carred out on the bass of some nd of orderng of explanatory varables, e.g., those varables can be deleted frst whch have smaller value of t- rato. In another example, suppose the expermenter s not nterested n all the parameters. In such cases, one can get the estmators of the parameters of nterest whch have smaller mean squared errors than the varance of OLSE of full vector by droppng some varables. If some varables are elmnated, then ths may reduce the predctve power of the model. Sometmes there s no assurance that how the model wll exhbt less multcollnearty. 3. Use some relevant pror nformaton One may search for some relevant pror nformaton about the regresson coeffcents. Ths may lead to specfcaton of estmates of some coeffcents. More general stuaton ncludes the specfcaton of some exact lnear restrctons and stochastc lnear restrctons. The procedures le restrcted regresson and mxed regresson can be used for ths purpose. The relevance and correctness of nformaton plays an mportant role n such analyss but t s dffcult to ensure t n practce. For example, the estmates derved n U.K. may not be vald n Inda.

4. Employ generalzed nverse ( X ' X) < If ran, then the generalzed nverse can be used to fnd the nverse of. Then can be estmated by ˆ β = ( X ' X) X ' y. X ' X β 4 In such case, the estmates wll not be unque except n the case of use of Moore-Penrose nverse of Dfferent methods of fndng generalzed nverse may gve dfferent results. So appled worers wll get dfferent results. Moreover, t s also not nown that whch method of fndng generalzed nverse s optmum. ( X ' X). 5. Use of prncpal component regresson The prncpal component regresson s based on the technque of prncpal component analyss. The explanatory varables are transformed nto a new set of orthogonal varables called as prncpal components. Usually ths technque s used for reducng the dmensonalty of data by retanng some levels of varablty of explanatory varables whch s expressed by the varablty n study varable. The prncpal components nvolves the determnaton of a set of lnear combnatons of explanatory varables such that they retan the total varablty of the system and these lnear combnatons are mutually ndependent of each other. Such obtaned prncpal components are raned n the order of ther mportance. The mportance beng udged n terms of varablty explaned by a prncpal component relatve to the total varablty n the system. The procedure then nvolves elmnatng some of the prncpal components whch contrbute n explanng relatvely less varaton. After elmnaton of the least mportant prncpal components, the set up of multple regresson s used by replacng the explanatory varables wth prncpal components.

5 Then study varable s regressed aganst the set of selected prncpal components usng ordnary least squares method. Snce all the prncpal components are orthogonal, they are mutually ndependent and so OLS s used wthout any problem. Once the estmates of regresson coeffcents for the reduced set of orthogonal varables (prncpal components) have been obtaned, they are mathematcally transformed nto a new set of estmated regresson coeffcents that correspond to the orgnal correlated set of varables. These new estmated coeffcents are the prncpal components estmators of regresson coeffcents. Suppose there are explanatory varables X, X2,..., X. Consder the lnear functon of X, X2,.., X le Z Z 2 = = = = ax bx etc. The constants a, a2,..., a are determned such that the varance of s maxmzed subect to the normalzng condton 2 = 2 that a =. The constant b, b,..., b are determned such that the varance of Z2 s maxmzed subect to the normalty condton that = b 2 = and s ndependent of the frst prncpal component. Z

6 We contnue wth such process and obtan such lnear combnatons such that they are orthogonal to ther precedng lnear combnatons and satsfy the normalty condton. Then we obtan ther varances. Suppose such lnear combnatons are Z, Z2,.., Z and for them, Var( Z) > Var( Z2) >... > Var( Z ). The lnear combnaton havng the varance s the frst prncpal component. The lnear combnaton havng the second largest varance s the second largest prncpal component and so on. These prncpal components have the property that Var( Z) = Var( X ). Also, the X, X2,..., X are correlated = = but Z, Z2,.., Z are orthogonal or uncorrelated. So there wll be zero multcollnearty among Z, Z2,.., Z. The problem of multcollnearty arses because X, X2,..., X are not ndependent. Snce the prncpal components based on X, X2,..., X are mutually ndependent, so they can be used as explanatory varables and such regresson wll combat the multcollnearty. Let λ, λ2,..., λ be the egenvalues of X ' X, Λ= dag( λ s dagonal matrx, V s a orthogonal, λ2,..., λ ) matrx whose columns are the egenvectors assocated wth λ, λ2,..., λ. Consder the canoncal form of the lnear model where y = Xβ + ε = XVV ' β + ε = Zα + ε Z = XV, α = V ' β, V ' X ' XV = Z ' Z = Λ Columns of ( ) Z = Z, Z2,..., Z defne a new set of explanatory varables whch are called as prncpal component.

7 The OLSE of α s ˆ ( Z' Z) ' α = = Λ Z' y Z y and ts covarance matrx s V( ˆ α) = σ ( Z' Z) = σ Λ 2 2,,...,. 2 = σ dag λ λ2 λ λ th Note that s the varance of prncpal component and Z' Z= ZZ = Λ. A small egenvalue of X X means that the lnear relatonshp between the orgnal explanatory varable exst and the varance of correspondng orthogonal regresson coeffcent s large whch ndcates that the multcollnearty exsts. If one or more are small, then t ndcates that multcollnearty s present. = = λ

8 Retanment of prncpal components The new set of varables,.e., prncpal components are orthogonal, and they retan the same magntude of varance as of orgnal set. If multcollnearty s severe, then there wll be at least one small value of egenvalue. The elmnaton of one or more prncpal components assocated wth smallest egenvalues wll reduce the total varance n the model. Moreover, the prncpal components responsble for creatng multcollnearty wll be removed and the resultng model wll be apprecably mproved. The prncpal component matrx Z = Z, Z2,..., Z wth Z, Z2,..., Z contans exactly the same nformaton as the orgnal data n X n the sense that the total varablty n X and Z s same. The dfference between them s that the orgnal data are arranged nto a set of new varables whch are uncorrelated wth each other and can be raned wth respect to the th magntude of ther egenvalues. The column vector correspondng to the largest accounts for the largest proporton of the varaton n the orgnal data. Thus the Z s are ndexed so that λ and s the > λ2 >... > λ > 0 Z varance of. [ ] Z A strategy of elmnaton of prncpal components s to begn by dscardng the component assocated wth the smallest egenvalue. The dea behnd to do ths s that the prncpal component wth smallest egenvalue s contrbutng the least varance and so s least nformatve. λ λ

Usng ths procedure, prncpal components are elmnated untl the remanng components explan some preselected varance s terms of percentage of total varance. For example, f 90% of total varance s needed, and suppose r prncpal components are elmnated whch means that( r) prncpal components contrbute 90% of the total varaton, then r s selected to satsfy = = λ > 0.90. λ 9 Varous strateges to choose requred number of prncpal components are also avalable n the lterature. Suppose after usng such a rule, the r prncpal components are elmnated. Now only (- r) components wll be used for regresson. So Z matrx s parttoned as ( ) ( ) Z= Z Z = XV V r r where submatrx Z r s of order n x r and contans the prncpal components to be elmnated. The submatrx Z -r s of order n x ( r) and contans the prncpal components to be retaned. The reduced model obtaned after the elmnaton of r prncpal components can be expressed as y = Z α + ε*. The random error component s represented as ust to dstngush wth. The reduced coeffcents contan the coeffcents assocated wth retaned Z s. So ε * (,,..., ) Z = Z Z Z 2 (,,..., ) α = α α α 2 ( ) V = V, V,..., V. 2 ε

Usng OLS on the model wth retaned prncpal components, the OLSE of ' ' ˆ = ( ZZ) Zy. α. Now t s transformed bac to orgnal explanatory varables as follows: α = V ' β α = V ˆ β = V ' ˆ α β pc whch s the prncpal component regresson estmator of Ths method mproves the effcency as well as combats multcollnearty. β. α r s 0