Chapter 10 Building the Regression Model II: Diagnostics

Similar documents
Contents. 1 Review of Residuals. 2 Detecting Outliers. 3 Influential Observations. 4 Multicollinearity and its Effects

Chapter 11 Building the Regression Model II:

Final Review. Yang Feng. Yang Feng (Columbia University) Final Review 1 / 58

Regression Review. Statistics 149. Spring Copyright c 2006 by Mark E. Irwin

STAT 4385 Topic 06: Model Diagnostics

Review: Second Half of Course Stat 704: Data Analysis I, Fall 2014

Topic 18: Model Selection and Diagnostics

Multiple Linear Regression

Outline. Remedial Measures) Extra Sums of Squares Standardized Version of the Multiple Regression Model

Chapter 14 Logistic Regression, Poisson Regression, and Generalized Linear Models

Regression Diagnostics Procedures

Linear Regression. Applied Linear Regression Models (Kutner, Nachtsheim, Neter, Li) hsuhl (NUK) SDA Regression 1 / 34

Multicollinearity occurs when two or more predictors in the model are correlated and provide redundant information about the response.

Remedial Measures, Brown-Forsythe test, F test

Model Selection. Frank Wood. December 10, 2009

REGRESSION DIAGNOSTICS AND REMEDIAL MEASURES

Need for Several Predictor Variables

Chapter 13 Introduction to Nonlinear Regression( 非線性迴歸 )

Regression Model Building

Lecture One: A Quick Review/Overview on Regular Linear Regression Models

The Model Building Process Part I: Checking Model Assumptions Best Practice

Regression Diagnostics for Survey Data

Beam Example: Identifying Influential Observations using the Hat Matrix

Chapter 2 Inferences in Regression and Correlation Analysis

Statistics GIDP Ph.D. Qualifying Exam Methodology May 26 9:00am-1:00pm

The Model Building Process Part I: Checking Model Assumptions Best Practice (Version 1.1)

STATISTICS 174: APPLIED STATISTICS FINAL EXAM DECEMBER 10, 2002

Dr. Maddah ENMG 617 EM Statistics 11/28/12. Multiple Regression (3) (Chapter 15, Hines)

MATH 644: Regression Analysis Methods

1) Answer the following questions as true (T) or false (F) by circling the appropriate letter.

Psychology Seminar Psych 406 Dr. Jeffrey Leitzel

Math 423/533: The Main Theoretical Topics

Diagnostics and Remedial Measures

Quantitative Methods I: Regression diagnostics

Sections 7.1, 7.2, 7.4, & 7.6

COMPREHENSIVE WRITTEN EXAMINATION, PAPER III FRIDAY AUGUST 26, 2005, 9:00 A.M. 1:00 P.M. STATISTICS 174 QUESTION

No other aids are allowed. For example you are not allowed to have any other textbook or past exams.

Simple Linear Regression

Remedial Measures for Multiple Linear Regression Models

STA 4210 Practise set 2a

6.1 Introduction. Regression Model:

Formal Statement of Simple Linear Regression Model

Ridge Regression. Summary. Sample StatFolio: ridge reg.sgp. STATGRAPHICS Rev. 10/1/2014

Lecture 1: Linear Models and Applications

Statistics GIDP Ph.D. Qualifying Exam Methodology May 26 9:00am-1:00pm

Regression diagnostics

CHAPTER 5. Outlier Detection in Multivariate Data

Chapter 1 Linear Regression with One Predictor Variable

Regression Diagnostics

y response variable x 1, x 2,, x k -- a set of explanatory variables

Prediction Intervals in the Presence of Outliers

Stat 5100 Handout #26: Variations on OLS Linear Regression (Ch. 11, 13)

Regression coefficients may even have a different sign from the expected.

Available online at (Elixir International Journal) Statistics. Elixir Statistics 49 (2012)

Leverage. the response is in line with the other values, or the high leverage has caused the fitted model to be pulled toward the observed response.

holding all other predictors constant

STATISTICS 479 Exam II (100 points)

Chapter 2 Multiple Regression (Part 4)

Unit 10: Simple Linear Regression and Correlation

The Masking and Swamping Effects Using the Planted Mean-Shift Outliers Models

UNIVERSITY OF MASSACHUSETTS. Department of Mathematics and Statistics. Basic Exam - Applied Statistics. Tuesday, January 17, 2017

Chapter 6 Multiple Regression

Simple Linear Regression

Generalized Linear Models

The Effect of a Single Point on Correlation and Slope

Simple Regression Model Setup Estimation Inference Prediction. Model Diagnostic. Multiple Regression. Model Setup and Estimation.

ECON 450 Development Economics

Detecting and Assessing Data Outliers and Leverage Points

Regression Analysis for Data Containing Outliers and High Leverage Points

Regression Model Specification in R/Splus and Model Diagnostics. Daniel B. Carr

Lecture 10 Multiple Linear Regression

Lecture 15 Multiple regression I Chapter 6 Set 2 Least Square Estimation The quadratic form to be minimized is

Lecture 4: Regression Analysis

6. Multiple Linear Regression

Department of Mathematics The University of Toledo. Master of Science Degree Comprehensive Examination Applied Statistics.

9. Linear Regression and Correlation

3. Diagnostics and Remedial Measures

Biostatistics 380 Multiple Regression 1. Multiple Regression

The Algorithm for Multiple Outliers Detection Against Masking and Swamping Effects

9 Correlation and Regression

MLR Model Checking. Author: Nicholas G Reich, Jeff Goldsmith. This material is part of the statsteachr project

Multiple Regression. Dr. Frank Wood. Frank Wood, Linear Regression Models Lecture 12, Slide 1

Multiple Linear Regression

Any of 27 linear and nonlinear models may be fit. The output parallels that of the Simple Regression procedure.

STAT Checking Model Assumptions

Chapter 9 Other Topics on Factorial and Fractional Factorial Designs

Single and multiple linear regression analysis

Regression Analysis V... More Model Building: Including Qualitative Predictors, Model Searching, Model "Checking"/Diagnostics

Regression Analysis V... More Model Building: Including Qualitative Predictors, Model Searching, Model "Checking"/Diagnostics

LINEAR REGRESSION ANALYSIS. MODULE XVI Lecture Exercises

Unit 11: Multiple Linear Regression

STA 4210 Practise set 2b

Lecture 2. The Simple Linear Regression Model: Matrix Approach

Chapter 12: Multiple Regression

CHAPTER6 LINEAR REGRESSION

Statistical Modelling in Stata 5: Linear Models

Chapter 2 Multiple Regression I (Part 1)

Business Statistics. Tommaso Proietti. Linear Regression. DEF - Università di Roma 'Tor Vergata'

Linear Algebra Review

K. Model Diagnostics. residuals ˆɛ ij = Y ij ˆµ i N = Y ij Ȳ i semi-studentized residuals ω ij = ˆɛ ij. studentized deleted residuals ɛ ij =

Transcription:

Chapter 10 Building the Regression Model II: Diagnostics 許湘伶 Applied Linear Regression Models (Kutner, Nachtsheim, Neter, Li) hsuhl (NUK) LR Chap 10 1 / 41

10.1 Model Adequacy for a Predictor Variable-Added Variable Plots a number of refined( 推敲 ) diagnostics for checking the adequacy of a regression model detecting improper functional form for a predictor variable outliers( 離群值 ) influential( 有影響的 ) observations multicollinearity hsuhl (NUK) LR Chap 10 2 / 41

10.1 Model Adequacy for a Predictor Variable-Added Variable Plots Added Variable Plots Previous Chap. 3,6: check whether a curvature effect for that variable is required in the model the residual plots vs. the predictor variables: determine whether it would be helpful to add one or more of these variables to the model Figure : (Chap. 3)Residual Plots for Possible Omission of Important Predictor Variable-Productivity Example. hsuhl (NUK) LR Chap 10 3 / 41

10.1 Model Adequacy for a Predictor Variable-Added Variable Plots Added Variable Plots (cont.) Limitation: not properly show the nature of the marginal effect of a predictor variable, given the other predictor variables in the model Added-variable plots(partial regression plots; adjusted variable plots;): provide graphic information about the marginal importance of a predictor variable X k, given the other predictor variables already in the model In an added-variable plot, both Y and X k under consideration are regressed against the other predictor variables and residuals are obtained for each. hsuhl (NUK) LR Chap 10 4 / 41

10.1 Model Adequacy for a Predictor Variable-Added Variable Plots Added Variable Plots (cont.) The plot of these residuals: 1 the marginal importance of this variable in reducing the residual variability 2 provide information about the nature of the marginal regression relation for X k under consideration for possible inclusion in the regression model hsuhl (NUK) LR Chap 10 5 / 41

10.1 Model Adequacy for a Predictor Variable-Added Variable Plots Added Variable Plots (cont.) Illustration: the regression effect for X 1, given that X 2 is already in the model (Chap. 7, p. 270 271 partial determination) Ŷ i (X 2 ) = b 0 + b 2 X i2 e i (Y X 2 ) = Y i Ŷ i (X 2 ) regress X 1 on X 2 : ˆX i1 (X 2 ) = b 0 + b 2X i2 e i (X 1 X 2 ) = X i1 ˆX i1 (X 2 ) ( R 2 = R 2 Y 1 2 : e i(y X 2 ) is regressed on e i (X 1 X 2 )) added-variable for predictor variable X 1 : Plot e(y X 2 ) vs e(x 1 X 2 ) hsuhl (NUK) LR Chap 10 6 / 41

10.1 Model Adequacy for a Predictor Variable-Added Variable Plots Added Variable Plots (cont.) Figure : Prototype Added-Variable Plots (a) X 1 contains no additional information; no helpful to add X 1 (b) a linear band with a nonzero slope; X 1 may be a helpful addition to the regression model already containing X 2 (c) a curvilinear band; X 1 may be helpful and suggesting the possible nature of the curvature effect hsuhl (NUK) LR Chap 10 7 / 41

10.1 Model Adequacy for a Predictor Variable-Added Variable Plots Added Variable Plots (cont.) Added-variable plots: providing information about the possible nature of the marginal relationship for a predictor variable, given the other variables already in the regression model the strength of the relationship useful for uncovering( 發現 ) outlying data points that may have a strong influence in estimating the relationship of the predictor variable X k to the response variable hsuhl (NUK) LR Chap 10 8 / 41

10.1 Model Adequacy for a Predictor Variable-Added Variable Plots Added Variable Plots (cont.) (a) SSE(X 2 ) (b) SSE(X1, X2) Difference (SSE(X 2 ) SSE(X 1, X 2 )): SSR(X 1 X 2 ); provides information about the marginal strength of the linear relation of X 1 to the response variable, given that X 2 is in the model hsuhl (NUK) LR Chap 10 9 / 41

10.1 Model Adequacy for a Predictor Variable-Added Variable Plots Added Variable Plots (cont.) Table 10.1 Average Annual Income Risk Aversion Amount of Life Insurance Carried Manager (thousand dollars) Score (thousand dollars) i X i1 X i2 Y i 1 45.01 6 91 2 57.20 4 162 3 26.85 5 11 16 46.13 4 91 17 30.37 3 14 18 39.06 5 63 r 12 = 0.254 Ŷ = 205.72 + 6.2880X 1 + 4.738X 2 Residual plot: a linear relation for X 1 is not appropriate in the model already containing X 2 hsuhl (NUK) LR Chap 10 10 / 41

10.1 Model Adequacy for a Predictor Variable-Added Variable Plots Added Variable Plots (cont.) ex<-read.table("ch10ta01.txt",header=f) colnames(ex)<-c("x1","x2","y") attach(ex) fit<-lm(y~x1+x2) par(mfrow=c(1,3)) plot(x1,resid(fit),pch=16) abline(0,0,lty=2,col="gray") ## Method 1: plot(resid(lm(y~x2)) ~ resid(lm(x1~x2)),col="blue",pch=16, xlab="e(x_1 X_2)", ylab="e(y X_2)") abline(lm(resid(lm(y~x2))~resid(lm(x1~x2))),col="red") ## Method 2: using avplot() library(car) avplot( model=lm( Y~X1+X2 ), variable=x1 ) hsuhl (NUK) LR Chap 10 11 / 41

10.1 Model Adequacy for a Predictor Variable-Added Variable Plots Added Variable Plots (cont.) Ŷ (X 2 ) = 50.70 + 15.54X 2 ; ˆX 1 (X 2 ) = 40.779 + 1.718X 2 Plots: through (0,0); b 1 = 6.2880; suggest the curvilinear relation between Y and X 1 X 2 is strongly positive; a slight concave upward shape R 2 Y 1 2 = 0.984 Incorporating a curvilinear effect for X 1 hsuhl (NUK) LR Chap 10 12 / 41

10.1 Model Adequacy for a Predictor Variable-Added Variable Plots Added Variable Plots (cont.) Comments: An added-variable plot only suggests the nature of the functional relation in which a predictor variable should be added to the regression model but does not provide an analytic expression of the relation. Added-variable plots need to be used with caution for identifying the nature of the marginal effect of a predictor variable. may not show the proper form of the marginal effect of a predictor variable if the functional relations for some or all of the predictor variables already in the regression model are misspecified the relations of the predictor variable to the response variable are complex high multicollinearity among the predictor variables hsuhl (NUK) LR Chap 10 13 / 41

10.1 Model Adequacy for a Predictor Variable-Added Variable Plots Added Variable Plots (cont.) Any fitted multiple regression function can be obtained from a sequence of fitted partial regressions. Ex: having e(y X 2 ), e(x 1 X 2 ) e(ŷ X 2) = 6.2880[e(X 1 X 2 )] [Ŷ Ŷ (X 2 )] = 6.2880[X 1 ˆX 1 (X 2 )] Ŷ = 205.72 + 6.2880X 1 + 4.737X 2 (Ŷ (X 2 ) = 50.70 + 15.54X 2 ; ˆX 1 (X 2 ) = 40.779 + 1.718X 2 ) hsuhl (NUK) LR Chap 10 14 / 41

10.2 Identifying Outlying Y Observations- Studentized Deleted Residuals Identifying Outlying Y Observations Outlying or Extreme: the observations for these cases are well separated from the remainder of the data large residuals; have dramatic effects hsuhl (NUK) LR Chap 10 15 / 41

10.2 Identifying Outlying Y Observations- Studentized Deleted Residuals Identifying Outlying Y Observations (cont.) Outlying: Y value, X values or both Not all outlying cases have a strong influence on the fitted regression function. A basic step: determine if the regression model under consideration is heavily influenced by one or a few cases in the data set hsuhl (NUK) LR Chap 10 16 / 41

10.2 Identifying Outlying Y Observations- Studentized Deleted Residuals Identifying Outlying Y Observations (cont.) Two refined measures for identifying cases with outlying Y observations: Residuals, Semistudentized Residuals: e i = Y i Ŷi; e i = e i MSE Hat matrix: H = X(X X) 1 X Ŷ = HY e = (I H)Y σ 2 {e} = σ 2 (I H) σ 2 {e i } = σ 2 (1 h ii ), h ii : the ith elements on the diagonal of H σ{e i, e j } = h ij σ 2, i j s 2 {e i } = MSE(1 h ii ), s{e i, e j } = h ij MSE hsuhl (NUK) LR Chap 10 17 / 41

10.2 Identifying Outlying Y Observations- Studentized Deleted Residuals Identifying Outlying Y Observations (cont.) h ii = X i(x X) 1 X i, X i = [1 X i,1 X i,p 1 ] Small data set: n = 4 Ŷ = 80.93 5.84X 1 + 11.32X 2 MSE = 574.9 s 2 {e 1 } = 574.9(1 0.3877) = 352.0 hsuhl (NUK) LR Chap 10 18 / 41

10.2 Identifying Outlying Y Observations- Studentized Deleted Residuals Identifying Outlying Y Observations (cont.) Deleted Residuals: The difference between Y i and Ŷi(i): (PRESS prediction error) delted residual: d i = Y i Ŷi(i) = e i 1 h ii h ii : the larger will be the deleted residuals as compared to the ordinary residual the estimated variance of d i : s 2 {d i } = MSE (i) (1 + X i(x (i)x (i) ) 1 X i ) = MSE (i) 1 h ii d i t((n 1) p) s{d i } hsuhl (NUK) LR Chap 10 19 / 41

10.2 Identifying Outlying Y Observations- Studentized Deleted Residuals Identifying Outlying Y Observations (cont.) Studentized Deleted Residuals: t i = ( d i s{d i } = e i MSE(i) (1 h ii ) (n p)mse = (n p 1)MSE (i) + e2 i 1 h ii [ ] 1/2 n p 1 t i = e i SSE(1 h ii ) ei 2 h ii : the larger will be the deleted residuals as compared to the ordinary residual ) hsuhl (NUK) LR Chap 10 20 / 41

10.2 Identifying Outlying Y Observations- Studentized Deleted Residuals Identifying Outlying Y Observations (cont.) the estimated variance of d i : s 2 {d i } = MSE (i) (1 + X i(x (i)x (i) ) 1 X i ) = MSE (i) 1 h ii d i t((n 1) p) s{d i } Test for Outliers: whose studentized deleted residuals are large in absolute value If the regression model is appropriate, so that no case is outlying. Each t i t(n p 1). t i : the appropriate Bonferroni critical value: t(1 α/2n; n p 1) hsuhl (NUK) LR Chap 10 21 / 41

10.2 Identifying Outlying Y Observations- Studentized Deleted Residuals Identifying Outlying Y Observations (cont.) Body fat example t 13 < 3.252 = t(1 α/2n; n p 1) (α =.10) The Bonferroni procedure provides a conservative( 保守的 ) test for the presence of an outlier. hsuhl (NUK) LR Chap 10 22 / 41

Identifying Outlying X Observations-Hat Matrix Leverage Values Identifying Outlying X Observations Using H for identifying outlying X 0 h ii 1 ni=1 h ii = p h ii : called leverage; measure the distance between X i and X large h ii X i distant from the center of all Xs hsuhl (NUK) LR Chap 10 23 / 41

Identifying Outlying X Observations-Hat Matrix Leverage Values Identifying Outlying X Observations (cont.) If X i is outlying has a large leverage h ii Ŷ i : a linear combination of Y (Ŷ = HY ) h ii : the weight of Y i the larger is h ii, the more important is Y i in determining Ŷi The larger is h ii, the smaller is σ 2 {e i }. h ii = 1 σ 2 {e i } = 0 Rule: 1 2 hii h ii > 2 h = 2 n = 2 p n ( 適用 : 2p n 1) { very high leverage:hii > 0.5 moderate leverage:h ii : 0.2 0.5 hsuhl (NUK) LR Chap 10 24 / 41

Identifying Outlying X Observations-Hat Matrix Leverage Values Identifying Outlying X Observations (cont.) Body fat example 2p/n = 0.30 h 33 = 0.372; h 15,15 = 0.333 hsuhl (NUK) LR Chap 10 25 / 41

Identifying Outlying X Observations-Hat Matrix Leverage Values Identifying Outlying X Observations (cont.) ex<-read.table("ch07ta01.txt",header=f) colnames(ex)<-c("x1","x2","x3","y") attach(ex) fit<-lm(y~x1+x2+x3) n<-length(y); p = 3 plot(x2~x1, pch=16 ) text(x1+0.5, X2, labels=as.character(1:length(x1)),col="red") hii<-hatvalues(fit) index<-hii>2*p/n points( X1[index], X2[index], cex=2.0, col= blue ) hsuhl (NUK) LR Chap 10 26 / 41

Identifying Influential Cases-DFFITS, Cook s Distance, and DFBETAS Measures DFFITS meansure Influence on single fitted value: DFFITS-measure the influence that case i has on Ŷi (DFFITS) i = ( Ŷi Ŷ i(i) hii = t i MSE(i) h ii 1 h ii ) 1/2 Rule: { DFFITS > 1 for small to medium data sets DFFITS > 2 p/n for large data sets If X i is an outlier and has a high h ii, (DFFITS) i will tend to be large absolutely. hsuhl (NUK) LR Chap 10 27 / 41

Identifying Influential Cases-DFFITS, Cook s Distance, and DFBETAS Measures Cook s distance Influence on all fitted value: Cook s distance-consider the influence of the ith case on all n fitted values nj=1 (Ŷ i Ŷ j(i) ) 2 D i = = (Ŷ Ŷ (i)) (Ŷ Ŷ (i)) pmse pmse [ ] = e2 i h ii pmse (1 h ii ) 2 1 the size of e i 2 the leverage h ii 3 e i or h ii D i Rule: D i F (p, n p) { little influence : P(F (p, n p) Di ) > 0.1 or 0.2 major influence : P(F (p, n p) D i ) > 0.5 hsuhl (NUK) LR Chap 10 28 / 41

Identifying Influential Cases-DFFITS, Cook s Distance, and DFBETAS Measures Cook s distance (cont.) hsuhl (NUK) LR Chap 10 29 / 41

Identifying Influential Cases-DFFITS, Cook s Distance, and DFBETAS Measures DFBETAS Influence on regression coefficients: DFBETAS-the difference between b k and b k(i) Rule: (DFBETAS) k(i) = b k b k(i) MSE(i) c kk, k = 0, 1,..., p 1 where c kk : the kth diagonal element of (X X) 1 { DFBETAS > 1 for small to medium data sets DFBETAS > 2 n for large data sets hsuhl (NUK) LR Chap 10 30 / 41

Identifying Influential Cases-DFFITS, Cook s Distance, and DFBETAS Measures DFBETAS (cont.) # Body fat example (Table 10.4) > influence.measures(fit) Influence measures of lm(formula = Y ~ X1 + X2) : dfb.1_ dfb.x1 dfb.x2 dffit cov.r cook.d hat inf 1-3.05e-01-1.31e-01 2.32e-01-3.66e-01 1.361 4.60e-02 0.2010 2 1.73e-01 1.15e-01-1.43e-01 3.84e-01 0.844 4.55e-02 0.0589 3-8.47e-01-1.18e+00 1.07e+00-1.27e+00 1.189 4.90e-01 0.3719 * 4-1.02e-01-2.94e-01 1.96e-01-4.76e-01 0.977 7.22e-02 0.1109 5-6.37e-05-3.05e-05 5.02e-05-7.29e-05 1.595 1.88e-09 0.2480 * 6 3.97e-02 4.01e-02-4.43e-02-5.67e-02 1.371 1.14e-03 0.1286 7-7.75e-02-1.56e-02 5.43e-02 1.28e-01 1.397 5.76e-03 0.1555 8 2.61e-01 3.91e-01-3.32e-01 5.75e-01 0.780 9.79e-02 0.0963 9-1.51e-01-2.95e-01 2.47e-01 4.02e-01 1.081 5.31e-02 0.1146 10 2.38e-01 2.45e-01-2.69e-01-3.64e-01 1.110 4.40e-02 0.1102 11-9.02e-03 1.71e-02-2.48e-03 5.05e-02 1.359 9.04e-04 0.1203 12-1.30e-01 2.25e-02 7.00e-02 3.23e-01 1.152 3.52e-02 0.1093 13 1.19e-01 5.92e-01-3.89e-01-8.51e-01 0.827 2.12e-01 0.1784 14 4.52e-01 1.13e-01-2.98e-01 6.36e-01 0.937 1.25e-01 0.1480 15-3.00e-03-1.25e-01 6.88e-02 1.89e-01 1.775 1.26e-02 0.3332 * 16 9.31e-03 4.31e-02-2.51e-02 8.38e-02 1.309 2.47e-03 0.0953 17 7.95e-02 5.50e-02-7.61e-02-1.18e-01 1.312 4.93e-03 0.1056 18 1.32e-01 7.53e-02-1.16e-01-1.66e-01 1.462 9.64e-03 0.1968 19-1.30e-01-4.07e-03 6.44e-02-3.15e-01 1.002 3.24e-02 0.0670 20 1.02e-02 2.29e-03-3.31e-03 9.40e-02 1.224 3.10e-03 0.0501 hsuhl (NUK) LR Chap 10 31 / 41

Identifying Influential Cases-DFFITS, Cook s Distance, and DFBETAS Measures DFBETAS (cont.) Some final comments: Analysis of outlying and influential cases: a necessary component of good regression analysis neither automatic nor foolproof( 極簡單的 ) require good judgment by the analyst Methods described: ineffective Extensions of the single-case diagnostic procedures: computational requirements hsuhl (NUK) LR Chap 10 32 / 41

Multicollinearity Diagnostics-VIF Variance Inflation( 膨脹 ) Factor-VIF Some problems: multicollinearity Adding or deleting X: change the regression coefficients Extra sum of squares: depending upon which other X k s are already included in the model X k highly correlated with each other s{b k } the estimated regression coefficients individually may not be statistically significant hsuhl (NUK) LR Chap 10 33 / 41

Multicollinearity Diagnostics-VIF Variance Inflation( 膨脹 ) Factor-VIF (cont.) informal diagnostics for multicollinearity: 1 Large changes in b k when X k is added or deleted, or when an observation is altered( 改變 ) or deleted 2 Nonsignificant results in individual tests on the regression coefficients for important predictor variables. 3 Estimated regression coefficients with an algebraic sign that is the opposite of that expected from theoretical consideration or prior experience. 4 large r XX 5 Wide confidence interval for β k hsuhl (NUK) LR Chap 10 34 / 41

Multicollinearity Diagnostics-VIF Variance Inflation( 膨脹 ) Factor-VIF (cont.) Important limitations: do not provide quantitative measurements may not identify the nature of the multicollinearity sometimes the observed behavior may occur without multicollinearity being present hsuhl (NUK) LR Chap 10 35 / 41

Multicollinearity Diagnostics-VIF Variance Inflation( 膨脹 ) Factor-VIF (cont.) Variance Inflation Factor (VIF)( 變異數膨脹因子 ) a formal method: detecting multicollinearity; widely accepted measure how much the variances of b k s are inflated( 膨脹 ) as compared to when the predictor variables are not linearly related. Illustration: Variance-covarianve matrix of b: σ 2 {b} = σ 2 (X X) 1 hsuhl (NUK) LR Chap 10 36 / 41

Multicollinearity Diagnostics-VIF Variance Inflation( 膨脹 ) Factor-VIF (cont.) Using the standardized regression model: Variance-covarianve matrix of b : σ 2 {b } = (σ ) 2 r 1 XX (σ ) 2 : the errir term variance for the transformed model (VIF ) k : the kth diagonal element of r 1 XX σ 2 {b k} = (σ ) 2 (VIF ) k = (σ ) 2 1 R 2 k VIF for b k: (VIF ) k = (1 R 2 k ) 1, k = 1, 2,..., p 1 R 2 k : X k is regressed on the p 2 other X k s R 2 k = 0 (VIF ) k = 1: X k is not linearly related to X k s R 2 k 0 (VIF ) k > 1: indicate inflated variance for b k as a result of the intercorrelations among the X variables hsuhl (NUK) LR Chap 10 37 / 41

Multicollinearity Diagnostics-VIF Variance Inflation( 膨脹 ) Factor-VIF (cont.) Perfect linear association with X k s R 2 k = 1 (VIF ) k and σ 2 {b k} are unbounded Rule: largest VIF value among all Xs as an indicator of the severity( 嚴重 ) of multicollinearity max{vif 1,..., VIF p 1 } > 10 hsuhl (NUK) LR Chap 10 38 / 41

Multicollinearity Diagnostics-VIF Variance Inflation( 膨脹 ) Factor-VIF (cont.) If VIF > 1 serious multicollinearity problems p 1 E (b k βk) 2 = (σ ) 2 k=1 p (VIF ) k k=1 large VIF larger differences between b k and β k If no linearly R 2 k 0 (VIF ) k = 1 p 1 E (b k βk) 2 = (σ ) 2 (p 1) k=1 VIF = (σ ) 2 p k=1 (VIF ) k (σ )(p 1) = p k=1 (VIF ) k (p 1) hsuhl (NUK) LR Chap 10 39 / 41

Multicollinearity Diagnostics-VIF Variance Inflation( 膨脹 ) Factor-VIF (cont.) Figure : VIF-Body Fat Example with three Xs VIF 3 = 105 r 2 13 = 0.458 2 = 0.209764, r 2 23 = 0.085 2 : not large X 3 : R 2 3 = 0.990; strong related to X 1, X 2 hsuhl (NUK) LR Chap 10 40 / 41

Multicollinearity Diagnostics-VIF Variance Inflation( 膨脹 ) Factor-VIF (cont.) Comments Some program: using 1/(VIF ) k = 1 R 2 k < 0.01 (0.001, 0.001) Limitation: cannot distinguish between several simultaneous multicollinearities Other methods: more complex than VIF hsuhl (NUK) LR Chap 10 41 / 41