Beam Example: Identifying Influential Observations using the Hat Matrix

Similar documents
Pumpkin Example: Flaws in Diagnostics: Correcting Models

R version ( ) Copyright (C) 2009 The R Foundation for Statistical Computing ISBN

Spectrophotometer Example: One-Factor Random Effects ANOVA

Straw Example: Randomized Block ANOVA

Natural language support but running in an English locale

(1) (AE) (BE) + (AB) + (CE) (AC) (BE) + (ABCE) +(DE) (AD) (BD) + (ABDE) + (CD) (ACDE) (BCDE) + (ABCD)

R version ( ) Copyright (C) 2009 The R Foundation for Statistical Computing ISBN

Leverage. the response is in line with the other values, or the high leverage has caused the fitted model to be pulled toward the observed response.

Carapace Example: Confidence Intervals in the Wilcoxon Sign-rank Test

Horse Kick Example: Chi-Squared Test for Goodness of Fit with Unknown Parameters

Math Treibergs. Turbine Blade Example: 2 6 Factorial Name: Example Experiment with 1 4 Replication February 22, 2016

Math Treibergs. Peanut Oil Data: 2 5 Factorial design with 1/2 Replication. Name: Example April 22, Data File Used in this Analysis:

Gas Transport Example: Confidence Intervals in the Wilcoxon Rank-Sum Test

Regression Review. Statistics 149. Spring Copyright c 2006 by Mark E. Irwin

UNIVERSITY OF MASSACHUSETTS. Department of Mathematics and Statistics. Basic Exam - Applied Statistics. Tuesday, January 17, 2017

R version ( ) Copyright (C) 2009 The R Foundation for Statistical Computing ISBN

Regression Diagnostics

Lecture 1: Linear Models and Applications

Basic Data Analysis. Stephen Turnbull Business Administration and Public Policy Lecture 9: June 6, Abstract. Regression and R.

Problem 4.4 and 6.16 of Devore discuss the Rayleigh Distribution, whose pdf is x, x > 0; x x 0. x 2 f(x; θ) dx x=0. pe p dp.

STAT 4385 Topic 06: Model Diagnostics

Munitions Example: Negative Binomial to Predict Number of Accidents

Contents. 1 Review of Residuals. 2 Detecting Outliers. 3 Influential Observations. 4 Multicollinearity and its Effects

(U 338-E) 2018 Nuclear Decommissioning Cost Triennial Proceeding A Workpapers

STATISTICS 174: APPLIED STATISTICS FINAL EXAM DECEMBER 10, 2002

R version ( ) Copyright (C) 2009 The R Foundation for Statistical Computing ISBN

f Simulation Example: Simulating p-values of Two Sample Variance Test. Name: Example June 26, 2011 Math Treibergs

MATH 644: Regression Analysis Methods

Chaper 5: Matrix Approach to Simple Linear Regression. Matrix: A m by n matrix B is a grid of numbers with m rows and n columns. B = b 11 b m1 ...

Introduction and Single Predictor Regression. Correlation

14 Multiple Linear Regression

Chapter 3 - Linear Regression

X = S 2 = 1 n. X i. i=1

MS&E 226: Small Data

R version ( ) Copyright (C) 2009 The R Foundation for Statistical Computing ISBN

Lecture 18: Simple Linear Regression

MODELS WITHOUT AN INTERCEPT

ST430 Exam 1 with Answers

Introduction to Linear Regression Rebecca C. Steorts September 15, 2015

Detecting and Assessing Data Outliers and Leverage Points

STAT 350: Summer Semester Midterm 1: Solutions

Math 2311 Written Homework 6 (Sections )

IES 612/STA 4-573/STA Winter 2008 Week 1--IES 612-STA STA doc

CAS MA575 Linear Models

Psychology 405: Psychometric Theory

Tests of Linear Restrictions

Lab 3 A Quick Introduction to Multiple Linear Regression Psychology The Multiple Linear Regression Model

STATISTICS 479 Exam II (100 points)

Chapter 10 Building the Regression Model II: Diagnostics

22s:152 Applied Linear Regression. Take random samples from each of m populations.

Dealing with Heteroskedasticity

6.1 Introduction. Regression Model:

> modlyq <- lm(ly poly(x,2,raw=true)) > summary(modlyq) Call: lm(formula = ly poly(x, 2, raw = TRUE))

Regression Model Specification in R/Splus and Model Diagnostics. Daniel B. Carr

Multiple Linear Regression

No other aids are allowed. For example you are not allowed to have any other textbook or past exams.

Lecture 4: Regression Analysis

a. YOU MAY USE ONE 8.5 X11 TWO-SIDED CHEAT SHEET AND YOUR TEXTBOOK (OR COPY THEREOF).

Lecture 10. Factorial experiments (2-way ANOVA etc)

STATISTICS 110/201 PRACTICE FINAL EXAM

Multiple Regression: Example

Regression diagnostics

Lecture 6 Multiple Linear Regression, cont.

Statistics GIDP Ph.D. Qualifying Exam Methodology May 26 9:00am-1:00pm

Introduction to Mixed Models in R

Multiple Predictor Variables: ANOVA

COMPREHENSIVE WRITTEN EXAMINATION, PAPER III FRIDAY AUGUST 26, 2005, 9:00 A.M. 1:00 P.M. STATISTICS 174 QUESTION

Figure 1: The fitted line using the shipment route-number of ampules data. STAT5044: Regression and ANOVA The Solution of Homework #2 Inyoung Kim

Stat 5102 Final Exam May 14, 2015

Tests and Confidence Intervals with R *

REGRESSION OUTLIERS AND INFLUENTIAL OBSERVATIONS USING FATHOM

Lecture 2. Simple linear regression

Generating OLS Results Manually via R

CHAPTER 5. Outlier Detection in Multivariate Data

22s:152 Applied Linear Regression. There are a couple commonly used models for a one-way ANOVA with m groups. Chapter 8: ANOVA

Analytics 512: Homework # 2 Tim Ahn February 9, 2016

lm statistics Chris Parrish

R Output for Linear Models using functions lm(), gls() & glm()

22s:152 Applied Linear Regression

Matrices and vectors A matrix is a rectangular array of numbers. Here s an example: A =

MA 575 Linear Models: Cedric E. Ginestet, Boston University Mixed Effects Estimation, Residuals Diagnostics Week 11, Lecture 1

Weighted Least Squares

Inference for Regression

Econometrics 1. Lecture 8: Linear Regression (2) 黄嘉平

ST430 Exam 2 Solutions

Linear Modelling: Simple Regression

R 2 and F -Tests and ANOVA

Lecture 2. The Simple Linear Regression Model: Matrix Approach

Density Temp vs Ratio. temp

Regression Diagnostics for Survey Data

SCHOOL OF MATHEMATICS AND STATISTICS

Statistics GIDP Ph.D. Qualifying Exam Methodology May 26 9:00am-1:00pm

Multiple Linear Regression

Lectures on Simple Linear Regression Stat 431, Summer 2012

MLR Model Checking. Author: Nicholas G Reich, Jeff Goldsmith. This material is part of the statsteachr project

Lecture 4 Multiple linear regression

Variance Decomposition and Goodness of Fit

Chapter 12: Multiple Linear Regression

Nature vs. nurture? Lecture 18 - Regression: Inference, Outliers, and Intervals. Regression Output. Conditions for inference.

13 Simple Linear Regression

Transcription:

Math 3080. Treibergs Beam Example: Identifying Influential Observations using the Hat Matrix Name: Example March 22, 204 This R c program explores influential observations and their detection using the hat matrix. We use Devore s data although we do not recover the same numbers he gets. This data was taken from Devore, Probability and Statistics for Enigneering and the Sciencws, 8th ed., Brooks/Cole, Boston, 202. data is taken from the article Testing for the Inclusion of Variables in Linear Regression.. in Technometrics 966, which was reanalyzed by Hoaglin & Welsch, The Hat Matrix... Amer. Statistician, 978. The strength of a beam y is to be regressed on specific gravity x and moisture content x 2. Recall, that to find the least squares fit to the linear model y i = β 0 + β x i, + β 2 x i,2 + ɛ i where ɛ i are n IID N(0, σ 2 ) random variables, we solve for the vector of estimated coefficients ˆβ = (X T X) X T y, where k is the number of variables, ˆβ is the (k + ) matrix of estimated coefficients, X is the n (k + ) dimensional model matrix consisting of columns of ones and independent variables and ɛ and y are n dimensional martices of noise and observed responses. (See my Colleges Example, where this is derived.) The hat matrix is given by H = X (X T X) X T It gives the fitted values as a function of the observed values ŷ = Hy. On average, the diagonals of the hat matrix h ii tell the sensitivity of the ith fitted value to the ith observed value. These are referred to as leverage values. Thus large hat diagonals signal influential observations. Since the average of the h ii s is (k + )/n, Devore suggests that any diagonal hat entries larger than twice the average, h ii 2(k + )/n, should be identified as potentially influential. Further Devore looks at the sensitivity to omission of the ith observation on the computations of the ˆβ s. If we let ˆβ j be the estimated coefficient using all observations and β j the estimate after deleting the ith observation, then a measure of the effect of the ith observation on the jth beta is ˆβ j β j s ˆβj. if this is large, on the order of one (difference on the order of a standard deviation) then this is indicative of an influential observation. Some statisticians regard larger than two as large. The actual quantity that is computed differs from this formula slightly and is called dffbeta, whch can be generated by the software. Similarly the software computes df f it for each observation, which is the difference in the fitted value relative to the standard deviation due to the omission of the ith observation. The studentized residuals (as opposed to the standardized) are computed by leaving out the ith observation to for the fitted value, which should look like the standardized residual unless the point is influential. The Cook s distance is another measure that averages all df beta s and has the same range as the square of the betas. Influential observations should be mentioned in summarizing your data analysis. First, if outliers and influential can be identified as typographical or measurement errors, they can be removed from the sample. Otherwise you have to deal with the data as it is. If the fitted line does not change much after removing the influential points then just fit with all data points. But if the coefficients change, then a range of coefficients should be reported. Omitting influential points will underreport the variation.

Data Set Used in this Analysis : # Math 3080- Beam Data March 29, 204 # Treibergs # # From Devore, "Probability and Statistics for Engineering and the # Sciences," 8th ed., Brooks/Cole, Boston, example 3.24. # From the article "Testing for the Inclusion of Variables in Linear # Regression.." in Technometrics 966, which was reanalyzed by Hoaglin & # Welsch, "The Hat Matrix," Amer. Statistician, 978. # # variables # x specific gravity # x2 moisture content # y strength "x" "x2" "y".499..4.558 8.9 2.74.604 8.8 3.3.4 8.9.5.550 8.8 2.38.528 9.9 2.60.48 0.7.3.480 0.5.70.406 0.5.02.467 0.7.4 R Session: R version 2.3. (20-07-08) Copyright (C) 20 The R Foundation for Statistical Computing ISBN 3-90005-07-0 Platform: i386-apple-darwin9.8.0/i386 (32-bit) R is free software and comes with ABSOLUTELY NO WARRANTY. You are welcome to redistribute it under certain conditions. Type license() or licence() for distribution details. Natural language support but running in an English locale R is a collaborative project with many contributors. Type contributors() for more information and citation() on how to cite R or R packages in publications. Type demo() for some demos, help() for on-line help, or help.start() for an HTML browser interface to help. Type q() to quit R. [R.app GUI.4 (5874) i386-apple-darwin9.8.0] 2

[History restored from /Users/andrejstreibergs/.Rapp.history] tt=read.table("m3082databeam.txt",header=t) attach(tt) tt x x2 y 0.499..4 2 0.558 8.9 2.74 3 0.604 8.8 3.3 4 0.4 8.9.5 5 0.550 8.8 2.38 6 0.528 9.9 2.60 7 0.48 0.7.3 8 0.480 0.5.70 9 0.406 0.5.02 0 0.467 0.7.4 ################## PRINT PAIRWISE SCARTTERPLOTS FOR THEIS DATA ######### pairs(tt,gap=0) ################## RUN THE LINEAR MODEL OF THIS DATA ################### f = lm(y ~ x + x2 ) summary(f); anova(f) Call: lm(formula = y ~ x + x2) Residuals: Min Q Median 3Q Max -0.38607-0.0598 0.0294 0.07932 0.45560 Coefficients: Estimate Std. Error t value Pr( t ) (Intercept).3683.5985 7.2 0.00092 *** x 7.660.4984 5.2 0.0038 ** x2-0.330 0.093-3.020 0.09382 * Signif. codes: 0 *** 0.00 ** 0.0 * 0.05. 0. Residual standard error: 0.2605 on 7 degrees of freedom Multiple R-squared: 0.905,Adjusted R-squared: 0.8849 F-statistic: 35.6 on 2 and 7 DF, p-value: 0.000244 Analysis of Variance Table Response: y Df Sum Sq Mean Sq F value Pr(F) x 4.259 4.259 62.040 0.000003 *** x2 0.692 0.692 9.23 0.093822 * Residuals 7 0.4752 0.0679 --- Signif. codes: 0 *** 0.00 ** 0.0 * 0.05. 0. ######### NOTE THAT THIS DIFFERS FROM TABLE 3.2 IN DEVORE ########## ######### I WAS UNABLE DEVORE S CENTERING OR STANDARDIZATION ########## 3

######### COMPUTE THE MATRICES "BY HAND" ############################## X = model.matrix(f); X (Intercept) x x2 0.499. 2 0.558 8.9 3 0.604 8.8 4 0.4 8.9 5 0.550 8.8 6 0.528 9.9 7 0.48 0.7 8 0.480 0.5 9 0.406 0.5 0 0.467 0.7 attr(,"assign") [] 0 2 ############ HERE IS A WAY TO COMPUTE INVERSE MATRICES ########## ############ START WITH AN EXAMPLE OF A 3X3 SYMMETRIC MATRIX ###### ExMx=matrix(c(6,6,4,6,8,2,4,2,0),ncol=3); ExMx [,] [,2] [,3] [,] 6 6 4 [2,] 6 8 2 [3,] 4 2 0 ExMxInv = chol2inv(chol(exmx)); ExMxInv [,] [,2] [,3] [,].875-0.825-0.325 [2,] -0.825 0.6875 0.875 [3,] -0.325 0.875 0.875 ############# CHECK IT IS THE INVERSE ######################## ExMx %*% ExMxInv; ExMxInv %*% ExMx [,] [,2] [,3] [,].000000e+00-7.7756e-6 -.0223e-6 [2,] 2.220446e-6.000000e+00.665335e-6 [3,] 0.000000e+00 2.220446e-6.000000e+00 [,] [,2] [,3] [,].000000e+00 2.220446e-6 0.000000e+00 [2,] -7.7756e-6.000000e+00 2.220446e-6 [3,] -.0223e-6.665335e-6.000000e+00 ############## USE MATRIX ARITHMETIC TO RECOVER HAT-BETA ##### XTX = t(x) %*% X; XTX (Intercept) x x2 (Intercept) 0.000 4.92000 98.8000 x 4.92 2.463435 48.379 x2 98.800 48.37900 984.0000 4

XTXInv = chol2inv(chol(xtx)); XTXInv [,] [,2] [,3] [,] 37.639824-28.82426-2.3640486 [2,] -28.82426 33.075676.2697253 [3,] -2.364049.269725 0.76034 #################### CHECK IT IS THE INVERSE ######### XTXInv %*% XTX (Intercept) x x2 [,].000000e+00-7.05427e-4-9.094947e-3 [2,] -.42085e-4.000000e+00 0.000000e+00 [3,] 0.000000e+00 3.55274e-5.000000e+00 ################### NOW USE HAT-BETA = (XTX)^(-) XT Y ##### XTX = t(x) %*% X; XTX (Intercept) x x2 (Intercept) 0.000 4.92000 98.8000 x 4.92 2.463435 48.379 x2 98.800 48.37900 984.0000 XTXInv = chol2inv(chol(xtx)); XTXInv [,] [,2] [,3] [,] 37.639824-28.82426-2.3640486 [2,] -28.82426 33.075676.2697253 [3,] -2.364049.269725 0.76034 #################### CHECK IT IS THE INVERSE ######### XTXInv %*% XTX (Intercept) x x2 [,].000000e+00-7.05427e-4-9.094947e-3 [2,] -.42085e-4.000000e+00 0.000000e+00 [3,] 0.000000e+00 3.55274e-5.000000e+00 ################ FORMULA FOR HAT-BETA ################ cbind(coefficients(f), XTXInv%*% t(x) %*% Y) [,] [,2] (Intercept).368386.368386 x 7.660449 7.660449 x2-0.330494-0.330494 ############## HAT MATRIX ############################ Hat = X %*% XTXInv %*% t(x); Hat 2 3 4 5 6 0.384960960-0.0092906 0.05747392-0.2639022-0.03849958 0.6827506 2-0.0092906 0.24870706 0.300694 0.20435 0.250027 0.380252 3 0.05747392 0.300694 0.4258974-0.048788 0.28677648 0.826845 4-0.2639026 0.204346-0.0487876 0.68843926 0.706730-0.04648074 5-0.03849958 0.25002707 0.286776478 0.706730 0.2574245 0.78965 6 0.68275060 0.380257 0.8268453-0.04648074 0.78965 0.4452200 7 0.559632-0.04257466-0.203305 0.6507325-0.03590573 0.05039582 8 0.27078893 0.03360238 0.04202728-0.02328874 0.022846 0.576935 9 0.085559649-0.03564040-0.30382773 0.2672920-0.093 0.0260232 0 0.238678698 0.003674987 0.00232067-0.02733838-0.00926060 0.0982357 5

7 8 9 0 0.55963 0.2707889 0.08555965 0.238678698 2-0.0425747 0.03360232-0.0356404 0.003674987 3-0.20330 0.04202722-0.3038277 0.00232067 4 0.6507325-0.02328874 0.2672920-0.027338385 5-0.03590573 0.022846-0.093-0.009260600 6 0.05039582 0.576935 0.0260232 0.0982357 7 0.24567547 0.482977 0.25254000 0.76598560 8 0.482977 0.534595 0.24820 0.678346 9 0.25254000 0.24820 0.27730348 0.557288 0 0.7659856 0.678346 0.55729 0.8693635 ############### NOTE THAT THIS IS DIFFERENT THAN p.583 ############# ############### CHECK THAT IT IS THE HAT MATRIX. IT RECOVERS ############# ############### THE FITTED VALUES HAT-Y = HAT * Y. ############# cbind( fitted(f), Hat %*% y) [,] [,2].52607.52607 2 2.70435 2.70435 3 3.08973 3.08973 4.5783.5783 5 2.67608 2.67608 6 2.4440 2.4440 7.03766.03766 8.57862.57862 9.077.077 0.430.430 ########## WE CAN EXTRACT THE DIAGONAL OF HAT FROM THE MODEL ######## cbind(hatvalues(f), diag(hat)) [,] [,2] 0.384960 0.384960 2 0.248707 0.248707 3 0.425892 0.425892 4 0.6884393 0.6884393 5 0.257425 0.257425 6 0.445220 0.445220 7 0.2456755 0.2456755 8 0.534592 0.534592 9 0.2773035 0.2773035 0 0.869364 0.869364 ######## THE CRITICAL VALUES ARE DOUBLE OR TRIPLE AVERAGE VALUES ##### n=length(x); k=2; DoubleMeanHat=2*(k+)/n;TripleMeanHat=3*(k+)/n; DoubleMeanHat; TripleMeanHat [] 0.6 [] 0.9 ############# SO THE FOURTH OBSERVATION IS POTENTIALLY INFLUENTIAL #### 6

############# RUN THE REGRESSION WITHOUT OBS. 4 ####################### f5=lm(y~x+x2,data=tt, subset=-4); summary(f5) Call: lm(formula = y ~ x + x2, data = tt, subset = -4) Residuals: Min Q Median 3Q Max -0.33339-0.05037 0.027 0.0565 0.46579 Coefficients: Estimate Std. Error t value Pr( t ) (Intercept) 2.407 2.907 4.269 0.00527 ** x 6.7992 2.566 2.702 0.03549 * x2-0.3905 0.794-2.77 0.07237. --- Signif. codes: 0 *** 0.00 ** 0.0 * 0.05. 0. Residual standard error: 0.277 on 6 degrees of freedom Multiple R-squared: 0.908,Adjusted R-squared: 0.88 F-statistic: 30.65 on 2 and 6 DF, p-value: 0.0007089 ###### NORMALIZED DIFFERENCE OF BETAS WITH AND WITHOUT OBS. 4 #### coefficients(f)-coefficients(f5) (Intercept) x x2 -.042339 0.86093069 0.06039984 ####### COVARIANCE MATRIX FOR THE REGRESSION #################### vcov(f) (Intercept) x x2 (Intercept) 2.555398 -.956525-0.6048095 x -.956522 2.24530739 0.0869397 x2-0.604809 0.0869397 0.094989 ############ THE SD S OF ESTIMATED BETAS IS SQRT OF DIAGONAL #### sbetas = sqrt(diag(vcov(f))); sbetas (Intercept) x x2.5984805.4984350 0.09355 ########## DIFFERENCES OF BETAS WITH AND WITHOUT OBS 4 $$$$$$$$$$ diffbetas = coefficients(f)-coefficients(f5); diffbetas (Intercept) x x2 -.042339 0.86093069 0.06039984 sdiffbetas=diffbetas/sbetas sdiffbetas (Intercept) x x2-0.652082 0.5745532 0.5525275 7

#### ONE COULD COMPARE THE EFFECT OF REMOVING ONE OBSERVATION #### FOR EACH OF THE ESTIMATED BETAS. THE CODE USED BY R #### DIFFERS SLIGHTLY FROM THE COMPUTATION SUGGESTED IN DEVORE. #### LOOK AT THE N=4 ROW IN THE PRINTOUT OF INFLUENCE MEASURES. #### THE dfb (SHORT FOR dfbeta) GIVES THE CHANGE OF THE ESTIMATED #### PARAMETER RELATIVE TO ITS STANDARD DEVIATION. #### dffits ARE THE RELATIVE CHANGE OF THE FITTED VALUE DUE TO THE #### OMISSION OF THE DATA POINT. #### COOK S DISTANCE cookd IS A JOINT MEASURE OF THE COMPONENTS OF #### THE dfbetas. THE SQUARE ROOT OF COOK S DISTANCE IS IN THE SAME #### SCALE AS dfbetas. THE SUMMARY IS AVAILABLE USING influence.measures. if=influence.measures(f); if Influence measures of lm(formula = y ~ x + x2) : dfb._ dfb.x dfb.x2 dffit cov.r cook.d hat inf.54932-0.984725 -.70e+00 -.97706 0.304 7.45e-0 0.385 * 2 0.0425 0.027472-3.58e-02 0.08424 2.09 2.75e-03 0.249 3-0.02278 0.09897-2.80e-02 0.5692 2.656 9.52e-03 0.43 * 4-0.6345 0.54059 5.20e-0-0.65684 4.630.62e-0 0.688 * 5-0.26223-0.54524 4.54e-0-0.8297 0.908 2.0e-0 0.257 6-0.43299 0.570478 3.7e-0.0284 0.28 2.0e-0 0.45 7 0.0232-0.0795 5.27e-02 0.282.958.8e-02 0.246 8-0.08605 0.034932.6e-0 0.20335.677.55e-02 0.53 9 0.00736-0.04502 -.76e-05 0.023 2.96.77e-04 0.277 0 0.00239-0.000482-3.52e-03-0.00568.953.26e-05 0.87 ############# PLOT DIAGNOSTICS ######################### layout(matrix(c(,3,2,4),ncol=2)) plot(f, which=:4) plot(rstandard(f));abline(h=c(0,-2,2),lty=c(2,3,3)) plot(rstudent(f));abline(h=c(0,-2,2),lty=c(2,3,3)) plot(dffits(f),type="b",xlab="index") matplot(dfbetas(f),type=c("b","b","b"),pch=c("0","","2"),xlab="index") 8

9.0 9.5 0.0 0.5.0 9.0 9.5 0.0 0.5.0 x x2 0.40 0.45 0.50 0.55 0.60 y.0.5 2.0 2.5 3.0 0.40 0.45 0.50 0.55 0.60.0.5 2.0 2.5 3.0 9

Residuals vs Fitted Normal Q-Q Residuals -0.4-0.2 0.0 0.2 0.4 6 5 Standardized residuals -2-0 2 5 6.0.5 2.0 2.5 3.0 Fitted values -.5 -.0-0.5 0.0 0.5.0.5 Theoretical Quantiles Scale-Location Cook's distance Standardized residuals 0.0 0.4 0.8.2 6 5 Cook's distance 0.0 0.2 0.4 0.6 0.8 5 6.0.5 2.0 2.5 3.0 2 4 6 8 0 Fitted values Obs. number 0

rstandard(f) -2-0 2 rstudent(f) -2-0 2 2 4 6 8 0 Index 2 4 6 8 0 Index dffits(f) -2.0 -.0 0.0.0 dfbetas(f) -.5-0.5 0.5.5 0 2 0 0 2 2 2 2 0 0 2 0 02 0 2 0 2 0 2 2 4 6 8 0 Index 2 4 6 8 0 Index