Chapter 11 Building the Regression Model II:
|
|
- Abigail Jennings
- 5 years ago
- Views:
Transcription
1 Chapter 11 Building the Regression Model II: Remedial Measures( 補救措施 ) 許湘伶 Applied Linear Regression Models (Kutner, Nachtsheim, Neter, Li) hsuhl (NUK) LR Chap 11 1 / 48
2 11.1 WLS remedial measures may need to be taken: a regression model is not appropriate several cases are very influential previous: transformations to linearize the regression relation the error distributions more nearly normal make the variances of the error terms more nearly equal hsuhl (NUK) LR Chap 11 2 / 48
3 11.1 WLS remedial measures In this chapter- remedial measure to deal with: unequal error variances a high degree of multicollinearity influential observations two methods for nonparametric regression: lowess regression trees bootstrapping: for evaluating the precision of the complex estimators hsuhl (NUK) LR Chap 11 3 / 48
4 11.1 Unequal Error Variances Remedial Measures-Weighted Least Squares Weighted Least Squares (WLS) Chap. 3,6: transformation of Y - reducing or eliminating unequal variances difficulty: may create an inappropriate regression relationship weighted least squares( 加權最小平方法 ): when an appropriate regression relationship has been found but the variances of the error terms are unequal hsuhl (NUK) LR Chap 11 4 / 48
5 11.1 Unequal Error Variances Remedial Measures-Weighted Least Squares Weighted Least Squares (WLS) (cont.) Y i = β 0 + β 1 X i1 + + β p 1 X i,p 1 + ε i, i = 1,..., n parameters: β 0,..., β p 1 known constants: X i1,..., X i,p 1 ε i indep. N(0, σ 2 i ) σ 2 {ε} n n = σ σ σn 2 b = (X X) 1 X Y unbias, consistent but don t have minimum variance hsuhl (NUK) LR Chap 11 5 / 48
6 11.1 Unequal Error Variances Remedial Measures-Weighted Least Squares Weighted Least Squares (WLS) (cont.) Consider: Observations with small variances provide more reliable information about the regression function than those with large variances. Error Variances Know (w i = 1 )(unrealistic 不切實際的 ) σi 2 Error Variances Know up to Proportionality Constant (w i = k 1 ) σi 2 Error Variances Unknown: estimation of variance function or standard deviation function (w i = 1 (ŝ i, w ) 2 i = 1ˆν i ) hsuhl (NUK) LR Chap 11 6 / 48
7 11.1 Unequal Error Variances Remedial Measures-Weighted Least Squares WLS-error variances known Methods of maximum likelihood to obtain estimators: indep. ε i N(0, σi 2 ) and w i = 1 σi 2 L(β) = = [ n i=1 n i=1 ( wi 2π 1 (2πσi 2 ] ) 1/2 b= arg max L(β) β [ exp 1 )1/2 2σi 2 [ exp b = arg min β Q w (Q w = 1 2 (Y i β 0 β 1 X i1 β p 1 X i,p 1 ) 2 ] ] n w i (Y i β 0 β 1 X i1 β p 1 X i,p 1 ) 2 i=1 n w i (Y i β 0 β 1 X i1 β p 1 X i,p 1 ) 2 ) i=1 Called weighted least squares criterion (min Q w ) w i = 1 OLS hsuhl (NUK) LR Chap 11 7 / 48
8 11.1 Unequal Error Variances Remedial Measures-Weighted Least Squares WLS-error variances known (cont.) w i = 1 : reflects the amount of information contained in the σi 2 observation Y i vary i large w i less W = n n w w w n The normal equations: (X W X)b w = X W Y The weighted least squares (MLE) for β: σ 2 {b w } p p b w = (X W X) 1 X W Y = (X W X) 1 ( σ 2 {Y } = W 1 ) hsuhl (NUK) LR Chap 11 8 / 48
9 11.1 Unequal Error Variances Remedial Measures-Weighted Least Squares WLS-error variances known (cont.) b w : unbiased, consistent, have minimum variance among unbiased linear estimators When the weighted are known, b w generally exhibits less variability than b. hsuhl (NUK) LR Chap 11 9 / 48
10 11.1 Unequal Error Variances Remedial Measures-Weighted Least Squares WLS-error variances known up to proportionality constant w i = k 1 σ 2 i σ 2 {b w } p p k: unknwon s 2 {b w } p p where MSE w = b w = (X W X) 1 X W Y = k(x W X) 1 = MSE w (X W X) 1 wi (Y i Ŷ i ) 2 n p = (unaffected) wi e 2 i n p MSE w : an estimator of the proportionality constant k hsuhl (NUK) LR Chap / 48
11 11.1 Unequal Error Variances Remedial Measures-Weighted Least Squares WLS-error variances unknown Estimation of Variance Function or Standard Deviation Function One rarely has knowledge of the variances σ 2 i. estimate of the variances σ 2 i = E{ε 2 i } (E{ε i }) 2 = E{ε 2 i } Estimator of σi 2 : the squared residual ei 2 Estimator of σ i = σi 2 : the absolute residual e i hsuhl (NUK) LR Chap / 48
12 11.1 Unequal Error Variances Remedial Measures-Weighted Least Squares WLS-error variances unknown (cont.) the estimation process: 1 Fit the regression model by OLS and analyze e i 2 Estimate the variance function or the standard deviation function by regressing ei 2 or e i on the appropriate predictor(s). 3 ŝ i for standard deviation function, ˆν i for the variance function: to obtain w i 4 obtain b w by using w i (Sometimes, iterated for several times to reach stabilize iteratively reweighted lease squares) hsuhl (NUK) LR Chap / 48
13 11.1 Unequal Error Variances Remedial Measures-Weighted Least Squares WLS-error variances unknown (cont.) Reference: figure from Faraway s Linear Models with R (2005, p. 59) hsuhl (NUK) LR Chap / 48
14 11.1 Unequal Error Variances Remedial Measures-Weighted Least Squares WLS-error variances unknown (cont.) Some possible variance and standard deviation functions: 1 plot e i vs. X 1 : a megaphone( 大聲公 ) shape e i regresses on X 1 2 plot e i vs. Ŷ : a megaphone( 大聲公 ) shape e i regresses on Ŷ 3 plot e 2 i vs. X 3 : upward tendency e 2 i regresses on X 3 4 plot e i vs. X 2 : increase rapidly with X 2 and increases more slowly e i regresses on X 2, X 2 2 w i = 1 (ŝ i ) 2 : ŝ i-fitted value from standard function w i = 1ˆν i : ˆν i -fitted value from variance function the weight matrix W b w = (X W X) 1 X W Y hsuhl (NUK) LR Chap / 48
15 11.1 Unequal Error Variances Remedial Measures-Weighted Least Squares WLS-error variances unknown (cont.) Using of Replicated or Near Replicates Replicate observations are made at each combination of levels of Xs It the number of replications (or near replications) is large, w i may be obtained directly from the sample variances of Y observations at each combination of levels on X variables each case in a replicate group receives the same weight with this method Confidence interval: (similar but approximate) b k ± t(1 α/2; n p)s{b k } (6.50) hsuhl (NUK) LR Chap / 48
16 11.1 Unequal Error Variances Remedial Measures-Weighted Least Squares WLS-error variances unknown (cont.) Using of Ordinary Least Squares with Unequal Error Variances: b = (X X) 1 X Y σ 2 {b} = (X X) 1 (X σ 2 {ε}x)(x X) 1 S 2 {b} = (X X) 1 (X S 0 {ε}x)(x X) 1 where S 0 = diag{e1, 2 e2, 2..., en} 2 hsuhl (NUK) LR Chap / 48
17 11.1 Unequal Error Variances Remedial Measures-Weighted Least Squares blood pressure example 54 subjects: women hsuhl (NUK) LR Chap / 48
18 11.1 Unequal Error Variances Remedial Measures-Weighted Least Squares blood pressure example (cont.) linear regression function by unweighted least squares: Ŷ = (3.994) ( ) X hsuhl (NUK) LR Chap / 48
19 11.1 Unequal Error Variances Remedial Measures-Weighted Least Squares blood pressure example (cont.) e i regresses on X: ŝ = X ŝ 1 = w 1 = 1/(ŝ 1 ) 2 = WLS: Ŷ = X C.I. for β 1 : α = 0.05; s{b w1 } = ± t(0.975; 52) ( ) β =2.007 hsuhl (NUK) LR Chap / 48
20 11.1 Unequal Error Variances Remedial Measures-Weighted Least Squares blood pressure example (cont.) unequal variances: heteroscedasticity( 異質性 ); equal variances:homoscedasticity( 等變異性 ) R 2 does not have a clear-cut meaning for WLS WLS may be view as OLS of transformed variables: W 1/2 Y Y w Y = Xβ + ε, = W 1/2 Xβ X w β E{ε} = 0, σ 2 {ε} = W 1 + W 1/2 ε ε w E{ε w } = 0; σ 2 {ε w } = I; b w = (X wx w ) 1 X wy w = (X W X) 1 X W Y hsuhl (NUK) LR Chap / 48
21 11.1 Unequal Error Variances Remedial Measures-Weighted Least Squares blood pressure example (cont.) The weighted least squares normal equations: wi Y i = b w0 wi + b w1 wi X i wi X i Y i = b w0 wi X i + b w1 wi X 2 i The weighted least squares estimators b w0 and b w1 in (11.9) are: b w1 = b w0 = wi X i Y i wi X i wi Y i wi wi Xi 2 ( w i X i ) 2 wi wi Y i b w1 wi X i wi hsuhl (NUK) LR Chap / 48
22 11.2 Ridge Regression Ridge Regression( 脊迴歸 ) ridge regression: a method of overcoming serious multicollinearity problem by modifying the method of least squares. allowing biased estimators of the regression functions When an estimator has only a small bias and is substantially more precise than an unbiased estimator, it may well be the preferred estimator since it will have a larger probability of being close to the true parameter value. hsuhl (NUK) LR Chap / 48
23 11.2 Ridge Regression Ridge Regression( 脊迴歸 ) (cont.) Estimator b is unbiased but imprecise Estimator b R is much more precise but has a small bias The probability that b R falls near β is much greater than that for b hsuhl (NUK) LR Chap / 48
24 11.2 Ridge Regression Ridge Regression( 脊迴歸 ) (cont.) A measure of the combined effect of bias and sampling variation: The mean squared error (MSE): E{b R β} 2 = σ 2 {b R } + (E{b R } β) 2 Transformed variables: Y i = Y i ( 1 Yi Ȳ n 1 s Y ), X ik = = β 1X i1 + β 2X i2 + β p 1X i,p 1 + ε i r XX b = r YX ( ) 1 Xik X k n 1 s k hsuhl (NUK) LR Chap / 48
25 11.2 Ridge Regression Ridge Regression( 脊迴歸 ) (cont.) The ridge standardized regression estimators are obtained by: (r XX + ci)b R = r YX b R = (r XX + ci) 1 r YX Biasing constant c 0: reflects the amount of bias c = 0 OLS; c > 0 bias but more stable (less variable) than OLS estimators hsuhl (NUK) LR Chap / 48
26 11.2 Ridge Regression Ridge Regression( 脊迴歸 ) (cont.) Choice of Biasing Constant c: bias component (E{b R } β) of E{b R β} 2 as c while the variance (σ 2 {b R }) component becomes smaller Always a c s.t. b R has a smaller total MSE than b difficulty: the optimum value of c varies from one application to another and is unknown Using Ridge trace( 脊軌跡 ) and (VIF ) k (c) hsuhl (NUK) LR Chap / 48
27 11.2 Ridge Regression Ridge Regression( 脊迴歸 ) (cont.) Ridge trace:a simultaneous plot of the values of the (p 1) b R for different c; 0 c 1 Choose c: the Ridge trace starts to become stable and VIF has become sufficiently small. hsuhl (NUK) LR Chap / 48
28 11.2 Ridge Regression Ridge Regression( 脊迴歸 ) (cont.) The normal equation for the ridge estimators: (1 + c)b1 R + r 12 b2 R + + r 1,p 1 bp 1 R = r Y 1 r 21 b1 R + (1 + c)b2 R + + r 2,p 1 bp 1 R = r Y 2. r p 1,1 b1 R + r p 1,2 b2 R + + (1 + c)bp 1 R = r Y,p 1 VIF for bk R are defined analogously to those for OLS b k : measures how large is the variance of bk R relative to what the variance would be if X s were uncorrelated. hsuhl (NUK) LR Chap / 48
29 11.2 Ridge Regression Ridge Regression( 脊迴歸 ) (cont.) VIF for b R k : the diagonal elements of the following matrix (r XX + ci) 1 r XX (r XX + ci) 1 Ridge regression estimates can be obtained by the method of penalized least squares. Q = n i=1 [ Y i (β1x i1 + + βp 1X i,p 1) ] p c (βj ) 2 sometimes referred to as shrinkage esimators Limitation: ridge regression is that ordinary inference procedures are not applicable and exact distribution properties are not known. (bootstrapping ) j=1 hsuhl (NUK) LR Chap / 48
30 11.2 Ridge Regression Ridge Regression( 脊迴歸 ) (cont.) (a) Table 11.2 (b) Table hsuhl (NUK) LR Chap / 48
31 11.2 Ridge Regression Ridge Regression( 脊迴歸 ) (cont.) c = 0.02 Ŷ = X X X 3 hsuhl (NUK) LR Chap / 48
32 11.2 Ridge Regression Ridge Regression( 脊迴歸 ) (cont.) LASSO: Least absolute shrinkage( 收縮 ) and selection operator( 最小絕對緊縮與選擇算子 ) ˆβ = arg min β n p 1 (Y i i=1 j=0 p 1 X i jβ j ) 2 + λ j=0 β j hsuhl (NUK) LR Chap / 48
33 11.2 Ridge Regression Ridge Regression( 脊迴歸 ) (cont.) hsuhl (NUK) LR Chap / 48
34 11.3 Robust Regression Robust Regression To examine whether an outlying case is the result of a recording error, breakdown( 故障 ) of a measurement instrument, or the like. Erroneous data can be corrected corrected Erroneous data cannot be corrected discarded( 拋棄 ) Many time: impossible to tell for certain whether the observations for an outlying case are erroneous not be discarded If an outlying influential case is not clearly erroneous to examine the adequacy of the model hsuhl (NUK) LR Chap / 48
35 11.3 Robust Regression Robust Regression (cont.) Numerous robust regression procedures have been developed. LAR or LAD regression: Least absolute residuals or least absolute deviations( 最小絕對離差 ) minimum L 1 norm regression Criterion: min L 1 = min β β { n Y i (β 0 + β 1 X i1 + + β p 1 x i,p 1 ) i=1 LAR: less emphasis on outlying observations than does the method of least squares linear programming method: to obtain the LAR estimators LAR method: the residuals will not sum to zero } hsuhl (NUK) LR Chap / 48
36 11.3 Robust Regression Robust Regression (cont.) LMS regression: Least median of squares median { [Y i (β 0 + β 1 X i1 + + β p 1 x i,p 1 )] 2} IRLS robust regression: Iteratively reweighted least squares method hsuhl (NUK) LR Chap / 48
37 11.4 Regression Tree Regression Tree Powerful, conceptually simple method of nonparametric regression 1 the range of the predictor is partitioned into segments 2 within each segment the estimated regression fit is given by the mean of the response in the segment hsuhl (NUK) LR Chap / 48
38 11.4 Regression Tree Regression Tree (cont.) hsuhl (NUK) LR Chap / 48
39 11.4 Regression Tree Regression Tree (cont.) hsuhl (NUK) LR Chap / 48
40 11.4 Regression Tree Regression Tree (cont.) r = 2: the best point is choose to min SSE = SSE(R 21 ) + SSE(R 22 ) SSE(R rj ) = (Y i Ȳ Rjk ) 2 SSE(R rj ): the sum of squared residuals in region R rj hsuhl (NUK) LR Chap / 48
41 11.4 Regression Tree Regression Tree (cont.) hsuhl (NUK) LR Chap / 48
42 11.4 Regression Tree Regression Tree (cont.) Determining the number of Regions R hsuhl (NUK) LR Chap / 48
43 11.5 Bootstrapping Bootstrapping In many nonstandard situations, (ex: nonconstant error variances estimated by iteratively reweighted least squares), standard methods for evaluating the precision may not be available or may only be approximately applicable when the sample size is large Bootstrapping: provide estimates of the precision of sample estimated for these complex cases hsuhl (NUK) LR Chap / 48
44 11.5 Bootstrapping Bootstrapping (cont.) general procedure: obtain b 1 by some procedure and wish to evaluate the precision of b 1 by bootstrap method The bootstrap method calls for( 需要 ) the selection from the observed sample data of a random sample of size n with replacement. the bootstrap sample may contain some duplicate( 複製的 ) data from the original sample and omit some other data in the original sample calculates the estimated regression coefficients from the bootstrap sample b 1 repeated a large number of times The estimated standard deviation of all of b 1 s {b 1} an estimate of the variability of the sampling distribution of b 1 hsuhl (NUK) LR Chap / 48
45 11.5 Bootstrapping Bootstrapping (cont.) Bootstrap Sampling: two basic ways When the regression function being fitted is a good model for the data, the error terms have constant variance, and fixed X sampling is appropriate. e i from the original fitting are regarded as the sample data to be sampled with replacement After a bootstrap sample with n: e 1,..., e n Y i = Ŷ i + e i Y regression on Xs b 1 hsuhl (NUK) LR Chap / 48
46 11.5 Bootstrapping Bootstrapping (cont.) When there is some doubt about the adequacy of the regression function being fitted, the error variance are not constant, and/or random X sampling is appropriate. (X i, Y i ) in the original sample are considered to be the data to be sampled with replacement After a bootstrap sample with n: n pairs: (X 1, Y 1 ),..., (X n, Y n ) Y regression on X s b 1 hsuhl (NUK) LR Chap / 48
47 11.5 Bootstrapping Bootstrapping (cont.) bootstrap samples are adequate One can observe the variability of the bootstrap estimates by s {b 1} as the number of bootstrap samples is increased. Bootstrap Confidence Intervals From the bootstrap distribution of b 1, find the α/2 and 1 α/2 quantiles b 1(α/2) and b 1(1 α/2) Percentiles from b 1 : b 1(α/2) β 1 b 1(1 α/2) reflection method: require a larger number of bootstrap samples 2b 1 b 1(α/2) β 1 2b 1 b 1(α/2) hsuhl (NUK) LR Chap / 48
48 11.3 & 11.4 Other remedial measures Other remedial measures Section 11.3 Robust regression Least absolute residuals (LAR)-minimum L 1 -norma regression Section 11.4 Nonparametric regression Lowess Method Regression Trees hsuhl (NUK) LR Chap / 48
Remedial Measures for Multiple Linear Regression Models
Remedial Measures for Multiple Linear Regression Models Yang Feng http://www.stat.columbia.edu/~yangfeng Yang Feng (Columbia University) Remedial Measures for Multiple Linear Regression Models 1 / 25 Outline
More informationChapter 13 Introduction to Nonlinear Regression( 非線性迴歸 )
Chapter 13 Introduction to Nonlinear Regression( 非線性迴歸 ) and Neural Networks( 類神經網路 ) 許湘伶 Applied Linear Regression Models (Kutner, Nachtsheim, Neter, Li) hsuhl (NUK) LR Chap 10 1 / 35 13 Examples of nonlinear
More informationFinal Review. Yang Feng. Yang Feng (Columbia University) Final Review 1 / 58
Final Review Yang Feng http://www.stat.columbia.edu/~yangfeng Yang Feng (Columbia University) Final Review 1 / 58 Outline 1 Multiple Linear Regression (Estimation, Inference) 2 Special Topics for Multiple
More informationStat 5100 Handout #26: Variations on OLS Linear Regression (Ch. 11, 13)
Stat 5100 Handout #26: Variations on OLS Linear Regression (Ch. 11, 13) 1. Weighted Least Squares (textbook 11.1) Recall regression model Y = β 0 + β 1 X 1 +... + β p 1 X p 1 + ε in matrix form: (Ch. 5,
More informationChapter 14 Logistic Regression, Poisson Regression, and Generalized Linear Models
Chapter 14 Logistic Regression, Poisson Regression, and Generalized Linear Models 許湘伶 Applied Linear Regression Models (Kutner, Nachtsheim, Neter, Li) hsuhl (NUK) LR Chap 10 1 / 29 14.1 Regression Models
More informationLinear Regression. Applied Linear Regression Models (Kutner, Nachtsheim, Neter, Li) hsuhl (NUK) SDA Regression 1 / 34
Linear Regression 許湘伶 Applied Linear Regression Models (Kutner, Nachtsheim, Neter, Li) hsuhl (NUK) SDA Regression 1 / 34 Regression analysis is a statistical methodology that utilizes the relation between
More informationChapter 10 Building the Regression Model II: Diagnostics
Chapter 10 Building the Regression Model II: Diagnostics 許湘伶 Applied Linear Regression Models (Kutner, Nachtsheim, Neter, Li) hsuhl (NUK) LR Chap 10 1 / 41 10.1 Model Adequacy for a Predictor Variable-Added
More informationChapter 2 Inferences in Regression and Correlation Analysis
Chapter 2 Inferences in Regression and Correlation Analysis 許湘伶 Applied Linear Regression Models (Kutner, Nachtsheim, Neter, Li) hsuhl (NUK) LR Chap 2 1 / 102 Inferences concerning the regression parameters
More informationRegression Models - Introduction
Regression Models - Introduction In regression models there are two types of variables that are studied: A dependent variable, Y, also called response variable. It is modeled as random. An independent
More informationReview: Second Half of Course Stat 704: Data Analysis I, Fall 2014
Review: Second Half of Course Stat 704: Data Analysis I, Fall 2014 Tim Hanson, Ph.D. University of South Carolina T. Hanson (USC) Stat 704: Data Analysis I, Fall 2014 1 / 13 Chapter 8: Polynomials & Interactions
More informationRegression Models - Introduction
Regression Models - Introduction In regression models, two types of variables that are studied: A dependent variable, Y, also called response variable. It is modeled as random. An independent variable,
More informationChapter 1 Linear Regression with One Predictor Variable
Chapter 1 Linear Regression with One Predictor Variable 許湘伶 Applied Linear Regression Models (Kutner, Nachtsheim, Neter, Li) hsuhl (NUK) LR Chap 1 1 / 41 Regression analysis is a statistical methodology
More informationContents. 1 Review of Residuals. 2 Detecting Outliers. 3 Influential Observations. 4 Multicollinearity and its Effects
Contents 1 Review of Residuals 2 Detecting Outliers 3 Influential Observations 4 Multicollinearity and its Effects W. Zhou (Colorado State University) STAT 540 July 6th, 2015 1 / 32 Model Diagnostics:
More informationWeighted Least Squares
Weighted Least Squares ST 430/514 Recall the linear regression equation E(Y ) = β 0 + β 1 x 1 + β 2 x 2 + + β k x k We have estimated the parameters β 0, β 1, β 2,..., β k by minimizing the sum of squared
More informationLecture 14 Simple Linear Regression
Lecture 4 Simple Linear Regression Ordinary Least Squares (OLS) Consider the following simple linear regression model where, for each unit i, Y i is the dependent variable (response). X i is the independent
More informationREGRESSION DIAGNOSTICS AND REMEDIAL MEASURES
REGRESSION DIAGNOSTICS AND REMEDIAL MEASURES Lalmohan Bhar I.A.S.R.I., Library Avenue, Pusa, New Delhi 110 01 lmbhar@iasri.res.in 1. Introduction Regression analysis is a statistical methodology that utilizes
More informationFormal Statement of Simple Linear Regression Model
Formal Statement of Simple Linear Regression Model Y i = β 0 + β 1 X i + ɛ i Y i value of the response variable in the i th trial β 0 and β 1 are parameters X i is a known constant, the value of the predictor
More informationOutline. Review regression diagnostics Remedial measures Weighted regression Ridge regression Robust regression Bootstrapping
Topic 19: Remedies Outline Review regression diagnostics Remedial measures Weighted regression Ridge regression Robust regression Bootstrapping Regression Diagnostics Summary Check normality of the residuals
More informationOutline. Remedial Measures) Extra Sums of Squares Standardized Version of the Multiple Regression Model
Outline 1 Multiple Linear Regression (Estimation, Inference, Diagnostics and Remedial Measures) 2 Special Topics for Multiple Regression Extra Sums of Squares Standardized Version of the Multiple Regression
More informationMath 423/533: The Main Theoretical Topics
Math 423/533: The Main Theoretical Topics Notation sample size n, data index i number of predictors, p (p = 2 for simple linear regression) y i : response for individual i x i = (x i1,..., x ip ) (1 p)
More informationDiagnostics and Remedial Measures
Diagnostics and Remedial Measures Yang Feng http://www.stat.columbia.edu/~yangfeng Yang Feng (Columbia University) Diagnostics and Remedial Measures 1 / 72 Remedial Measures How do we know that the regression
More informationSimple and Multiple Linear Regression
Sta. 113 Chapter 12 and 13 of Devore March 12, 2010 Table of contents 1 Simple Linear Regression 2 Model Simple Linear Regression A simple linear regression model is given by Y = β 0 + β 1 x + ɛ where
More informationSTAT5044: Regression and Anova. Inyoung Kim
STAT5044: Regression and Anova Inyoung Kim 2 / 47 Outline 1 Regression 2 Simple Linear regression 3 Basic concepts in regression 4 How to estimate unknown parameters 5 Properties of Least Squares Estimators:
More informationWeighted Least Squares
Weighted Least Squares The standard linear model assumes that Var(ε i ) = σ 2 for i = 1,..., n. As we have seen, however, there are instances where Var(Y X = x i ) = Var(ε i ) = σ2 w i. Here w 1,..., w
More informationQuantitative Methods I: Regression diagnostics
Quantitative Methods I: Regression University College Dublin 10 December 2014 1 Assumptions and errors 2 3 4 Outline Assumptions and errors 1 Assumptions and errors 2 3 4 Assumptions: specification Linear
More informationISyE 691 Data mining and analytics
ISyE 691 Data mining and analytics Regression Instructor: Prof. Kaibo Liu Department of Industrial and Systems Engineering UW-Madison Email: kliu8@wisc.edu Office: Room 3017 (Mechanical Engineering Building)
More informationRegression Analysis for Data Containing Outliers and High Leverage Points
Alabama Journal of Mathematics 39 (2015) ISSN 2373-0404 Regression Analysis for Data Containing Outliers and High Leverage Points Asim Kumer Dey Department of Mathematics Lamar University Md. Amir Hossain
More informationApplied Regression Analysis
Applied Regression Analysis Chapter 3 Multiple Linear Regression Hongcheng Li April, 6, 2013 Recall simple linear regression 1 Recall simple linear regression 2 Parameter Estimation 3 Interpretations of
More informationMulticollinearity and A Ridge Parameter Estimation Approach
Journal of Modern Applied Statistical Methods Volume 15 Issue Article 5 11-1-016 Multicollinearity and A Ridge Parameter Estimation Approach Ghadban Khalaf King Khalid University, albadran50@yahoo.com
More informationOrdinary Least Squares Regression
Ordinary Least Squares Regression Goals for this unit More on notation and terminology OLS scalar versus matrix derivation Some Preliminaries In this class we will be learning to analyze Cross Section
More informationPAijpam.eu M ESTIMATION, S ESTIMATION, AND MM ESTIMATION IN ROBUST REGRESSION
International Journal of Pure and Applied Mathematics Volume 91 No. 3 2014, 349-360 ISSN: 1311-8080 (printed version); ISSN: 1314-3395 (on-line version) url: http://www.ijpam.eu doi: http://dx.doi.org/10.12732/ijpam.v91i3.7
More informationMS&E 226: Small Data. Lecture 11: Maximum likelihood (v2) Ramesh Johari
MS&E 226: Small Data Lecture 11: Maximum likelihood (v2) Ramesh Johari ramesh.johari@stanford.edu 1 / 18 The likelihood function 2 / 18 Estimating the parameter This lecture develops the methodology behind
More informationData Mining Stat 588
Data Mining Stat 588 Lecture 02: Linear Methods for Regression Department of Statistics & Biostatistics Rutgers University September 13 2011 Regression Problem Quantitative generic output variable Y. Generic
More informationLecture 18 Miscellaneous Topics in Multiple Regression
Lecture 18 Miscellaneous Topics in Multiple Regression STAT 512 Spring 2011 Background Reading KNNL: 8.1-8.5,10.1, 11, 12 18-1 Topic Overview Polynomial Models (8.1) Interaction Models (8.2) Qualitative
More informationWeighted Least Squares
Weighted Least Squares The standard linear model assumes that Var(ε i ) = σ 2 for i = 1,..., n. As we have seen, however, there are instances where Var(Y X = x i ) = Var(ε i ) = σ2 w i. Here w 1,..., w
More informationLinear regression methods
Linear regression methods Most of our intuition about statistical methods stem from linear regression. For observations i = 1,..., n, the model is Y i = p X ij β j + ε i, j=1 where Y i is the response
More informationChapter 2 Multiple Regression I (Part 1)
Chapter 2 Multiple Regression I (Part 1) 1 Regression several predictor variables The response Y depends on several predictor variables X 1,, X p response {}}{ Y predictor variables {}}{ X 1, X 2,, X p
More informationChapter 1: Linear Regression with One Predictor Variable also known as: Simple Linear Regression Bivariate Linear Regression
BSTT523: Kutner et al., Chapter 1 1 Chapter 1: Linear Regression with One Predictor Variable also known as: Simple Linear Regression Bivariate Linear Regression Introduction: Functional relation between
More informationStatistics 203: Introduction to Regression and Analysis of Variance Course review
Statistics 203: Introduction to Regression and Analysis of Variance Course review Jonathan Taylor - p. 1/?? Today Review / overview of what we learned. - p. 2/?? General themes in regression models Specifying
More informationDr. Maddah ENMG 617 EM Statistics 11/28/12. Multiple Regression (3) (Chapter 15, Hines)
Dr. Maddah ENMG 617 EM Statistics 11/28/12 Multiple Regression (3) (Chapter 15, Hines) Problems in multiple regression: Multicollinearity This arises when the independent variables x 1, x 2,, x k, are
More informationStatistical View of Least Squares
Basic Ideas Some Examples Least Squares May 22, 2007 Basic Ideas Simple Linear Regression Basic Ideas Some Examples Least Squares Suppose we have two variables x and y Basic Ideas Simple Linear Regression
More informationStatistics and Econometrics I
Statistics and Econometrics I Point Estimation Shiu-Sheng Chen Department of Economics National Taiwan University September 13, 2016 Shiu-Sheng Chen (NTU Econ) Statistics and Econometrics I September 13,
More informationMultiple Linear Regression
Multiple Linear Regression University of California, San Diego Instructor: Ery Arias-Castro http://math.ucsd.edu/~eariasca/teaching.html 1 / 42 Passenger car mileage Consider the carmpg dataset taken from
More informationSimple Regression Model Setup Estimation Inference Prediction. Model Diagnostic. Multiple Regression. Model Setup and Estimation.
Statistical Computation Math 475 Jimin Ding Department of Mathematics Washington University in St. Louis www.math.wustl.edu/ jmding/math475/index.html October 10, 2013 Ridge Part IV October 10, 2013 1
More informationVariable Selection and Model Building
LINEAR REGRESSION ANALYSIS MODULE XIII Lecture - 37 Variable Selection and Model Building Dr. Shalabh Department of Mathematics and Statistics Indian Institute of Technology Kanpur The complete regression
More informationEcon 510 B. Brown Spring 2014 Final Exam Answers
Econ 510 B. Brown Spring 2014 Final Exam Answers Answer five of the following questions. You must answer question 7. The question are weighted equally. You have 2.5 hours. You may use a calculator. Brevity
More informationThe Model Building Process Part I: Checking Model Assumptions Best Practice
The Model Building Process Part I: Checking Model Assumptions Best Practice Authored by: Sarah Burke, PhD 31 July 2017 The goal of the STAT T&E COE is to assist in developing rigorous, defensible test
More informationHigh-dimensional regression modeling
High-dimensional regression modeling David Causeur Department of Statistics and Computer Science Agrocampus Ouest IRMAR CNRS UMR 6625 http://www.agrocampus-ouest.fr/math/causeur/ Course objectives Making
More informationInference in Regression Analysis
Inference in Regression Analysis Dr. Frank Wood Frank Wood, fwood@stat.columbia.edu Linear Regression Models Lecture 4, Slide 1 Today: Normal Error Regression Model Y i = β 0 + β 1 X i + ǫ i Y i value
More informationRegression Diagnostics for Survey Data
Regression Diagnostics for Survey Data Richard Valliant Joint Program in Survey Methodology, University of Maryland and University of Michigan USA Jianzhu Li (Westat), Dan Liao (JPSM) 1 Introduction Topics
More informationLecture 24: Weighted and Generalized Least Squares
Lecture 24: Weighted and Generalized Least Squares 1 Weighted Least Squares When we use ordinary least squares to estimate linear regression, we minimize the mean squared error: MSE(b) = 1 n (Y i X i β)
More informationLecture 3: Linear Models. Bruce Walsh lecture notes Uppsala EQG course version 28 Jan 2012
Lecture 3: Linear Models Bruce Walsh lecture notes Uppsala EQG course version 28 Jan 2012 1 Quick Review of the Major Points The general linear model can be written as y = X! + e y = vector of observed
More informationStatistics 203: Introduction to Regression and Analysis of Variance Penalized models
Statistics 203: Introduction to Regression and Analysis of Variance Penalized models Jonathan Taylor - p. 1/15 Today s class Bias-Variance tradeoff. Penalized regression. Cross-validation. - p. 2/15 Bias-variance
More informationBusiness Statistics. Tommaso Proietti. Linear Regression. DEF - Università di Roma 'Tor Vergata'
Business Statistics Tommaso Proietti DEF - Università di Roma 'Tor Vergata' Linear Regression Specication Let Y be a univariate quantitative response variable. We model Y as follows: Y = f(x) + ε where
More informationOn dealing with spatially correlated residuals in remote sensing and GIS
On dealing with spatially correlated residuals in remote sensing and GIS Nicholas A. S. Hamm 1, Peter M. Atkinson and Edward J. Milton 3 School of Geography University of Southampton Southampton SO17 3AT
More informationSimple Linear Regression
Simple Linear Regression In simple linear regression we are concerned about the relationship between two variables, X and Y. There are two components to such a relationship. 1. The strength of the relationship.
More informationDay 4: Shrinkage Estimators
Day 4: Shrinkage Estimators Kenneth Benoit Data Mining and Statistical Learning March 9, 2015 n versus p (aka k) Classical regression framework: n > p. Without this inequality, the OLS coefficients have
More informationLecture 10 Multiple Linear Regression
Lecture 10 Multiple Linear Regression STAT 512 Spring 2011 Background Reading KNNL: 6.1-6.5 10-1 Topic Overview Multiple Linear Regression Model 10-2 Data for Multiple Regression Y i is the response variable
More informationSimple Linear Regression Analysis
LINEAR REGRESSION ANALYSIS MODULE II Lecture - 6 Simple Linear Regression Analysis Dr. Shalabh Department of Mathematics and Statistics Indian Institute of Technology Kanpur Prediction of values of study
More informationCh 2: Simple Linear Regression
Ch 2: Simple Linear Regression 1. Simple Linear Regression Model A simple regression model with a single regressor x is y = β 0 + β 1 x + ɛ, where we assume that the error ɛ is independent random component
More informationRegression. Oscar García
Regression Oscar García Regression methods are fundamental in Forest Mensuration For a more concise and general presentation, we shall first review some matrix concepts 1 Matrices An order n m matrix is
More informationOn Modifications to Linking Variance Estimators in the Fay-Herriot Model that Induce Robustness
Statistics and Applications {ISSN 2452-7395 (online)} Volume 16 No. 1, 2018 (New Series), pp 289-303 On Modifications to Linking Variance Estimators in the Fay-Herriot Model that Induce Robustness Snigdhansu
More informationRemedial Measures, Brown-Forsythe test, F test
Remedial Measures, Brown-Forsythe test, F test Dr. Frank Wood Frank Wood, fwood@stat.columbia.edu Linear Regression Models Lecture 7, Slide 1 Remedial Measures How do we know that the regression function
More informationInstructions: Closed book, notes, and no electronic devices. Points (out of 200) in parentheses
ISQS 5349 Final Spring 2011 Instructions: Closed book, notes, and no electronic devices. Points (out of 200) in parentheses 1. (10) What is the definition of a regression model that we have used throughout
More informationSemester 2, 2015/2016
ECN 3202 APPLIED ECONOMETRICS 5. HETEROSKEDASTICITY Mr. Sydney Armstrong Lecturer 1 The University of Guyana 1 Semester 2, 2015/2016 WHAT IS HETEROSKEDASTICITY? The multiple linear regression model can
More informationRidge regression. Patrick Breheny. February 8. Penalized regression Ridge regression Bayesian interpretation
Patrick Breheny February 8 Patrick Breheny High-Dimensional Data Analysis (BIOS 7600) 1/27 Introduction Basic idea Standardization Large-scale testing is, of course, a big area and we could keep talking
More informationSTAT 704 Sections IRLS and Bootstrap
STAT 704 Sections 11.4-11.5. IRLS and John Grego Department of Statistics, University of South Carolina Stat 704: Data Analysis I 1 / 14 LOWESS IRLS LOWESS LOWESS (LOcally WEighted Scatterplot Smoothing)
More informationLecture 16 Solving GLMs via IRWLS
Lecture 16 Solving GLMs via IRWLS 09 November 2015 Taylor B. Arnold Yale Statistics STAT 312/612 Notes problem set 5 posted; due next class problem set 6, November 18th Goals for today fixed PCA example
More informationRobust Variable Selection Methods for Grouped Data. Kristin Lee Seamon Lilly
Robust Variable Selection Methods for Grouped Data by Kristin Lee Seamon Lilly A dissertation submitted to the Graduate Faculty of Auburn University in partial fulfillment of the requirements for the Degree
More informationMulticollinearity occurs when two or more predictors in the model are correlated and provide redundant information about the response.
Multicollinearity Read Section 7.5 in textbook. Multicollinearity occurs when two or more predictors in the model are correlated and provide redundant information about the response. Example of multicollinear
More informationMultiple Regression Analysis. Part III. Multiple Regression Analysis
Part III Multiple Regression Analysis As of Sep 26, 2017 1 Multiple Regression Analysis Estimation Matrix form Goodness-of-Fit R-square Adjusted R-square Expected values of the OLS estimators Irrelevant
More informationLecture 2: Linear Models. Bruce Walsh lecture notes Seattle SISG -Mixed Model Course version 23 June 2011
Lecture 2: Linear Models Bruce Walsh lecture notes Seattle SISG -Mixed Model Course version 23 June 2011 1 Quick Review of the Major Points The general linear model can be written as y = X! + e y = vector
More informationEconometrics Multiple Regression Analysis: Heteroskedasticity
Econometrics Multiple Regression Analysis: João Valle e Azevedo Faculdade de Economia Universidade Nova de Lisboa Spring Semester João Valle e Azevedo (FEUNL) Econometrics Lisbon, April 2011 1 / 19 Properties
More informationAvailable online at (Elixir International Journal) Statistics. Elixir Statistics 49 (2012)
10108 Available online at www.elixirpublishers.com (Elixir International Journal) Statistics Elixir Statistics 49 (2012) 10108-10112 The detention and correction of multicollinearity effects in a multiple
More informationEstimating σ 2. We can do simple prediction of Y and estimation of the mean of Y at any value of X.
Estimating σ 2 We can do simple prediction of Y and estimation of the mean of Y at any value of X. To perform inferences about our regression line, we must estimate σ 2, the variance of the error term.
More informationEffects of Outliers and Multicollinearity on Some Estimators of Linear Regression Model
204 Effects of Outliers and Multicollinearity on Some Estimators of Linear Regression Model S. A. Ibrahim 1 ; W. B. Yahya 2 1 Department of Physical Sciences, Al-Hikmah University, Ilorin, Nigeria. e-mail:
More informationThe Model Building Process Part I: Checking Model Assumptions Best Practice (Version 1.1)
The Model Building Process Part I: Checking Model Assumptions Best Practice (Version 1.1) Authored by: Sarah Burke, PhD Version 1: 31 July 2017 Version 1.1: 24 October 2017 The goal of the STAT T&E COE
More informationIndian Statistical Institute
Indian Statistical Institute Introductory Computer programming Robust Regression methods with high breakdown point Author: Roll No: MD1701 February 24, 2018 Contents 1 Introduction 2 2 Criteria for evaluating
More informationImproved Liu Estimators for the Poisson Regression Model
www.ccsenet.org/isp International Journal of Statistics and Probability Vol., No. ; May 202 Improved Liu Estimators for the Poisson Regression Model Kristofer Mansson B. M. Golam Kibria Corresponding author
More informationHeteroskedasticity. Part VII. Heteroskedasticity
Part VII Heteroskedasticity As of Oct 15, 2015 1 Heteroskedasticity Consequences Heteroskedasticity-robust inference Testing for Heteroskedasticity Weighted Least Squares (WLS) Feasible generalized Least
More informationMath 5305 Notes. Diagnostics and Remedial Measures. Jesse Crawford. Department of Mathematics Tarleton State University
Math 5305 Notes Diagnostics and Remedial Measures Jesse Crawford Department of Mathematics Tarleton State University (Tarleton State University) Diagnostics and Remedial Measures 1 / 44 Model Assumptions
More informationRidge Regression. Summary. Sample StatFolio: ridge reg.sgp. STATGRAPHICS Rev. 10/1/2014
Ridge Regression Summary... 1 Data Input... 4 Analysis Summary... 5 Analysis Options... 6 Ridge Trace... 7 Regression Coefficients... 8 Standardized Regression Coefficients... 9 Observed versus Predicted...
More informationAdvanced Econometrics I
Lecture Notes Autumn 2010 Dr. Getinet Haile, University of Mannheim 1. Introduction Introduction & CLRM, Autumn Term 2010 1 What is econometrics? Econometrics = economic statistics economic theory mathematics
More informationAnalysis of Cross-Sectional Data
Analysis of Cross-Sectional Data Kevin Sheppard http://www.kevinsheppard.com Oxford MFE This version: November 8, 2017 November 13 14, 2017 Outline Econometric models Specification that can be analyzed
More informationLeast Squares Estimation-Finite-Sample Properties
Least Squares Estimation-Finite-Sample Properties Ping Yu School of Economics and Finance The University of Hong Kong Ping Yu (HKU) Finite-Sample 1 / 29 Terminology and Assumptions 1 Terminology and Assumptions
More informationThe general linear regression with k explanatory variables is just an extension of the simple regression as follows
3. Multiple Regression Analysis The general linear regression with k explanatory variables is just an extension of the simple regression as follows (1) y i = β 0 + β 1 x i1 + + β k x ik + u i. Because
More informationMS&E 226. In-Class Midterm Examination Solutions Small Data October 20, 2015
MS&E 226 In-Class Midterm Examination Solutions Small Data October 20, 2015 PROBLEM 1. Alice uses ordinary least squares to fit a linear regression model on a dataset containing outcome data Y and covariates
More informationChapter 4 Experiments with Blocking Factors
Chapter 4 Experiments with Blocking Factors 許湘伶 Design and Analysis of Experiments (Douglas C. Montgomery) hsuhl (NUK) DAE Chap. 4 1 / 54 The Randomized Complete Block Design (RCBD; 隨機化完全集區設計 ) 1 Variability
More informationMa 3/103: Lecture 24 Linear Regression I: Estimation
Ma 3/103: Lecture 24 Linear Regression I: Estimation March 3, 2017 KC Border Linear Regression I March 3, 2017 1 / 32 Regression analysis Regression analysis Estimate and test E(Y X) = f (X). f is the
More informationStatistics 910, #5 1. Regression Methods
Statistics 910, #5 1 Overview Regression Methods 1. Idea: effects of dependence 2. Examples of estimation (in R) 3. Review of regression 4. Comparisons and relative efficiencies Idea Decomposition Well-known
More informationAdvanced Quantitative Methods: ordinary least squares
Advanced Quantitative Methods: Ordinary Least Squares University College Dublin 31 January 2012 1 2 3 4 5 Terminology y is the dependent variable referred to also (by Greene) as a regressand X are the
More informationLinear Regression. September 27, Chapter 3. Chapter 3 September 27, / 77
Linear Regression Chapter 3 September 27, 2016 Chapter 3 September 27, 2016 1 / 77 1 3.1. Simple linear regression 2 3.2 Multiple linear regression 3 3.3. The least squares estimation 4 3.4. The statistical
More informationA Short Introduction to Curve Fitting and Regression by Brad Morantz
A Short Introduction to Curve Fitting and Regression by Brad Morantz bradscientist@machine-cognition.com Overview What can regression do for me? Example Model building Error Metrics OLS Regression Robust
More informationSTAT 135 Lab 13 (Review) Linear Regression, Multivariate Random Variables, Prediction, Logistic Regression and the δ-method.
STAT 135 Lab 13 (Review) Linear Regression, Multivariate Random Variables, Prediction, Logistic Regression and the δ-method. Rebecca Barter May 5, 2015 Linear Regression Review Linear Regression Review
More informationIntroduction to Linear regression analysis. Part 2. Model comparisons
Introduction to Linear regression analysis Part Model comparisons 1 ANOVA for regression Total variation in Y SS Total = Variation explained by regression with X SS Regression + Residual variation SS Residual
More informationMath 3330: Solution to midterm Exam
Math 3330: Solution to midterm Exam Question 1: (14 marks) Suppose the regression model is y i = β 0 + β 1 x i + ε i, i = 1,, n, where ε i are iid Normal distribution N(0, σ 2 ). a. (2 marks) Compute the
More informationIEOR 165 Lecture 7 1 Bias-Variance Tradeoff
IEOR 165 Lecture 7 Bias-Variance Tradeoff 1 Bias-Variance Tradeoff Consider the case of parametric regression with β R, and suppose we would like to analyze the error of the estimate ˆβ in comparison to
More informationLeast Absolute Value vs. Least Squares Estimation and Inference Procedures in Regression Models with Asymmetric Error Distributions
Journal of Modern Applied Statistical Methods Volume 8 Issue 1 Article 13 5-1-2009 Least Absolute Value vs. Least Squares Estimation and Inference Procedures in Regression Models with Asymmetric Error
More informationIntroductory Econometrics
Based on the textbook by Wooldridge: : A Modern Approach Robert M. Kunst robert.kunst@univie.ac.at University of Vienna and Institute for Advanced Studies Vienna November 23, 2013 Outline Introduction
More informationNeed for Several Predictor Variables
Multiple regression One of the most widely used tools in statistical analysis Matrix expressions for multiple regression are the same as for simple linear regression Need for Several Predictor Variables
More information