Chapter 11 Building the Regression Model II:

Size: px
Start display at page:

Download "Chapter 11 Building the Regression Model II:"

Transcription

1 Chapter 11 Building the Regression Model II: Remedial Measures( 補救措施 ) 許湘伶 Applied Linear Regression Models (Kutner, Nachtsheim, Neter, Li) hsuhl (NUK) LR Chap 11 1 / 48

2 11.1 WLS remedial measures may need to be taken: a regression model is not appropriate several cases are very influential previous: transformations to linearize the regression relation the error distributions more nearly normal make the variances of the error terms more nearly equal hsuhl (NUK) LR Chap 11 2 / 48

3 11.1 WLS remedial measures In this chapter- remedial measure to deal with: unequal error variances a high degree of multicollinearity influential observations two methods for nonparametric regression: lowess regression trees bootstrapping: for evaluating the precision of the complex estimators hsuhl (NUK) LR Chap 11 3 / 48

4 11.1 Unequal Error Variances Remedial Measures-Weighted Least Squares Weighted Least Squares (WLS) Chap. 3,6: transformation of Y - reducing or eliminating unequal variances difficulty: may create an inappropriate regression relationship weighted least squares( 加權最小平方法 ): when an appropriate regression relationship has been found but the variances of the error terms are unequal hsuhl (NUK) LR Chap 11 4 / 48

5 11.1 Unequal Error Variances Remedial Measures-Weighted Least Squares Weighted Least Squares (WLS) (cont.) Y i = β 0 + β 1 X i1 + + β p 1 X i,p 1 + ε i, i = 1,..., n parameters: β 0,..., β p 1 known constants: X i1,..., X i,p 1 ε i indep. N(0, σ 2 i ) σ 2 {ε} n n = σ σ σn 2 b = (X X) 1 X Y unbias, consistent but don t have minimum variance hsuhl (NUK) LR Chap 11 5 / 48

6 11.1 Unequal Error Variances Remedial Measures-Weighted Least Squares Weighted Least Squares (WLS) (cont.) Consider: Observations with small variances provide more reliable information about the regression function than those with large variances. Error Variances Know (w i = 1 )(unrealistic 不切實際的 ) σi 2 Error Variances Know up to Proportionality Constant (w i = k 1 ) σi 2 Error Variances Unknown: estimation of variance function or standard deviation function (w i = 1 (ŝ i, w ) 2 i = 1ˆν i ) hsuhl (NUK) LR Chap 11 6 / 48

7 11.1 Unequal Error Variances Remedial Measures-Weighted Least Squares WLS-error variances known Methods of maximum likelihood to obtain estimators: indep. ε i N(0, σi 2 ) and w i = 1 σi 2 L(β) = = [ n i=1 n i=1 ( wi 2π 1 (2πσi 2 ] ) 1/2 b= arg max L(β) β [ exp 1 )1/2 2σi 2 [ exp b = arg min β Q w (Q w = 1 2 (Y i β 0 β 1 X i1 β p 1 X i,p 1 ) 2 ] ] n w i (Y i β 0 β 1 X i1 β p 1 X i,p 1 ) 2 i=1 n w i (Y i β 0 β 1 X i1 β p 1 X i,p 1 ) 2 ) i=1 Called weighted least squares criterion (min Q w ) w i = 1 OLS hsuhl (NUK) LR Chap 11 7 / 48

8 11.1 Unequal Error Variances Remedial Measures-Weighted Least Squares WLS-error variances known (cont.) w i = 1 : reflects the amount of information contained in the σi 2 observation Y i vary i large w i less W = n n w w w n The normal equations: (X W X)b w = X W Y The weighted least squares (MLE) for β: σ 2 {b w } p p b w = (X W X) 1 X W Y = (X W X) 1 ( σ 2 {Y } = W 1 ) hsuhl (NUK) LR Chap 11 8 / 48

9 11.1 Unequal Error Variances Remedial Measures-Weighted Least Squares WLS-error variances known (cont.) b w : unbiased, consistent, have minimum variance among unbiased linear estimators When the weighted are known, b w generally exhibits less variability than b. hsuhl (NUK) LR Chap 11 9 / 48

10 11.1 Unequal Error Variances Remedial Measures-Weighted Least Squares WLS-error variances known up to proportionality constant w i = k 1 σ 2 i σ 2 {b w } p p k: unknwon s 2 {b w } p p where MSE w = b w = (X W X) 1 X W Y = k(x W X) 1 = MSE w (X W X) 1 wi (Y i Ŷ i ) 2 n p = (unaffected) wi e 2 i n p MSE w : an estimator of the proportionality constant k hsuhl (NUK) LR Chap / 48

11 11.1 Unequal Error Variances Remedial Measures-Weighted Least Squares WLS-error variances unknown Estimation of Variance Function or Standard Deviation Function One rarely has knowledge of the variances σ 2 i. estimate of the variances σ 2 i = E{ε 2 i } (E{ε i }) 2 = E{ε 2 i } Estimator of σi 2 : the squared residual ei 2 Estimator of σ i = σi 2 : the absolute residual e i hsuhl (NUK) LR Chap / 48

12 11.1 Unequal Error Variances Remedial Measures-Weighted Least Squares WLS-error variances unknown (cont.) the estimation process: 1 Fit the regression model by OLS and analyze e i 2 Estimate the variance function or the standard deviation function by regressing ei 2 or e i on the appropriate predictor(s). 3 ŝ i for standard deviation function, ˆν i for the variance function: to obtain w i 4 obtain b w by using w i (Sometimes, iterated for several times to reach stabilize iteratively reweighted lease squares) hsuhl (NUK) LR Chap / 48

13 11.1 Unequal Error Variances Remedial Measures-Weighted Least Squares WLS-error variances unknown (cont.) Reference: figure from Faraway s Linear Models with R (2005, p. 59) hsuhl (NUK) LR Chap / 48

14 11.1 Unequal Error Variances Remedial Measures-Weighted Least Squares WLS-error variances unknown (cont.) Some possible variance and standard deviation functions: 1 plot e i vs. X 1 : a megaphone( 大聲公 ) shape e i regresses on X 1 2 plot e i vs. Ŷ : a megaphone( 大聲公 ) shape e i regresses on Ŷ 3 plot e 2 i vs. X 3 : upward tendency e 2 i regresses on X 3 4 plot e i vs. X 2 : increase rapidly with X 2 and increases more slowly e i regresses on X 2, X 2 2 w i = 1 (ŝ i ) 2 : ŝ i-fitted value from standard function w i = 1ˆν i : ˆν i -fitted value from variance function the weight matrix W b w = (X W X) 1 X W Y hsuhl (NUK) LR Chap / 48

15 11.1 Unequal Error Variances Remedial Measures-Weighted Least Squares WLS-error variances unknown (cont.) Using of Replicated or Near Replicates Replicate observations are made at each combination of levels of Xs It the number of replications (or near replications) is large, w i may be obtained directly from the sample variances of Y observations at each combination of levels on X variables each case in a replicate group receives the same weight with this method Confidence interval: (similar but approximate) b k ± t(1 α/2; n p)s{b k } (6.50) hsuhl (NUK) LR Chap / 48

16 11.1 Unequal Error Variances Remedial Measures-Weighted Least Squares WLS-error variances unknown (cont.) Using of Ordinary Least Squares with Unequal Error Variances: b = (X X) 1 X Y σ 2 {b} = (X X) 1 (X σ 2 {ε}x)(x X) 1 S 2 {b} = (X X) 1 (X S 0 {ε}x)(x X) 1 where S 0 = diag{e1, 2 e2, 2..., en} 2 hsuhl (NUK) LR Chap / 48

17 11.1 Unequal Error Variances Remedial Measures-Weighted Least Squares blood pressure example 54 subjects: women hsuhl (NUK) LR Chap / 48

18 11.1 Unequal Error Variances Remedial Measures-Weighted Least Squares blood pressure example (cont.) linear regression function by unweighted least squares: Ŷ = (3.994) ( ) X hsuhl (NUK) LR Chap / 48

19 11.1 Unequal Error Variances Remedial Measures-Weighted Least Squares blood pressure example (cont.) e i regresses on X: ŝ = X ŝ 1 = w 1 = 1/(ŝ 1 ) 2 = WLS: Ŷ = X C.I. for β 1 : α = 0.05; s{b w1 } = ± t(0.975; 52) ( ) β =2.007 hsuhl (NUK) LR Chap / 48

20 11.1 Unequal Error Variances Remedial Measures-Weighted Least Squares blood pressure example (cont.) unequal variances: heteroscedasticity( 異質性 ); equal variances:homoscedasticity( 等變異性 ) R 2 does not have a clear-cut meaning for WLS WLS may be view as OLS of transformed variables: W 1/2 Y Y w Y = Xβ + ε, = W 1/2 Xβ X w β E{ε} = 0, σ 2 {ε} = W 1 + W 1/2 ε ε w E{ε w } = 0; σ 2 {ε w } = I; b w = (X wx w ) 1 X wy w = (X W X) 1 X W Y hsuhl (NUK) LR Chap / 48

21 11.1 Unequal Error Variances Remedial Measures-Weighted Least Squares blood pressure example (cont.) The weighted least squares normal equations: wi Y i = b w0 wi + b w1 wi X i wi X i Y i = b w0 wi X i + b w1 wi X 2 i The weighted least squares estimators b w0 and b w1 in (11.9) are: b w1 = b w0 = wi X i Y i wi X i wi Y i wi wi Xi 2 ( w i X i ) 2 wi wi Y i b w1 wi X i wi hsuhl (NUK) LR Chap / 48

22 11.2 Ridge Regression Ridge Regression( 脊迴歸 ) ridge regression: a method of overcoming serious multicollinearity problem by modifying the method of least squares. allowing biased estimators of the regression functions When an estimator has only a small bias and is substantially more precise than an unbiased estimator, it may well be the preferred estimator since it will have a larger probability of being close to the true parameter value. hsuhl (NUK) LR Chap / 48

23 11.2 Ridge Regression Ridge Regression( 脊迴歸 ) (cont.) Estimator b is unbiased but imprecise Estimator b R is much more precise but has a small bias The probability that b R falls near β is much greater than that for b hsuhl (NUK) LR Chap / 48

24 11.2 Ridge Regression Ridge Regression( 脊迴歸 ) (cont.) A measure of the combined effect of bias and sampling variation: The mean squared error (MSE): E{b R β} 2 = σ 2 {b R } + (E{b R } β) 2 Transformed variables: Y i = Y i ( 1 Yi Ȳ n 1 s Y ), X ik = = β 1X i1 + β 2X i2 + β p 1X i,p 1 + ε i r XX b = r YX ( ) 1 Xik X k n 1 s k hsuhl (NUK) LR Chap / 48

25 11.2 Ridge Regression Ridge Regression( 脊迴歸 ) (cont.) The ridge standardized regression estimators are obtained by: (r XX + ci)b R = r YX b R = (r XX + ci) 1 r YX Biasing constant c 0: reflects the amount of bias c = 0 OLS; c > 0 bias but more stable (less variable) than OLS estimators hsuhl (NUK) LR Chap / 48

26 11.2 Ridge Regression Ridge Regression( 脊迴歸 ) (cont.) Choice of Biasing Constant c: bias component (E{b R } β) of E{b R β} 2 as c while the variance (σ 2 {b R }) component becomes smaller Always a c s.t. b R has a smaller total MSE than b difficulty: the optimum value of c varies from one application to another and is unknown Using Ridge trace( 脊軌跡 ) and (VIF ) k (c) hsuhl (NUK) LR Chap / 48

27 11.2 Ridge Regression Ridge Regression( 脊迴歸 ) (cont.) Ridge trace:a simultaneous plot of the values of the (p 1) b R for different c; 0 c 1 Choose c: the Ridge trace starts to become stable and VIF has become sufficiently small. hsuhl (NUK) LR Chap / 48

28 11.2 Ridge Regression Ridge Regression( 脊迴歸 ) (cont.) The normal equation for the ridge estimators: (1 + c)b1 R + r 12 b2 R + + r 1,p 1 bp 1 R = r Y 1 r 21 b1 R + (1 + c)b2 R + + r 2,p 1 bp 1 R = r Y 2. r p 1,1 b1 R + r p 1,2 b2 R + + (1 + c)bp 1 R = r Y,p 1 VIF for bk R are defined analogously to those for OLS b k : measures how large is the variance of bk R relative to what the variance would be if X s were uncorrelated. hsuhl (NUK) LR Chap / 48

29 11.2 Ridge Regression Ridge Regression( 脊迴歸 ) (cont.) VIF for b R k : the diagonal elements of the following matrix (r XX + ci) 1 r XX (r XX + ci) 1 Ridge regression estimates can be obtained by the method of penalized least squares. Q = n i=1 [ Y i (β1x i1 + + βp 1X i,p 1) ] p c (βj ) 2 sometimes referred to as shrinkage esimators Limitation: ridge regression is that ordinary inference procedures are not applicable and exact distribution properties are not known. (bootstrapping ) j=1 hsuhl (NUK) LR Chap / 48

30 11.2 Ridge Regression Ridge Regression( 脊迴歸 ) (cont.) (a) Table 11.2 (b) Table hsuhl (NUK) LR Chap / 48

31 11.2 Ridge Regression Ridge Regression( 脊迴歸 ) (cont.) c = 0.02 Ŷ = X X X 3 hsuhl (NUK) LR Chap / 48

32 11.2 Ridge Regression Ridge Regression( 脊迴歸 ) (cont.) LASSO: Least absolute shrinkage( 收縮 ) and selection operator( 最小絕對緊縮與選擇算子 ) ˆβ = arg min β n p 1 (Y i i=1 j=0 p 1 X i jβ j ) 2 + λ j=0 β j hsuhl (NUK) LR Chap / 48

33 11.2 Ridge Regression Ridge Regression( 脊迴歸 ) (cont.) hsuhl (NUK) LR Chap / 48

34 11.3 Robust Regression Robust Regression To examine whether an outlying case is the result of a recording error, breakdown( 故障 ) of a measurement instrument, or the like. Erroneous data can be corrected corrected Erroneous data cannot be corrected discarded( 拋棄 ) Many time: impossible to tell for certain whether the observations for an outlying case are erroneous not be discarded If an outlying influential case is not clearly erroneous to examine the adequacy of the model hsuhl (NUK) LR Chap / 48

35 11.3 Robust Regression Robust Regression (cont.) Numerous robust regression procedures have been developed. LAR or LAD regression: Least absolute residuals or least absolute deviations( 最小絕對離差 ) minimum L 1 norm regression Criterion: min L 1 = min β β { n Y i (β 0 + β 1 X i1 + + β p 1 x i,p 1 ) i=1 LAR: less emphasis on outlying observations than does the method of least squares linear programming method: to obtain the LAR estimators LAR method: the residuals will not sum to zero } hsuhl (NUK) LR Chap / 48

36 11.3 Robust Regression Robust Regression (cont.) LMS regression: Least median of squares median { [Y i (β 0 + β 1 X i1 + + β p 1 x i,p 1 )] 2} IRLS robust regression: Iteratively reweighted least squares method hsuhl (NUK) LR Chap / 48

37 11.4 Regression Tree Regression Tree Powerful, conceptually simple method of nonparametric regression 1 the range of the predictor is partitioned into segments 2 within each segment the estimated regression fit is given by the mean of the response in the segment hsuhl (NUK) LR Chap / 48

38 11.4 Regression Tree Regression Tree (cont.) hsuhl (NUK) LR Chap / 48

39 11.4 Regression Tree Regression Tree (cont.) hsuhl (NUK) LR Chap / 48

40 11.4 Regression Tree Regression Tree (cont.) r = 2: the best point is choose to min SSE = SSE(R 21 ) + SSE(R 22 ) SSE(R rj ) = (Y i Ȳ Rjk ) 2 SSE(R rj ): the sum of squared residuals in region R rj hsuhl (NUK) LR Chap / 48

41 11.4 Regression Tree Regression Tree (cont.) hsuhl (NUK) LR Chap / 48

42 11.4 Regression Tree Regression Tree (cont.) Determining the number of Regions R hsuhl (NUK) LR Chap / 48

43 11.5 Bootstrapping Bootstrapping In many nonstandard situations, (ex: nonconstant error variances estimated by iteratively reweighted least squares), standard methods for evaluating the precision may not be available or may only be approximately applicable when the sample size is large Bootstrapping: provide estimates of the precision of sample estimated for these complex cases hsuhl (NUK) LR Chap / 48

44 11.5 Bootstrapping Bootstrapping (cont.) general procedure: obtain b 1 by some procedure and wish to evaluate the precision of b 1 by bootstrap method The bootstrap method calls for( 需要 ) the selection from the observed sample data of a random sample of size n with replacement. the bootstrap sample may contain some duplicate( 複製的 ) data from the original sample and omit some other data in the original sample calculates the estimated regression coefficients from the bootstrap sample b 1 repeated a large number of times The estimated standard deviation of all of b 1 s {b 1} an estimate of the variability of the sampling distribution of b 1 hsuhl (NUK) LR Chap / 48

45 11.5 Bootstrapping Bootstrapping (cont.) Bootstrap Sampling: two basic ways When the regression function being fitted is a good model for the data, the error terms have constant variance, and fixed X sampling is appropriate. e i from the original fitting are regarded as the sample data to be sampled with replacement After a bootstrap sample with n: e 1,..., e n Y i = Ŷ i + e i Y regression on Xs b 1 hsuhl (NUK) LR Chap / 48

46 11.5 Bootstrapping Bootstrapping (cont.) When there is some doubt about the adequacy of the regression function being fitted, the error variance are not constant, and/or random X sampling is appropriate. (X i, Y i ) in the original sample are considered to be the data to be sampled with replacement After a bootstrap sample with n: n pairs: (X 1, Y 1 ),..., (X n, Y n ) Y regression on X s b 1 hsuhl (NUK) LR Chap / 48

47 11.5 Bootstrapping Bootstrapping (cont.) bootstrap samples are adequate One can observe the variability of the bootstrap estimates by s {b 1} as the number of bootstrap samples is increased. Bootstrap Confidence Intervals From the bootstrap distribution of b 1, find the α/2 and 1 α/2 quantiles b 1(α/2) and b 1(1 α/2) Percentiles from b 1 : b 1(α/2) β 1 b 1(1 α/2) reflection method: require a larger number of bootstrap samples 2b 1 b 1(α/2) β 1 2b 1 b 1(α/2) hsuhl (NUK) LR Chap / 48

48 11.3 & 11.4 Other remedial measures Other remedial measures Section 11.3 Robust regression Least absolute residuals (LAR)-minimum L 1 -norma regression Section 11.4 Nonparametric regression Lowess Method Regression Trees hsuhl (NUK) LR Chap / 48

Remedial Measures for Multiple Linear Regression Models

Remedial Measures for Multiple Linear Regression Models Remedial Measures for Multiple Linear Regression Models Yang Feng http://www.stat.columbia.edu/~yangfeng Yang Feng (Columbia University) Remedial Measures for Multiple Linear Regression Models 1 / 25 Outline

More information

Chapter 13 Introduction to Nonlinear Regression( 非線性迴歸 )

Chapter 13 Introduction to Nonlinear Regression( 非線性迴歸 ) Chapter 13 Introduction to Nonlinear Regression( 非線性迴歸 ) and Neural Networks( 類神經網路 ) 許湘伶 Applied Linear Regression Models (Kutner, Nachtsheim, Neter, Li) hsuhl (NUK) LR Chap 10 1 / 35 13 Examples of nonlinear

More information

Final Review. Yang Feng. Yang Feng (Columbia University) Final Review 1 / 58

Final Review. Yang Feng.   Yang Feng (Columbia University) Final Review 1 / 58 Final Review Yang Feng http://www.stat.columbia.edu/~yangfeng Yang Feng (Columbia University) Final Review 1 / 58 Outline 1 Multiple Linear Regression (Estimation, Inference) 2 Special Topics for Multiple

More information

Stat 5100 Handout #26: Variations on OLS Linear Regression (Ch. 11, 13)

Stat 5100 Handout #26: Variations on OLS Linear Regression (Ch. 11, 13) Stat 5100 Handout #26: Variations on OLS Linear Regression (Ch. 11, 13) 1. Weighted Least Squares (textbook 11.1) Recall regression model Y = β 0 + β 1 X 1 +... + β p 1 X p 1 + ε in matrix form: (Ch. 5,

More information

Chapter 14 Logistic Regression, Poisson Regression, and Generalized Linear Models

Chapter 14 Logistic Regression, Poisson Regression, and Generalized Linear Models Chapter 14 Logistic Regression, Poisson Regression, and Generalized Linear Models 許湘伶 Applied Linear Regression Models (Kutner, Nachtsheim, Neter, Li) hsuhl (NUK) LR Chap 10 1 / 29 14.1 Regression Models

More information

Linear Regression. Applied Linear Regression Models (Kutner, Nachtsheim, Neter, Li) hsuhl (NUK) SDA Regression 1 / 34

Linear Regression. Applied Linear Regression Models (Kutner, Nachtsheim, Neter, Li) hsuhl (NUK) SDA Regression 1 / 34 Linear Regression 許湘伶 Applied Linear Regression Models (Kutner, Nachtsheim, Neter, Li) hsuhl (NUK) SDA Regression 1 / 34 Regression analysis is a statistical methodology that utilizes the relation between

More information

Chapter 10 Building the Regression Model II: Diagnostics

Chapter 10 Building the Regression Model II: Diagnostics Chapter 10 Building the Regression Model II: Diagnostics 許湘伶 Applied Linear Regression Models (Kutner, Nachtsheim, Neter, Li) hsuhl (NUK) LR Chap 10 1 / 41 10.1 Model Adequacy for a Predictor Variable-Added

More information

Chapter 2 Inferences in Regression and Correlation Analysis

Chapter 2 Inferences in Regression and Correlation Analysis Chapter 2 Inferences in Regression and Correlation Analysis 許湘伶 Applied Linear Regression Models (Kutner, Nachtsheim, Neter, Li) hsuhl (NUK) LR Chap 2 1 / 102 Inferences concerning the regression parameters

More information

Regression Models - Introduction

Regression Models - Introduction Regression Models - Introduction In regression models there are two types of variables that are studied: A dependent variable, Y, also called response variable. It is modeled as random. An independent

More information

Review: Second Half of Course Stat 704: Data Analysis I, Fall 2014

Review: Second Half of Course Stat 704: Data Analysis I, Fall 2014 Review: Second Half of Course Stat 704: Data Analysis I, Fall 2014 Tim Hanson, Ph.D. University of South Carolina T. Hanson (USC) Stat 704: Data Analysis I, Fall 2014 1 / 13 Chapter 8: Polynomials & Interactions

More information

Regression Models - Introduction

Regression Models - Introduction Regression Models - Introduction In regression models, two types of variables that are studied: A dependent variable, Y, also called response variable. It is modeled as random. An independent variable,

More information

Chapter 1 Linear Regression with One Predictor Variable

Chapter 1 Linear Regression with One Predictor Variable Chapter 1 Linear Regression with One Predictor Variable 許湘伶 Applied Linear Regression Models (Kutner, Nachtsheim, Neter, Li) hsuhl (NUK) LR Chap 1 1 / 41 Regression analysis is a statistical methodology

More information

Contents. 1 Review of Residuals. 2 Detecting Outliers. 3 Influential Observations. 4 Multicollinearity and its Effects

Contents. 1 Review of Residuals. 2 Detecting Outliers. 3 Influential Observations. 4 Multicollinearity and its Effects Contents 1 Review of Residuals 2 Detecting Outliers 3 Influential Observations 4 Multicollinearity and its Effects W. Zhou (Colorado State University) STAT 540 July 6th, 2015 1 / 32 Model Diagnostics:

More information

Weighted Least Squares

Weighted Least Squares Weighted Least Squares ST 430/514 Recall the linear regression equation E(Y ) = β 0 + β 1 x 1 + β 2 x 2 + + β k x k We have estimated the parameters β 0, β 1, β 2,..., β k by minimizing the sum of squared

More information

Lecture 14 Simple Linear Regression

Lecture 14 Simple Linear Regression Lecture 4 Simple Linear Regression Ordinary Least Squares (OLS) Consider the following simple linear regression model where, for each unit i, Y i is the dependent variable (response). X i is the independent

More information

REGRESSION DIAGNOSTICS AND REMEDIAL MEASURES

REGRESSION DIAGNOSTICS AND REMEDIAL MEASURES REGRESSION DIAGNOSTICS AND REMEDIAL MEASURES Lalmohan Bhar I.A.S.R.I., Library Avenue, Pusa, New Delhi 110 01 lmbhar@iasri.res.in 1. Introduction Regression analysis is a statistical methodology that utilizes

More information

Formal Statement of Simple Linear Regression Model

Formal Statement of Simple Linear Regression Model Formal Statement of Simple Linear Regression Model Y i = β 0 + β 1 X i + ɛ i Y i value of the response variable in the i th trial β 0 and β 1 are parameters X i is a known constant, the value of the predictor

More information

Outline. Review regression diagnostics Remedial measures Weighted regression Ridge regression Robust regression Bootstrapping

Outline. Review regression diagnostics Remedial measures Weighted regression Ridge regression Robust regression Bootstrapping Topic 19: Remedies Outline Review regression diagnostics Remedial measures Weighted regression Ridge regression Robust regression Bootstrapping Regression Diagnostics Summary Check normality of the residuals

More information

Outline. Remedial Measures) Extra Sums of Squares Standardized Version of the Multiple Regression Model

Outline. Remedial Measures) Extra Sums of Squares Standardized Version of the Multiple Regression Model Outline 1 Multiple Linear Regression (Estimation, Inference, Diagnostics and Remedial Measures) 2 Special Topics for Multiple Regression Extra Sums of Squares Standardized Version of the Multiple Regression

More information

Math 423/533: The Main Theoretical Topics

Math 423/533: The Main Theoretical Topics Math 423/533: The Main Theoretical Topics Notation sample size n, data index i number of predictors, p (p = 2 for simple linear regression) y i : response for individual i x i = (x i1,..., x ip ) (1 p)

More information

Diagnostics and Remedial Measures

Diagnostics and Remedial Measures Diagnostics and Remedial Measures Yang Feng http://www.stat.columbia.edu/~yangfeng Yang Feng (Columbia University) Diagnostics and Remedial Measures 1 / 72 Remedial Measures How do we know that the regression

More information

Simple and Multiple Linear Regression

Simple and Multiple Linear Regression Sta. 113 Chapter 12 and 13 of Devore March 12, 2010 Table of contents 1 Simple Linear Regression 2 Model Simple Linear Regression A simple linear regression model is given by Y = β 0 + β 1 x + ɛ where

More information

STAT5044: Regression and Anova. Inyoung Kim

STAT5044: Regression and Anova. Inyoung Kim STAT5044: Regression and Anova Inyoung Kim 2 / 47 Outline 1 Regression 2 Simple Linear regression 3 Basic concepts in regression 4 How to estimate unknown parameters 5 Properties of Least Squares Estimators:

More information

Weighted Least Squares

Weighted Least Squares Weighted Least Squares The standard linear model assumes that Var(ε i ) = σ 2 for i = 1,..., n. As we have seen, however, there are instances where Var(Y X = x i ) = Var(ε i ) = σ2 w i. Here w 1,..., w

More information

Quantitative Methods I: Regression diagnostics

Quantitative Methods I: Regression diagnostics Quantitative Methods I: Regression University College Dublin 10 December 2014 1 Assumptions and errors 2 3 4 Outline Assumptions and errors 1 Assumptions and errors 2 3 4 Assumptions: specification Linear

More information

ISyE 691 Data mining and analytics

ISyE 691 Data mining and analytics ISyE 691 Data mining and analytics Regression Instructor: Prof. Kaibo Liu Department of Industrial and Systems Engineering UW-Madison Email: kliu8@wisc.edu Office: Room 3017 (Mechanical Engineering Building)

More information

Regression Analysis for Data Containing Outliers and High Leverage Points

Regression Analysis for Data Containing Outliers and High Leverage Points Alabama Journal of Mathematics 39 (2015) ISSN 2373-0404 Regression Analysis for Data Containing Outliers and High Leverage Points Asim Kumer Dey Department of Mathematics Lamar University Md. Amir Hossain

More information

Applied Regression Analysis

Applied Regression Analysis Applied Regression Analysis Chapter 3 Multiple Linear Regression Hongcheng Li April, 6, 2013 Recall simple linear regression 1 Recall simple linear regression 2 Parameter Estimation 3 Interpretations of

More information

Multicollinearity and A Ridge Parameter Estimation Approach

Multicollinearity and A Ridge Parameter Estimation Approach Journal of Modern Applied Statistical Methods Volume 15 Issue Article 5 11-1-016 Multicollinearity and A Ridge Parameter Estimation Approach Ghadban Khalaf King Khalid University, albadran50@yahoo.com

More information

Ordinary Least Squares Regression

Ordinary Least Squares Regression Ordinary Least Squares Regression Goals for this unit More on notation and terminology OLS scalar versus matrix derivation Some Preliminaries In this class we will be learning to analyze Cross Section

More information

PAijpam.eu M ESTIMATION, S ESTIMATION, AND MM ESTIMATION IN ROBUST REGRESSION

PAijpam.eu M ESTIMATION, S ESTIMATION, AND MM ESTIMATION IN ROBUST REGRESSION International Journal of Pure and Applied Mathematics Volume 91 No. 3 2014, 349-360 ISSN: 1311-8080 (printed version); ISSN: 1314-3395 (on-line version) url: http://www.ijpam.eu doi: http://dx.doi.org/10.12732/ijpam.v91i3.7

More information

MS&E 226: Small Data. Lecture 11: Maximum likelihood (v2) Ramesh Johari

MS&E 226: Small Data. Lecture 11: Maximum likelihood (v2) Ramesh Johari MS&E 226: Small Data Lecture 11: Maximum likelihood (v2) Ramesh Johari ramesh.johari@stanford.edu 1 / 18 The likelihood function 2 / 18 Estimating the parameter This lecture develops the methodology behind

More information

Data Mining Stat 588

Data Mining Stat 588 Data Mining Stat 588 Lecture 02: Linear Methods for Regression Department of Statistics & Biostatistics Rutgers University September 13 2011 Regression Problem Quantitative generic output variable Y. Generic

More information

Lecture 18 Miscellaneous Topics in Multiple Regression

Lecture 18 Miscellaneous Topics in Multiple Regression Lecture 18 Miscellaneous Topics in Multiple Regression STAT 512 Spring 2011 Background Reading KNNL: 8.1-8.5,10.1, 11, 12 18-1 Topic Overview Polynomial Models (8.1) Interaction Models (8.2) Qualitative

More information

Weighted Least Squares

Weighted Least Squares Weighted Least Squares The standard linear model assumes that Var(ε i ) = σ 2 for i = 1,..., n. As we have seen, however, there are instances where Var(Y X = x i ) = Var(ε i ) = σ2 w i. Here w 1,..., w

More information

Linear regression methods

Linear regression methods Linear regression methods Most of our intuition about statistical methods stem from linear regression. For observations i = 1,..., n, the model is Y i = p X ij β j + ε i, j=1 where Y i is the response

More information

Chapter 2 Multiple Regression I (Part 1)

Chapter 2 Multiple Regression I (Part 1) Chapter 2 Multiple Regression I (Part 1) 1 Regression several predictor variables The response Y depends on several predictor variables X 1,, X p response {}}{ Y predictor variables {}}{ X 1, X 2,, X p

More information

Chapter 1: Linear Regression with One Predictor Variable also known as: Simple Linear Regression Bivariate Linear Regression

Chapter 1: Linear Regression with One Predictor Variable also known as: Simple Linear Regression Bivariate Linear Regression BSTT523: Kutner et al., Chapter 1 1 Chapter 1: Linear Regression with One Predictor Variable also known as: Simple Linear Regression Bivariate Linear Regression Introduction: Functional relation between

More information

Statistics 203: Introduction to Regression and Analysis of Variance Course review

Statistics 203: Introduction to Regression and Analysis of Variance Course review Statistics 203: Introduction to Regression and Analysis of Variance Course review Jonathan Taylor - p. 1/?? Today Review / overview of what we learned. - p. 2/?? General themes in regression models Specifying

More information

Dr. Maddah ENMG 617 EM Statistics 11/28/12. Multiple Regression (3) (Chapter 15, Hines)

Dr. Maddah ENMG 617 EM Statistics 11/28/12. Multiple Regression (3) (Chapter 15, Hines) Dr. Maddah ENMG 617 EM Statistics 11/28/12 Multiple Regression (3) (Chapter 15, Hines) Problems in multiple regression: Multicollinearity This arises when the independent variables x 1, x 2,, x k, are

More information

Statistical View of Least Squares

Statistical View of Least Squares Basic Ideas Some Examples Least Squares May 22, 2007 Basic Ideas Simple Linear Regression Basic Ideas Some Examples Least Squares Suppose we have two variables x and y Basic Ideas Simple Linear Regression

More information

Statistics and Econometrics I

Statistics and Econometrics I Statistics and Econometrics I Point Estimation Shiu-Sheng Chen Department of Economics National Taiwan University September 13, 2016 Shiu-Sheng Chen (NTU Econ) Statistics and Econometrics I September 13,

More information

Multiple Linear Regression

Multiple Linear Regression Multiple Linear Regression University of California, San Diego Instructor: Ery Arias-Castro http://math.ucsd.edu/~eariasca/teaching.html 1 / 42 Passenger car mileage Consider the carmpg dataset taken from

More information

Simple Regression Model Setup Estimation Inference Prediction. Model Diagnostic. Multiple Regression. Model Setup and Estimation.

Simple Regression Model Setup Estimation Inference Prediction. Model Diagnostic. Multiple Regression. Model Setup and Estimation. Statistical Computation Math 475 Jimin Ding Department of Mathematics Washington University in St. Louis www.math.wustl.edu/ jmding/math475/index.html October 10, 2013 Ridge Part IV October 10, 2013 1

More information

Variable Selection and Model Building

Variable Selection and Model Building LINEAR REGRESSION ANALYSIS MODULE XIII Lecture - 37 Variable Selection and Model Building Dr. Shalabh Department of Mathematics and Statistics Indian Institute of Technology Kanpur The complete regression

More information

Econ 510 B. Brown Spring 2014 Final Exam Answers

Econ 510 B. Brown Spring 2014 Final Exam Answers Econ 510 B. Brown Spring 2014 Final Exam Answers Answer five of the following questions. You must answer question 7. The question are weighted equally. You have 2.5 hours. You may use a calculator. Brevity

More information

The Model Building Process Part I: Checking Model Assumptions Best Practice

The Model Building Process Part I: Checking Model Assumptions Best Practice The Model Building Process Part I: Checking Model Assumptions Best Practice Authored by: Sarah Burke, PhD 31 July 2017 The goal of the STAT T&E COE is to assist in developing rigorous, defensible test

More information

High-dimensional regression modeling

High-dimensional regression modeling High-dimensional regression modeling David Causeur Department of Statistics and Computer Science Agrocampus Ouest IRMAR CNRS UMR 6625 http://www.agrocampus-ouest.fr/math/causeur/ Course objectives Making

More information

Inference in Regression Analysis

Inference in Regression Analysis Inference in Regression Analysis Dr. Frank Wood Frank Wood, fwood@stat.columbia.edu Linear Regression Models Lecture 4, Slide 1 Today: Normal Error Regression Model Y i = β 0 + β 1 X i + ǫ i Y i value

More information

Regression Diagnostics for Survey Data

Regression Diagnostics for Survey Data Regression Diagnostics for Survey Data Richard Valliant Joint Program in Survey Methodology, University of Maryland and University of Michigan USA Jianzhu Li (Westat), Dan Liao (JPSM) 1 Introduction Topics

More information

Lecture 24: Weighted and Generalized Least Squares

Lecture 24: Weighted and Generalized Least Squares Lecture 24: Weighted and Generalized Least Squares 1 Weighted Least Squares When we use ordinary least squares to estimate linear regression, we minimize the mean squared error: MSE(b) = 1 n (Y i X i β)

More information

Lecture 3: Linear Models. Bruce Walsh lecture notes Uppsala EQG course version 28 Jan 2012

Lecture 3: Linear Models. Bruce Walsh lecture notes Uppsala EQG course version 28 Jan 2012 Lecture 3: Linear Models Bruce Walsh lecture notes Uppsala EQG course version 28 Jan 2012 1 Quick Review of the Major Points The general linear model can be written as y = X! + e y = vector of observed

More information

Statistics 203: Introduction to Regression and Analysis of Variance Penalized models

Statistics 203: Introduction to Regression and Analysis of Variance Penalized models Statistics 203: Introduction to Regression and Analysis of Variance Penalized models Jonathan Taylor - p. 1/15 Today s class Bias-Variance tradeoff. Penalized regression. Cross-validation. - p. 2/15 Bias-variance

More information

Business Statistics. Tommaso Proietti. Linear Regression. DEF - Università di Roma 'Tor Vergata'

Business Statistics. Tommaso Proietti. Linear Regression. DEF - Università di Roma 'Tor Vergata' Business Statistics Tommaso Proietti DEF - Università di Roma 'Tor Vergata' Linear Regression Specication Let Y be a univariate quantitative response variable. We model Y as follows: Y = f(x) + ε where

More information

On dealing with spatially correlated residuals in remote sensing and GIS

On dealing with spatially correlated residuals in remote sensing and GIS On dealing with spatially correlated residuals in remote sensing and GIS Nicholas A. S. Hamm 1, Peter M. Atkinson and Edward J. Milton 3 School of Geography University of Southampton Southampton SO17 3AT

More information

Simple Linear Regression

Simple Linear Regression Simple Linear Regression In simple linear regression we are concerned about the relationship between two variables, X and Y. There are two components to such a relationship. 1. The strength of the relationship.

More information

Day 4: Shrinkage Estimators

Day 4: Shrinkage Estimators Day 4: Shrinkage Estimators Kenneth Benoit Data Mining and Statistical Learning March 9, 2015 n versus p (aka k) Classical regression framework: n > p. Without this inequality, the OLS coefficients have

More information

Lecture 10 Multiple Linear Regression

Lecture 10 Multiple Linear Regression Lecture 10 Multiple Linear Regression STAT 512 Spring 2011 Background Reading KNNL: 6.1-6.5 10-1 Topic Overview Multiple Linear Regression Model 10-2 Data for Multiple Regression Y i is the response variable

More information

Simple Linear Regression Analysis

Simple Linear Regression Analysis LINEAR REGRESSION ANALYSIS MODULE II Lecture - 6 Simple Linear Regression Analysis Dr. Shalabh Department of Mathematics and Statistics Indian Institute of Technology Kanpur Prediction of values of study

More information

Ch 2: Simple Linear Regression

Ch 2: Simple Linear Regression Ch 2: Simple Linear Regression 1. Simple Linear Regression Model A simple regression model with a single regressor x is y = β 0 + β 1 x + ɛ, where we assume that the error ɛ is independent random component

More information

Regression. Oscar García

Regression. Oscar García Regression Oscar García Regression methods are fundamental in Forest Mensuration For a more concise and general presentation, we shall first review some matrix concepts 1 Matrices An order n m matrix is

More information

On Modifications to Linking Variance Estimators in the Fay-Herriot Model that Induce Robustness

On Modifications to Linking Variance Estimators in the Fay-Herriot Model that Induce Robustness Statistics and Applications {ISSN 2452-7395 (online)} Volume 16 No. 1, 2018 (New Series), pp 289-303 On Modifications to Linking Variance Estimators in the Fay-Herriot Model that Induce Robustness Snigdhansu

More information

Remedial Measures, Brown-Forsythe test, F test

Remedial Measures, Brown-Forsythe test, F test Remedial Measures, Brown-Forsythe test, F test Dr. Frank Wood Frank Wood, fwood@stat.columbia.edu Linear Regression Models Lecture 7, Slide 1 Remedial Measures How do we know that the regression function

More information

Instructions: Closed book, notes, and no electronic devices. Points (out of 200) in parentheses

Instructions: Closed book, notes, and no electronic devices. Points (out of 200) in parentheses ISQS 5349 Final Spring 2011 Instructions: Closed book, notes, and no electronic devices. Points (out of 200) in parentheses 1. (10) What is the definition of a regression model that we have used throughout

More information

Semester 2, 2015/2016

Semester 2, 2015/2016 ECN 3202 APPLIED ECONOMETRICS 5. HETEROSKEDASTICITY Mr. Sydney Armstrong Lecturer 1 The University of Guyana 1 Semester 2, 2015/2016 WHAT IS HETEROSKEDASTICITY? The multiple linear regression model can

More information

Ridge regression. Patrick Breheny. February 8. Penalized regression Ridge regression Bayesian interpretation

Ridge regression. Patrick Breheny. February 8. Penalized regression Ridge regression Bayesian interpretation Patrick Breheny February 8 Patrick Breheny High-Dimensional Data Analysis (BIOS 7600) 1/27 Introduction Basic idea Standardization Large-scale testing is, of course, a big area and we could keep talking

More information

STAT 704 Sections IRLS and Bootstrap

STAT 704 Sections IRLS and Bootstrap STAT 704 Sections 11.4-11.5. IRLS and John Grego Department of Statistics, University of South Carolina Stat 704: Data Analysis I 1 / 14 LOWESS IRLS LOWESS LOWESS (LOcally WEighted Scatterplot Smoothing)

More information

Lecture 16 Solving GLMs via IRWLS

Lecture 16 Solving GLMs via IRWLS Lecture 16 Solving GLMs via IRWLS 09 November 2015 Taylor B. Arnold Yale Statistics STAT 312/612 Notes problem set 5 posted; due next class problem set 6, November 18th Goals for today fixed PCA example

More information

Robust Variable Selection Methods for Grouped Data. Kristin Lee Seamon Lilly

Robust Variable Selection Methods for Grouped Data. Kristin Lee Seamon Lilly Robust Variable Selection Methods for Grouped Data by Kristin Lee Seamon Lilly A dissertation submitted to the Graduate Faculty of Auburn University in partial fulfillment of the requirements for the Degree

More information

Multicollinearity occurs when two or more predictors in the model are correlated and provide redundant information about the response.

Multicollinearity occurs when two or more predictors in the model are correlated and provide redundant information about the response. Multicollinearity Read Section 7.5 in textbook. Multicollinearity occurs when two or more predictors in the model are correlated and provide redundant information about the response. Example of multicollinear

More information

Multiple Regression Analysis. Part III. Multiple Regression Analysis

Multiple Regression Analysis. Part III. Multiple Regression Analysis Part III Multiple Regression Analysis As of Sep 26, 2017 1 Multiple Regression Analysis Estimation Matrix form Goodness-of-Fit R-square Adjusted R-square Expected values of the OLS estimators Irrelevant

More information

Lecture 2: Linear Models. Bruce Walsh lecture notes Seattle SISG -Mixed Model Course version 23 June 2011

Lecture 2: Linear Models. Bruce Walsh lecture notes Seattle SISG -Mixed Model Course version 23 June 2011 Lecture 2: Linear Models Bruce Walsh lecture notes Seattle SISG -Mixed Model Course version 23 June 2011 1 Quick Review of the Major Points The general linear model can be written as y = X! + e y = vector

More information

Econometrics Multiple Regression Analysis: Heteroskedasticity

Econometrics Multiple Regression Analysis: Heteroskedasticity Econometrics Multiple Regression Analysis: João Valle e Azevedo Faculdade de Economia Universidade Nova de Lisboa Spring Semester João Valle e Azevedo (FEUNL) Econometrics Lisbon, April 2011 1 / 19 Properties

More information

Available online at (Elixir International Journal) Statistics. Elixir Statistics 49 (2012)

Available online at   (Elixir International Journal) Statistics. Elixir Statistics 49 (2012) 10108 Available online at www.elixirpublishers.com (Elixir International Journal) Statistics Elixir Statistics 49 (2012) 10108-10112 The detention and correction of multicollinearity effects in a multiple

More information

Estimating σ 2. We can do simple prediction of Y and estimation of the mean of Y at any value of X.

Estimating σ 2. We can do simple prediction of Y and estimation of the mean of Y at any value of X. Estimating σ 2 We can do simple prediction of Y and estimation of the mean of Y at any value of X. To perform inferences about our regression line, we must estimate σ 2, the variance of the error term.

More information

Effects of Outliers and Multicollinearity on Some Estimators of Linear Regression Model

Effects of Outliers and Multicollinearity on Some Estimators of Linear Regression Model 204 Effects of Outliers and Multicollinearity on Some Estimators of Linear Regression Model S. A. Ibrahim 1 ; W. B. Yahya 2 1 Department of Physical Sciences, Al-Hikmah University, Ilorin, Nigeria. e-mail:

More information

The Model Building Process Part I: Checking Model Assumptions Best Practice (Version 1.1)

The Model Building Process Part I: Checking Model Assumptions Best Practice (Version 1.1) The Model Building Process Part I: Checking Model Assumptions Best Practice (Version 1.1) Authored by: Sarah Burke, PhD Version 1: 31 July 2017 Version 1.1: 24 October 2017 The goal of the STAT T&E COE

More information

Indian Statistical Institute

Indian Statistical Institute Indian Statistical Institute Introductory Computer programming Robust Regression methods with high breakdown point Author: Roll No: MD1701 February 24, 2018 Contents 1 Introduction 2 2 Criteria for evaluating

More information

Improved Liu Estimators for the Poisson Regression Model

Improved Liu Estimators for the Poisson Regression Model www.ccsenet.org/isp International Journal of Statistics and Probability Vol., No. ; May 202 Improved Liu Estimators for the Poisson Regression Model Kristofer Mansson B. M. Golam Kibria Corresponding author

More information

Heteroskedasticity. Part VII. Heteroskedasticity

Heteroskedasticity. Part VII. Heteroskedasticity Part VII Heteroskedasticity As of Oct 15, 2015 1 Heteroskedasticity Consequences Heteroskedasticity-robust inference Testing for Heteroskedasticity Weighted Least Squares (WLS) Feasible generalized Least

More information

Math 5305 Notes. Diagnostics and Remedial Measures. Jesse Crawford. Department of Mathematics Tarleton State University

Math 5305 Notes. Diagnostics and Remedial Measures. Jesse Crawford. Department of Mathematics Tarleton State University Math 5305 Notes Diagnostics and Remedial Measures Jesse Crawford Department of Mathematics Tarleton State University (Tarleton State University) Diagnostics and Remedial Measures 1 / 44 Model Assumptions

More information

Ridge Regression. Summary. Sample StatFolio: ridge reg.sgp. STATGRAPHICS Rev. 10/1/2014

Ridge Regression. Summary. Sample StatFolio: ridge reg.sgp. STATGRAPHICS Rev. 10/1/2014 Ridge Regression Summary... 1 Data Input... 4 Analysis Summary... 5 Analysis Options... 6 Ridge Trace... 7 Regression Coefficients... 8 Standardized Regression Coefficients... 9 Observed versus Predicted...

More information

Advanced Econometrics I

Advanced Econometrics I Lecture Notes Autumn 2010 Dr. Getinet Haile, University of Mannheim 1. Introduction Introduction & CLRM, Autumn Term 2010 1 What is econometrics? Econometrics = economic statistics economic theory mathematics

More information

Analysis of Cross-Sectional Data

Analysis of Cross-Sectional Data Analysis of Cross-Sectional Data Kevin Sheppard http://www.kevinsheppard.com Oxford MFE This version: November 8, 2017 November 13 14, 2017 Outline Econometric models Specification that can be analyzed

More information

Least Squares Estimation-Finite-Sample Properties

Least Squares Estimation-Finite-Sample Properties Least Squares Estimation-Finite-Sample Properties Ping Yu School of Economics and Finance The University of Hong Kong Ping Yu (HKU) Finite-Sample 1 / 29 Terminology and Assumptions 1 Terminology and Assumptions

More information

The general linear regression with k explanatory variables is just an extension of the simple regression as follows

The general linear regression with k explanatory variables is just an extension of the simple regression as follows 3. Multiple Regression Analysis The general linear regression with k explanatory variables is just an extension of the simple regression as follows (1) y i = β 0 + β 1 x i1 + + β k x ik + u i. Because

More information

MS&E 226. In-Class Midterm Examination Solutions Small Data October 20, 2015

MS&E 226. In-Class Midterm Examination Solutions Small Data October 20, 2015 MS&E 226 In-Class Midterm Examination Solutions Small Data October 20, 2015 PROBLEM 1. Alice uses ordinary least squares to fit a linear regression model on a dataset containing outcome data Y and covariates

More information

Chapter 4 Experiments with Blocking Factors

Chapter 4 Experiments with Blocking Factors Chapter 4 Experiments with Blocking Factors 許湘伶 Design and Analysis of Experiments (Douglas C. Montgomery) hsuhl (NUK) DAE Chap. 4 1 / 54 The Randomized Complete Block Design (RCBD; 隨機化完全集區設計 ) 1 Variability

More information

Ma 3/103: Lecture 24 Linear Regression I: Estimation

Ma 3/103: Lecture 24 Linear Regression I: Estimation Ma 3/103: Lecture 24 Linear Regression I: Estimation March 3, 2017 KC Border Linear Regression I March 3, 2017 1 / 32 Regression analysis Regression analysis Estimate and test E(Y X) = f (X). f is the

More information

Statistics 910, #5 1. Regression Methods

Statistics 910, #5 1. Regression Methods Statistics 910, #5 1 Overview Regression Methods 1. Idea: effects of dependence 2. Examples of estimation (in R) 3. Review of regression 4. Comparisons and relative efficiencies Idea Decomposition Well-known

More information

Advanced Quantitative Methods: ordinary least squares

Advanced Quantitative Methods: ordinary least squares Advanced Quantitative Methods: Ordinary Least Squares University College Dublin 31 January 2012 1 2 3 4 5 Terminology y is the dependent variable referred to also (by Greene) as a regressand X are the

More information

Linear Regression. September 27, Chapter 3. Chapter 3 September 27, / 77

Linear Regression. September 27, Chapter 3. Chapter 3 September 27, / 77 Linear Regression Chapter 3 September 27, 2016 Chapter 3 September 27, 2016 1 / 77 1 3.1. Simple linear regression 2 3.2 Multiple linear regression 3 3.3. The least squares estimation 4 3.4. The statistical

More information

A Short Introduction to Curve Fitting and Regression by Brad Morantz

A Short Introduction to Curve Fitting and Regression by Brad Morantz A Short Introduction to Curve Fitting and Regression by Brad Morantz bradscientist@machine-cognition.com Overview What can regression do for me? Example Model building Error Metrics OLS Regression Robust

More information

STAT 135 Lab 13 (Review) Linear Regression, Multivariate Random Variables, Prediction, Logistic Regression and the δ-method.

STAT 135 Lab 13 (Review) Linear Regression, Multivariate Random Variables, Prediction, Logistic Regression and the δ-method. STAT 135 Lab 13 (Review) Linear Regression, Multivariate Random Variables, Prediction, Logistic Regression and the δ-method. Rebecca Barter May 5, 2015 Linear Regression Review Linear Regression Review

More information

Introduction to Linear regression analysis. Part 2. Model comparisons

Introduction to Linear regression analysis. Part 2. Model comparisons Introduction to Linear regression analysis Part Model comparisons 1 ANOVA for regression Total variation in Y SS Total = Variation explained by regression with X SS Regression + Residual variation SS Residual

More information

Math 3330: Solution to midterm Exam

Math 3330: Solution to midterm Exam Math 3330: Solution to midterm Exam Question 1: (14 marks) Suppose the regression model is y i = β 0 + β 1 x i + ε i, i = 1,, n, where ε i are iid Normal distribution N(0, σ 2 ). a. (2 marks) Compute the

More information

IEOR 165 Lecture 7 1 Bias-Variance Tradeoff

IEOR 165 Lecture 7 1 Bias-Variance Tradeoff IEOR 165 Lecture 7 Bias-Variance Tradeoff 1 Bias-Variance Tradeoff Consider the case of parametric regression with β R, and suppose we would like to analyze the error of the estimate ˆβ in comparison to

More information

Least Absolute Value vs. Least Squares Estimation and Inference Procedures in Regression Models with Asymmetric Error Distributions

Least Absolute Value vs. Least Squares Estimation and Inference Procedures in Regression Models with Asymmetric Error Distributions Journal of Modern Applied Statistical Methods Volume 8 Issue 1 Article 13 5-1-2009 Least Absolute Value vs. Least Squares Estimation and Inference Procedures in Regression Models with Asymmetric Error

More information

Introductory Econometrics

Introductory Econometrics Based on the textbook by Wooldridge: : A Modern Approach Robert M. Kunst robert.kunst@univie.ac.at University of Vienna and Institute for Advanced Studies Vienna November 23, 2013 Outline Introduction

More information

Need for Several Predictor Variables

Need for Several Predictor Variables Multiple regression One of the most widely used tools in statistical analysis Matrix expressions for multiple regression are the same as for simple linear regression Need for Several Predictor Variables

More information