Linear regression. Linear regression is a simple approach to supervised learning. It assumes that the dependence of Y on X 1,X 2,...X p is linear.

Similar documents
Statistical Methods for Data Mining

Lecture 6: Linear Regression (continued)

Lecture 6: Linear Regression

Machine Learning Linear Regression. Prof. Matteo Matteucci

Statistical Machine Learning I

Multiple (non) linear regression. Department of Computer Science, Czech Technical University in Prague

Linear Regression In God we trust, all others bring data. William Edwards Deming

Machine Learning for OR & FE

Data Analysis 1 LINEAR REGRESSION. Chapter 03

Linear Model Selection and Regularization

LECTURE 03: LINEAR REGRESSION PT. 1. September 18, 2017 SDS 293: Machine Learning

Final Overview. Introduction to ML. Marek Petrik 4/25/2017

Regression, Ridge Regression, Lasso

Chapter 4: Regression Models

Machine Learning for OR & FE

Lecture 5: Clustering, Linear Regression

LECTURE 04: LINEAR REGRESSION PT. 2. September 20, 2017 SDS 293: Machine Learning

Lecture 5: Clustering, Linear Regression

Day 4: Shrinkage Estimators

MS-C1620 Statistical inference

Linear model selection and regularization

Python 데이터분석 보충자료. 윤형기

Simple Linear Regression for the Climate Data

Course in Data Science

Lecture 5: Clustering, Linear Regression

Simple Linear Regression for the MPG Data

Chapter 1 Statistical Inference

Chapter 9 Regression. 9.1 Simple linear regression Linear models Least squares Predictions and residuals.

Exam Applied Statistical Regression. Good Luck!

Business Statistics. Lecture 9: Simple Regression

Final Review. Yang Feng. Yang Feng (Columbia University) Final Review 1 / 58

ST430 Exam 2 Solutions

Data Mining Stat 588

Chapter 4. Regression Models. Learning Objectives

Chapter 3 Multiple Regression Complete Example

STATS216v Introduction to Statistical Learning Stanford University, Summer Midterm Exam (Solutions) Duration: 1 hours

Unit 6 - Introduction to linear regression

Model Selection. Frank Wood. December 10, 2009

Introduction to Machine Learning and Cross-Validation

Dimension Reduction Methods

MLR Model Selection. Author: Nicholas G Reich, Jeff Goldsmith. This material is part of the statsteachr project

Simple Linear Regression. Y = f (X) + ε. Regression as a term. Linear Regression. Basic structure. Ivo Ugrina. September 29, 2016

Linear Regression. September 27, Chapter 3. Chapter 3 September 27, / 77

From statistics to data science. BAE 815 (Fall 2017) Dr. Zifei Liu

Multiple Regression Methods

Machine Learning Linear Classification. Prof. Matteo Matteucci

CHAPTER 4 & 5 Linear Regression with One Regressor. Kazu Matsuda IBEC PHBU 430 Econometrics

Stat/F&W Ecol/Hort 572 Review Points Ané, Spring 2010

Chapter 14 Student Lecture Notes 14-1

Statistical Methods for Data Mining

Math 423/533: The Main Theoretical Topics

ECON 497 Midterm Spring

Business Statistics. Tommaso Proietti. Linear Regression. DEF - Università di Roma 'Tor Vergata'

holding all other predictors constant

9. Linear Regression and Correlation

Predictive Analytics : QM901.1x Prof U Dinesh Kumar, IIMB. All Rights Reserved, Indian Institute of Management Bangalore

STATISTICS 110/201 PRACTICE FINAL EXAM

Linear Models 1. Isfahan University of Technology Fall Semester, 2014

regression analysis is a type of inferential statistics which tells us whether relationships between two or more variables exist

Class Notes: Week 8. Probit versus Logit Link Functions and Count Data

Linear Regression. Furthermore, it is simple.

CS6220: DATA MINING TECHNIQUES

Introduction to Statistical modeling: handout for Math 489/583

5. Let W follow a normal distribution with mean of μ and the variance of 1. Then, the pdf of W is

CS6220: DATA MINING TECHNIQUES

y response variable x 1, x 2,, x k -- a set of explanatory variables

Unit 6 - Simple linear regression

Introduction to Regression

SF2930: REGRESION ANALYSIS LECTURE 1 SIMPLE LINEAR REGRESSION.

Midterm 2 - Solutions

General Linear Model (Chapter 4)

10. Alternative case influence statistics

Linear Regression with Multiple Regressors

Chapter 10 Nonlinear Models

Activity #12: More regression topics: LOWESS; polynomial, nonlinear, robust, quantile; ANOVA as regression

Harvard University. Rigorous Research in Engineering Education

Introduction to Logistic Regression

Biostatistics-Lecture 16 Model Selection. Ruibin Xi Peking University School of Mathematical Sciences

4. Nonlinear regression functions

Regression Models. Chapter 4. Introduction. Introduction. Introduction

Bayesian Analysis LEARNING OBJECTIVES. Calculating Revised Probabilities. Calculating Revised Probabilities. Calculating Revised Probabilities

Midterm 2 - Solutions

Direct Learning: Linear Regression. Donglin Zeng, Department of Biostatistics, University of North Carolina

Week 1 Lectures. Part I: Introduction to data mining Part II: Regression

Business Statistics. Lecture 10: Course Review

Terminology for Statistical Data

Test 3 Practice Test A. NOTE: Ignore Q10 (not covered)

Multiple Regression. Peerapat Wongchaiwat, Ph.D.

Multiple Linear Regression

Trendlines Simple Linear Regression Multiple Linear Regression Systematic Model Building Practical Issues

Simple Linear Regression: One Qualitative IV

Lecture 7: Modeling Krak(en)

How the mean changes depends on the other variable. Plots can show what s happening...

bivariate correlation bivariate regression multiple regression

Lecture Slides. Elementary Statistics Tenth Edition. by Mario F. Triola. and the Triola Statistics Series. Slide 1

Linear Models (continued)

Mathematical Notation Math Introduction to Applied Statistics

22s:152 Applied Linear Regression

Chapter 27 Summary Inferences for Regression

Linear Regression with Multiple Regressors

Transcription:

Linear regression Linear regression is a simple approach to supervised learning. It assumes that the dependence of Y on X 1,X 2,...X p is linear. 1/48

Linear regression Linear regression is a simple approach to supervised learning. It assumes that the dependence of Y on X 1,X 2,...X p is linear. True regression functions are never linear! f(x) 3 4 5 6 7 2 4 6 8 X 1/48

Linear regression Linear regression is a simple approach to supervised learning. It assumes that the dependence of Y on X 1,X 2,...X p is linear. True regression functions are never linear! f(x) 3 4 5 6 7 2 4 6 8 although it may seem overly simplistic, linear regression is extremely useful both conceptually and practically. X 1/48

Linear regression for the advertising data Consider the advertising data shown on the next slide. Questions we might ask: Is there a relationship between advertising budget and sales? How strong is the relationship between advertising budget and sales? Which media contribute to sales? How accurately can we predict future sales? Is the relationship linear? Is there synergy among the advertising media? 2/48

Advertising data Sales 5 10 15 20 25 Sales 5 10 15 20 25 Sales 5 10 15 20 25 0 50 100 200 300 TV 0 10 20 30 40 50 Radio 0 20 40 60 80 100 Newspaper 3/48

Simple linear regression using a single predictor X. We assume a model Y = 0 + 1 X +, where 0 and 1 are two unknown constants that represent the intercept and slope, also known as coe cients or parameters, and is the error term. Given some estimates ˆ0 and ˆ1 for the model coe we predict future sales using ŷ = ˆ0 + ˆ1x, cients, where ŷ indicates a prediction of Y on the basis of X = x. The hat symbol denotes an estimated value. 4/48

Estimation of the parameters by least squares Let ŷ i = ˆ0 + ˆ1x i be the prediction for Y based on the ith value of X. Thene i = y i ŷ i represents the ith residual 5/48

Estimation of the parameters by least squares Let ŷ i = ˆ0 + ˆ1x i be the prediction for Y based on the ith value of X. Thene i = y i ŷ i represents the ith residual We define the residual sum of squares (RSS) as RSS = e 2 1 + e 2 2 + + e 2 n, or equivalently as RSS = (y 1 ˆ0 ˆ1x 1 ) 2 +(y 2 ˆ0 ˆ1x 2 ) 2 +...+(y n ˆ0 ˆ1x n ) 2. 5/48

Estimation of the parameters by least squares Let ŷ i = ˆ0 + ˆ1x i be the prediction for Y based on the ith value of X. Thene i = y i ŷ i represents the ith residual We define the residual sum of squares (RSS) as RSS = e 2 1 + e 2 2 + + e 2 n, or equivalently as RSS = (y 1 ˆ0 ˆ1x 1 ) 2 +(y 2 ˆ0 ˆ1x 2 ) 2 +...+(y n ˆ0 ˆ1x n ) 2. The least squares approach chooses ˆ0 and ˆ1 to minimize the RSS. The minimizing values can be shown to be P n i=1 ˆ1 = (x i x)(y i ȳ) P n i=1 (x i x) 2, ˆ0 =ȳ ˆ1 x, where ȳ 1 n P n i=1 y i and x 1 n P n i=1 x i are the sample means. 5/48

P P Example: advertising data Sales 5 10 15 20 25 0 50 100 150 200 250 300 TV FIGURE 3.1. For the Advertising data, the least squares fit for the regression The least squares fit for the regression of sales onto TV. of sales onto TV is shown. The fit is found by minimizing the sum of squared In this errors. case Eacha grey linear segment fit captures represents an error, theand essence the fit makes ofa compromise by averaging their squares. In this case a linear fit captures the essence of the relationship, although the relationship, it is somewhat although it is somewhat deficient in the left of left the plot. of the plot. Figure 3.1 displays the simple linear regression fit to the Advertising data, where ˆ0 =7.03 and ˆ1 =0.0475. In other words, according to this 6/48

Assessing the Accuracy of the Coe cient Estimates The standard error of an estimator reflects how it varies under repeated sampling. We have SE( ˆ1) 2 = apple 2 P n i=1 (x i x) 2, SE( ˆ0) 2 1 = 2 n + x P 2 n i=1 (x i x) 2, where 2 = Var( ) 7/48

Assessing the Accuracy of the Coe cient Estimates The standard error of an estimator reflects how it varies under repeated sampling. We have SE( ˆ1) 2 = apple 2 P n i=1 (x i x) 2, SE( ˆ0) 2 1 = 2 n + x P 2 n i=1 (x i x) 2, where 2 = Var( ) These standard errors can be used to compute confidence intervals. A 95% confidence interval is defined as a range of values such that with 95% probability, the range will contain the true unknown value of the parameter. It has the form ˆ1 ± 2 SE( ˆ1). 7/48

Confidence intervals continued That is, there is approximately a 95% chance that the interval h ˆ1 2 SE( ˆ1), ˆ1 +2 SE( ˆ1)i will contain the true value of 1 (under a scenario where we got repeated samples like the present sample) 8/48

Confidence intervals continued That is, there is approximately a 95% chance that the interval h ˆ1 2 SE( ˆ1), ˆ1 +2 SE( ˆ1)i will contain the true value of 1 (under a scenario where we got repeated samples like the present sample) For the advertising data, the 95% confidence interval for 1 is [0.042, 0.053] 8/48

Hypothesis testing Standard errors can also be used to perform hypothesis tests on the coe cients. The most common hypothesis test involves testing the null hypothesis of H 0 : There is no relationship between X and Y versus the alternative hypothesis H A : There is some relationship between X and Y. 9/48

Hypothesis testing Standard errors can also be used to perform hypothesis tests on the coe cients. The most common hypothesis test involves testing the null hypothesis of H 0 : There is no relationship between X and Y versus the alternative hypothesis H A : There is some relationship between X and Y. Mathematically, this corresponds to testing H 0 : 1 =0 versus H A : 1 6= 0, since if 1 =0thenthemodelreducestoY = 0 +, and X is not associated with Y. 9/48

Hypothesis testing continued To test the null hypothesis, we compute a t-statistic, given by t = ˆ1 0 SE( ˆ1), This will have a t-distribution with n freedom, assuming 1 = 0. 2 degrees of Using statistical software, it is easy to compute the probability of observing any value equal to t or larger. We call this probability the p-value. 10 / 48

Results for the advertising data Coe cient Std. Error t-statistic p-value Intercept 7.0325 0.4578 15.36 < 0.0001 TV 0.0475 0.0027 17.67 < 0.0001 11 / 48

Assessing the Overall Accuracy of the Model We compute the Residual Standard Error v r 1 RSE = n 2 RSS = u t 1 nx (y i ŷ i ) n 2 2, where the residual sum-of-squares is RSS = P n i=1 (y i ŷ i ) 2. i=1 12 / 48

Assessing the Overall Accuracy of the Model We compute the Residual Standard Error v r 1 RSE = n 2 RSS = u t 1 nx (y i ŷ i ) n 2 2, where the residual sum-of-squares is RSS = P n i=1 (y i ŷ i ) 2. R-squared or fraction of variance explained is R 2 = TSS RSS =1 TSS i=1 RSS TSS where TSS = P n i=1 (y i ȳ) 2 is the total sum of squares. 12 / 48

Assessing the Overall Accuracy of the Model We compute the Residual Standard Error v r 1 RSE = n 2 RSS = u t 1 nx (y i ŷ i ) n 2 2, where the residual sum-of-squares is RSS = P n i=1 (y i ŷ i ) 2. R-squared or fraction of variance explained is R 2 = TSS RSS =1 TSS i=1 RSS TSS where TSS = P n i=1 (y i ȳ) 2 is the total sum of squares. It can be shown that in this simple linear regression setting that R 2 = r 2,wherer is the correlation between X and Y : P n i=1 r = (x i x)(y i y) pp n i=1 (x i x) 2p P n i=1 (y i y). 2 12 / 48

Advertising data results Quantity Value Residual Standard Error 3.26 R 2 0.612 F-statistic 312.1 13 / 48

Multiple Linear Regression Here our model is Y = 0 + 1 X 1 + 2 X 2 + + p X p +, We interpret j as the average e ect on Y of a one unit increase in X j, holding all other predictors fixed. Inthe advertising example, the model becomes sales = 0 + 1 TV + 2 radio + 3 newspaper +. 14 / 48

Interpreting regression coe cients The ideal scenario is when the predictors are uncorrelated abalanced design: - Each coe cient can be estimated and tested separately. - Interpretations such as a unit change in X j is associated with a j change in Y, while all the other variables stay fixed, are possible. Correlations amongst predictors cause problems: - The variance of all coe cients tends to increase, sometimes dramatically - Interpretations become hazardous when X j changes, everything else changes. Claims of causality should be avoided for observational data. 15 / 48

The woes of (interpreting) regression coe cients Data Analysis and Regression Mosteller and Tukey 1977 a regression coe cient j estimates the expected change in Y per unit change in X j, with all other predictors held fixed. But predictors usually change together! 16 / 48

The woes of (interpreting) regression coe cients Data Analysis and Regression Mosteller and Tukey 1977 a regression coe cient j estimates the expected change in Y per unit change in X j, with all other predictors held fixed. But predictors usually change together! Example: Y total amount of change in your pocket; X 1 = # of coins; X 2 = # of pennies, nickels and dimes. By itself, regression coe cient of Y on X 2 will be > 0. But how about with X 1 in model? 16 / 48

The woes of (interpreting) regression coe cients Data Analysis and Regression Mosteller and Tukey 1977 a regression coe cient j estimates the expected change in Y per unit change in X j, with all other predictors held fixed. But predictors usually change together! Example: Y total amount of change in your pocket; X 1 = # of coins; X 2 = # of pennies, nickels and dimes. By itself, regression coe cient of Y on X 2 will be > 0. But how about with X 1 in model? Y = number of tackles by a football player in a season; W and H are his weight and height. Fitted regression model is Ŷ = b 0 +.50W.10H. How do we interpret ˆ2 < 0? 16 / 48

Two quotes by famous Statisticians Essentially, all models are wrong, but some are useful George Box 17 / 48

Two quotes by famous Statisticians Essentially, all models are wrong, but some are useful George Box The only way to find out what will happen when a complex system is disturbed is to disturb the system, not merely to observe it passively Fred Mosteller and John Tukey, paraphrasing George Box 17 / 48

Estimation and Prediction for Multiple Regression Given estimates ˆ0, ˆ1,... ˆp, we can make predictions using the formula ŷ = ˆ0 + ˆ1x 1 + ˆ2x 2 + + ˆpx p. We estimate 0, 1,..., p as the values that minimize the sum of squared residuals RSS = = nx (y i ŷ i ) 2 i=1 nx i=1 (y i ˆ0 ˆ1x i1 ˆ2x i2 ˆpx ip ) 2. This is done using standard statistical software. The values ˆ0, ˆ1,..., ˆp that minimize RSS are the multiple least squares regression coe cient estimates. 18 / 48

Y X 2 X 1 19 / 48

Results for advertising data Coe cient Std. Error t-statistic p-value Intercept 2.939 0.3119 9.42 < 0.0001 TV 0.046 0.0014 32.81 < 0.0001 radio 0.189 0.0086 21.89 < 0.0001 newspaper -0.001 0.0059-0.18 0.8599 Correlations: TV radio newspaper sales TV 1.0000 0.0548 0.0567 0.7822 radio 1.0000 0.3541 0.5762 newspaper 1.0000 0.2283 sales 1.0000 20 / 48

Some important questions 1. Is at least one of the predictors X 1,X 2,...,X p useful in predicting the response? 21 / 48

Some important questions 1. Is at least one of the predictors X 1,X 2,...,X p useful in predicting the response? 2. Do all the predictors help to explain Y, or is only a subset of the predictors useful? 21 / 48

Some important questions 1. Is at least one of the predictors X 1,X 2,...,X p useful in predicting the response? 2. Do all the predictors help to explain Y, or is only a subset of the predictors useful? 3. How well does the model fit the data? 21 / 48

Some important questions 1. Is at least one of the predictors X 1,X 2,...,X p useful in predicting the response? 2. Do all the predictors help to explain Y, or is only a subset of the predictors useful? 3. How well does the model fit the data? 4. Given a set of predictor values, what response value should we predict, and how accurate is our prediction? 21 / 48

Is at least one predictor useful? For the first question, we can use the F-statistic F = (TSS RSS)/p RSS/(n p 1) F p,n p 1 Quantity Value Residual Standard Error 1.69 R 2 0.897 F-statistic 570 22 / 48

Deciding on the important variables The most direct approach is called all subsets or best subsets regression: we compute the least squares fit for all possible subsets and then choose between them based on some criterion that balances training error with model size. 23 / 48

Deciding on the important variables The most direct approach is called all subsets or best subsets regression: we compute the least squares fit for all possible subsets and then choose between them based on some criterion that balances training error with model size. However we often can t examine all possible models, since they are 2 p of them; for example when p = 40 there are over a billion models! Instead we need an automated approach that searches through a subset of them. We discuss two commonly use approaches next. 23 / 48

Forward selection Begin with the null model a model that contains an intercept but no predictors. Fit p simple linear regressions and add to the null model the variable that results in the lowest RSS. Add to that model the variable that results in the lowest RSS amongst all two-variable models. Continue until some stopping rule is satisfied, for example when all remaining variables have a p-value above some threshold. 24 / 48

Backward selection Start with all variables in the model. Remove the variable with the largest p-value that is, the variable that is the least statistically significant. The new (p 1)-variable model is fit, and the variable with the largest p-value is removed. Continue until a stopping rule is reached. For instance, we may stop when all remaining variables have a significant p-value defined by some significance threshold. 25 / 48

Model selection continued Later we discuss more systematic criteria for choosing an optimal member in the path of models produced by forward or backward stepwise selection. These include Mallow s C p, Akaike information criterion (AIC), Bayesian information criterion (BIC), adjusted R 2 and Cross-validation (CV). 26 / 48

Other Considerations in the Regression Model Qualitative Predictors Some predictors are not quantitative but are qualitative, taking a discrete set of values. These are also called categorical predictors or factor variables. See for example the scatterplot matrix of the credit card data in the next slide. In addition to the 7 quantitative variables shown, there are four qualitative variables: gender, student (student status), status (marital status), and ethnicity (Caucasian, African American (AA) or Asian). 27 / 48

26 3. Linear Regression Credit Card Data 40 60 80 100 5 10 15 20 2000 8000 14000 1500 20 80 100 0 500 Balance 6 8 20 40 60 Age 15 20 2 4 Cards 100 150 5 10 Education 8000 14000 50 Income 600 1000 2000 Limit 200 Rating 0 500 1500 2 4 6 8 50 100 150 200 600 1000 FIGURE 3.6. The Credit data set contains information about balance, age, 28 / 48

Qualitative Predictors continued Example: investigate di erences in credit card balance between males and females, ignoring the other variables. We create a new variable ( 1 if ith person is female x i = 0 if ith person is male Resulting model: y i = 0 + 1 x i + i = ( 0 + 1 + i if ith person is female 0 + i if ith person is male. Intrepretation? 29 / 48

Credit card data continued Results for gender model: Coe cient Std. Error t-statistic p-value Intercept 509.80 33.13 15.389 < 0.0001 gender[female] 19.73 46.05 0.429 0.6690 30 / 48

Qualitative predictors with more than two levels With more than two levels, we create additional dummy variables. For example, for the ethnicity variable we create two dummy variables. The first could be ( 1 if ith person is Asian x i1 = 0 if ith person is not Asian, and the second could be ( 1 if ith person is Caucasian x i2 = 0 if ith person is not Caucasian. 31 / 48

Qualitative predictors with more than two levels continued. Then both of these variables can be used in the regression equation, in order to obtain the model 8 >< 0 + 1 + i if ith person is Asian y i = 0+ 1x i1+ 2x i2+ i = 0 + 2 + i if ith person is Caucasian >: 0 + i if ith person is AA. There will always be one fewer dummy variable than the number of levels. The level with no dummy variable African American in this example is known as the baseline. 32 / 48

Results for ethnicity Coe cient Std. Error t-statistic p-value Intercept 531.00 46.32 11.464 < 0.0001 ethnicity[asian] -18.69 65.02-0.287 0.7740 ethnicity[caucasian] -12.50 56.68-0.221 0.8260 33 / 48

Extensions of the Linear Model Removing the additive assumption: interactions and nonlinearity Interactions: In our previous analysis of the Advertising data, we assumed that the e ect on sales of increasing one advertising medium is independent of the amount spent on the other media. For example, the linear model \sales = 0 + 1 TV + 2 radio + 3 newspaper states that the average e ect on sales of a one-unit increase in TV is always 1, regardless of the amount spent on radio. 34 / 48

Interactions continued But suppose that spending money on radio advertising actually increases the e ectiveness of TV advertising, so that the slope term for TV should increase as radio increases. In this situation, given a fixed budget of $100, 000, spending half on radio and half on TV may increase sales more than allocating the entire amount to either TV or to radio. In marketing, this is known as a synergy e ect, and in statistics it is referred to as an interaction e ect. 35 / 48

Interaction in the Advertising data? Sales TV Radio When levels FIGURE of 3.5. either For the TV Advertising or radio data, a linear areregression low, fit then to sales the usingtrue sales TV and radio as predictors. From pattern of the residuals, we can see that are lowerthere than is a pronounced predicted non-linear byrelationship the linear in the data. model. But when advertising is split between the two media, then the model tends to underestimate sales. which simplifies to (3.15) for a simple linear regression. Thus, models with more variables can have higher RSE if the decrease in RSS is small relative to the increase in p. In addition to looking at the RSE and R 2 statistics just discussed, it can be useful to plot the data. Graphical summaries can reveal problems with 36 / 48

Modelling interactions Advertising data Model takes the form sales = 0 + 1 TV + 2 radio + 3 (radio TV)+ Results: = 0 +( 1 + 3 radio) TV + 2 radio +. Coe cient Std. Error t-statistic p-value Intercept 6.7502 0.248 27.23 < 0.0001 TV 0.0191 0.002 12.70 < 0.0001 radio 0.0289 0.009 3.24 0.0014 TV radio 0.0011 0.000 20.73 < 0.0001 37 / 48

Interpretation The results in this table suggests that interactions are important. The p-value for the interaction term TV radio is extremely low, indicating that there is strong evidence for H A : 3 6= 0. The R 2 for the interaction model is 96.8%, compared to only 89.7% for the model that predicts sales using TV and radio without an interaction term. 38 / 48

Interpretation continued This means that (96.8 89.7)/(100 89.7) = 69% of the variability in sales that remains after fitting the additive model has been explained by the interaction term. The coe cient estimates in the table suggest that an increase in TV advertising of $1, 000 is associated with increased sales of ( ˆ1 + ˆ3 radio) 1000 = 19 + 1.1 radio units. An increase in radio advertising of $1, 000 will be associated with an increase in sales of ( ˆ2 + ˆ3 TV) 1000 = 29 + 1.1 TV units. 39 / 48

Hierarchy Sometimes it is the case that an interaction term has a very small p-value, but the associated main e ects (in this case, TV and radio) do not. The hierarchy principle: If we include an interaction in a model, we should also include the main e ects, even if the p-values associated with their coe cients are not significant. 40 / 48

Hierarchy continued The rationale for this principle is that interactions are hard to interpret in a model without main e ects their meaning is changed. Specifically, the interaction terms also contain main e ects, if the model has no main e ect terms. 41 / 48

Interactions between qualitative and quantitative variables Consider the Credit data set, and suppose that we wish to predict balance using income (quantitative) and student (qualitative). Without an interaction term, the model takes the form ( 2 if ith person is a student balance i 0 + 1 income i + 0 if ith person is not a student ( 0 + 2 if ith person is a student = 1 income i + 0 if ith person is not a student. 42 / 48

With interactions, it takes the form ( 2 + 3 income i if student balance i 0 + 1 income i + 0 if not student ( ( 0 + 2)+( 1 + 3) income i if student = 0 + 1 income i if not student 43 / 48

Balance 200 600 1000 1400 Balance 200 600 1000 1400 student non student 0 50 100 150 0 50 100 150 Income Income Credit data; Left: no interaction between income and student. Right: with an interaction term between income and student. 44 / 48

Non-linear e ects of predictors polynomial regression on Auto data Miles per gallon 10 20 30 40 50 Linear Degree 2 Degree 5 50 100 150 200 Horsepower 45 / 48

The figure suggests that mpg = 0 + 1 horsepower + 2 horsepower 2 + may provide a better fit. Coe cient Std. Error t-statistic p-value Intercept 56.9001 1.8004 31.6 < 0.0001 horsepower -0.4662 0.0311-15.0 < 0.0001 horsepower 2 0.0012 0.0001 10.1 < 0.0001 46 / 48

What we did not cover Outliers Non-constant variance of error terms High leverage points Collinearity See text Section 3.33 47 / 48

Generalizations of the Linear Model In much of the rest of this course, we discuss methods that expand the scope of linear models and how they are fit: Classification problems: logistic regression, support vector machines Non-linearity: kernel smoothing, splines and generalized additive models; nearest neighbor methods. Interactions: Tree-based methods, bagging, random forests and boosting (these also capture non-linearities) Regularized fitting: Ridge regression and lasso 48 / 48