Introduction to Linear Regression

Similar documents
Key Algebraic Results in Linear Regression

Regression and the 2-Sample t

Lecture 18: Simple Linear Regression

Business Statistics. Lecture 9: Simple Regression

STAT 3022 Spring 2007

Regression and correlation. Correlation & Regression, I. Regression & correlation. Regression vs. correlation. Involve bivariate, paired data, X & Y

Diagnostics and Transformations Part 2

UNIVERSITY OF MASSACHUSETTS. Department of Mathematics and Statistics. Basic Exam - Applied Statistics. Tuesday, January 17, 2017

1 The Classic Bivariate Least Squares Model

Understanding p Values

Lab 3 A Quick Introduction to Multiple Linear Regression Psychology The Multiple Linear Regression Model

AMS 7 Correlation and Regression Lecture 8

Unit 6 - Introduction to linear regression

Simple Linear Regression

LECTURE 15: SIMPLE LINEAR REGRESSION I

Intro to Linear Regression

The Matrix Algebra of Sample Statistics

Intro to Linear Regression

Data Analysis and Statistical Methods Statistics 651

Random Vectors, Random Matrices, and Matrix Expected Value

Chapter 12 - Part I: Correlation Analysis

9. Linear Regression and Correlation

Introduction and Single Predictor Regression. Correlation

Regression, Part I. - In correlation, it would be irrelevant if we changed the axes on our graph.

Announcements: You can turn in homework until 6pm, slot on wall across from 2202 Bren. Make sure you use the correct slot! (Stats 8, closest to wall)

STAT 111 Recitation 7

Stat 411/511 ESTIMATING THE SLOPE AND INTERCEPT. Charlotte Wickham. stat511.cwick.co.nz. Nov

Unit 6 - Simple linear regression

Correlation. January 11, 2018

L21: Chapter 12: Linear regression

Linear Regression & Correlation

Psych 10 / Stats 60, Practice Problem Set 10 (Week 10 Material), Solutions

Simple linear regression

Inferences for Regression

Modeling Overdispersion

SLR output RLS. Refer to slr (code) on the Lecture Page of the class website.

Chapter 12: Linear regression II

Inference for Regression

Business Statistics. Lecture 10: Correlation and Linear Regression

We d like to know the equation of the line shown (the so called best fit or regression line).

Review of Basic Probability Theory

Myths about the Analysis of Change

The Tetrad Criterion

Section 3: Simple Linear Regression

appstats27.notebook April 06, 2017

Quantitative Understanding in Biology Module II: Model Parameter Estimation Lecture I: Linear Correlation and Regression

Instrumental Variables

BIOSTATISTICS NURS 3324

Path Diagrams. James H. Steiger. Department of Psychology and Human Development Vanderbilt University

ST430 Exam 1 with Answers

Chapter 27 Summary Inferences for Regression

Example: 1982 State SAT Scores (First year state by state data available)

Linear Regression. In this lecture we will study a particular type of regression model: the linear regression model

Lecture 4 Multiple linear regression

MATH 1150 Chapter 2 Notation and Terminology

Dealing with Heteroskedasticity

Lecture 30. DATA 8 Summer Regression Inference

Exam 3 Practice Questions Psych , Fall 9

MATH 644: Regression Analysis Methods

Chapter 16. Simple Linear Regression and dcorrelation

Variance. Standard deviation VAR = = value. Unbiased SD = SD = 10/23/2011. Functional Connectivity Correlation and Regression.

Correlation and regression

1 Multiple Regression

Nonlinear Regression

Biostatistics and Design of Experiments Prof. Mukesh Doble Department of Biotechnology Indian Institute of Technology, Madras

Instrumental Variables

STA 302f16 Assignment Five 1

HOLLOMAN S AP STATISTICS BVD CHAPTER 08, PAGE 1 OF 11. Figure 1 - Variation in the Response Variable

Statistics Introductory Correlation

ES-2 Lecture: More Least-squares Fitting. Spring 2017

Week 11 Sample Means, CLT, Correlation

Correlation & Simple Regression

Chapter 7. Linear Regression (Pt. 1) 7.1 Introduction. 7.2 The Least-Squares Regression Line

Correlation and Linear Regression

Can you tell the relationship between students SAT scores and their college grades?

Principal components

Scatter plot of data from the study. Linear Regression

Gregory Carey, 1998 Regression & Path Analysis - 1 MULTIPLE REGRESSION AND PATH ANALYSIS

Chapter 12 - Lecture 2 Inferences about regression coefficient

Quantitative Bivariate Data

Chapter 1 Review of Equations and Inequalities

22 Approximations - the method of least squares (1)

STATISTICS 110/201 PRACTICE FINAL EXAM

Applied Regression Modeling: A Business Approach Chapter 2: Simple Linear Regression Sections

Linear Regression and Correlation

Chapter 4 Describing the Relation between Two Variables

Linear Regression Model. Badr Missaoui

Density Temp vs Ratio. temp

Interpreting Regression

Applied Regression Analysis

STA Module 5 Regression and Correlation. Learning Objectives. Learning Objectives (Cont.) Upon completing this module, you should be able to:

UNIVERSITY OF TORONTO SCARBOROUGH Department of Computer and Mathematical Sciences Midterm Test, October 2013

Regression, part II. I. What does it all mean? A) Notice that so far all we ve done is math.

APPENDIX 1 BASIC STATISTICS. Summarizing Data

Review of Multiple Regression

Lecture 19 Multiple (Linear) Regression

Business Analytics and Data Mining Modeling Using R Prof. Gaurav Dixit Department of Management Studies Indian Institute of Technology, Roorkee

Data Analysis and Statistical Methods Statistics 651

Chapter 3 Multiple Regression Complete Example

Scatter plot of data from the study. Linear Regression

Transcription:

Introduction to Linear Regression James H. Steiger Department of Psychology and Human Development Vanderbilt University James H. Steiger (Vanderbilt University) Introduction to Linear Regression 1 / 46

Introduction to Linear Regression 1 Introduction 2 Fitting a Straight Line Introduction Characteristics of a Straight Line Regression Notation The Least Squares Solution 3 Predicting Height from Shoe Size Creating a Fit Object Examining Summary Statistics Drawing the Regression Line Using the Regression Line 4 Partial Correlation An Example 5 Review Questions Question 1 Question 2 Question 3 James H. Steiger (Vanderbilt University) Introduction to Linear Regression 2 / 46

Introduction Introduction In this module, we discuss an extremely important technique in statistics Linear Regression. Linear regression is very closely related to correlation, and is extremely useful in a wide range of areas. James H. Steiger (Vanderbilt University) Introduction to Linear Regression 3 / 46

Introduction Introduction We begin by loading some data relating height to shoe size and drawing the scatterplot for the male data. > all.heights <- read.csv("shoesize.csv") > male.data <- all.heights[all.heights$gender == "M", ] #Select males > attach(male.data) #Make Variables Available James H. Steiger (Vanderbilt University) Introduction to Linear Regression 4 / 46

Introduction Introduction Next, we draw the scatterplot. The points align themselves in a linear pattern. > # Draw scatterplot > plot(size, Height, xlab = "Shoe Size", ylab = "Height in Inches") Height in Inches 65 70 75 80 8 10 12 14 Shoe Size James H. Steiger (Vanderbilt University) Introduction to Linear Regression 5 / 46

Introduction Introduction > cor(size, Height) [1] 0.7677 The correlation is an impressive 0.77. But how can we characterize the relationship between shoe size and height? In this case, linear regression is going to prove very useful. James H. Steiger (Vanderbilt University) Introduction to Linear Regression 6 / 46

Fitting a Straight Line Introduction Fitting a Straight Line Introduction If data are scattered around a straight line, then the relationship between the two variables can be thought of as being represented by that straight line, with some noise or error thrown in. We know that the correlation coefficient is a measure of how well the points will fit a straight line. But which straight line is best? James H. Steiger (Vanderbilt University) Introduction to Linear Regression 7 / 46

Fitting a Straight Line Introduction Fitting a Straight Line Introduction The key to understanding this is to realize the following: 1 Any straight line can be characterized by just two parameters, a slope and an intercept, and the equation for the straight line is Y = b 0 + b 1 Xa, where b 1 is the slope and b 0 is the intercept. 2 Any point can be characterized relative to a particular line in terms of two quantities: (a) where its X falls on a line, and (b) how far its Y is from the line in the vertical direction. Let s examine each of these preceding points. James H. Steiger (Vanderbilt University) Introduction to Linear Regression 8 / 46

Fitting a Straight Line Characteristics of a Straight Line Fitting a Straight Line Characteristics of a Straight Line The slope multiplies X, and so any change in X is multiplied by the slope and passed on to Y. Consequently, the slope represents the rise over the run, the amount by which Y increases for each unit increase in X. The intercept is, of course, the value of Y when X = 0. So if you have the slope and intercept, you have the line. James H. Steiger (Vanderbilt University) Introduction to Linear Regression 9 / 46

Fitting a Straight Line Characteristics of a Straight Line Fitting a Straight Line Characteristics of a Straight Line Suppose we draw a line any line in a plane. Then consider a point any point with respect to that line. What can we say? Let s use a concrete example. Suppose I draw the straight line whose equation is Y = 1.04X + 0.2 in a plane, and then plot the point (2, 3) by going over to 2 on the X -axis, then up to 3 on the Y -axis. James H. Steiger (Vanderbilt University) Introduction to Linear Regression 10 / 46

Fitting a Straight Line Characteristics of a Straight Line Fitting a Straight Line Characteristics of a Straight Line Y 0 1 2 3 4 5 (2,3) 0 1 2 3 4 5 X James H. Steiger (Vanderbilt University) Introduction to Linear Regression 11 / 46

Fitting a Straight Line Characteristics of a Straight Line Fitting a Straight Line Characteristics of a Straight Line Now suppose I were to try to use the straight line to predict the Y value of the point only from a knowledge of the X value of that point. The X value of the point is 2. If I substitute 2 for X in the formula Y = 1.04X + 0.2, I get Y = 2.28. This value lies on the line, directly above X. I ll draw that point on the scatterplot in blue. James H. Steiger (Vanderbilt University) Introduction to Linear Regression 12 / 46

Fitting a Straight Line Characteristics of a Straight Line Fitting a Straight Line Characteristics of a Straight Line Y 0 1 2 3 4 5 0 1 2 3 4 5 X James H. Steiger (Vanderbilt University) Introduction to Linear Regression 13 / 46

Fitting a Straight Line Characteristics of a Straight Line Fitting a Straight Line Characteristics of a Straight Line The Y value for the blue point is called the predicted value of Y, and is denoted Ŷ. Unless the actual point falls on the line, there will be some error in this prediction. The error is the discrepancy in the vertical direction from the line to the point. James H. Steiger (Vanderbilt University) Introduction to Linear Regression 14 / 46

Fitting a Straight Line Characteristics of a Straight Line Fitting a Straight Line Characteristics of a Straight Line Y 0 1 2 3 4 5 Y Y^ E 0 1 2 3 4 5 X James H. Steiger (Vanderbilt University) Introduction to Linear Regression 15 / 46

Fitting a Straight Line Regression Notation Fitting a Straight Line Regression Notation Now, let t generalize! We have just shown that, for any point with coordinates (X i, Y i ), relative to any line Y = b 0 + b 1 X, I may write Ŷ i = b 0 + b 1 X i (1) and Y i = Ŷ i + E i (2) with E i defined tautologically as E i = Y i Ŷ i (3) But we are not looking for any line. We are looking for the best line. And we have many points, not just one. And, by the way, what is the best line, and how do we find it? James H. Steiger (Vanderbilt University) Introduction to Linear Regression 16 / 46

Fitting a Straight Line The Least Squares Solution Fitting a Straight Line The Least Squares Solution It turns out, there are many possible ways of characterizing how well a line fits a set of points. However, one approach seems quite reasonable, and has many absolutely beautiful mathematical properties. This is the least squares criterion and the least squares solution for b 1 and b 0. James H. Steiger (Vanderbilt University) Introduction to Linear Regression 17 / 46

Fitting a Straight Line The Least Squares Solution Fitting a Straight Line The Least Squares Solution The least squares criterion states, the best-fitting line for a set of points is that line which minimizes the sum of squares of the E i for the entire set of points. Remember, the data points are there, plotted in the plane, nailed down, as it were. The only thing free to vary is the line, and it is characterized by just two parameters, the slope and intercept. For any slope b 1 and intercept b 0 I might choose, I can compute the sum of squared errors. And for any data set, the sum of squared errors is uniquely defined by that slope and intercept. The sum of squared errors is thus a function of b 1 and b 0. What we really have is a problem in minimizing a function of two unknowns. This is a routine problem in first-year calculus. We won t go through the proof of the least squares solution, we ll simply give you the result. James H. Steiger (Vanderbilt University) Introduction to Linear Regression 18 / 46

Fitting a Straight Line The Least Squares Solution Fitting a Straight Line The Least Squares Solution The solution to the least squares criterion is as follows b 1 = r y,x s y s x = s y,x s 2 x (4) and b 0 = M y b 1 M x (5) Note: If X and Y are both in Z score form, then b 1 = r y,x and b 0 = 0. Thus, once we remove the metric from the numbers, the very intimate connection between correlation and regression is revealed! James H. Steiger (Vanderbilt University) Introduction to Linear Regression 19 / 46

Predicting Height from Shoe Size Creating a Fit Object Predicting Height from Shoe Size Creating a Fit Object We could easily construct the slope and intercept of our regression line from summary statistics. But R actually has a facility to perform the entire analysis very quickly and automatically. You begin by producing a linear model fit object with the following syntax. > fit.object <- lm(height ~ Size) R is an object oriented language. That is, objects can contain data and when general functions are applied to an object, the object knows what to do. We ll demonstrate on the next slide. James H. Steiger (Vanderbilt University) Introduction to Linear Regression 20 / 46

Predicting Height from Shoe Size Examining Summary Statistics Predicting Height from Shoe Size Examining Summary Statistics R has a generic function called summary. Look what happens when we apply it to our fit object. > summary(fit.object) Call: lm(formula = Height ~ Size) Residuals: Min 1Q Median 3Q Max -7.289-1.112 0.066 1.356 5.824 Coefficients: Estimate Std. Error t value Pr(> t ) (Intercept) 52.5460 1.0556 49.8 <2e-16 *** Size 1.6453 0.0928 17.7 <2e-16 *** --- Signif. codes: 0 '***' 0.001 '**' 0.01 '*' 0.05 '.' 0.1 ' ' 1 Residual standard error: 2.02 on 219 degrees of freedom Multiple R-squared: 0.589, Adjusted R-squared: 0.588 F-statistic: 314 on 1 and 219 DF, p-value: <2e-16 James H. Steiger (Vanderbilt University) Introduction to Linear Regression 21 / 46

Predicting Height from Shoe Size Examining Summary Statistics Predicting Height from Shoe Size Examining Summary Statistics The coefficients for the intercept and slope are perhaps the most important part of the output. Here we see that the slope of the line is 1.6453 and the intercept is 52.5460. James H. Steiger (Vanderbilt University) Introduction to Linear Regression 22 / 46

Predicting Height from Shoe Size Examining Summary Statistics Predicting Height from Shoe Size Examining Summary Statistics Along with the estimates themselves, the program provides estimated standard errors of the coefficients, along with t statistics for testing the hypothesis that the coefficient is zero. James H. Steiger (Vanderbilt University) Introduction to Linear Regression 23 / 46

Predicting Height from Shoe Size Examining Summary Statistics Predicting Height from Shoe Size Examining Summary Statistics The program prints the R 2 value, also known as the coefficient of determination. When there is only one predictor, as in this case, the R 2 value is just r 2 x,y, the square of the correlation between height and shoe size. The adjusted R 2 value is an approximately unbiased estimator. With only one predictor, this can essentially be ignored, but with many predictors, it can be much lower than the standard R 2 estimate. The F -statistic tests that R 2 = 0 When there is only one predictor, it is the square of the t-statistic for testing that r x,y = 0. James H. Steiger (Vanderbilt University) Introduction to Linear Regression 24 / 46

Predicting Height from Shoe Size Examining Summary Statistics Predicting Height from Shoe Size Examining Summary Statistics James H. Steiger (Vanderbilt University) Introduction to Linear Regression 25 / 46

Predicting Height from Shoe Size Drawing the Regression Line Predicting Height from Shoe Size Drawing the Regression Line Now we draw the scatterplot with the best-fitting straight line. Notice how we draw the scatterplot first with the plot command, then draw the regression line in red with the abline command. > # draw scatterplot > plot(size, Height) > # draw regression line in red > abline(fit.object, col = "red") James H. Steiger (Vanderbilt University) Introduction to Linear Regression 26 / 46

Predicting Height from Shoe Size Drawing the Regression Line Predicting Height from Shoe Size Drawing the Regression Line Height 65 70 75 80 8 10 12 14 Size James H. Steiger (Vanderbilt University) Introduction to Linear Regression 27 / 46

Predicting Height from Shoe Size Using the Regression Line Predicting Height from Shoe Size Using the Regression Line We can now use the regression line to estimate a male student s height from his shoe size. Suppose a student s shoe size is 13. What is his predicted height? Ŷ = b 1 X + b 0 = (1.6453)(13) + 52.5460 = 73.9349 The predicted height is a bit less than 6 feet 2 inches. Of course, we know that not every student who has a size 13 show will have a height of 73.93. Some will be taller than that, some will be shorter. Is there something more we can say? James H. Steiger (Vanderbilt University) Introduction to Linear Regression 28 / 46

Predicting Height from Shoe Size Thinking about Residuals Predicting Height from Shoe Size Thinking about Residuals The predicted value Ŷ = 73.93 actually represents the average height of people with a shoe size of 13. According to the most commonly used linear regression model, people with a shoe size of 13 actually have a normal distribution with a mean of 73.93, and a standard deviation called the standard error of estimate. This quantity goes by several names, and in R output is called the residual standard error. An estimate of this quantity is included in the R regression output produced by the summary function. James H. Steiger (Vanderbilt University) Introduction to Linear Regression 29 / 46

Predicting Height from Shoe Size Thinking about Residuals Predicting Height from Shoe Size Thinking about Residuals James H. Steiger (Vanderbilt University) Introduction to Linear Regression 30 / 46

Predicting Height from Shoe Size Thinking about Residuals Predicting Height from Shoe Size Thinking about Residuals In the population, the standard error of estimate is calculated from the following formula σ e = 1 ρ 2 x,y σ y (6) In the sample, we estimate the standard error of estimate with the following formula n 1 s e = 1 rx,y n 2 2 s y (7) James H. Steiger (Vanderbilt University) Introduction to Linear Regression 31 / 46

Partial Correlation An Example Partial Correlation An Example Residuals can be thought of as The part of Y that is left over after that which can be predicted from X is partialled out. This notion has led to the concept of partial correlation. Let s introduce this notion in connection with an example. Suppose we gathered data on house fires in the Nashville area over the past month. We have data on two variables damage done by the fire, in thousands of dollars (Damage) and the number of fire trucks sent to the fire by the fire department (Trucks). Here are the data for the last 10 fires. James H. Steiger (Vanderbilt University) Introduction to Linear Regression 32 / 46

Partial Correlation An Example Partial Correlation An Example Trucks Damage 1 0 8 2 0 9 3 1 33 4 1 38 5 1 27 6 2 70 7 2 94 8 2 83 9 3 133 10 3 135 James H. Steiger (Vanderbilt University) Introduction to Linear Regression 33 / 46

Partial Correlation An Example Partial Correlation An Example Plotting the regression line, we see that there is indeed, a strong linear relationship between the number of fire trucks sent to a fire, and the damage done by the fire. > plot(trucks, Damage) > abline(lm(damage ~ Trucks), col = "red") James H. Steiger (Vanderbilt University) Introduction to Linear Regression 34 / 46

Partial Correlation An Example Partial Correlation An Example Damage 20 40 60 80 100 120 140 0.0 0.5 1.0 1.5 2.0 2.5 3.0 Trucks James H. Steiger (Vanderbilt University) Introduction to Linear Regression 35 / 46

Partial Correlation An Example Partial Correlation An Example The correlation between Trucks and Damage is 0.9779. Does this mean that the damage done by fire can be reduced by sending fewer trucks? Of course not. It turns out that the house fire records include another piece of information. Based on a complex rating system, each housefire has a rating based on the size of the conflagration. These ratings are in a variable called FireSize. On purely substantive and logical grounds, we might suspect that rather than fire trucks causing the damage, that this third variable, FireSize, causes both more damage to be done and more fire trucks to be sent. How can we investigate this notion statistically? James H. Steiger (Vanderbilt University) Introduction to Linear Regression 36 / 46

Partial Correlation An Example Partial Correlation An Example Suppose we predict Trucks from FireSize. The residuals represent that part of Trucks that isn t attributable to Firesize. Call these residuals E Trucks FireSize. Then suppose we predict Damage from Firesize. The residuals represent that part of Damage that cannot be predicted from FireSize. Call these residuals E Damage Firesize. The correlation between these two residual variables is called the partial correlation between Trucks and Damage with FireSize partialled out, and is denoted r Trucks,Damage FireSize. James H. Steiger (Vanderbilt University) Introduction to Linear Regression 37 / 46

Partial Correlation An Example Partial Correlation An Example There are several ways we can compute this partial correlation. One way is to compute the two residual variables discussed above, and then compute the correlation between them. > fit.1 <- lm(trucks ~ FireSize) > fit.2 <- lm(damage ~ FireSize) > E.1 <- residuals(fit.1) > E.2 <- residuals(fit.2) > cor(e.1, E.2) [1] -0.2163 James H. Steiger (Vanderbilt University) Introduction to Linear Regression 38 / 46

Partial Correlation An Example Partial Correlation An Example Another way is to use the textbook formula r x,y w = r x,y r x,w r y,w (1 rx,w 2 )(1 ry,w 2 ) (8) > r.xy <- cor(trucks, Damage) > r.xw <- cor(trucks, FireSize) > r.yw <- cor(damage, FireSize) > r.xy.given.w <- (r.xy - r.xw * r.yw)/sqrt((1 - r.xw^2) * (1 - r.yw^2)) > r.xy.given.w [1] -0.2163 James H. Steiger (Vanderbilt University) Introduction to Linear Regression 39 / 46

Partial Correlation An Example Partial Correlation An Example The partial correlation is 0.216. Once size of fire is accounted for, there is a negative correlation between number of fire trucks sent to the fire and damage done by the fire. James H. Steiger (Vanderbilt University) Introduction to Linear Regression 40 / 46

Review Questions Question 1 Review Questions Question 1 Recall that Ŷ = b 1 X + b 0, with b 1 = ρ YX σ Y /σ X. The predicted scores in Ŷ have a variance. Prove, using the laws of linear transformation, that this variance may be calculated as σ 2 Ŷ = ρ2 YX σ2 Y (9) James H. Steiger (Vanderbilt University) Introduction to Linear Regression 41 / 46

Review Questions Question 1 Review Questions Question 1 Answer. The laws of linear transformation state that, if Y = ax + b, then SY 2 = a2 SX 2. In other words, additive constants can be ignored, and multiplicative constants come straight through squared in the variance. Translating this idea to the Greek letter notation of the current problem, we find σ 2 Ŷ = σ 2 b 1 X +b 0 (10) = b1σ 2 X 2 (11) ( ρyx σ 2 ) Y = σx 2 (12) σ 2 X = ρ 2 YX σ2 Y (13) James H. Steiger (Vanderbilt University) Introduction to Linear Regression 42 / 46

Review Questions Question 2 Review Questions Question 2 In the lecture notes on linear combinations, we demonstrate that the covariance of two linear combinations may be derived by taking the algebraic product of the two linear combination expressions, and then applying a straightforward conversion rule. Using this approach, show that the covariance between Y and Ŷ is equal to the variance of Ŷ derived in the preceding problem. Hint. The covariance of Y and Ŷ is equal to the covariance of Y and b 1 X + b 0. James H. Steiger (Vanderbilt University) Introduction to Linear Regression 43 / 46

Review Questions Question 2 Review Questions Question 2 Answer. σ Y,b1 X +b 0 = σ Y,b1 X because additive constants never affect covariances. Now, applying the multiplicative rule, we find that σ Y,b1 X = b 1 σ Y,X (14) = ρ Y,X σ Y σ X ρ Y,X σ Y σ X (15) = ρ 2 Y,X σ2 Y (16) James H. Steiger (Vanderbilt University) Introduction to Linear Regression 44 / 46

Review Questions Question 3 Review Questions Question 3 Sarah got a Z score of +2.00 on the first midterm. If midterm 1 and midterm 2 are correlated 0.75 and the relationship is linear, what is the predicted Z-score of Sarah on the second exam? James H. Steiger (Vanderbilt University) Introduction to Linear Regression 45 / 46

Review Questions Question 3 Review Questions Question 3 Answer. When scores are in Z-score form, the formula for a predicted score is Ŷ = ρ Y,X X, and so Sarah s predicted score is (2)(0.75) = 1.50. This is a classic example of regression toward the mean. James H. Steiger (Vanderbilt University) Introduction to Linear Regression 46 / 46