Simple Linear Regression

Similar documents
Ch 2: Simple Linear Regression

Multiple Linear Regression

Applied Regression. Applied Regression. Chapter 2 Simple Linear Regression. Hongcheng Li. April, 6, 2013

Measuring the fit of the model - SSR

Ch 3: Multiple Linear Regression

Correlation Analysis

Simple Linear Regression

Estimating σ 2. We can do simple prediction of Y and estimation of the mean of Y at any value of X.

Lecture 14 Simple Linear Regression

Linear Regression Model. Badr Missaoui

ECO220Y Simple Regression: Testing the Slope

Chapter 12 - Lecture 2 Inferences about regression coefficient

STAT 3A03 Applied Regression With SAS Fall 2017

Chapter 14 Simple Linear Regression (A)

Linear models and their mathematical foundations: Simple linear regression

Linear Models and Estimation by Least Squares

Simple Linear Regression

Inference for Regression

Lecture 6 Multiple Linear Regression, cont.

Applied Regression Analysis

STAT420 Midterm Exam. University of Illinois Urbana-Champaign October 19 (Friday), :00 4:15p. SOLUTIONS (Yellow)

Math 3330: Solution to midterm Exam

TMA4255 Applied Statistics V2016 (5)

STAT5044: Regression and Anova. Inyoung Kim

Chapter 1: Linear Regression with One Predictor Variable also known as: Simple Linear Regression Bivariate Linear Regression

Simple and Multiple Linear Regression

Inferences for Regression

Weighted Least Squares

Review of probability and statistics 1 / 31

Sample Problems. Note: If you find the following statements true, you should briefly prove them. If you find them false, you should correct them.

Lecture 18: Simple Linear Regression

STAT 511. Lecture : Simple linear regression Devore: Section Prof. Michael Levine. December 3, Levine STAT 511

13 Simple Linear Regression

MAT2377. Rafa l Kulik. Version 2015/November/26. Rafa l Kulik

Regression Review. Statistics 149. Spring Copyright c 2006 by Mark E. Irwin

LECTURE 6. Introduction to Econometrics. Hypothesis testing & Goodness of fit

Multivariate Regression Analysis

Chapter 14. Linear least squares

Inference for Regression Simple Linear Regression

Linear regression. We have that the estimated mean in linear regression is. ˆµ Y X=x = ˆβ 0 + ˆβ 1 x. The standard error of ˆµ Y X=x is.

Lectures on Simple Linear Regression Stat 431, Summer 2012

Applied Econometrics (QEM)

F3: Classical normal linear rgression model distribution, interval estimation and hypothesis testing

Simple Linear Regression

Statistical View of Least Squares

Regression Analysis II

Correlation and the Analysis of Variance Approach to Simple Linear Regression

Simple linear regression

Chapter 1 Linear Regression with One Predictor

Review of Econometrics

Homework 2: Simple Linear Regression

Categorical Predictor Variables

Overview Scatter Plot Example

Section 3: Simple Linear Regression

Master s Written Examination - Solution

STAT2012 Statistical Tests 23 Regression analysis: method of least squares

Multiple Regression. Inference for Multiple Regression and A Case Study. IPS Chapters 11.1 and W.H. Freeman and Company

ECON The Simple Regression Model

The Simple Linear Regression Model

Inference in Regression Analysis

Section 4.6 Simple Linear Regression

Linear Regression. Simple linear regression model determines the relationship between one dependent variable (y) and one independent variable (x).

Statistics for Managers using Microsoft Excel 6 th Edition

Chapter 14 Student Lecture Notes Department of Quantitative Methods & Information Systems. Business Statistics. Chapter 14 Multiple Regression

Ma 3/103: Lecture 25 Linear Regression II: Hypothesis Testing and ANOVA

Problems. Suppose both models are fitted to the same data. Show that SS Res, A SS Res, B

Chapte The McGraw-Hill Companies, Inc. All rights reserved.

Correlation and Regression

TABLES AND FORMULAS FOR MOORE Basic Practice of Statistics

Business Statistics. Chapter 14 Introduction to Linear Regression and Correlation Analysis QMIS 220. Dr. Mohammad Zainal

Multivariate Regression

Oct Simple linear regression. Minimum mean square error prediction. Univariate. regression. Calculating intercept and slope

STAT 4385 Topic 03: Simple Linear Regression

Lecture 16 - Correlation and Regression

df=degrees of freedom = n - 1

R 2 and F -Tests and ANOVA

Confidence Intervals, Testing and ANOVA Summary

ECON 450 Development Economics

AMS 315/576 Lecture Notes. Chapter 11. Simple Linear Regression

MS&E 226: Small Data

From last time... The equations

Inference for Regression Inference about the Regression Model and Using the Regression Line

The Multiple Regression Model

Estadística II Chapter 4: Simple linear regression

2. Regression Review

Simple Regression Model Setup Estimation Inference Prediction. Model Diagnostic. Multiple Regression. Model Setup and Estimation.

Econ 3790: Business and Economics Statistics. Instructor: Yogesh Uppal

Basic Business Statistics 6 th Edition

Chapter 4: Regression Models

Simple linear regression

Formal Statement of Simple Linear Regression Model

Multivariate Regression (Chapter 10)

MATH 644: Regression Analysis Methods

Econometrics I KS. Module 1: Bivariate Linear Regression. Alexander Ahammer. This version: March 12, 2018

where x and ȳ are the sample means of x 1,, x n

Business Statistics. Tommaso Proietti. Linear Regression. DEF - Università di Roma 'Tor Vergata'

BNAD 276 Lecture 10 Simple Linear Regression Model

Scatter plot of data from the study. Linear Regression

Simple Linear Regression

Statistics for Engineers Lecture 9 Linear Regression

Transcription:

Simple Linear Regression In simple linear regression we are concerned about the relationship between two variables, X and Y. There are two components to such a relationship. 1. The strength of the relationship. 2. The direction of the relationship. We shall also be interested in making inferences about the relationship. We will be assuming here that the relationship between X and Y is linear (or has been linearized through transformation). 29

Covariance Suppose that (X, Y ) is a bivariate random vector. An important characteristic of the joint distribution of this random vector is the covariance. The theoretical definition of the covariance is Cov(X, Y ) = E [(X µ X )(Y µ Y )] where µ X = E[X] and µ Y = E[Y ]. The sign of Cov(X, Y ) tells us the direction of the relationship between X and Y. 30

Correlation A problem with the covariance is that it s value depends on the units in which X and Y are measured and so is not a good measure of strength of the relationship. To remove the effect of units we standardise the variables. Z X = X µ X Var(X) Z Y = Y µ Y Var(Y ) Both Z X and Z Y have mean 0 and variance 1. The covariance between Z X and Z Y ρ X,Y = Cov(Z X, Z Y ) = is called the correlation. Cov(X, Y ). Var(X)Var(Y ) 31

Correlation (ctd) It can be shown that 1 ρ X,Y 1. The sign of ρ is the same as the sign of the covariance and so gives the direction of the relationship. When the relationship is perfectly linear then ρ = 1. If the two variables are independent then ρ = 0. NOTE The inverse of this does not hold The strength of the relationship between the variables can be assessed by ρ (or ρ 2 ). 32

The Sample Correlation The sample estimate of ρ based on the random sample (x 1, y 1 ), (x 2, y 2 ),..., (x n, y n ) is (xi x)(y r X,Y = i y) (xi x) 2 = Sxy. (y i y) 2 Sxx Syy This sample quantity also satisfies that 1 < r < 1. It is important to note that r only measures the strength of the linear relationship between the two observed variables. When a linear relationship between X and Y is plausible, r gives us a very good indication of the strength and direction of that relationship. 33

A test for ρ = 0 When ρ = 0 and the joint distribution of (X, Y ) is bivariate normal it can be shown that t = r n 2 1 r 2 has a Student s t distribution with n 2 degrees of freedom. We can therefore test H 0 : ρ = 0 V H 1 : ρ 0 by calculating the observed value of t (call this t obs ). A p-value for the test can then be found as p = 2P(t n 2 t obs ) 34

Simple Linear Regression Correlation is an attribute of the joint distribution of (X, Y ). Cov(X, Y ) = Cov(Y, X) and ρ X,Y = ρ Y,X. Regression is based on the conditional distribution of Y given X. Regression is generally not symmetric so it matters which variable is called Y (the response) and which is called X (the covariate). Since X is considered fixed in regression, it is not necessary that it be a random variable at all. 35

The Simple Linear Model The assumptions of the model are 1. A linear relationship Y = β 0 + β 1 X + ε exists between X and Y. 2. E[ε X = x] = 0 and Var(ε X = x) = σ 2 for every x. 3. ε 1,..., ε n is a random sample from a N(0, σ 2 ) distribution. In terms of Y this means that the conditional distribution Y X = x = N(β 0 + β 1 x, σ 2 ) 36

Fitted Values and Errors Suppose that we have a dataset (y 1, x 1 ), (y 2, x 2 ),..., (y n, x n ). Our interest is in using our model to predict values of Y for any given value of X = x. If we know the values of β 0 and β 1 then the fitted value for the observation y i would be β 0 + β 1 x i. The error in the fitted value can be measured by the vertical distance ε i = y i β 0 β 1 x i We would like to make these errors as small as possible. 37

Estimating the Parameters Since we do not want negative errors to cancel out positive errors we use the sum of squared errors S(β 0, β 1 ) = n i=1 ε 2 i = n i=1 (y i β 0 β 1 x i ) 2 to be an overall measure of the fit of the line. The Method of Least Squares is an estimation method which estimates β 0 and β 1 as those values which minimize S(β 0, β 1 ). 38

Least Squares Estimates Theorem 1 Suppose we have a dataset (y 1, x 1 ), (y 2, x 2 ),..., (y n, x n ) for which the simple linear model holds. Then the least squares estimates of β 1 and β 0 are given by (yi y)(x ˆβ 1 = i x) (xi x) 2 = Sxy Sxx ˆβ 0 = y ˆβ 1 x The Least Squares Regression Line is then Ŷ = ˆβ 0 + ˆβ 1 X. 39

Residuals For each observation in our dataset we can compute the fitted value ŷ i = ˆβ 0 + ˆβ 1 x i i = 1,..., n. The vertical distance from the observed y i to the fitted value is called the residual e i = y i ŷ i = y i ˆβ 0 ˆβ 1 x i i = 1,..., n. The residuals can be thought of as predicted values of the unknown errors ε 1,..., ε n. 40

Properties of the Least Squares Line The least squares line always passes through the point (x, y). The estimated slope ˆβ 1 always has the same sign as the sample correlation between (x 1,..., x n ) and (y 1,..., y n ). The sum of the residuals is 0. The sum of the squares of the e i s is called the Residual Sum of Squares or Sum of Squared Errors (SSE). An unbiased estimator of σ 2 is ˆσ 2 e 2 = i n 2 = SSE n 2. 41

Theoretical Properties Theorem 2 Suppose that the linear model holds and that ˆβ 0 and ˆβ 1 are the least squares estimators given in Theorem 1. Then 1. E[ˆβ 0 x 1,..., x n ] = β 0 and E[ˆβ 1 x 1,..., x n ] = β 1. [ ] 2. 1 Var(ˆβ 0 x 1,..., x n ) = n + x2 σ 2 Sxx Var(ˆβ 1 x 1,..., x n ) = σ2 Sxx 3. Cov(ˆβ 0, ˆβ 1 ) = x Sxx σ2. 42

Standard Errors The variances in Theorem 2 depend on the unknown value of σ 2. If we replace σ 2 with the unbiased estimator ˆσ 2 and take the square root we get the standard errors of the estimators. s.e.(ˆβ 0 ) = s.e.(ˆβ 1 ) = [ 1 n + x2 Sxx ˆσ 2 Sxx ] ˆσ 2 The software will always report both the estimates and their standard errors. 43

Sampling Distributions The following theorem will not be proven in this course but the result is very important. Theorem 3 Suppose that the linear model Y i = β 0 + β 1 x i + ε i holds and ε 1,..., ε n are iid N(0, σ 2 ). Then ˆβ 0 β 0 s.e.(ˆβ 0 ) t n 2 ˆβ 1 β 1 s.e.(ˆβ 1 ) t n 2 44

Confidence Intervals We can use Theorem 3 to find confidence intervals for β 0 and β 1. Let t (n 2.α/2) be the (1 α/2) percentile of the t distribution with n 2 degrees of freedom. Table A.2 in your textbook gives the values of this percentile for various values of α/2 and degrees of freedom n 2. Confidence intervals for β 0 and β 1 are then ˆβ 0 ± t (n 2.α/2) s.e.(ˆβ 0 ) ˆβ 1 ± t (n 2.α/2) s.e.(ˆβ 1 ) 45

Hypothesis Testing Suppose that we wish to test the hypotheses H 0 : β 1 = β 0 1 V H 1 : β 1 β 0 1. From Theorem 3 we see that, if H 0 is true then t 1 = ˆβ 1 β 0 1 s.e.(ˆβ 1 ) t n 2. We can therefore find the observed value of t 1 and calculate the p-value, p 1 = 2P(t n 2 t 1,obs ). When p 1 is very small we reject H 0. We can do similar to test hypotheses about the intercept β 0. 46

Hypothesis Testing (ctd) Of particular interest are the tests of H 0 0 : β 0 = 0 V H 0 1 : β 0 0 H 1 0 : β 1 = 0 V H 1 1 : β 1 0 The test statistics for these two tests are the ratios of the estimates to their standard errors. Software generally includes the values of these test statistics as well as the associated p-values. 47

The ANOVA Table Another part of the software output is the ANOVA table. For example you might have Analysis of Variance Sum of Mean Source DF Squares Square F Value Pr > F Model 1 27420 27420 943.20 <.0001 Error 12 348.84837 29.07070 Corrected Total 13 27768 48

Sums of Squares The sums of squares are Total Sum of Squares (SST) = Error Sum of Squares (SSE) = Model Sum of Squares (SSR) = n i=1 n i=1 n i=1 (y i y) 2 (y i ŷ i ) 2 (ŷ i y) 2 It is not hard to see that n i=1 (y i y) 2 = n i=1 (ŷ i y) 2 + n i=1 (y i ŷ i ) 2. 49

The ANOVA Table (ctd) The Total degrees of freedom are always n 1. For simple regression the degrees of freedom used by the model is 1. Hence the Error degrees of freedom is n 1 1 = n 2. The Mean Square column is the Sum of Squares divided by the degrees of freedom. Note that the Error Mean Square (MSE) is ˆσ 2. 50

The F test for the Model The F Value in the ANOVA table is the ratio of the Model Mean Square over MSE. If the model is not useful in predicting Y then this ratio has an F distribution with 1 and n 2 degrees of freedom. We can get a p-value for the test that the model is not useful by looking at the tail probability that an F 1,n 2 random variable is greater than the observed F value. In simple regression, this is equivalent to the test of β 1 = 0 against β 1 0. 51

The Coefficient of Determination The total sum of squares is a measure of the variability in y 1,..., y n without taking the covariate into account. The error sum of squares is the amount of variability left after fitting a linear regression for the covariate. The model sum of squares is the amount of variability explained by the model. The proportion of the variability explained by the model is R 2 = SSR SST = 1 SSE SST In simple regression R 2 is the square of the sample correlation between x 1,..., x n and y 1,..., y n. 52

Predictions Suppose that x 0 is some new value of x for which we want to do prediction. There are two types of prediction that are of interest to us 1. Estimation of µ 0 = E[Y X = x 0 ]. 2. Prediction of a Y value for an individual with X = x 0. We can use our fitted regression model to do both of these. 53

Estimation of µ 0 By the linear model the true value of µ 0 is β 0 + β 1 x 0. An obvious estimator of µ 0 is then ˆµ 0 = ˆβ 0 + ˆβ 1 x 0. From Theorem 2 we see that ˆµ 0 is unbiased and ( 1 Var(ˆµ 0 ) = n + (x 0 x) 2 ) σ 2. Sxx A confidence interval for µ 0 is ˆµ 0 ± t (n 2,α/2) s.e.(ˆµ 0 ) 54

Predicting an Individual Value The value of Y for an individual with X = x 0 is Y 0 = β 0 + β 1 x 0 + ε 0 We can plug in the estimators ˆβ 0 and ˆβ 1 and take ε 0 to be equal to its mean (E[ε 0 ] = 0) to get the predicted value ŷ 0 = ˆβ 0 + ˆβ 1 x o = ˆµ 0 The Variance of Ŷ 0, however is Var(Ŷ 0 ) = Var(ˆµ 0 ) + Var(ε 0 ) = ( 1 + 1 n + (x 0 x) 2 Sxx ) σ 2. A prediction interval for an individual with x = x 0 is then ŷ 0 ± t (n 2,α/2) s.e.(ŷ 0 ) 55