Estimating σ 2. We can do simple prediction of Y and estimation of the mean of Y at any value of X.

Similar documents
Correlation Analysis

STAT Chapter 11: Regression

Ch 2: Simple Linear Regression

Chapter 16. Simple Linear Regression and dcorrelation

Linear Regression. Simple linear regression model determines the relationship between one dependent variable (y) and one independent variable (x).

Statistics for Managers using Microsoft Excel 6 th Edition

Chapter 16. Simple Linear Regression and Correlation

Keller: Stats for Mgmt & Econ, 7th Ed July 17, 2006

Measuring the fit of the model - SSR

Inferences for Regression

Business Statistics. Chapter 14 Introduction to Linear Regression and Correlation Analysis QMIS 220. Dr. Mohammad Zainal

Simple Linear Regression

Lecture 3: Inference in SLR

Inference for the Regression Coefficient

df=degrees of freedom = n - 1

Inference for Regression Simple Linear Regression

Regression Models - Introduction

Lecture 10 Multiple Linear Regression

Lectures on Simple Linear Regression Stat 431, Summer 2012

Chapter 4: Regression Models

Correlation and the Analysis of Variance Approach to Simple Linear Regression

Mathematics for Economics MA course

Chapter 13. Multiple Regression and Model Building

Basic Business Statistics 6 th Edition

Chapter 14 Simple Linear Regression (A)

Multiple Linear Regression

The Multiple Regression Model

Linear models and their mathematical foundations: Simple linear regression

Chapter 4. Regression Models. Learning Objectives

: The model hypothesizes a relationship between the variables. The simplest probabilistic model: or.

y ˆ i = ˆ " T u i ( i th fitted value or i th fit)

Chapter Learning Objectives. Regression Analysis. Correlation. Simple Linear Regression. Chapter 12. Simple Linear Regression

Formal Statement of Simple Linear Regression Model

Chapter 12 - Lecture 2 Inferences about regression coefficient

Data Analysis and Statistical Methods Statistics 651

CHAPTER EIGHT Linear Regression

Ch 3: Multiple Linear Regression

Regression Models. Chapter 4. Introduction. Introduction. Introduction

Simple linear regression

Unit 10: Simple Linear Regression and Correlation

Lecture 1 Linear Regression with One Predictor Variable.p2

Chapter 14 Student Lecture Notes Department of Quantitative Methods & Information Systems. Business Statistics. Chapter 14 Multiple Regression

9. Linear Regression and Correlation

Multiple Regression. Inference for Multiple Regression and A Case Study. IPS Chapters 11.1 and W.H. Freeman and Company

Chapter 7 Student Lecture Notes 7-1

The Simple Linear Regression Model

Psychology 282 Lecture #4 Outline Inferences in SLR

Regression Analysis. Regression: Methodology for studying the relationship among two or more variables

AMS 315/576 Lecture Notes. Chapter 11. Simple Linear Regression

Simple Linear Regression

Basic Business Statistics, 10/e

STA 108 Applied Linear Models: Regression Analysis Spring Solution for Homework #6

ECON The Simple Regression Model

3. Diagnostics and Remedial Measures

SSR = The sum of squared errors measures how much Y varies around the regression line n. It happily turns out that SSR + SSE = SSTO.

Section 3: Simple Linear Regression

Inference for Regression Inference about the Regression Model and Using the Regression Line

Lecture 11: Simple Linear Regression

Inference for Regression Inference about the Regression Model and Using the Regression Line, with Details. Section 10.1, 2, 3

Simple and Multiple Linear Regression

Simple Linear Regression. Material from Devore s book (Ed 8), and Cengagebrain.com

Variance. Standard deviation VAR = = value. Unbiased SD = SD = 10/23/2011. Functional Connectivity Correlation and Regression.

Homework 2: Simple Linear Regression

Review of Statistics 101

Chapter 13 Student Lecture Notes Department of Quantitative Methods & Information Systems. Business Statistics

Chapter 14 Student Lecture Notes 14-1

Estadística II Chapter 4: Simple linear regression

Lecture 2 Linear Regression: A Model for the Mean. Sharyn O Halloran

Overview Scatter Plot Example

Applied Regression Analysis

The simple linear regression model discussed in Chapter 13 was written as

SIMPLE REGRESSION ANALYSIS. Business Statistics

Lecture notes on Regression & SAS example demonstration

Applied Regression Modeling: A Business Approach Chapter 2: Simple Linear Regression Sections

Outline. Remedial Measures) Extra Sums of Squares Standardized Version of the Multiple Regression Model

Multiple Regression Analysis: Heteroskedasticity

STA121: Applied Regression Analysis

Simple Linear Regression

Lecture 2 Simple Linear Regression STAT 512 Spring 2011 Background Reading KNNL: Chapter 1

Simple Linear Regression

Chapter 3 Multiple Regression Complete Example

Simple Linear Regression

401 Review. 6. Power analysis for one/two-sample hypothesis tests and for correlation analysis.

Statistical Techniques II EXST7015 Simple Linear Regression

STAT 511. Lecture : Simple linear regression Devore: Section Prof. Michael Levine. December 3, Levine STAT 511

Inference for Regression

Lecture 9: Linear Regression

Chapter 3. Diagnostics and Remedial Measures

Section 4: Multiple Linear Regression

Scatter plot of data from the study. Linear Regression

ECO220Y Simple Regression: Testing the Slope

Lecture 9 SLR in Matrix Form

K. Model Diagnostics. residuals ˆɛ ij = Y ij ˆµ i N = Y ij Ȳ i semi-studentized residuals ω ij = ˆɛ ij. studentized deleted residuals ɛ ij =

Applied Econometrics (QEM)

Econ 3790: Business and Economics Statistics. Instructor: Yogesh Uppal

Density Temp vs Ratio. temp

6. Multiple Linear Regression

Ordinary Least Squares Regression Explained: Vartanian

DESAIN EKSPERIMEN Analysis of Variances (ANOVA) Semester Genap 2017/2018 Jurusan Teknik Industri Universitas Brawijaya

Chapter 8 Heteroskedasticity

Transcription:

Estimating σ 2 We can do simple prediction of Y and estimation of the mean of Y at any value of X. To perform inferences about our regression line, we must estimate σ 2, the variance of the error term. For a random variable Y, the estimated variance is: In regression, the estimated variance of Y (and also of ε) is: ( Yˆ) 2 Y is called the error (residual) sum of squares (SSE). It has n 2 degrees of freedom. The ratio MSE = SSE / df is called the mean squared error.

MSE is an unbiased estimate of the error variance σ 2. Also, MSE serves as an estimate of the error standard deviation σ. Partitioning Sums of Squares If we did not use X in our model, our estimate for the mean of Y would be: Picture: For each data point: Y Y = difference between observed Y and sample mean Y-value Y Yˆ = difference between observed Y and predicted Y-value Yˆ Y = difference between predicted Y and sample mean Y-value It can be shown:

TSS = overall variation in the Y-values SSR = variation in Y accounted for by regression line SSE = extra variation beyond what the regression relationship accounts for Computational Formulas: ( Y ) 2 TSS = S YY = Y n SSR = (S XY ) 2 / S XX = βˆ1s 2 XY SSE = S YY (S XY ) 2 / S XX = SYY 1 βˆ S XY Case (1): If SSR is a large part of TSS, the regression line accounts for a lot of the variation in Y. Case (2): If SSE is a large part of TSS, the regression line is leaving a great deal of variation unaccounted for. ANOVA test for β 1 If the SLR model is useless in explaining the variation in Y, then Y is just as good at estimating the mean of Y as Yˆ is. => true β 1 is zero and X doesn t belong in model Corresponds to case (2) above.

But if (1) is true, and the SLR model explains a lot of the variation in Y, we would conclude β 1 0. How to compare SSR to SSE to determine if (1) or (2) is true? Divide by their degrees of freedom. For the SLR model: We test: If MSR much bigger than MSE, conclude H a. Otherwise we cannot conclude H a. The ratio F* = MSR / MSE has an F distribution with df = (1, n 2) when H 0 is true. Thus we reject H 0 when where α is the significance level of our hypothesis test.

t-test of H 0 : β 1 = 0 Note: β 1 is a parameter (a fixed but unknown value) ˆ β The estimate 1 is a random variable (a statistic calculated from sample data). Therefore ˆ β 1 has a sampling distribution: ˆ β is an unbiased estimator of β1. ˆ β 1 1 estimates β1 with greater precision when: the true variance of Y is small. the sample size is large. the X-values in the sample are spread out. Standardizing, we see that: Problem: σ 2 is typically unknown. We estimate it with MSE. Then:

To test H 0 : β 1 = 0, we use the test statistic: Advantages of t-test over F-test: (1) Can test whether the true slope equals any specified value (not just 0). Example: To test H 0 : β 1 = 10, we use: (2) Can also use t-test for a one-tailed test, where: H a : β 1 < 0 or H a : β 1 > 0. H a Reject H 0 if:

(3) The value estimate. MSE S measures the precision of XX Confidence Interval for β 1 βˆ 1 as an βˆ 1 The sampling distribution of provides a confidence interval for the true slope β 1 : Example (House price data): Recall: S YY = 93232.142, S XY = 1275.494, S XX = 22.743 Our estimate of σ 2 is MSE = SSE / (n 2) SSE = MSE = and recall To test H 0 : β 1 = 0 vs. H a : β 1 0 (at α = 0.05)

Table A2: t.025 (56) 2.004. With 95% confidence, the true slope falls in the interval Interpretation: Inference about the Response Variable We may wish to: (1) Estimate the mean value of Y for a particular value of X. Example: (2) Predict the value of Y for a particular value of X. Example: The point estimates for (1) and (2) are the same: The value of the estimated regression function at X = 1.75. Example:

Variability associated with estimates for (1) and (2) is quite different. Var [ Eˆ( Y X )] = Var Yˆ [ pred ] = Since σ 2 is unknown, we estimate σ 2 with MSE: CI for E(Y X) at x*: Prediction Interval for Y value of a new observation with X = x*: Example: 95% CI for mean selling price for houses of 1750 square feet:

Example: 95% PI for selling price of a new house of 1750 square feet: βˆ Correlation 1 tells us something about whether there is a linear relationship between Y and X. Its value depends on the units of measurement for the variables. The correlation coefficient r and the coefficient of determination r 2 are unit-free numerical measures of the linear association between two variables. r = (measures strength and direction of linear relationship) r always between -1 and 1: r > 0

r < 0 r = 0 r near -1 or 1 r near 0 Correlation coefficient (1) makes no distinction between independent and dependent variables, and (2) requires variables to be numerical. Examples: House data: Note that s s estimated slope. Y X r = 1 ˆβ so r always has the same sign as the The population correlation coefficient is denoted ρ. Test of H 0 : ρ = 0 is equivalent to test of H 0 : β 1 = 0 in SLR (p-value will be the same) Software will give us r and the p-value for testing H 0 : ρ = 0 vs. H a : ρ 0. To test whether ρ is some nonzero value, need to use transformation see p. 318.

The square of r, denoted r 2, also measures strength of linear relationship. Definition: r 2 = SSR / TSS. Interpretation of r 2 : It is the proportion of overall sample variability in Y that is explained by its linear relationship with X. 2 ( n 2) r Note: In SLR, F = 2. 1 r Hence: large r 2 large F statistic significant linear relationship between Y and X. Example (House price data): Interpretation: Regression Diagnostics We assumed various things about the random error term. How do we check whether these assumptions are satisfied? The (unobservable) error term for each point is:

As estimated errors we use the residuals for each data point: Residual plots allow us to check for four types of violations of our assumptions: (1) The model is misspecified (linear trend between Y and X incorrect) (2) Non-constant error variance (spread of errors changes for different values of X) (3) Outliers exist (data values which do not fit overall trend) (4) Non-normal errors (error term is not (approx.) normally distributed) A residual plot plots the residuals Y Yˆ against the predicted values Yˆ. If this residual plot shows random scatter, this is good. If there is some notable pattern, there is a possible violation of our model assumptions. Pattern Violation

We can verify whether the errors are approximately normal with a Q-Q plot of the residuals. If Q-Q plot is roughly a straight line the errors may be assumed to be normal. Example (House data): Remedies for Violations Transforming Variables When the residual plot shows megaphone shape (non-constant error variance) opening to the right, we can use a variance-stabilizing transformation of Y. Picture: Let Y * = log( Y ) or Y * = Y and use Y* as the dependent variable.

These transformations tend to reduce the spread at high values of Yˆ. Transformations of Y may also help when the error distribution appears non-normal. Transformations of X and/or of Y can help if the residual plot shows evidence of a nonlinear trend. Depending on the situation, one or more of these transformations may be useful: Drawback: Interpretations, predictions, etc., are now in terms of the transformed variables. We must reverse the transformations to get a meaningful prediction. Example (Surgical data):