residual. (Note that usually in descriptions of regression analysis, upper-case

Similar documents
Linear Regression with One Regressor

Simple Linear Regression

12.2 Estimating Model parameters Assumptions: ox and y are related according to the simple linear regression model

ENGI 3423 Simple Linear Regression Page 12-01

Chapter 13 Student Lecture Notes 13-1

Chapter Business Statistics: A First Course Fifth Edition. Learning Objectives. Correlation vs. Regression. In this chapter, you learn:

Objectives of Multiple Regression

The equation is sometimes presented in form Y = a + b x. This is reasonable, but it s not the notation we use.

Multiple Linear Regression Analysis

Lecture 8: Linear Regression

Multiple Regression. More than 2 variables! Grade on Final. Multiple Regression 11/21/2012. Exam 2 Grades. Exam 2 Re-grades

STA302/1001-Fall 2008 Midterm Test October 21, 2008

b. There appears to be a positive relationship between X and Y; that is, as X increases, so does Y.

Statistics MINITAB - Lab 5

Lecture 2: Linear Least Squares Regression

ECON 482 / WH Hong The Simple Regression Model 1. Definition of the Simple Regression Model

Example: Multiple linear regression. Least squares regression. Repetition: Simple linear regression. Tron Anders Moger

CLASS NOTES. for. PBAF 528: Quantitative Methods II SPRING Instructor: Jean Swanson. Daniel J. Evans School of Public Affairs


Chapter Two. An Introduction to Regression ( )

Simple Linear Regression

Multiple Choice Test. Chapter Adequacy of Models for Regression

: At least two means differ SST

Simple Linear Regression and Correlation.

Simple Linear Regression - Scalar Form

Ordinary Least Squares Regression. Simple Regression. Algebra and Assumptions.

Probability and. Lecture 13: and Correlation

Simple Linear Regression and Correlation. Applied Statistics and Probability for Engineers. Chapter 11 Simple Linear Regression and Correlation

Linear Regression Siana Halim

Lecture 7. Confidence Intervals and Hypothesis Tests in the Simple CLR Model

Chapter 14 Logistic Regression Models

ESS Line Fitting

Lecture Notes Types of economic variables

Statistics: Unlocking the Power of Data Lock 5

Transforming Numerical Methods Education for the STEM Undergraduate Torque (N-m)

Statistics. Correlational. Dr. Ayman Eldeib. Simple Linear Regression and Correlation. SBE 304: Linear Regression & Correlation 1/3/2018

Chapter 2 Supplemental Text Material

Lecture 1: Introduction to Regression

Summary of the lecture in Biostatistics

Johns Hopkins University Department of Biostatistics Math Review for Introductory Courses

Chapter 2 Simple Linear Regression

Johns Hopkins University Department of Biostatistics Math Review for Introductory Courses

Statistics Review Part 3. Hypothesis Tests, Regression

Econometric Methods. Review of Estimation

Midterm Exam 1, section 1 (Solution) Thursday, February hour, 15 minutes

COV. Violation of constant variance of ε i s but they are still independent. The error term (ε) is said to be heteroscedastic.

Chapter 8. Inferences about More Than Two Population Central Values

REVIEW OF SIMPLE LINEAR REGRESSION SIMPLE LINEAR REGRESSION

Midterm Exam 1, section 2 (Solution) Thursday, February hour, 15 minutes

STA 108 Applied Linear Models: Regression Analysis Spring Solution for Homework #1

ENGI 4421 Propagation of Error Page 8-01

Chapter 3 Multiple Linear Regression Model

Applied Statistics and Probability for Engineers, 5 th edition February 23, b) y ˆ = (85) =

Correlation and Simple Linear Regression

Linear Regression. Hsiao-Lung Chan Dept Electrical Engineering Chang Gung University, Taiwan

Regression. Linear Regression. A Simple Data Display. A Batch of Data. The Mean is 220. A Value of 474. STAT Handout Module 15 1 st of June 2009

C. Statistics. X = n geometric the n th root of the product of numerical data ln X GM = or ln GM = X 2. X n X 1

Econ 388 R. Butler 2016 rev Lecture 5 Multivariate 2 I. Partitioned Regression and Partial Regression Table 1: Projections everywhere

Fundamentals of Regression Analysis

Handout #8. X\Y f(x) 0 1/16 1/ / /16 3/ / /16 3/16 0 3/ /16 1/16 1/8 g(y) 1/16 1/4 3/8 1/4 1/16 1

best estimate (mean) for X uncertainty or error in the measurement (systematic, random or statistical) best

Lecture 1: Introduction to Regression

Example. Row Hydrogen Carbon

QA 622 AUTUMN QUARTER ACADEMIC YEAR

Topic 4: Simple Correlation and Regression Analysis

STATISTICAL PROPERTIES OF LEAST SQUARES ESTIMATORS. x, where. = y - ˆ " 1

Functions of Random Variables

Third handout: On the Gini Index

Simple Linear Regression Analysis

4. Standard Regression Model and Spatial Dependence Tests

X X X E[ ] E X E X. is the ()m n where the ( i,)th. j element is the mean of the ( i,)th., then

THE ROYAL STATISTICAL SOCIETY HIGHER CERTIFICATE

y y ˆ i i Difference between

Linear Regression. Hsiao-Lung Chan Dept Electrical Engineering Chang Gung University, Taiwan

Wu-Hausman Test: But if X and ε are independent, βˆ. ECON 324 Page 1

LINEAR REGRESSION ANALYSIS

Handout #6. X\Y f(x) 0 1/16 1/ / /16 3/ / /16 3/16 0 3/ /16 1/16 1/8 g(y) 1/16 1/4 3/8 1/4 1/16 1

ECONOMETRIC THEORY. MODULE VIII Lecture - 26 Heteroskedasticity

Reaction Time VS. Drug Percentage Subject Amount of Drug Times % Reaction Time in Seconds 1 Mary John Carl Sara William 5 4

Recall MLR 5 Homskedasticity error u has the same variance given any values of the explanatory variables Var(u x1,...,xk) = 2 or E(UU ) = 2 I

Previous lecture. Lecture 8. Learning outcomes of this lecture. Today. Statistical test and Scales of measurement. Correlation

UNIVERSITY OF TORONTO AT SCARBOROUGH. Sample Exam STAC67. Duration - 3 hours

Sum Mean n

Analyzing Two-Dimensional Data. Analyzing Two-Dimensional Data

Multivariate Transformation of Variables and Maximum Likelihood Estimation

Maximum Likelihood Estimation

ENGI 4421 Joint Probability Distributions Page Joint Probability Distributions [Navidi sections 2.5 and 2.6; Devore sections

TESTS BASED ON MAXIMUM LIKELIHOOD

Topic 9. Regression and Correlation

Chapter 13, Part A Analysis of Variance and Experimental Design. Introduction to Analysis of Variance. Introduction to Analysis of Variance

This is the linearity assumption. The expected value of Y is really a straight-line function of x.

UNIVERSITY OF OSLO DEPARTMENT OF ECONOMICS

BIOSTATISTICS TOPIC 8: ANALYSIS OF CORRELATIONS I. SIMPLE LINEAR REGRESSION

Based on Neter, Wasserman and Whitemore: Applied Statistics, Chapter 18, pp

STA 105-M BASIC STATISTICS (This is a multiple choice paper.)

CHAPTER VI Statistical Analysis of Experimental Data

Understand the properties of the parameter estimates, how we compute confidence intervals for them and how we test hypotheses about them.

Homework Solution (#5)

Module 7: Probability and Statistics

hp calculators HP 30S Statistics Averages and Standard Deviations Average and Standard Deviation Practice Finding Averages and Standard Deviations

Transcription:

Regresso Aalyss Regresso aalyss fts or derves a model that descres the varato of a respose (or depedet ) varale as a fucto of oe or more predctor (or depedet ) varales. The geeral regresso model s oe of several that share the same asc coceptual model data = systematc compoet + rregular compoet where the systematc compoet s predctale or explaale y the predctor varales, ad s represeted y the regresso model, whle the rregular compoet s regarded as ose or predcto errors varatos the respose varale that ca ot e accouted for y the predctor varales. The regresso equato The specfc varte, lear, regresso model s Y X 0 where Y s the c value of the depedet or respose varale, 0 ad are the coeffcets of the regresso le (also kow as the slope ( ) ad tercept ( 0 )), s the -th depedet or predctor varale, ad s the predcto error, or ose or resdual. (Note that usually descrptos of regresso aalyss, upper-case X X s ad Y s stad for raw data values, whle lower case x s ad y s stad for devatos of X ad Y aout ther respectve meas,.e. x X X y Y Y There are several alteratve ways of wrtg the regresso equato or model true model, o error: Y 0 X, true model, wth error: Y 0 X, true model, o suscrpts: Y 0 X, true model, wth error: Y ax e (where a ad are the regresso coeffcets ad e s the resduals,

estmated model Y 0 X e, where 0 ad are estmates of 0 ad, ad e s a estmate of, estmated model Y ˆ ˆ 0 X e, where 0 ˆ ad ˆ are estmates of 0 ad, ad e s a estmate of. Other varales ad quattes There are a umer of other quattes that are mportat regresso aalyss, cludg: the ftted or predcted values of the respose varale Y (called y-hat ) Yˆ 0 X, Y ( X X) the resduals or predcto errors e Y Yˆ the sums of squared devatos ad ther cross products X ( ) x SY ( Y Y) y, ad XY xy S X X S ( X X)( Y Y) ad the resdual sum of squares e SSE

Fttg the regresso equato (.e. estmatg parameters) The regresso equato s ftted y choosg the values of 0 ad such a way that the sum of squares of the predcto errors, S, are mmzed,.e. M S e ( Y X ) 0 The specfc values of 0 ad that mmze S could e foud teratvely, or y tral ad error, ut t s kow that the followg ordary least-squares (OLS) estmates of 0 ad do fact mmze S: S XY SX ( X X) x Y X. 0 ( X X)( Y Y) x y, ad Goodess-of-ft statstcs The goodess of ft of the regresso equato, or a measure of the stregth of the relatoshp etwee Y ad X ca e descred several ways. As aalyss of varace, the sum of squares of the depedet varale Y ca e decomposed to two compoets TotalSS RegrSS ErrorSS ˆ ˆ Y Y Y Y Y Y ( ) ( ) ( ), Where TotalSS s the total sum of squares (of devatos of dvdual depedet varale values of the mea), RegrSS s the regresso sum of squares or that compoet of the total sum of squares explaed y the regresso equato, ad ErrorSS s the resdual sum of squares, or the sum of squares of the resdual,

e A F-statstc that ca e used to test the ull hypothess that the relatoshp etwee the predctor ad respose varales s ot sgfcat s MSRegr F MSError ( Y Y) /( k) ˆ ( Y Y ) /( k) The deomator of ths expresso, ˆ ( Y Y ) /( k), s also kow as the measquare error of the regresso, ad s sometmes represeted y s. The square root of the mea-square error, s, s called the stadard error of the regresso, ad provdes a measure of ucertaty the estmates of Y produced y the regresso equato. I geeral, the larger the F, the stroger the relatoshp. Aother measure of the stregth of the relatoshp etwee the respose ad predctor varale s the explaed varace (a proporto, ut sometmes expressed as a percetage), also kow as the coeffcet of determato, or R RegrSS TotalSS ErrorSS TotalSS ( Yˆ Y) ( Y Y) ( Yˆ Yˆ) ( Y Y) Sgfcace of the regresso coeffcets There are a umer of other quattes that are useful terpretg a regresso equato. These clude stadard errors for the slope ad tercept R

X se 0 s S X ( ), ad se( ) s / S X Usg these stadard errors, t-statstcs that ca e used to test hypotheses aout the regresso coeffcets ca e costructed: where 0 ad 0 0 t ( 0), ad se( 0 ) ( t ) se( ) are hyptheszed values of the regresso coeffcets, whch are usually take to e 0, so that large values of the t-statstcs wll sgal that 0 ad values that are sgfcat (.e. ot zero). The stadard error or stadard devato of the predcted value of the respose varale, Ŷ, gve a partcular value of the predcted varale, X, s ˆ ( X X ) se( Y X ) s SX