Multiple Linear Regression Analysis

Similar documents
Chapter 3 Multiple Linear Regression Model

Chapter 14 Logistic Regression Models

Example: Multiple linear regression. Least squares regression. Repetition: Simple linear regression. Tron Anders Moger

ENGI 3423 Simple Linear Regression Page 12-01

ECONOMETRIC THEORY. MODULE VIII Lecture - 26 Heteroskedasticity

Probability and. Lecture 13: and Correlation

b. There appears to be a positive relationship between X and Y; that is, as X increases, so does Y.

STA302/1001-Fall 2008 Midterm Test October 21, 2008

Simple Linear Regression Analysis

LINEAR REGRESSION ANALYSIS

Simple Linear Regression

Objectives of Multiple Regression

Multiple Regression. More than 2 variables! Grade on Final. Multiple Regression 11/21/2012. Exam 2 Grades. Exam 2 Re-grades

( ) = ( ) ( ) Chapter 13 Asymptotic Theory and Stochastic Regressors. Stochastic regressors model

Lecture 7. Confidence Intervals and Hypothesis Tests in the Simple CLR Model

Chapter 13 Student Lecture Notes 13-1

Ordinary Least Squares Regression. Simple Regression. Algebra and Assumptions.

{ }{ ( )} (, ) = ( ) ( ) ( ) Chapter 14 Exercises in Sampling Theory. Exercise 1 (Simple random sampling): Solution:

residual. (Note that usually in descriptions of regression analysis, upper-case

Lecture 8: Linear Regression

Simple Linear Regression

Midterm Exam 1, section 1 (Solution) Thursday, February hour, 15 minutes

12.2 Estimating Model parameters Assumptions: ox and y are related according to the simple linear regression model

Statistics MINITAB - Lab 5

Econometric Methods. Review of Estimation

Dr. Shalabh. Indian Institute of Technology Kanpur

: At least two means differ SST

Simple Linear Regression and Correlation. Applied Statistics and Probability for Engineers. Chapter 11 Simple Linear Regression and Correlation

THE ROYAL STATISTICAL SOCIETY HIGHER CERTIFICATE

4. Standard Regression Model and Spatial Dependence Tests

Comparison of Dual to Ratio-Cum-Product Estimators of Population Mean

Linear Regression with One Regressor

Chapter 3 Sampling For Proportions and Percentages

Midterm Exam 1, section 2 (Solution) Thursday, February hour, 15 minutes

Chapter Business Statistics: A First Course Fifth Edition. Learning Objectives. Correlation vs. Regression. In this chapter, you learn:

Summary of the lecture in Biostatistics

STA 108 Applied Linear Models: Regression Analysis Spring Solution for Homework #1

UNIVERSITY OF OSLO DEPARTMENT OF ECONOMICS

Dr. Shalabh Department of Mathematics and Statistics Indian Institute of Technology Kanpur

ECON 482 / WH Hong The Simple Regression Model 1. Definition of the Simple Regression Model

Simulation Output Analysis

Sampling Theory MODULE X LECTURE - 35 TWO STAGE SAMPLING (SUB SAMPLING)


Correlation and Simple Linear Regression

Chapter Two. An Introduction to Regression ( )

Statistics: Unlocking the Power of Data Lock 5

Recall MLR 5 Homskedasticity error u has the same variance given any values of the explanatory variables Var(u x1,...,xk) = 2 or E(UU ) = 2 I

The equation is sometimes presented in form Y = a + b x. This is reasonable, but it s not the notation we use.

Chapter -2 Simple Random Sampling

Chapter 13, Part A Analysis of Variance and Experimental Design. Introduction to Analysis of Variance. Introduction to Analysis of Variance

Lecture Note to Rice Chapter 8

Statistics. Correlational. Dr. Ayman Eldeib. Simple Linear Regression and Correlation. SBE 304: Linear Regression & Correlation 1/3/2018

Chapter 11 The Analysis of Variance

UNIVERSITY OF OSLO DEPARTMENT OF ECONOMICS

Lecture Notes Types of economic variables

Chapter 2 Simple Linear Regression

Reaction Time VS. Drug Percentage Subject Amount of Drug Times % Reaction Time in Seconds 1 Mary John Carl Sara William 5 4

ε. Therefore, the estimate

UNIVERSITY OF TORONTO AT SCARBOROUGH. Sample Exam STAC67. Duration - 3 hours

Simple Linear Regression and Correlation.

Fundamentals of Regression Analysis

Multiple Choice Test. Chapter Adequacy of Models for Regression

Multivariate Transformation of Variables and Maximum Likelihood Estimation

REVIEW OF SIMPLE LINEAR REGRESSION SIMPLE LINEAR REGRESSION

STATISTICAL PROPERTIES OF LEAST SQUARES ESTIMATORS. x, where. = y - ˆ " 1

Module 7. Lecture 7: Statistical parameter estimation

Simple Linear Regression - Scalar Form

ESS Line Fitting

Chapter -2 Simple Random Sampling

MEASURES OF DISPERSION

Regression. Linear Regression. A Simple Data Display. A Batch of Data. The Mean is 220. A Value of 474. STAT Handout Module 15 1 st of June 2009

CLASS NOTES. for. PBAF 528: Quantitative Methods II SPRING Instructor: Jean Swanson. Daniel J. Evans School of Public Affairs

Econ 388 R. Butler 2016 rev Lecture 5 Multivariate 2 I. Partitioned Regression and Partial Regression Table 1: Projections everywhere

STK3100 and STK4100 Autumn 2017

Lecture 3 Probability review (cont d)

Chapter 2 Supplemental Text Material

Handout #8. X\Y f(x) 0 1/16 1/ / /16 3/ / /16 3/16 0 3/ /16 1/16 1/8 g(y) 1/16 1/4 3/8 1/4 1/16 1

best estimate (mean) for X uncertainty or error in the measurement (systematic, random or statistical) best

Mean is only appropriate for interval or ratio scales, not ordinal or nominal.

Dimensionality Reduction and Learning

STK3100 and STK4100 Autumn 2018

Example. Row Hydrogen Carbon

TESTS BASED ON MAXIMUM LIKELIHOOD

Functions of Random Variables

Quiz 1- Linear Regression Analysis (Based on Lectures 1-14)

Understand the properties of the parameter estimates, how we compute confidence intervals for them and how we test hypotheses about them.

University of Belgrade. Faculty of Mathematics. Master thesis Regression and Correlation

Sampling Theory MODULE V LECTURE - 14 RATIO AND PRODUCT METHODS OF ESTIMATION

Lecture 2: The Simple Regression Model

Class 13,14 June 17, 19, 2015

Chapter 5 Transformation and Weighting to Correct Model Inadequacies

STA 105-M BASIC STATISTICS (This is a multiple choice paper.)

ENGI 4421 Propagation of Error Page 8-01

Maximum Likelihood Estimation

CHAPTER 2. = y ˆ β x (.1022) So we can write

Special Instructions / Useful Data

Statistics Review Part 3. Hypothesis Tests, Regression

2SLS Estimates ECON In this case, begin with the assumption that E[ i

Lecture Notes Forecasting the process of estimating or predicting unknown situations

ENGI 4421 Joint Probability Distributions Page Joint Probability Distributions [Navidi sections 2.5 and 2.6; Devore sections

Transcription:

LINEA EGESSION ANALYSIS MODULE III Lecture - 4 Multple Lear egresso Aalyss Dr. Shalabh Departmet of Mathematcs ad Statstcs Ida Isttute of Techology Kapur

Cofdece terval estmato The cofdece tervals multple regresso model ca be costructed for dvdual regresso coeffcets as well as otly. We cosder both of them as follows: Cofdece terval o the dvdual regresso coeffcet Assumg ε ' s are detcally ad depedetly dstrbuted followg, we have y N X I ~ ( βσ, ) b N X X ~ ( βσ, ( ' ) ). Thus the margal dstrbuto of ay regresso coeffcet estmate ~ ( β, σ ) b N C = + N(0, σ ) y Xβ ε where s the th dagoal elemet of ( X ' X). Thus where C b β t = ~ t ( k) uder H 0, =,,... ˆ σ C SSre s yy ' bx ' ' y ˆ σ = =. k k

3 So the 00( α)% cofdece terval for β ( =,,..., k) s obtaed as follows: b β P tα t, k α = α ˆ σ C P b t C b t C So the cofdece terval s ˆ ˆ α σ β + α σ = α. b ˆ, ˆ tα σ C b + tα σ C.

4 Smultaeous cofdece tervals o regresso coeffcets A set of cofdece tervals that are true smultaeously wth probablty cofdece tervals. It s relatvely easy to defe a ot cofdece rego for Sce ( b β)' X ' X( b β) ~ F k MS re s k, k ( b β)' X ' X( b β) P Fα ( k, k) α. k MS = re s β ( α) multple regresso model. are called smultaeous or ot So a 00 ( α)% ot cofdece rego for all of the parameters s ( b β)' X' Xb ( β) ~ Fα ( k, k ) k MS res β whch descrbes a ellptcally shaped rego.

5 Coeffcet of determato ( ) ad adusted Let be the multple correlato coeffcet betwee y ad X, X,..., X k. The square of multple correlato coeffcet ( ) s called as coeffcet of determato. The value of commoly descrbes that how well the sample regresso le fts to the observed data. Ths s also treated as a measure of goodess of ft of the model. Assumg that the tercept term s preset the model as the where y = β+ βx + β3x3 +... + βk Xk + u, =,,..., = SS r e s SS T SS r e g = SS = SS SS = SS re g T ee ' ( y y) res T : sum of squares due to resduals, : total sum of squares, : sum of squares due to regresso. measure the explaatory power of the model whch tur reflects the goodess of ft of the model. It reflects the model adequacy the sese that how much s the explaatory power of explaatory varable.

Sce e e = y I X X X X y = y Hy ' ' ( ' ) ' ', ( y y) = y y, = = where y = y = ' y wth = (,,..., ) ', y = ( y, y,..., y) ' = 6 Thus where So ( y y) = y ' y ' yy ' = = y' y y' ' y = y' y y' ( ' ) ' y = y' I ( ' ) ' y = y ' Ay A= I ' ( ' ) '. y Hy =. y ' Ay The lmts of are 0 ad,.e., 0. = 0 dcates the poorest ft of the model. = dcates the best ft of the model. = 0.95 dcates that 95% of the varato y s explaed by the explaatory varables. I smple words, the model s 95% good. Smlarly ay other value of betwee 0 ad dcates the adequacy of ftted model.

7 Adusted If more explaatory varables are added to the model, the creases. I case the varables are rrelevat, the wll stll crease ad gves a overly optmstc pcture. Wth a purpose of correcto overly optmstc pcture, adusted, deoted as SSre s/( k) = SS / ( ) = k ( ). or ad s used whch s defed as We wll see later that ( - k) ad ( - ) are the degrees of freedom assocated wth the dstrbutos of SS res ad SS T. Moreover, the quattes ad are based o the ubased estmators of respectve varaces of e ad y s the cotext of aalyss of varace. T SS r e s SS T k The adusted wll decle f the addto of a extra varable produces too small a reducto ( ) for the crease s. k to compesate Aother lmtato of adusted s that t ca be egatve also. For example f = 0.84 = 0.08< 0 7 9 whch has o terpretato. k = 3, = 0, = 0.6, the

Lmtatos 8. If costat term s abset the model, the ca ot be defed. I such cases, ca be egatve. Some ad-hoc measures based o for regresso le through org have bee proposed the lterature.. s sestve to extreme values, so lacks robustess. 3. Cosder a stuato where we have followg two models: The questo s ow whch model s better? y = β + β X +... + β X + u, =,,.., k k log y = γ + γ X +... + γ X + v k k For the frst model, = = = ( y yˆ ) ( y y) ad for the secod model, a opto s to defe as (log y log ˆ y) =. (log y log y) = = As such ad are ot comparable. If stll, the two models are eeded to be compared, a better proposto to defe ca be as follows: * ( y ˆ at log y ) = 3 = ( y y) = * where y = logy. Now o comparso may gve a dea about the adequacy of the two models. ad 3

9 elatoshp of aalyss of varace test ad coeffcet of determato β 0 3 Assumg to be a tercept term, the for : = =... = = 0 the F-statstc aalyss of varace test s F = MS MS re g res where s the coeffcet of determato. H β β β k ( k) SS k SS k SS = = = SS reg reg T ( k ) SSres k SST SSreg k SS re g k = k SS re g T So F ad are closely related. Whe = 0, the F = 0. I lmt, whe =, F =. So both F ad vary drectly. Larger mples greater F value. That s why the F test uder aalyss of varace s termed as the measure of overall sgfcace of estmated regresso. It s also a test of sgfcace of. If F s hghly sgfcat, t mples that we ca reect H 0,.e. y s learly related to X s.

0 Predcto of values of study varable The predcto multple regresso model has two aspects. Predcto of average value of study varable or mea respose.. Predcto of actual value of study varable.. Predcto of average value of y We eed to predct E(y) at a gve The predctor as a pot estmate s p= xb= x( X' X) X' y ' ' 0 0 E( p) = x β. ' 0 So p s a ubased predctor for E( y). x = ( x, x,..., x )'. 0 0 0 0k Its varace s [ ] [ ] Var( p) = E p E( y) ' p E( y) = σ x ( X ' X) x. ' 0 0

The cofdece terval o the mea respose at a partcular pot, such as x0, x0,..., x0k ca be foud as follows: Defe x = ( x, x,..., x )'. 0 0 0 0k The ftted value at The x 0 s yˆ = xb. ' 0 0 E( yˆ ) = x β = E( y x ) ' 0 0 0 Var( yˆ ) = σ x ( X ' X ) x ' 0 0 0 yˆ E( y x ) = α 0 0 P tα t, ' α k ˆ σ x0( X ' X) x0 ' ' P yˆ0 t ˆ ˆ ˆ α σ x0( X ' X) x0 E( y x0) y0 + tα σ x0( X ' X) x0 = α. 0 0 0 x, x,..., x k 0 The 00( α)% cofdece terval o the mea respose at the pot,.e., E( y x ) s ' ' yˆ0 t ˆ ˆ ˆ α σ x0( X ' X) x0, y0 + tα σ x0( X ' X) x0.

. Predcto of actual value of y We eed to predct y at a gve x = ( x, x,..., x )'. 0 0 0 0k The predctor as a pot estmate s p f = xb ' 0 So p f E( p ) = x β f ' 0 s a ubased predctor for y. It's varace s ( ) Var( p ) = E ( p y)( p y)' f f f ' = σ + x0( X ' X) x 0. The 00( α)% cofdece terval for ths future observato s ' ' p ˆ ˆ f tα σ [ + x0( X ' X) x0], pf + tα σ [ + x0( X ' X) x0].