Properties of Least Squares

Similar documents
Econ Statistical Properties of the OLS estimator. Sanjaya DeSilva

Chapter 2 - The Simple Linear Regression Model S =0. e i is a random error. S β2 β. This is a minimization problem. Solution is a calculus exercise.

Chapter 3. Two-Variable Regression Model: The Problem of Estimation

Economics 130. Lecture 4 Simple Linear Regression Continued

Econ107 Applied Econometrics Topic 3: Classical Model (Studenmund, Chapter 4)

e i is a random error

Linear Regression Analysis: Terminology and Notation

Interval Estimation in the Classical Normal Linear Regression Model. 1. Introduction

ECONOMICS 351*-A Mid-Term Exam -- Fall Term 2000 Page 1 of 13 pages. QUEEN'S UNIVERSITY AT KINGSTON Department of Economics

Chapter 11: Simple Linear Regression and Correlation

Chapter 14 Simple Linear Regression

Statistics for Managers Using Microsoft Excel/SPSS Chapter 13 The Simple Linear Regression Model and Correlation

Outline. Zero Conditional mean. I. Motivation. 3. Multiple Regression Analysis: Estimation. Read Wooldridge (2013), Chapter 3.

STAT 3008 Applied Regression Analysis

Statistics for Economics & Business

x i1 =1 for all i (the constant ).

The Multiple Classical Linear Regression Model (CLRM): Specification and Assumptions. 1. Introduction

1. Inference on Regression Parameters a. Finding Mean, s.d and covariance amongst estimates. 2. Confidence Intervals and Working Hotelling Bands

Department of Quantitative Methods & Information Systems. Time Series and Their Components QMIS 320. Chapter 6

Chapter 4: Regression With One Regressor

Statistics for Business and Economics

Biostatistics 360 F&t Tests and Intervals in Regression 1

Problem of Estimation. Ordinary Least Squares (OLS) Ordinary Least Squares Method. Basic Econometrics in Transportation. Bivariate Regression Analysis

The Ordinary Least Squares (OLS) Estimator

Statistics MINITAB - Lab 2

17 - LINEAR REGRESSION II

j) = 1 (note sigma notation) ii. Continuous random variable (e.g. Normal distribution) 1. density function: f ( x) 0 and f ( x) dx = 1

/ n ) are compared. The logic is: if the two

β0 + β1xi and want to estimate the unknown

Linear regression. Regression Models. Chapter 11 Student Lecture Notes Regression Analysis is the

β0 + β1xi. You are interested in estimating the unknown parameters β

Exam. Econometrics - Exam 1

Basic Business Statistics, 10/e

β0 + β1xi. You are interested in estimating the unknown parameters β

Correlation and Regression. Correlation 9.1. Correlation. Chapter 9

Learning Objectives for Chapter 11

Econ107 Applied Econometrics Topic 9: Heteroskedasticity (Studenmund, Chapter 10)

Chapter 5: Hypothesis Tests, Confidence Intervals & Gauss-Markov Result

Lecture 6: Introduction to Linear Regression

Here is the rationale: If X and y have a strong positive relationship to one another, then ( x x) will tend to be positive when ( y y)

Y = β 0 + β 1 X 1 + β 2 X β k X k + ε

F8: Heteroscedasticity

Topic 7: Analysis of Variance

Comparison of Regression Lines

[The following data appear in Wooldridge Q2.3.] The table below contains the ACT score and college GPA for eight college students.

Chapter 13: Multiple Regression

ECONOMETRICS - FINAL EXAM, 3rd YEAR (GECO & GADE)

Lecture 4 Hypothesis Testing

Lecture Notes for STATISTICAL METHODS FOR BUSINESS II BMGT 212. Chapters 14, 15 & 16. Professor Ahmadi, Ph.D. Department of Management

Continuous vs. Discrete Goods

Resource Allocation and Decision Analysis (ECON 8010) Spring 2014 Foundations of Regression Analysis

STAT 3340 Assignment 1 solutions. 1. Find the equation of the line which passes through the points (1,1) and (4,5).

Biostatistics. Chapter 11 Simple Linear Correlation and Regression. Jing Li

28. SIMPLE LINEAR REGRESSION III

Lecture 3 Stat102, Spring 2007

ANSWERS CHAPTER 9. TIO 9.2: If the values are the same, the difference is 0, therefore the null hypothesis cannot be rejected.

Chapter 7 Generalized and Weighted Least Squares Estimation. In this method, the deviation between the observed and expected values of

18. SIMPLE LINEAR REGRESSION III

See Book Chapter 11 2 nd Edition (Chapter 10 1 st Edition)

Midterm Examination. Regression and Forecasting Models

Chapter 8 Indicator Variables

Professor Chris Murray. Midterm Exam

Lab 4: Two-level Random Intercept Model

Introduction to Dummy Variable Regressors. 1. An Example of Dummy Variable Regressors

Lecture 3 Specification

Correlation and Regression

Reduced slides. Introduction to Analysis of Variance (ANOVA) Part 1. Single factor

Lecture 9: Linear regression: centering, hypothesis testing, multiple covariates, and confounding

Lecture 9: Linear regression: centering, hypothesis testing, multiple covariates, and confounding

Limited Dependent Variables

Introduction to Regression

4 Analysis of Variance (ANOVA) 5 ANOVA. 5.1 Introduction. 5.2 Fixed Effects ANOVA

Negative Binomial Regression

Econometrics of Panel Data

CHAPER 11: HETEROSCEDASTICITY: WHAT HAPPENS WHEN ERROR VARIANCE IS NONCONSTANT?

a. (All your answers should be in the letter!

Now we relax this assumption and allow that the error variance depends on the independent variables, i.e., heteroskedasticity

Lecture 2: Prelude to the big shrink

Durban Watson for Testing the Lack-of-Fit of Polynomial Regression Models without Replications

DO NOT OPEN THE QUESTION PAPER UNTIL INSTRUCTED TO DO SO BY THE CHIEF INVIGILATOR. Introductory Econometrics 1 hour 30 minutes

Introduction to Analysis of Variance (ANOVA) Part 1

2016 Wiley. Study Session 2: Ethical and Professional Standards Application

The SAS program I used to obtain the analyses for my answers is given below.

Predictive Analytics : QM901.1x Prof U Dinesh Kumar, IIMB. All Rights Reserved, Indian Institute of Management Bangalore

Chapter 14 Simple Linear Regression Page 1. Introduction to regression analysis 14-2

First Year Examination Department of Statistics, University of Florida

Chapter 15 Student Lecture Notes 15-1

Modeling and Simulation NETW 707

x = , so that calculated

(c) Pongsa Pornchaiwiseskul, Faculty of Economics, Chulalongkorn University

4.3 Poisson Regression

LINEAR REGRESSION ANALYSIS. MODULE VIII Lecture Indicator Variables

Dr. Shalabh Department of Mathematics and Statistics Indian Institute of Technology Kanpur

STATISTICS QUESTIONS. Step by Step Solutions.

Inner Product. Euclidean Space. Orthonormal Basis. Orthogonal

Sociology 301. Bivariate Regression. Clarification. Regression. Liying Luo Last exam (Exam #4) is on May 17, in class.

Factor models with many assets: strong factors, weak factors, and the two-pass procedure

Primer on High-Order Moment Estimators

Statistics Chapter 4

PubH 7405: REGRESSION ANALYSIS. SLR: INFERENCES, Part II

Transcription:

Week 3 3.1 Smple Lnear Regresson Model 3. Propertes of Least Squares Estmators Y Y β 1 + β X + u weekly famly expendtures X weekly famly ncome For a gven level of x, the expected level of food expendtures wll be: E(Y X ) β 1 + β X Propertes of the Smple Lnear Regresson Model 3.3 The populaton parameters β 1 and β are unknown populaton constants. 3.4 1. Y β 1 + β X + u. E(u ) 0 <> E(Y ) β 1 + β X 3. var(u ) σ var(y ) 4. cov(u,uj j ) cov(y,y j ) 0 5. x not constant for every observaton 6. u ~N(0,σ ) <> Y ~N(β 1 +β x,σ ) The formulas that produce the sample estmates β 1 (or b 1 ) and or β (or b )are called the estmators of β 1 and β. When and β 1 β are used to represent the formulas rather than specfc values, they are called estmators of β 1 and β whch are random varables because they are dfferent from sample to sample

Estmators are Random Varables 3.5 If the least squares estmators β 1 and β are random varables, then what are ther means, varances, covarances and probablty dstrbutons? Compare the propertes of alternatve estmators wth the propertes of the least squares estmators. The Expected Values of β 1 and β The least squares formulas (estmators) n the smple regresson case: β nσx Y - ΣX ΣY nσx -(ΣX ) β 1 Y - β X Σx y Σx where Y ΣY / n and X ΣX / n x y X Y 3.6 X Y Substtute Y Into b formula to get: β 1 + β X + u β β + nσx u - ΣX Σu nσx -(ΣX( ) The mean of β s: 0 3.7 E(β ) β + nσx E(u )- ΣX ΣE(u ) nσx -(ΣX( ) Snce E(u ) 0, then E(β ) β. An Unbased Estmator The result E(β ) β means that the dstrbuton of β s centered at β. Snce the dstrbuton of β s centered at β,we say that β s an unbased estmator of β. 3.8

Wrong Model Specfcaton The unbasedness result on the prevous slde assumes that we are usng the correct model If the model s of the wrong form or there are mssng varables, then E(u ) 0, and E(β ) β For example: Y β 1 + β X + (β 3 X 3 +v) u β 3 X 3 +v E(u ) 0 3.9 Unbased Estmator of the Intercept 3.10 In a smlar manner, the estmator β 1 of the ntercept or constant term can be shown to be an unbased estmator of β 1 when the model s correctly specfed. E(β 1 ) β 1 Equvalent expressons for β : ( ) β Σ(x x )(Y Y ) Σ(x x) Σx y Σx 3.11 β β Var(β ) se(β ) 3.1 Expand and multply top and bottom by n: β nσx Y Σx ΣY nσx (Σx ) σ u

Estmatng the varance of the error term, σ u u Y β 1 β X n Σu 1 σ n k σ s an unbased estmator of σ degrees of freedom (df) n - # of parameters (k) df n - k 3.13 Varance of β Gven that both Y and u have varance σ, the varance of the estmator β s: σ var(β ) Σ(x x) β σ Σx x S(b) Se(b se(β 0 ) (8.50) /9.55 0.7809 0.8836 β s a functon of the Y values but var(β ) does not nvolve Y drectly. 3.14 Varance of β 1 Gven β 1 Y β x the varance of the estmator β 1 s: Σx var(β 1 ) σ σ Σx n Σ(x x) ( ) nσx se(β 1 ) (8.50) (35/0(9.55)) 87.38 9.34 3.15 Covarance of β 1 and β x cov(β 1,β ) σ x σ Σ(x x) Σ x 3.16 If x 0, slope can change wthout affectng the covarance of β 1 and β.

What factors determne varance and covarance? 3.17 1. σ : uncertanty about Y values uncertanty about β 1, β and ther relatonshp.. The more spread out the X values, the more confdence n β 1, β, etc. 3. The larger the sample sze, n, the smaller the varances and covarances. 4. The varance ββ 1 s large when the (squared) X values are far from zero (n ether drecton). 5. Changng the slope, β, has no effect on the ntercept, β 1, when the sample mean of X s zero. But f sample mean s postve, the covarance between β 1 and β wll be negatve, and vce versa. Propertes of Least Squares Resduals 1. Σu 0. Σu X 0 3. Σ(Y( -Y )( )(Y -Y)0 or> Σ u y 0 4. The lnear regresson lne must pass through the sample mean of X and Y. 3.18 Decomposton of sum of squares (Y - Y) (Y -Y ) + (Y -Y) To measure varaton: u Σ(Y -Y) Σ[(Y -Y ) + (Y -Y)] Σ(Y - Y) Σ(Y - Y ) + Σ(Y - Y) TSS Total sum of squares RSS (Σu ) Resdual sum of squares 3.19 ESS Explaned sum of squares r - Measure of goodness of ft Defne r r ESS TSS 1 - RSS TSS 1 - Σ(Y -Y) Σ(Y -Y) Σ u Σ(Y( -Y) 1 r 0 3.0

Alternatve r expresson r Σ(β x ) β ESS Σ(Y -Y) TSS Σ(Y -Y) S x S y β Σx Σx y Σx β Σx y 3.1 Σx (Σx y ) Σx r (Σx y ) Σx Note: corr(x,y) Then, r ± r (Σx y ) (Σx ) 1/ r 3. Notes: 3.3 Y As r 0 3.4 SRF Whch SRF? Y X http://www.ruf.rce.edu/~lane/stat_sm/comp_r/ndex.html As r 1 SRF SRF go through all ponts X

3.5 3.6 Gauss-Markov Theorem Gven the assumptons of classcal lnear regresson model, the ordnary least squares (OLS) estmators β 1 and β are the best lnear unbased estmators (BLUE) of β 1 and β. Ths means that β 1 and β have the smallest varance of all lnear unbased estmator of β 1 and β. Note: Gauss-Markov Theorem doesn t apply to non-lnear estmators 3.7 Unbased :The expected value of the estmator β k equals to the true value of β k Prob. (β ) E(β )β E(β )>β E(β )<β Based underestmate E(β ) Unbased E(β ) True value of β Based overestmate E(β ) 3.8 β

Probablty Dstrbuton of Least Squares Estmators 3.9 normally dstrbuted under The Central Lmt Theorem 3.30 β 1 ~ N β 1, β ~ N β, σ Σx nσx σ Σx If the Gauss-Markov assumptons hold, and sample sze, n, s suffcently large, then the least squares estmators, β 1 and β, have a dstrbuton that approxmates the normal ldstrbuton b wth hgreater accuracy the larger the value of sample sze, n. Effcency: β k s an effcent unbased estmator f for a gven sample sze, n, the varance of β k s the smallest one. Consstency: β k s a consstent estmator of β k f the probablty of β k - β k < ε s 1. It means that as the sample sze gets larger, β k wll become more accurately. (εa very small #) 3.31 The Least Squares Predctor, Y Gven a value of the explanatory varable, X, we would lke to predct a value of the dependent varable, Y, from the OLS estmated result. The least squares predctor s: Y β 1 + β X 3.3

Summary of BLUE estmators 3.33 Mean E(β 1 )β 1 Varance and E(β )β Σx nσx Var(β 1 )σ and Var(β ) Standard error or standard devaton se(β k ) var(β k ) σ Σx Estmated Error Varance σ T Σu 1 n E(σ ) σ 3.34 Standard Error of Regresson (Estmate), SEE σ σ T Σu 1 n k # of ndependent varables plus the constant term