Linear Models and Estimation by Least Squares

Similar documents
Ch 2: Simple Linear Regression

Linear models and their mathematical foundations: Simple linear regression

Simple Linear Regression

Measuring the fit of the model - SSR

Chapter 12 - Lecture 2 Inferences about regression coefficient

MAT2377. Rafa l Kulik. Version 2015/November/26. Rafa l Kulik

STA 2201/442 Assignment 2

Simple and Multiple Linear Regression

Introduction to Estimation Methods for Time Series models. Lecture 1

Lecture 15. Hypothesis testing in the linear model

Simple Linear Regression

AMS 315/576 Lecture Notes. Chapter 11. Simple Linear Regression

TMA4255 Applied Statistics V2016 (5)

Ph.D. Qualifying Exam Friday Saturday, January 6 7, 2017

BIOS 2083 Linear Models c Abdus S. Wahed

General Linear Model: Statistical Inference

Regression and Statistical Inference

where x and ȳ are the sample means of x 1,, x n

Problems. Suppose both models are fitted to the same data. Show that SS Res, A SS Res, B

Simple Linear Regression

Multiple Linear Regression

EC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix)

Applied Econometrics (QEM)

STAT5044: Regression and Anova. Inyoung Kim

Lecture 14 Simple Linear Regression

Estimating σ 2. We can do simple prediction of Y and estimation of the mean of Y at any value of X.

[y i α βx i ] 2 (2) Q = i=1

Covariance and Correlation

Bias Variance Trade-off

Ma 3/103: Lecture 25 Linear Regression II: Hypothesis Testing and ANOVA

Summer School in Statistics for Astronomers V June 1 - June 6, Regression. Mosuk Chow Statistics Department Penn State University.

Formal Statement of Simple Linear Regression Model

The Simple Linear Regression Model

Master s Written Examination

Linear Regression. Simple linear regression model determines the relationship between one dependent variable (y) and one independent variable (x).

Confidence Intervals, Testing and ANOVA Summary

Multivariate Regression

Statistical Hypothesis Testing

INTERVAL ESTIMATION AND HYPOTHESES TESTING

Inference in Regression Analysis

Applied Regression. Applied Regression. Chapter 2 Simple Linear Regression. Hongcheng Li. April, 6, 2013

Ch 3: Multiple Linear Regression

COPYRIGHT. Abraham, B. and Ledolter, J. Introduction to Regression Modeling Belmont, CA: Duxbury Press, 2006

Multiple Linear Regression

Lecture 3: Multiple Regression

Lecture 4 Multiple linear regression

Midterm 2 - Solutions

Chapte The McGraw-Hill Companies, Inc. All rights reserved.

ECO220Y Simple Regression: Testing the Slope

Final Exam - Solutions

ECON 4160, Autumn term Lecture 1

REGRESSION ANALYSIS AND INDICATOR VARIABLES

MFin Econometrics I Session 4: t-distribution, Simple Linear Regression, OLS assumptions and properties of OLS estimators

Ecn Analysis of Economic Data University of California - Davis February 23, 2010 Instructor: John Parman. Midterm 2. Name: ID Number: Section:

STAT 135 Lab 11 Tests for Categorical Data (Fisher s Exact test, χ 2 tests for Homogeneity and Independence) and Linear Regression

401 Review. 6. Power analysis for one/two-sample hypothesis tests and for correlation analysis.

Question. Hypothesis testing. Example. Answer: hypothesis. Test: true or not? Question. Average is not the mean! μ average. Random deviation or not?

Review of Econometrics

Applied Regression Analysis

y ˆ i = ˆ " T u i ( i th fitted value or i th fit)

Next is material on matrix rank. Please see the handout

Statistics 112 Simple Linear Regression Fuel Consumption Example March 1, 2004 E. Bura

Regression Analysis. BUS 735: Business Decision Making and Research. Learn how to detect relationships between ordinal and categorical variables.

WISE MA/PhD Programs Econometrics Instructor: Brett Graham Spring Semester, Academic Year Exam Version: A

Statement: With my signature I confirm that the solutions are the product of my own work. Name: Signature:.

Master s Written Examination - Solution

Oct Simple linear regression. Minimum mean square error prediction. Univariate. regression. Calculating intercept and slope

Lecture 6 Multiple Linear Regression, cont.

Stat 579: Generalized Linear Models and Extensions

14 Multiple Linear Regression

Peter Hoff Linear and multilinear models April 3, GLS for multivariate regression 5. 3 Covariance estimation for the GLM 8

Outline. Remedial Measures) Extra Sums of Squares Standardized Version of the Multiple Regression Model

Scatter plot of data from the study. Linear Regression

Inference for Regression

ECON 450 Development Economics

The Slow Convergence of OLS Estimators of α, β and Portfolio. β and Portfolio Weights under Long Memory Stochastic Volatility

Linear regression. We have that the estimated mean in linear regression is. ˆµ Y X=x = ˆβ 0 + ˆβ 1 x. The standard error of ˆµ Y X=x is.

17: INFERENCE FOR MULTIPLE REGRESSION. Inference for Individual Regression Coefficients

Topic 22 Analysis of Variance

Problem Selected Scores

Statistics Ph.D. Qualifying Exam: Part I October 18, 2003

Part III: Unstructured Data

2. A Review of Some Key Linear Models Results. Copyright c 2018 Dan Nettleton (Iowa State University) 2. Statistics / 28

WISE International Masters

Section 4.6 Simple Linear Regression

Econ 3790: Statistics Business and Economics. Instructor: Yogesh Uppal

Lecture 25: Review. Statistics 104. April 23, Colin Rundel

Simple Linear Regression Analysis

Problem 1 (20) Log-normal. f(x) Cauchy

STAT Chapter 11: Regression

The regression model with one fixed regressor cont d

LIST OF FORMULAS FOR STK1100 AND STK1110

NATIONAL UNIVERSITY OF SINGAPORE EXAMINATION. ST4233 Linear Models: Solutions. (Semester I: ) November/December, 2007 Time Allowed : 2 Hours

Math 180B Problem Set 3

STAT 511. Lecture : Simple linear regression Devore: Section Prof. Michael Levine. December 3, Levine STAT 511

STA 2101/442 Assignment 3 1

WISE MA/PhD Programs Econometrics Instructor: Brett Graham Spring Semester, Academic Year Exam Version: A

Linear Regression Model. Badr Missaoui

Lecture 10: Panel Data

Econometrics A. Simple linear model (2) Keio University, Faculty of Economics. Simon Clinet (Keio University) Econometrics A October 16, / 11

Transcription:

Linear Models and Estimation by Least Squares Jin-Lung Lin 1 Introduction Causal relation investigation lies in the heart of economics. Effect (Dependent variable) cause (Independent variable) Example: Demand for money depends upon income, interest rate, etc. Two most frequently estimation methods: Maximum likelihood estimate Least square estimate Linear Statistical Models Definition 11.1 A linear statistical model relating a random response Y to a set of independent variables x 1, x,..., x k is of the form Y i = β 0 + β 1 x 1i + β i x + + β k x ki + ɛ i i = 1,,, T E(ɛ t ) = 0, t V (ɛ t ) =, t E(ɛ t ɛ s ) = 0, t s E(x it ɛ t ) = i = 1,,, k0 Take expectation of both sides to obtain µ y = β 0 + β 1 µ x1 + + β k µ xk and subtract it from the regression model, (y t µ y ) = β 1 (x 1t µ x1 )+β (x i µ x )+ +β k (x ki µ xk )+ɛ t, t = 1,,, T where β 0, β 1,..., β k are unknown parameters, ɛ is a random variable, and x 1, x,..., x k are fixed variables. 1

3 The Method of Least Squares Simple Linear Regression Model Y = β 0 + β 1 x + ɛ ŷ i = ˆβ 0 + ˆβ 1 x i Minimize the sum of squared deviation: SSE = (y i ŷ i ) = (y i ˆβ 0 ˆβ 1 x i ) i=1 Least Squares Estimators for Simple Linear Regression Model 1. ˆβ1 = Sxy, where S xy = n i=1 (x i x)(y i ȳ), and = n i=1 (x i x).. ˆβ0 = ȳ ˆβ 1 x. Note that (Ŷ ȳ) = ˆβ 1 (x x) ˆβ 1 = w i y i ˆβ 0 = w i = i=1 v i y i i=1 x i x n i=1 (x i x) v i = 1 n xw i w i = 0 i wi = i=1 1 n i=1 (x i x) 4 Properties of the Least Squares Estimators- Simple Linear Regression 1. The estimators ˆβ 0 and ˆβ 1 are unbiased, that is, E( ˆβ i ) = β i, for i=0,1.

. V ( ˆβ 0 ) = c 00, where c 00 = x i n. 3. 3. V ( ˆβ 1 ) = c 11, where c 00 = 1. 4. Cov( ˆβ 0, ˆβ 1 ) = c 01, where c 01 = x. 5. An unbiased estimator of is S = SSE/(n ), where SSE = S yy ˆβ 1 S xy and S yy = (y i ȳ). If, in addition, the ɛ i, for i = 1,,..., n are normally distributed: 6. Both ˆβ 0 and ˆβ 1 are normal distributed. 7. The random variable (n )S freedom. has a χ distribution with n- degrees of 8. The statistic S is independent of both ˆβ 0 and ˆβ 1. 5 Inference Concerning the Parameters β i Test of Hypothesis for β i H 0 : H a : Test statistic: T = ˆβ i β i0 S c ii Rejection region: x i β i = β i0 β i > β i0 (upper tail alternative) β i < β i0 (lower tail alternative) β i β i0 (two tailed alternative) t > t α (upper tail alternative) t < t α (lower tail alternative) t > t α/ (two tailed alternative) where c 00 = n, and c 11 = 1 Notice that t α is based on (n-) degrees of freedom. A (1 α) 100 % Confidence Interval for β i. ˆβ ± t α/ S c ii where c 00 = x i n, c 11 = 1 3

6 Inference Concerning Linear Functions of the Model Parameters: Simple Linear Regression A Test for θ = a 0 β 0 + a 1 β 1 H 0 : H a : θ = θ 0 θ > θ 0 θ < θ 0 θ θ 0 Test statistic: T = Rejection region: ˆθ θ 0 S ( a 0 x /n+a i 1 a 0 a 1 x ) Sxx t > t α t < t α t t α/ Here t α is based upon n- degrees of freedom. A (1 α) 100 % Confidence Interval for θ = a 0 β 0 + a 1 β 1 ˆθ ± t α/ S ( a 0 x i /n + a 1 a 0 a 1 x ) where the tabulated t α/ is based upon n- degrees of freedom. A (1 α) 100 % Confidence Interval for E(Y ) = β 0 + β 1 x ˆβ 0 + ˆβ 1 x ± t α/ S 1/n + (x x) / where the tabulated t α/ is based upon n- degrees of freedom. 7 Predicting a Particular Value of Y Using Simple Linear Regression A (1 α) 100 % Prediction Interval for Y when x = x ˆβ 0 + ˆβ 1 x ± t α/ S 1 + 1/n + (x x) / 4

8 Correlation In Y = β 0 + β 1 x + ɛ, x could be a fixed regressor controlled by researchers or can be observed values of a random variable. For the former, E(Y ) = β 0 +β 1 x. For the latter, E(Y X = x) = β 0 +β 1 x. Typically, we often assume that (X, Y ) has a bivariate normal distribution with E(X) = µ x, E(Y ) = µ y, V (X) = x, V (Y ) = y, Cov(X, Y ) = σ xy = ρσ x σ y. In this case, E(Y X = x) = β 0 + β 1 x, where β 1 = σ y σ x ρ The MLE of correlation ρ is given by the sample correlation coefficient: r = = n i=1 (X i X)(Y i Ȳ ) n i=1 (X i X) n i=1 (Y i Ȳ ) X xy = ˆβ 1 Sxx S yy S yy Testing β 1 = 0 is equivalent to testing ρ = 0. ˆβ 1 t = S/ = r n 1 r Z-test (1/) log[(1 + r)/(1 r)] N((1/) log[(1 + ρ)/(1 ρ)], 1/(n 3)) Z = ( 1 1+r ) log( 1 r ) ( 1 ) log( 1+ρ 0 1 n 3 1 ρ 0 ) 9 Fitting the Linear Model by Using Matrices Least Squares Equations and Solutions for a General Linear Model Equations : (X X)ˆβ = X Y Solutions : ˆβ = (X X) 1 X Y SSE = Y Y ˆβ X Y 5

10 Linear Functions of the Model Parameters: Multiple Linear Regression Multiple Linear Regression Y i = β 0 + β 1 x 1i + + β k x ki + ɛ i, i = 1,,..., n. ˆβ = (X X) 1 X Y Properties of the Least Squares Estimators- Multiple Linear Regression 1. E( ˆβ i ) = β i, i = 0, 1,..., k.. V ( ˆβ i ) = c ii, where c ij (or, in this special case, c ii ) is the element in row i and column j (here, column i) of (X X) 1. 3. Cov( ˆβ i, ˆβ j ) = c ij 4. An unbiased estimator of is S = SSE/[n (k + 1)], where SSE = Y Y ˆβ X Y. 5. Each ˆβ i is normally distributed. 6. The random variable [n (k + 1)]S has a χ distribution with n-(k+1) degrees of freedom. 7. The statistics S and ˆβ i, for i = 0, 1,,..., k are independent. 11 Inference Concerning Linear Functions of the Model Parameters: Multiple Linear Regression A Test for a β H 0 : H a : a β = (a β) 0 a β > (a β) 0 a β < (a β) 0 a β (a β) 0 6

Test statistic: Rejection region: T = a ˆβ (a β) 0 S a (X X) 1 a t > t α t < t α t t α/ Here t α is based upon [n-(k+1)] degrees of freedom. A (1 α) 100 % Confidence Interval for a β a ˆβ ± tα/ S a (X X) 1 a 1 Predicting a Particular Value of Y Using Multiple Regression A (1 α) 100 % Prediction Interval for Y when x 1 = x 1, x = x,..., x k = x k where a = [1, x 1, x,..., x k ] a ˆβ ± tα/ S 1 + a (X X) 1 a 13 A test for H 0 : β g+1 = β g+ = = β k = 0 Reduced model (R): Complete model (C): Y = β 0 + β 1 x 1 + + β g x g + ɛ Y = β 0 + β 1 x 1 + + β g x g + β g+1 x g+1 + + β k x k ɛ SSE R = SSE C + (SSE R SSE C ) χ 3 = SSE R χ = SSE C χ 1 = SSE R SSE C 7

F = F > F α χ 1/(k g) χ /(n [k + 1]) = SSE R SSE C /(k g) SSE C /(n [k + 1]) 8