Chapter 7 Generalized and Weighted Least Squares Estimation. In this method, the deviation between the observed and expected values of

Similar documents
Chapter 2 - The Simple Linear Regression Model S =0. e i is a random error. S β2 β. This is a minimization problem. Solution is a calculus exercise.

LINEAR REGRESSION ANALYSIS. MODULE IX Lecture Multicollinearity

However, since P is a symmetric idempotent matrix, of P are either 0 or 1 [Eigen-values

Chapter 3. Two-Variable Regression Model: The Problem of Estimation

e i is a random error

The Ordinary Least Squares (OLS) Estimator

Economics 130. Lecture 4 Simple Linear Regression Continued

LINEAR REGRESSION ANALYSIS. MODULE IX Lecture Multicollinearity

Econ107 Applied Econometrics Topic 3: Classical Model (Studenmund, Chapter 4)

Econometrics of Panel Data

Econ Statistical Properties of the OLS estimator. Sanjaya DeSilva

x i1 =1 for all i (the constant ).

The Multiple Classical Linear Regression Model (CLRM): Specification and Assumptions. 1. Introduction

Systems of Equations (SUR, GMM, and 3SLS)

Dr. Shalabh Department of Mathematics and Statistics Indian Institute of Technology Kanpur

Linear Approximation with Regularization and Moving Least Squares

Fall 2012 Analysis of Experimental Measurements B. Eisenstein/rev. S. Errede

Outline. Zero Conditional mean. I. Motivation. 3. Multiple Regression Analysis: Estimation. Read Wooldridge (2013), Chapter 3.

Now we relax this assumption and allow that the error variance depends on the independent variables, i.e., heteroskedasticity

Primer on High-Order Moment Estimators

LECTURE 9 CANONICAL CORRELATION ANALYSIS

Dr. Shalabh Department of Mathematics and Statistics Indian Institute of Technology Kanpur

ECONOMICS 351*-A Mid-Term Exam -- Fall Term 2000 Page 1 of 13 pages. QUEEN'S UNIVERSITY AT KINGSTON Department of Economics

Maximum Likelihood Estimation of Binary Dependent Variables Models: Probit and Logit. 1. General Formulation of Binary Dependent Variables Models

β0 + β1xi. You are interested in estimating the unknown parameters β

Tests of Single Linear Coefficient Restrictions: t-tests and F-tests. 1. Basic Rules. 2. Testing Single Linear Coefficient Restrictions

Explaining the Stein Paradox

Estimation: Part 2. Chapter GREG estimation

Chat eld, C. and A.J.Collins, Introduction to multivariate analysis. Chapman & Hall, 1980

Chapter 11: Simple Linear Regression and Correlation

Inner Product. Euclidean Space. Orthonormal Basis. Orthogonal

10-701/ Machine Learning, Fall 2005 Homework 3

Dr. Shalabh Department of Mathematics and Statistics Indian Institute of Technology Kanpur

Lecture 19. Endogenous Regressors and Instrumental Variables

Properties of Least Squares

GMM Method (Single-equation) Pongsa Pornchaiwiseskul Faculty of Economics Chulalongkorn University

Chapter 12 Analysis of Covariance

a. (All your answers should be in the letter!

First Year Examination Department of Statistics, University of Florida

Maximum Likelihood Estimation of Binary Dependent Variables Models: Probit and Logit. 1. General Formulation of Binary Dependent Variables Models

Exam. Econometrics - Exam 1

Feb 14: Spatial analysis of data fields

Composite Hypotheses testing

Chapter 9: Statistical Inference and the Relationship between Two Variables

1. Inference on Regression Parameters a. Finding Mean, s.d and covariance amongst estimates. 2. Confidence Intervals and Working Hotelling Bands

APPENDIX A Some Linear Algebra

Econ107 Applied Econometrics Topic 9: Heteroskedasticity (Studenmund, Chapter 10)

j) = 1 (note sigma notation) ii. Continuous random variable (e.g. Normal distribution) 1. density function: f ( x) 0 and f ( x) dx = 1

where I = (n x n) diagonal identity matrix with diagonal elements = 1 and off-diagonal elements = 0; and σ 2 e = variance of (Y X).

3.1 Expectation of Functions of Several Random Variables. )' be a k-dimensional discrete or continuous random vector, with joint PMF p (, E X E X1 E X

β0 + β1xi. You are interested in estimating the unknown parameters β

[The following data appear in Wooldridge Q2.3.] The table below contains the ACT score and college GPA for eight college students.

n α j x j = 0 j=1 has a nontrivial solution. Here A is the n k matrix whose jth column is the vector for all t j=0

F8: Heteroscedasticity

8.4 COMPLEX VECTOR SPACES AND INNER PRODUCTS

A Comparative Study for Estimation Parameters in Panel Data Model

CIS526: Machine Learning Lecture 3 (Sept 16, 2003) Linear Regression. Preparation help: Xiaoying Huang. x 1 θ 1 output... θ M x M

2.3 Nilpotent endomorphisms

Number of cases Number of factors Number of covariates Number of levels of factor i. Value of the dependent variable for case k

Topic 5: Non-Linear Regression

Computation of Higher Order Moments from Two Multinomial Overdispersion Likelihood Models

Radar Trackers. Study Guide. All chapters, problems, examples and page numbers refer to Applied Optimal Estimation, A. Gelb, Ed.

Statistics for Economics & Business

Limited Dependent Variables

Linear Regression Analysis: Terminology and Notation

ISQS 6348 Final Open notes, no books. Points out of 100 in parentheses. Y 1 ε 2

WORKSHOP ON THE ANALYSIS OF NEURAL DATA 2001 MARINE BIOLOGICAL LABORATORY WOODS HOLE, MASSACHUSETTS

Professor Chris Murray. Midterm Exam

Yong Joon Ryang. 1. Introduction Consider the multicommodity transportation problem with convex quadratic cost function. 1 2 (x x0 ) T Q(x x 0 )

Resource Allocation and Decision Analysis (ECON 8010) Spring 2014 Foundations of Regression Analysis

Sampling Theory MODULE V LECTURE - 17 RATIO AND PRODUCT METHODS OF ESTIMATION

Linear Algebra and its Applications

SIO 224. m(r) =(ρ(r),k s (r),µ(r))

Lecture 3 Specification

Interval Estimation in the Classical Normal Linear Regression Model. 1. Introduction

Lecture 4 Hypothesis Testing

Chapter 5. Solution of System of Linear Equations. Module No. 6. Solution of Inconsistent and Ill Conditioned Systems

First day August 1, Problems and Solutions

Lecture 2: Prelude to the big shrink

β0 + β1xi and want to estimate the unknown

THE ROYAL STATISTICAL SOCIETY 2006 EXAMINATIONS SOLUTIONS HIGHER CERTIFICATE

Solutions Homework 4 March 5, 2018

Bose (1942) showed b t r 1 is a necessary condition. PROOF (Murty 1961): Assume t is a multiple of k, i.e. t nk, where n is an integer.

Statistics for Business and Economics

Lectures - Week 4 Matrix norms, Conditioning, Vector Spaces, Linear Independence, Spanning sets and Basis, Null space and Range of a Matrix

Econometrica Supplementary Material

4 Analysis of Variance (ANOVA) 5 ANOVA. 5.1 Introduction. 5.2 Fixed Effects ANOVA

Introduction to Regression

If we apply least squares to the transformed data we obtain. which yields the generalized least squares estimator of β, i.e.,

Interpreting Slope Coefficients in Multiple Linear Regression Models: An Example

ECONOMETRICS - FINAL EXAM, 3rd YEAR (GECO & GADE)

Statistics for Managers Using Microsoft Excel/SPSS Chapter 13 The Simple Linear Regression Model and Correlation

MEM 255 Introduction to Control Systems Review: Basics of Linear Algebra

since [1-( 0+ 1x1i+ 2x2 i)] [ 0+ 1x1i+ assumed to be a reasonable approximation

Example: (13320, 22140) =? Solution #1: The divisors of are 1, 2, 3, 4, 5, 6, 9, 10, 12, 15, 18, 20, 27, 30, 36, 41,

The Geometry of Logit and Probit

Exercise 1 The General Linear Model : Answers

Chapter 3 Describing Data Using Numerical Measures

The written Master s Examination

A Monte Carlo Study for Swamy s Estimate of Random Coefficient Panel Data Model

Transcription:

Chapter 7 Generalzed and Weghted Least Squares Estmaton The usual lnear regresson model assumes that all the random error components are dentcally and ndependently dstrbuted wth constant varance. When ths assumpton s volated, then ordnary least squares estmator of regresson coeffcent looses ts property of mnmum varance n the class of lnear and unbased estmators. The volaton of such assumpton can arse n anyone of the followng stuatons:. The varance of random error components s not constant.. The random error components are not ndependent. 3. The random error components do not have constant varance as well as they are not ndependent. In such cases, the covarance matrx of random error components does not reman n the form of an dentty matrx but can be consdered as any postve defnte matrx. Under such assumpton, the OLSE does not reman effcent as n the case of dentty covarance matrx. The generalzed or weghted least squares method s used n such stuatons to estmate the parameters of the model. In ths method, the devaton between the observed and expected values of y s multpled by a weght ω where ω s chosen to be nversely proportonal to the varance of y. For smple lnear regresson model, the weghted least squares functon s n ( ). S( β, β ) = ω y β β x The least squares normal equatons are obtaned by dfferentatng S( β, β) wth respect to β and β and equatng them to zero as n n n ˆ β ω + ˆ β ω x = ω y = = = n n n x + x = xy = = = ˆ β ω ˆ β ω ω. Soluton of these two normal equatons gve the weghted least squares estmate of β and β.

Generalzed least squares estmaton Suppose n usual multple regresson model = β + ε ε = ε = σ, y X wth E( ), V( ) I the assumpton V( ε) = σ I s volated and become V ( ε) = σ Ω where Ω s a known n n nonsngular, postve defnte and symmetrc matrx. Ths structure of Ω ncorporates both the cases. - when Ω s dagonal but wth unequal varances and - when Ω s not necessarly dagonal dependng on the presence of correlated errors, some of dagonal elements are nonzero. The OLSE of β s b= ( X ' X) X ' y In such cases OLSE gves unbased estmate but has more varablty as E( b) ( X ' X) X ' E( y) ( X ' X) X ' X = = β = β V b X X X V y X X X X X X X X X ( ) = ( ' ) ' ( ) ( ' ) = σ ( ' ) ' Ω ( ' ). Now we attempt to fnd better estmator as follows: Snce Ω s postve defnte, symmetrc, so there exsts a nonsngular matrx K such that. KK ' = Ω. Then n the model premutlply by y = Xβ + ε, K, ths gves K y = K Xβ + K or z = Bβ + g ε where and z K yb, K X, g K ε = = =. Now observe that Eg K Eε ) ( ) = ( ) =

{ }{ } Vg ( ) = E g Eg ( ) g Eg ( ) ' = E( gg ') = E K εε ' K' = K E( εε ') K' = σ K ΩK' = σ K KK ' K ' = σ I. Thus the elements of g have mean and they are uncorrelated. So ether mnmze S( β ) = g' g = ε' Ω ε and get normal equatons as - ˆ (X' Ω X) β = X ' Ω y ˆ β = X Ω X X Ω y or ( ' ) '. = Ω ( y Xβ)' ( y Xβ) Alternatvely, we can apply OLS to transformed model and obtan OLSE of β as ˆ ( BB ' ) ' β = Bz X K K X X K K y = ( ' ' ) ' ' X X X y = ( ' Ω ) ' Ω Ths s termed as generalzed least squares estmator (GLSE) of β. The estmaton error of GLSE s ˆ ( BB ' ) B'( B g) β = β + = β + BB ˆ β β = Bg ( ' ) ' or ( BB ' ) Bg '. Then E ˆ β β = BB BEg = ( ) ( ' ) ' ( ) whch shows that GLSE s an unbased estmator of β. The covarance matrx of GLSE s gven by { }{ } V( ˆ β) = E ˆ β E( ˆ β) ˆ β E( ˆ β) ' E ( BB ' ) BggB ' ' '( BB ' ) = = ( BB ' ) BEgg ' ( ') B'( BB ' ). 3

Snce so E( gg ') = K E( εε ') K ' = σ K ΩK' = σ K = σ I, KK ' K ' V( ˆ β) = σ ( BB ' ) BBBB ' ( ' ) = σ ( BB ' ) = σ ( X ' K' K X) = σ ( X ' Ω X). Now we prove that GLSE s the best lnear unbased estmator of β. The Gauss-Markov theorem for the case Var( ε ) = Ω The Gauss-Markov theorem establshes that the generalzed least-squares (GLS) estmator of ˆ β gven by β = ( X ' Ω X) X ' Ω y, s BLUE (best lnear unbased estmator). By best β, we mean that ˆβ mnmzes the varance for any lnear combnaton of the estmated coeffcents, ( ˆ β ) ( ' Ω ) ' Ω = ( X ' Ω X) X ' Ω E( y) = ( X ' Ω X) X ' Ω Xβ = β. E E X X X y Thus ˆβ s an unbased estmator of β. The covarance matrx of ˆβ s gven by V ˆ X Ω X X Ω V y X Ω X X Ω ( X ' Ω X) X ' Ω Ω ( X ' Ω X) X ' Ω ' ( X ' Ω X) X ' Ω Ω Ω X( X ' Ω X) = ( X ' Ω X). ( β ) ( ' ) ' ( ) ( ' ) ' ' 'βˆ. We note that Thus, Var( ' ˆ β) = ' Var( ˆ β) = Ω ' ( X ' X). 4

Let β be another unbased estmator of β that s a lnear combnaton of the data. Our goal, then, s to show that Var( ' β ) '( X ' Ω X ) wth at least one such that Var( ' β ) '( X ' X ) Ω. We frst note that we can wrte any other estmator of β that s a lnear combnaton of the data as β ( X ' Ω X) X ' Ω + B y+ b * where B s an p n matrx and estmator to form the alternatve estmate. Then * ( ) E( β ) = E ( X ' Ω X) X ' Ω + B y+ b o = ( X ' Ω X) X ' Ω + B E( y) + b * ( X ' Ω X ) X ' Ω + B XB + b = ( X ' Ω X) X ' Ω Xβ + BXβ + b = β + BX β + b. * * Consequently, β s unbased f and only f both * ( ) b * o s a p vector of constants that approprately adjusts the GLS V ( β ) = Var ( X ' Ω X ) X ' Ω + B y = X Ω X X Ω + B V y X Ω X X Ω + B = ( X ' Ω X) X ' Ω + B Ω ( X ' Ω X) X ' Ω + B ' ( X ' Ω X) X ' Ω + B Ω Ω X( X ' Ω X) + B' ( X ' Ω X) + BΩB' ( ' ) ' ( ) ( ' ) ' ' because BX =, whch mples that ( BX )' = X ' B ' =. Then Var( ' β) = ' V ( β) ( X X B B ) = ' ( ' Ω ) + Ω ' '( ' ) ' ' = X Ω X + BΩB = Var(' ˆ β ) + ' BΩB '. * b = and BX =. The covarance matrx of β s We note that Ω s a postve defnte matrx. Consequently, there exsts some nonsngulat matrx K such that Ω= K' K As a result, BΩ B ' = BK ' KB ' s at least postve semdefnte matrx; hence, ' BΩB'. Next note that we can defne * = KB '. As a result, p *' * * = ' BΩ B' = = whch must be strctly greater than for some unless B = lnear unbased estmator.. Thus, the GLS estmate of β s the best 5

Weghted least squares estmaton When ε 's are uncorrelated and have unequal varances, then ω V ( ε) σ σ = Ω= ω. ω n The estmaton procedure s usually called as weghted least squares. Let W = Ω then the weghted least squares estmator of β s obtaned by solvng normal equaton ( X ' WX ) ˆ β = X ' Wy whch gves ˆ ( X ' WX ) ' β = X Wy where ω, ω,..., ω n are called the weghts. The observatons wth large varances usual have smaller weghts than observatons wth small varance. 6