GLS and FGLS. Econ 671. Purdue University. Justin L. Tobias (Purdue) GLS and FGLS 1 / 22

Similar documents
Regression #3: Properties of OLS Estimator

Lecture 10: Panel Data

Regression #4: Properties of OLS Estimator (Part 2)

Regression #2. Econ 671. Purdue University. Justin L. Tobias (Purdue) Regression #2 1 / 24

Regression #5: Confidence Intervals and Hypothesis Testing (Part 1)

Panel Data Models. James L. Powell Department of Economics University of California, Berkeley

Hypothesis Testing. Econ 690. Purdue University. Justin L. Tobias (Purdue) Testing 1 / 33

Gibbs Sampling in Linear Models #1

Intermediate Econometrics

Panel Data Models. Chapter 5. Financial Econometrics. Michael Hauser WS17/18 1 / 63

1 Introduction to Generalized Least Squares

Properties of the least squares estimates

1 Motivation for Instrumental Variable (IV) Regression

Econometrics of Panel Data

ECON 5350 Class Notes Functional Form and Structural Change

Simple Linear Regression

Econometrics - 30C00200

Regression #8: Loose Ends

Introduction to Econometrics Final Examination Fall 2006 Answer Sheet

Multiple Regression Analysis

Econometrics. Week 4. Fall Institute of Economic Studies Faculty of Social Sciences Charles University in Prague

Zellner s Seemingly Unrelated Regressions Model. James L. Powell Department of Economics University of California, Berkeley

Introduction to Econometrics Midterm Examination Fall 2005 Answer Key

Casuality and Programme Evaluation

Seemingly Unrelated Regressions in Panel Models. Presented by Catherine Keppel, Michael Anreiter and Michael Greinecker

Gibbs Sampling in Linear Models #2

Cross Sectional Time Series: The Normal Model and Panel Corrected Standard Errors

Lesson 17: Vector AutoRegressive Models

Weighted Least Squares

Economics 582 Random Effects Estimation

Questions and Answers on Heteroskedasticity, Autocorrelation and Generalized Least Squares

Nonparametric Regression

Problem Set #6: OLS. Economics 835: Econometrics. Fall 2012

Categorical Predictor Variables

Fixed Effects Models for Panel Data. December 1, 2014

F9 F10: Autocorrelation

Econ 583 Final Exam Fall 2008

Advanced Econometrics

Statistics 910, #5 1. Regression Methods

For more information about how to cite these materials visit

10 Panel Data. Andrius Buteikis,

MS&E 226: Small Data

The BLP Method of Demand Curve Estimation in Industrial Organization

The Simple Linear Regression Model

Econometric Analysis of Cross Section and Panel Data

Spatial Regression. 15. Spatial Panels (3) Luc Anselin. Copyright 2017 by Luc Anselin, All Rights Reserved

Econometrics. Week 6. Fall Institute of Economic Studies Faculty of Social Sciences Charles University in Prague

MS&E 226: Small Data. Lecture 11: Maximum likelihood (v2) Ramesh Johari

Ordinary Least Squares Regression

Discrete Dependent Variable Models

Gibbs Sampling in Endogenous Variables Models

Eco517 Fall 2014 C. Sims FINAL EXAM

Multivariate Time Series

IEOR 165 Lecture 7 1 Bias-Variance Tradeoff

Time-Series Cross-Section Analysis

ECON The Simple Regression Model

Econometrics Multiple Regression Analysis: Heteroskedasticity

WU Weiterbildung. Linear Mixed Models

Reliability of inference (1 of 2 lectures)

1 The Multiple Regression Model: Freeing Up the Classical Assumptions

BIOS 2083 Linear Models c Abdus S. Wahed

Model Mis-specification

Auto correlation 2. Note: In general we can have AR(p) errors which implies p lagged terms in the error structure, i.e.,

ECO375 Tutorial 8 Instrumental Variables

ECON3150/4150 Spring 2015

7 Sensitivity Analysis

Instrumental Variables, Simultaneous and Systems of Equations

Repeated observations on the same cross-section of individual units. Important advantages relative to pure cross-section data

MAT2377. Rafa l Kulik. Version 2015/November/26. Rafa l Kulik

An overview of applied econometrics

The linear regression model: functional form and structural breaks

Testing Error Correction in Panel data

Bias Variance Trade-off

Lecture 24: Weighted and Generalized Least Squares

Applied Econometrics (MSc.) Lecture 3 Instrumental Variables

Econ 836 Final Exam. 2 w N 2 u N 2. 2 v N

Econ 582 Fixed Effects Estimation of Panel Data

Nonparametric Heteroscedastic Transformation Regression Models for Skewed Data, with an Application to Health Care Costs

14.74 Lecture 10: The returns to human capital: education

MA Advanced Econometrics: Applying Least Squares to Time Series

Lecture 12: Application of Maximum Likelihood Estimation:Truncation, Censoring, and Corner Solutions

Spatial inference. Spatial inference. Accounting for spatial correlation. Multivariate normal distributions

Applied Quantitative Methods II

6 Pattern Mixture Models

Groupe de lecture. Instrumental Variables Estimates of the Effect of Subsidized Training on the Quantiles of Trainee Earnings. Abadie, Angrist, Imbens

Heteroscedasticity. Jamie Monogan. Intermediate Political Methodology. University of Georgia. Jamie Monogan (UGA) Heteroscedasticity POLS / 11

Non-Spherical Errors

Chapter 6. Panel Data. Joan Llull. Quantitative Statistical Methods II Barcelona GSE

Chapter 15 Panel Data Models. Pooling Time-Series and Cross-Section Data

Heteroskedasticity. in the Error Component Model

Final Exam. Economics 835: Econometrics. Fall 2010

Instrumental Variables and Two-Stage Least Squares

Sensitivity of GLS estimators in random effects models

Weighted Least Squares

Scatter plot of data from the study. Linear Regression

Data Mining Stat 588

Dealing With Endogeneity

Problem Set 2

CLUSTER EFFECTS AND SIMULTANEITY IN MULTILEVEL MODELS

Generalized Least Squares Theory

Transcription:

GLS and FGLS Econ 671 Purdue University Justin L. Tobias (Purdue) GLS and FGLS 1 / 22

In this lecture we continue to discuss properties associated with the GLS estimator. In addition we discuss the practical issue of what to do when the covariance matrix is unknown (which, of course, it is), leading to a discussion of the Feasible Generalized Least Squares (FGLS) estimator. We close the lecture by providing a series of models where FGLS can be implemented, leaving most details of the implementation to other courses. Justin L. Tobias (Purdue) GLS and FGLS 2 / 22

We begin our GLS investigation by deriving the (finite sample) covariance matrix. To this end, we note: Recall that the variance of the OLS estimator in the presence of a general Ω was: Aitken s theorem tells us that the GLS variance is smaller. This is obvious, right? Justin L. Tobias (Purdue) GLS and FGLS 3 / 22

To establish this result, note: We claim that this difference can be written in the form: where The proof is simple - just multiply the expression out. Var( ˆβ OLS X ) is produced directly in the product, while the remaining terms combine to give Var( ˆβ GLS X ). The matrix AΩA is clearly positive definite since Ω is positive definite. Justin L. Tobias (Purdue) GLS and FGLS 4 / 22

GLS Asymptotics We do not discuss the asymptotic derivations of the GLS estimator here. Essentially, the methods applied for the OLS case can again be applied upon transforming the data. In this regard, we assume where V is finite and nonsingular. Asymptotic normality follows using similar arguments: Justin L. Tobias (Purdue) GLS and FGLS 5 / 22

Feasible Generalized Least Squares The assumption that Ω is known is, of course, a completely unrealistic one. In many situations (see the examples that follow), we either suppose, or the model naturally suggests, that Ω is comprised of a finite set of parameters, say α, and once α is known, Ω is also known. Suppose that a consistent estimator of α is available, and denote this estimator as ˆα. We could then replace Ω with ˆΩ = Ω(ˆα) and implement the Feasible Generalized Least Squares Estimator (FGLS): Justin L. Tobias (Purdue) GLS and FGLS 6 / 22

FGLS Asymptotics To work out the asymptotics of the FGLS estimator, a standard approach is to first demonstrate that FGLS is asymptotically equivalent to GLS, so that the distribution theory for GLS can be borrowed and applied to the FGLS estimator (which remains true, asymptotically). We say that two estimators are asymptotically equivalent if: Sufficient conditions for this to be true are: and Justin L. Tobias (Purdue) GLS and FGLS 7 / 22

A few things to note: The preference for FGLS over OLS is an asymptotic one. In fact, you can manufacture cases where the OLS estimator is preferable to FGLS in finite samples. Interestingly note that FGLS is asymptotically efficient (among the class of linear unbiased estimators) even though we only require a consistent estimator of Ω, not necessarily an efficient one. The finite sample properties of FGLS are quite difficult to work out, in general, as we use the data twice. Rather surprisingly, however, the FGLS estimator is often unbiased. Justin L. Tobias (Purdue) GLS and FGLS 8 / 22

FGLS Example #1 Consider the regression model: Here, D i is an observed, binary dummy variable. Think about this as representing two groups: for example, variances may potentially differ across men and women. How would estimation proceed here? To implement FGLS, we require consistent estimates of the variance parameters. How could we obtain such estimates? Justin L. Tobias (Purdue) GLS and FGLS 9 / 22

FGLS Example #1 Thus, in practice, we might proceed as follows: 1 Run OLS and obtain ˆβ. 2 Given ˆβ, estimate σ1 2 and σ2 2 slide. using the formula provided on the last 3 Calculate ˆβ FGLS as: ˆβ FGLS = (X ˆΩ 1 X ) 1 X ˆΩ 1 y, ˆΩ diag{di ˆσ 2 1 + (1 D i )ˆσ 2 2} n i=1. 4 Calculate Var( ˆβ FGLS X ) as: (X ˆΩ 1 X ) 1 Justin L. Tobias (Purdue) GLS and FGLS 10 / 22

FGLS Example #2: General Parametric Heteroscedasticity Consider the regression model: The z i are observed in the data set. Potentially these are the same as x i, though z i may contain more (or fewer) variables than are included in x i. The first element of z i is assumed to be an intercept. In your study, it is possible that the main interest lies in the conditional variability of y rather than (just) the mean. For example, it might be interesting to see how the variability of earnings is related to things like education, test scores, etc. This motivates the adoption of a parametric model to describe the heteroscedasticity. (In contrast to the White approach, where these are merely nuisance parameters). Justin L. Tobias (Purdue) GLS and FGLS 11 / 22

FGLS Example #2: General Parametric Heteroscedasticity How would we proceed here? Again, we recognize that ˆβ is consistent for β and, likewise ˆɛ 2 i To fix ideas (though this is not necessary), suppose p ɛ 2 i. Then, Thus, Justin L. Tobias (Purdue) GLS and FGLS 12 / 22

FGLS Example #2: General Parametric Heteroscedasticity This suggests that we can estimate α from an OLS regression of log(ˆɛ 2 i ) on z i. (Another useful alternative is nonlinear least-squares, though we do not discuss that here). The intercept parameter included in α, however, will be biased and inconsistent. This relates to the fact that, although the random variable ν i will have mean 1, the mean of u i = log ν i is not necessarily zero. In fact, when the errors are normally distributed, the mean of the log of the chi-square is approximately -1.27. Thus, if we are willing to make the normality assumption, we can fix this problem by simply adding 1.27 to the estimated intercept parameter. (But, does this really matter for purposes of point estimation?) Justin L. Tobias (Purdue) GLS and FGLS 13 / 22

FGLS Example #3: Random Effects Panel Models Consider the regression model: We now have 2 subscripts: i (for individual or unit) and t for time. The primary problem in these models is often to account for correlation in outcomes for a particular unit over time. The presence of the random effects α i accomplishes just that. Justin L. Tobias (Purdue) GLS and FGLS 14 / 22

FGLS Example #3: Random Effects Panel Models y it = x it β + α i + ɛ it, i = 1, 2,..., n, t = 1, 2,..., T. To this we additionally assume that E(α i X ) = 0 E(ɛ it X ) = 0 Var(α i X ) = σ 2 α Var(ɛ it X ) = σɛ 2 E(α i α j X ) = 0, j i E(ɛ it ɛ js X ) = 0 whenever i j or t s E(α i ɛ jt X ) = 0 i, j, t Justin L. Tobias (Purdue) GLS and FGLS 15 / 22

FGLS Example #3: Random Effects Panel Models Now, stack observations for individual i over t. Let y i1 y i2 y i =. y it What is Var(y i X ) under the given assumptions? It follows that the entire dependent variable vector, y, stacked across all individuals i in a similar manner has conditional covariance matrix: Justin L. Tobias (Purdue) GLS and FGLS 16 / 22

FGLS Example #3: Random Effects Panel Models So the panel model, like those before it, fits into the general framework of GLS estimation, where the covariance matrix depends only on a finite number of parameters (in this case, just 2 parameters). Methods for estimating σ 2 α and σ 2 ɛ exist, but are discussed in detail in 672. Justin L. Tobias (Purdue) GLS and FGLS 17 / 22

FGLS Example #4: Time Series Consider the regression model: (the variable y t 1 is called a lag of y t ). Note, then, by substitution: and continuing, so that provided Justin L. Tobias (Purdue) GLS and FGLS 18 / 22

FGLS Example #4: Time Series Likewise, In a similar manner, we can show: so that Again, this fits within the GLS structure. Justin L. Tobias (Purdue) GLS and FGLS 19 / 22

FGLS Example #5: SUR Models Suppose that the outcome for one unit at time t (perhaps the stock return for company 1 at time t) is: Likewise, the return for a second unit is given as: Should we run separate regressions to estimate β 1 and β 2? Justin L. Tobias (Purdue) GLS and FGLS 20 / 22

FGLS Example #5: SUR Models In particular, there may be contemporaneous correlation among the errors that could lead to (asymptotic) efficiency gains via the use of FGLS. That is, suppose: [ ɛ1t ɛ 2t ] [( iid 0 N 0 ) ( σ 2, 1 σ 12 σ 12 σ2 2 )] N (0, Σ). Justin L. Tobias (Purdue) GLS and FGLS 21 / 22

Thus, if we can consistently estimate the three parameters of Σ, there will be asymptotic efficiency gains to using FGLS instead of running separate regressions, equation-by-equation. Justin L. Tobias (Purdue) GLS and FGLS 22 / 22 FGLS Example #5: SUR Models In this case, we could stack the observations into one big regression model, where the resulting covariance matrix will not be diagonal with a constant variance. Specifically, we can write: or where