Zellner s Seemingly Unrelated Regressions Model. James L. Powell Department of Economics University of California, Berkeley

Similar documents
Review of Classical Least Squares. James L. Powell Department of Economics University of California, Berkeley

Panel Data Models. James L. Powell Department of Economics University of California, Berkeley

1 Outline. 1. Motivation. 2. SUR model. 3. Simultaneous equations. 4. Estimation

Seemingly Unrelated Regressions in Panel Models. Presented by Catherine Keppel, Michael Anreiter and Michael Greinecker

Multiple Equation GMM with Common Coefficients: Panel Data

Introductory Econometrics

Ch 3: Multiple Linear Regression

Economics 582 Random Effects Estimation

Models, Testing, and Correction of Heteroskedasticity. James L. Powell Department of Economics University of California, Berkeley

the error term could vary over the observations, in ways that are related

The Algebra of the Kronecker Product. Consider the matrix equation Y = AXB where

Section 6: Heteroskedasticity and Serial Correlation

Econ 582 Fixed Effects Estimation of Panel Data

Recent Advances in the Field of Trade Theory and Policy Analysis Using Micro-Level Data

GLS and FGLS. Econ 671. Purdue University. Justin L. Tobias (Purdue) GLS and FGLS 1 / 22

Instrumental Variables, Simultaneous and Systems of Equations

Intermediate Econometrics

Econometrics I KS. Module 2: Multivariate Linear Regression. Alexander Ahammer. This version: April 16, 2018

Heteroskedasticity. We now consider the implications of relaxing the assumption that the conditional

Econometrics Multiple Regression Analysis: Heteroskedasticity

7. GENERALIZED LEAST SQUARES (GLS)

Econometrics. Week 4. Fall Institute of Economic Studies Faculty of Social Sciences Charles University in Prague

GENERALISED LEAST SQUARES AND RELATED TOPICS

Multiple Regression Analysis

Ch 2: Simple Linear Regression

Econ 510 B. Brown Spring 2014 Final Exam Answers

Econ 583 Final Exam Fall 2008

Problems. Suppose both models are fitted to the same data. Show that SS Res, A SS Res, B

Peter Hoff Linear and multilinear models April 3, GLS for multivariate regression 5. 3 Covariance estimation for the GLM 8

Topic 7: Heteroskedasticity

Lecture 4: Heteroskedasticity

Restricted Maximum Likelihood in Linear Regression and Linear Mixed-Effects Model

Weighted Least Squares

y it = α i + β 0 ix it + ε it (0.1) The panel data estimators for the linear model are all standard, either the application of OLS or GLS.

Generalized Method of Moments: I. Chapter 9, R. Davidson and J.G. MacKinnon, Econometric Theory and Methods, 2004, Oxford.

Lecture: Simultaneous Equation Model (Wooldridge s Book Chapter 16)

Binary choice 3.3 Maximum likelihood estimation

Inverse of a Square Matrix. For an N N square matrix A, the inverse of A, 1

Econometrics - 30C00200

Increasing the Power of Specification Tests. November 18, 2018

Econometrics. Week 6. Fall Institute of Economic Studies Faculty of Social Sciences Charles University in Prague

Heteroskedasticity. Part VII. Heteroskedasticity

A COMPARISON OF HETEROSCEDASTICITY ROBUST STANDARD ERRORS AND NONPARAMETRIC GENERALIZED LEAST SQUARES

Generalized Least Squares Theory

Agricultural and Applied Economics 637 Applied Econometrics II

Simple Estimators for Semiparametric Multinomial Choice Models

14 Multiple Linear Regression

Econometrics of Panel Data

Estimating Seemingly Unrelated Regressions with First Order Autoregressive Disturbances

Multiple Linear Regression

LECTURE 11: GENERALIZED LEAST SQUARES (GLS) In this lecture, we will consider the model y = Xβ + ε retaining the assumption Ey = Xβ.

Introductory Econometrics

EC327: Advanced Econometrics, Spring 2007

Simple Linear Regression

Economics 536 Lecture 7. Introduction to Specification Testing in Dynamic Econometric Models

Problem Set #6: OLS. Economics 835: Econometrics. Fall 2012

Introductory Econometrics

Instrumental Variables

Linear Regression. In this problem sheet, we consider the problem of linear regression with p predictors and one intercept,

Robust Inference for Seemingly Unrelated Regression Models

(θ θ ), θ θ = 2 L(θ ) θ θ θ θ θ (θ )= H θθ (θ ) 1 d θ (θ )

Regression: Lecture 2

Quantitative Analysis of Financial Markets. Summary of Part II. Key Concepts & Formulas. Christopher Ting. November 11, 2017

Regression #3: Properties of OLS Estimator

Time Series Models and Inference. James L. Powell Department of Economics University of California, Berkeley

Spatial Regression. 15. Spatial Panels (3) Luc Anselin. Copyright 2017 by Luc Anselin, All Rights Reserved

Multivariate Regression Generalized Likelihood Ratio Tests for FMRI Activation

Chaper 5: Matrix Approach to Simple Linear Regression. Matrix: A m by n matrix B is a grid of numbers with m rows and n columns. B = b 11 b m1 ...

Dealing With Endogeneity

Linear Models and Estimation by Least Squares

PANEL DATA RANDOM AND FIXED EFFECTS MODEL. Professor Menelaos Karanasos. December Panel Data (Institute) PANEL DATA December / 1

Nonstationary Panels

Questions and Answers on Heteroskedasticity, Autocorrelation and Generalized Least Squares

Heteroskedasticity-Robust Inference in Finite Samples

Graduate Econometrics Lecture 4: Heteroskedasticity

Chapter 6. Panel Data. Joan Llull. Quantitative Statistical Methods II Barcelona GSE

Unit roots in vector time series. Scalar autoregression True model: y t 1 y t1 2 y t2 p y tp t Estimated model: y t c y t1 1 y t1 2 y t2

Spring 2017 Econ 574 Roger Koenker. Lecture 14 GEE-GMM

Financial Econometrics Prof. Massimo Guidolin

Linear Models in Econometrics

Spatial inference. Spatial inference. Accounting for spatial correlation. Multivariate normal distributions

Multiple Linear Regression

Review of Econometrics

Topic 10: Panel Data Analysis

. a m1 a mn. a 1 a 2 a = a n

Least Squares Estimation-Finite-Sample Properties

UNIVERSITY OF OSLO DEPARTMENT OF ECONOMICS

Bootstrapping Heteroskedasticity Consistent Covariance Matrix Estimator

Heteroskedasticity and Autocorrelation

AMS-207: Bayesian Statistics

Non-Spherical Errors

This model of the conditional expectation is linear in the parameters. A more practical and relaxed attitude towards linear regression is to say that

LECTURE 2 LINEAR REGRESSION MODEL AND OLS

A Practitioner s Guide to Cluster-Robust Inference

Applied Econometrics (QEM)

Properties of the least squares estimates

Econometrics I Lecture 3: The Simple Linear Regression Model

Note on Bivariate Regression: Connecting Practice and Theory. Konstantin Kashin

Econometrics of Panel Data

CLUSTER EFFECTS AND SIMULTANEITY IN MULTILEVEL MODELS

Transcription:

Zellner s Seemingly Unrelated Regressions Model James L. Powell Department of Economics University of California, Berkeley Overview The seemingly unrelated regressions (SUR) model, proposed by Zellner, can be viewed as a special case of the generalized regression model E(y) β, V(y) σ 2 Ω; however, it does not share all of the features or problems of other leading special cases (e.g., models of heteroskedasticity or serial correlation). While, like those models, the matrix Ω generally involves unknown parameters which must be estimated, the usual estimators for the covariance matrix of the least squares estimator ˆβ LS are valid, so that the usual inference procedures based on normal theory are valid if the dependent variable y is multinormal or if the sample size N is large and suitable limit theorems are applicable. Also, unlike those other models, there is little reason to test the null hypothesis H : Ω I; the form of Ω is straightforward and its parameters are easy to estimate consistently, so a feasible version of Aitken s GLS estimator is an attractive alternative to the asymptotically-inefficient LS estimator. The basic SUR model assumes that, for each individual observation i, there are M dependent variables y i,,y ij,,y im available, each with its own linear regression model: y ij x ijβ j + ε ij, i,,n, or, with the usual stacking of observations over i, y j j β j + ε j for j,,m, where y j and ε j are N-vectors and j is an N K j matrix, where K j dim(β j ) is the number of regressors for the j th regression. The standard conditions for the classical regression model are assumed to hold for each j : namely, E(y j ) j β j, V(y j ) σ jj I N,

with j nonstochastic and rank( j ) K j. Under these conditions, and the additional condition of multinormality of y j, the usual inference theory is valid for the classical LS estimator of β j, applied separately to each equation. However, the SUR model permits nonzero covariance between the error terms ε ij and ε ij for a given individual i across equations j and k, i.e., Cov(ε ij,ε ik )σ ij while assuming Cov(ε ij,ε i k) if i 6 i. This can be expressed more compactly in matrix form: C(ε j, ε k )σ jk I N. It is the potential nonzero covariance across equations j and k that allows for an improvement in efficiency of GLS relative to the classical LS estimator of each β j. Kronecker Product Notation Zellner s insight was that, like the usual stacking of the individual dependent variables y ij into an N-vector y j, those latter vectors can themselves be stacked into an MN-dimensional vector y, with a corresponding arrangement for the error terms, coefficient vectors, and regressors: y ε y y 2 (MN ), ε ε 2 (MN ), β (K ) y M ε M β β 2 β M, and with (MN K) 2 M K M K j. j, With this notation, and the individual assumptions for each equation j, it follows that E(y) β 2

and V(y) (MN MN) σ I N σ 2 I N σ M I N σ 2 I N σ 22 I N σ M I N σ MM I N This nonscalar covariance matrix is a particular mixture of the matrix σ σ 2 σ M Σ σ 2 σ 22 (M M) σ M σ MM and the (N N) identity matrix I N. A notation system for such combinations was proposed by the otherwise-despicable mathematician Kronecker, the so-called Kronecker product notation; for two matrices A [a ij ](i,,l, j,m) and B, the Kronecker product of A and B is defined as With this notation, clearly for the stacked SUR model. A B a B a 2 B a M B a 2 B a 22 B a L B a LM B V(y) Σ I N Kronecker products satisfy a distributive rule, which will come in handy later:.. (A B)(C D) AC BD, assuming all matrix products are well defined. From this rule follows another for inverses of Kronecker products: (A B) A B, assuming both A and B are invertible. 3

Least Squares and Generalized Least Squares With the foregoing notation, the classical least squares estimator for the vector β can be expressed as ˆβ LS y ( ) y ( 2 2) 2 y 2 ( M M) M y M In contrast, the GLS estimator of β (assuming Σ is known) is. ˆβ GLS (Σ I N ) (Σ I N ) y Σ I N Σ I N y σ ( ) σ 2 ( 2) σ M ( M) σ 2 ( 2 ) σ 22 ( 2 2) σ M ( M ) σ MM ( M M) 2 M P j σj y j P j σ2j y j P j σmj y j, where σ ij is defined to be the element in the i th row and j th column of Σ, i.e., Σ σ ij. To get a better idea of what is going on with the GLS estimator, consider the special case M 2, with ˆβ LS µ b b 2 σ 22 and ˆβ GLS µ ˆβ ˆβ 2 ; then it can be shown that the GLS estimators ˆβ and ˆβ 2 satisfy the two equations µ σ2 ˆβ b y 2 2ˆβ 2, ˆβ 2 b 2 µ σ2 σ 2 2 2 y ˆβ. Thus the GLS estimators can be viewed as adjusted versions of classical LS, where the adjustment involves the regression of the GLS residuals from the other equation on the regressors from each equation. As noted by Luce in JASA, 964, the GLS estimator for this model can be calculated sequentially by including appropriately-reweighted residuals from all other equations as additional regressors for each equation. Another important special case is when the matrix Σ is diagonal, i.e., σ ij if i 6 j. In this case, 4

since Σ diag[/σ ii ], it follows that ˆβ GLS σ ( ) σ 22 ( 2 2) σ MM ( M M) ( ) y ( 2 2) 2 y 2 ( M M) M y M ˆβ LS. σ y σ 22 2 y 2 σ MM M y M Not surprisingly, then, if there is no covariance across equations in the error terms, there is no prospect for an efficiency improvement in the GLS estimator relative to LS, applied equation by equation. Still another important special case is when the matrix of regressors is identical for each equation, i.e., j for some N K matrix, with K K/M. Here the stacked matrix takes the form (MN K) (I M ), and the GLS estimator also reduces to classical LS: ˆβ GLS (Σ I N ) (Σ I N ) y (I M ) Σ I N (IM ) (IM ) Σ I N y Σ Σ y y I M y 2 ˆβ LS. ( ) y ( ) y 2 ( ) y M y M Some intuition for this reduction can be obtained by considering Luce s result that GLS can be obtained iteratively, by starting from classical LS estimators and including the residuals from other equations as 5

regressors for each equation. Since the LS residuals are, by construction, orthogonal to the common matrix of regressors, their inclusion in each equation will not affect the LS estimates of the β j coefficients. An extension of this argument implies that, if each matrix of regressors j has a submatrix in common for example, if all equations have an intercept term then the GLS coefficients corresponding to those common regressors will be identical to their LS counterparts. Feasible GLS An obvious estimator of the unknown covariance matrix Σ V(y) [σ ij ] would be ˆΣ [ˆσ ij ], with ˆσ jk N y j j ˆβ j y k kˆβ k ; while these estimators are not unbiased for σ jk, they are consistent under the usual conditions, and obtaining unbiased estimators for σ jk when j 6 k involves more than a simple degrees of freedom adjustment. Again imposing reasonable regularity conditions, it can be shown that the feasible GLS estimator ˆβ FGLS ˆΣ I N ˆΣ I N y is asymptotically equivalent to the infeasible GLS estimator which assumes Σ is known: N ˆβFGLS ˆβ GLS p. Thus N ˆβ FGLS β d N (, V), where V µ plim N Σ I N µ plim N I N plim ˆV, so inference on the parameter vector β can be carried out using the approximate normality of ˆβ FGLS : A ˆβ FGLS N β, ˆV. 6