The Equality of OLS and GLS Estimators in the

Similar documents
A note on the equality of the BLUPs for new observations under two linear models

A CLASS OF THE BEST LINEAR UNBIASED ESTIMATORS IN THE GENERAL LINEAR MODEL

New insights into best linear unbiased estimation and the optimality of least-squares

Reduced rank regression in cointegrated models

Projektpartner. Sonderforschungsbereich 386, Paper 163 (1999) Online unter:

Chapter 3 Least Squares Solution of y = A x 3.1 Introduction We turn to a problem that is dual to the overconstrained estimation problems considered s

Linear Regression and Its Applications

The exact bias of S 2 in linear panel regressions with spatial autocorrelation SFB 823. Discussion Paper. Christoph Hanck, Walter Krämer

PROBLEMS AND SOLUTIONS

Econometria. Estimation and hypotheses testing in the uni-equational linear regression model: cross-section data. Luca Fanelli. University of Bologna

Regression and Statistical Inference

OPTIMAL DESIGNS FOR COMPLETE INTERACTION. By Rainer Schwabe. Freie Universitat Berlin

(1.) For any subset P S we denote by L(P ) the abelian group of integral relations between elements of P, i.e. L(P ) := ker Z P! span Z P S S : For ea

QUASI-UNIFORMLY POSITIVE OPERATORS IN KREIN SPACE. Denitizable operators in Krein spaces have spectral properties similar to those

Statistics 910, #5 1. Regression Methods

Lecture 6: Geometry of OLS Estimation of Linear Regession

On V-orthogonal projectors associated with a semi-norm

On equality and proportionality of ordinary least squares, weighted least squares and best linear unbiased estimators in the general linear model

YORK UNIVERSITY. Faculty of Science Department of Mathematics and Statistics MATH M Test #1. July 11, 2013 Solutions

VICTORIA POWERS AND THORSTEN W ORMANN Suppose there exists a real, symmetric, psd matrix B such that f = x B x T and rank t. Since B is real symmetric

Evaluation of Measurement Comparisons Using Generalised Least Squares: the Role of Participants Estimates of Systematic Error

Economics 472. Lecture 10. where we will refer to y t as a m-vector of endogenous variables, x t as a q-vector of exogenous variables,

Group Theory. 1. Show that Φ maps a conjugacy class of G into a conjugacy class of G.

Linear Algebra: Characteristic Value Problem

290 J.M. Carnicer, J.M. Pe~na basis (u 1 ; : : : ; u n ) consisting of minimally supported elements, yet also has a basis (v 1 ; : : : ; v n ) which f

ISOLATED SEMIDEFINITE SOLUTIONS OF THE CONTINUOUS-TIME ALGEBRAIC RICCATI EQUATION

Estimation under a general partitioned linear model

Lecture 1: Review of linear algebra

16 Chapter 3. Separation Properties, Principal Pivot Transforms, Classes... for all j 2 J is said to be a subcomplementary vector of variables for (3.

Solutions of a constrained Hermitian matrix-valued function optimization problem with applications

More on generalized inverses of partitioned matrices with Banachiewicz-Schur forms

Time Invariant Variables and Panel Data Models : A Generalised Frisch- Waugh Theorem and its Implications

Vector spaces. DS-GA 1013 / MATH-GA 2824 Optimization-based Data Analysis.

Properties of the least squares estimates

Example Bases and Basic Feasible Solutions 63 Let q = >: ; > and M = >: ;2 > and consider the LCP (q M). The class of ; ;2 complementary cones

1. Introduction Let the least value of an objective function F (x), x2r n, be required, where F (x) can be calculated for any vector of variables x2r

Rank equalities for idempotent and involutory matrices

Chapter 5 Prediction of Random Variables

Calculus and linear algebra for biomedical engineering Week 3: Matrices, linear systems of equations, and the Gauss algorithm

Wavelets and Linear Algebra

N.G.Bean, D.A.Green and P.G.Taylor. University of Adelaide. Adelaide. Abstract. process of an MMPP/M/1 queue is not a MAP unless the queue is a

Large Sample Properties of Estimators in the Classical Linear Regression Model

Peter Hoff Linear and multilinear models April 3, GLS for multivariate regression 5. 3 Covariance estimation for the GLM 8

Positive definiteness of tridiagonal matrices via the numerical range

Experimental evidence showing that stochastic subspace identication methods may fail 1

Chapter 3 Transformations

New concepts: Span of a vector set, matrix column space (range) Linearly dependent set of vectors Matrix null space

Tactical Decompositions of Steiner Systems and Orbits of Projective Groups

ECONOMETRICS. Bruce E. Hansen. c2000, 2001, 2002, 2003, University of Wisconsin

460 HOLGER DETTE AND WILLIAM J STUDDEN order to examine how a given design behaves in the model g` with respect to the D-optimality criterion one uses

Radoslav Harman. Department of Applied Mathematics and Statistics

Chapter 2 Linear Transformations

4.1 Eigenvalues, Eigenvectors, and The Characteristic Polynomial

Contents. 2.1 Vectors in R n. Linear Algebra (part 2) : Vector Spaces (by Evan Dummit, 2017, v. 2.50) 2 Vector Spaces

Problem Set 9 Due: In class Tuesday, Nov. 27 Late papers will be accepted until 12:00 on Thursday (at the beginning of class).

Linear Algebra (part 1) : Vector Spaces (by Evan Dummit, 2017, v. 1.07) 1.1 The Formal Denition of a Vector Space

304 A^VÇÚO 1n ò while the commonly employed loss function for the precision of estimation is squared error loss function ( β β) ( β β) (1.3) or weight

A Simple Proof of Fiedler's Conjecture Concerning Orthogonal Matrices

642:550, Summer 2004, Supplement 6 The Perron-Frobenius Theorem. Summer 2004

Quantitative Analysis of Financial Markets. Summary of Part II. Key Concepts & Formulas. Christopher Ting. November 11, 2017

The Algebraic Multigrid Projection for Eigenvalue Problems; Backrotations and Multigrid Fixed Points. Sorin Costiner and Shlomo Ta'asan

Sociedad de Estadística e Investigación Operativa

18.06 Professor Edelman Quiz 3 December 5, 2011

1. Let m 1 and n 1 be two natural numbers such that m > n. Which of the following is/are true?

Two Results About The Matrix Exponential

1 Vectors. Notes for Bindel, Spring 2017 Numerical Analysis (CS 4220)

Stochastic Design Criteria in Linear Models

Math 407: Linear Optimization

Math 1180, Notes, 14 1 C. v 1 v n v 2. C A ; w n. A and w = v i w i : v w = i=1

ON STATISTICAL INFERENCE UNDER ASYMMETRIC LOSS. Abstract. We introduce a wide class of asymmetric loss functions and show how to obtain

A Finite Element Method for an Ill-Posed Problem. Martin-Luther-Universitat, Fachbereich Mathematik/Informatik,Postfach 8, D Halle, Abstract

Scaled and adjusted restricted tests in. multi-sample analysis of moment structures. Albert Satorra. Universitat Pompeu Fabra.

1. The Polar Decomposition

Generalized Shifted Inverse Iterations on Grassmann Manifolds 1

Linear Algebra. Christos Michalopoulos. September 24, NTU, Department of Economics

[3] (b) Find a reduced row-echelon matrix row-equivalent to ,1 2 2

Research Article Stabilization Analysis and Synthesis of Discrete-Time Descriptor Markov Jump Systems with Partially Unknown Transition Probabilities

Lecture Notes 1: Vector spaces

1 Outline Part I: Linear Programming (LP) Interior-Point Approach 1. Simplex Approach Comparison Part II: Semidenite Programming (SDP) Concludin

Augustin: Some Basic Results on the Extension of Quasi-Likelihood Based Measurement Error Correction to Multivariate and Flexible Structural Models

Marcia Gumpertz and Sastry G. Pantula Department of Statistics North Carolina State University Raleigh, NC

Ridge analysis of mixture response surfaces

The equalities of ordinary least-squares estimators and best linear unbiased estimators for the restricted linear model

University of Missouri. In Partial Fulllment LINDSEY M. WOODLAND MAY 2015

Properties of Matrices and Operations on Matrices

identity matrix, shortened I the jth column of I; the jth standard basis vector matrix A with its elements a ij

PROOF OF TWO MATRIX THEOREMS VIA TRIANGULAR FACTORIZATIONS ROY MATHIAS

NORMS ON SPACE OF MATRICES

On Algebraic and Semialgebraic Groups and Semigroups

On splitting of the normalizer of a maximal torus in groups of Lie type

On the Schur Complement of Diagonally Dominant Matrices

New concepts: rank-nullity theorem Inverse matrix Gauss-Jordan algorithm to nd inverse

Re-sampling and exchangeable arrays University Ave. November Revised January Summary

Linear Algebra: Linear Systems and Matrices - Quadratic Forms and Deniteness - Eigenvalues and Markov Chains

identify only a few factors (possibly none) with dispersion eects and a larger number of factors with location eects on the transformed response. Tagu

GMM estimation of spatial panels

Coins with arbitrary weights. Abstract. Given a set of m coins out of a collection of coins of k unknown distinct weights, we wish to

Conditioning of the Entries in the Stationary Vector of a Google-Type Matrix. Steve Kirkland University of Regina

Some inequalities for sum and product of positive semide nite matrices

Chapter 6. Panel Data. Joan Llull. Quantitative Statistical Methods II Barcelona GSE

Transcription:

The Equality of OLS and GLS Estimators in the Linear Regression Model When the Disturbances are Spatially Correlated Butte Gotu 1 Department of Statistics, University of Dortmund Vogelpothsweg 87, 44221 Dortmund, Germany Abstract: Necessary and sucient conditions for the equality of ordinary least squares and generalized least squares estimators in the linear regression model with rst-order spatial error processes are given. Key words: Ordinary least squares, Generalized least squares, Best linear unbiased estimator, Spatial error process, Spatial correlation. 1 Introduction Consider the linear regression model for spatial correlation y = X + u ; u = C ; (1) where y is a T 1 observable random vector, X is a T k matrix of known constants with full column rank k, is a k 1 vector of unknown parameters, is a T 1 random vector with expectation zero and covariance matrix Cov() = 2 I (I is the T -dimensional identity matrix and 2 an unknown positive scalar). C denotes a T T matrix such that the product CC 0 is positive denite. The ordinary least squares (OLS) and the generalized least squares (GLS) 1 This work was partly supported by the Deutsche Forschungsgemeinschaft (DFG), Graduiertenkolleg \Angewandte Statistik". 1

estimators of the vector of unknown parameters in model (1) are given by ^ = (X 0 X)?1 X 0 y and ~ = (X 0 V?1 X)?1 X 0 V?1 y, respectively with covariance matrices Cov( ^) = 2 (X 0 X)?1 X 0 V X(X 0 X)?1, Cov( ) ~ = 2 (X 0 V?1 X)?1, where V = CC 0. When the covariance of the disturbance vector u is not a scalar multiple of the identity matrix, that is Cov(u) 6= 2 I as in model (1), it is well known that the GLS estimator provides the best linear unbiased estimator (BLUE) of in contrast to OLS. Since Cov(u) usually involves unknown parameters like spatial correlation coecient, it is natural to ask when both estimators coincide so that the OLS estimator can be applied without loss of eciency. Many of the criteria developed for the purpose of checking the equality of least squares estimators are not operational because of the unknown parameters involved (see Puntanen and Styan, 1989). In this paper, conditions under rst-order spatial error processes which can be veried in practice by using spatial weights matrix with known nonnegative weights and the matrix X of known constants are developed. The rst group of conditions is based on the invariance property of the column space of the matrix X under V (Kruskal, 1968), whereas the second one uses the symmetry of the product P X V (Zyskind, 1967), with P X = X(X 0 X)?1 X 0. 2 Equality of OLS and GLS estimators In assessing the conditions for the equality of OLS and GLS estimators, the structure of the covariance of the disturbance vector u plays an important role. So, we start by giving possible structures of Cov(u) under rst-order spatial error processes. Let the components of u follow a rst-order spatial autoregressive (AR(1)) process TX u i = w ij u j + i j=1 2

or, in matrix form u = W u + ; (2) where denotes a spatial correlation coecient for a given area partitioned into T nonoverlapping regions R i, i = 1; ; T. W is a weights matrix with known nonnegative weights dened by (see Cli and Ord, 1981, pp. 17-19) w ij 8 > < >: > 0 ; if R i and R j are neighbours (i 6= j) = 0 ; otherwise : The element w ij of the weights matrix indicates the strength of the eect of region R j on region R i. Under rst-order spatial moving average (MA(1)) process the components of u follow the pattern TX u i = w ij j + i j=1 or, in matrix form u = W + : (3) Equations (2) and (3) can be written as u = (I? W )?1 and u = (I + W ) (4) respectively, where in AR(1) case the matrix I? W must be nonsingular. From (1) and (4), we get four possible structures of Cov(u) = 2V for rstorder spatial error process: V = 8 >< >: (I + W )(I + W 0 ) : MA(1) (I + W ) : MA(1)? conditional (I? W )?1 (I? W 0 )?1 (5) : AR(1) (I? W )?1 : AR(1)? conditional : Note that the possible values of must be identied to ensure that V positive denite. is In the following we investigate conditions for the equality of OLS and GLS estimators by applying the result: two unbiased estimators coincide almost 3

surely if and only if their covariances are equal (see Puntanen and Styan, 1989, p. 154). This means, OLS and GLS are equal if and only if their covariances are equal. Let R(X) denote a k-dimensional space spanned by the columns of X. The well known Kruskal's (1968) column space condition for the equality of OLS estimator ^ and GLS estimator ~ in model (1) states that both estimators coincide if and only if R(V X) = R(X) ; (6) where V is assumed to be a nonsingular matrix. In order to apply Kruskal's condition, the value of the unknown parameter in the Matrix V must be given in addition to X. In practice typically will be unknown and one needs a more applicable condition to check the equality. Based on Kruskal's theorem Kramer and Donninger (1987) give a sucient condition which can be veried in practice when the disturbances follow a rst-order spatial autoregressive process. Baksalary (1988) generalizes this result for rst-order spatial error processes as follows. Theorem 1 Let W be a T T weights matrix and V be a T T positive denite matrix of the form V = (I + W 0 )(I + W ) or V = (I + W )(I + W 0 ) ; where 6= 0 is a scalar. If R(W X) R(X) and R(W 0 X) R(X), then ^ = ~. Proof: The conditions R(W X) R(X) and R(W 0 X) R(X) imply that R((I + W )X) = R(X) and R((I + W 0 )X) = R(X) irrespective of. From this we get R(V X) = R((I + W 0 )(I + W )X) = R(X) 4

and the equality of the estimators follows from Kruskal's theorem. 3 The following sucient condition for the equality under a specic matrix V is also based on condition (6). Theorem 2 Let b 1 and b 2 be T 1 vectors, and let V be a T T positive denite matrix of the pattern V = ci + b 1 b 0 2 + b 2 b 0 1 with a scalar c. If b 1 2 R(X) and b 2 2 R(X), then R(V X) = R(X): Proof: See Mathew, 1984, pp. 207-208. 3 By combining the results in theorems 1 and 2 the following sucient condition for the equality of OLS and GLS estimators can be formulated. Corollary 1 Let d be a T 1 vector and V be a T T positive denite matrix of the pattern V = c1i + c2 W + c3 d d 0 ; where c1; c2; c3 are scalars, and W is a T T matrix. If R(W X) R(X) and d 2 R(X), then ^ =. ~ Proof: The proof follows from Theorems 1 and 2. 3 Simple examples show that the conditions of the above results are not necessary for the equality of OLS and GLS estimators (see Baksalary, 1988 and Gotu, 1997). The theorem below, based on the result given by Baksalary (1988), provides necessary and sucient conditions. Theorem 3 Let W be a T T weights matrix and V be a T T matrix of the form V = (I + W )(I + W 0 ) : 5

Further, let be given by: = f 6= 0 : V positive denite and jj < 1g: Then the following conditions are equivalent: (i) R(V X) = R(X) for all 2. (ii) R(V X) = R(X) for two dierent 1; 2 2. (iii) R((W + W 0 )X) R(X) and R(W 0 W X) R(X). Proof: (i) =) (ii): The condition R(V X) = R(X) for 6= 0 holds if and only if If (7) is valid for all 2, then (ii) =) (iii): R((W + W 0 + W W 0 )X) R(X): (7) R((W + W 0 + 1W W 0 )X) R(X) R((W + W 0 + 2W W 0 )X) R(X): (8) From equation (8) we get R((1? 2)W W 0 X) R(X). This implies R(W W 0 X) R(X), and R((W 0 + W )X) R(X) follows from (7). (iii) =) (i): Follows direct from (7). 3 Remarks: The matrix V is positive denite if I + W is nonsingular and the nonsingularity of I + W holds if there exists a matrix-norm which satises the inequality jj jjw jj < 1 (see Horn and Johnson, 1985, p. 301). For any given weights matrix W with row sums equal to one, the maximum row sum matrix-norm is equal to one, so the matrix I + W is nonsingular for jj < 1. 6

Let A be a symmetric matrix. Then R(AX) R(X) if and only if P X A = AP X. This means condition (iii) is equivalent to (W +W 0 )P X = P X (W + W 0 ) and W W 0 P X = P X W W 0. Theorem 3 applies also for V matrix of the form V = ((I? W 0 )(I? W ))?1 ; because R(V X) = R(X) () R(V?1 X) = R(X) : If OLS and GLS estimators are equal for two dierent values of, that is R(V X) = R(X) for dierent 1; 2 2, then from the equivalence of (i) and (ii) follows that both estimators are equal for all 2. Condition (iii) can be applied to check the equality of OLS and GLS without specifying the value of. For V matrix of the form (I?W )?1 or I+W, where W is symmetric, condition (iii) should be restated as R(W X) R(X). Let W1 and W2 be T T weights matrices, and D1 and D2 be T T diagonal matrices with full rank. Suppose that W1 0 D?1 1 = D 1?1 W1 and D2W 0 2 = W2D2. If V is of the pattern (I? W1)?1 D1 or (I + W2)D2, condition (iii) should, accordingly, be restated as R(D?1 1 X) R(X) and R(D?1 1 W1X) R(X); R(D2X) R(X) and R(D2W2X) R(X). In the following, conditions for the equality of least squares estimators for a subvector of will be discussed. Suppose that X1 and X2 are submatrices of X, and 1 and 2 be subvectors of. Further, let ^2 and 2 ~ be the respective subvectors of ^ and. ~ Splitting model (1) into y = X11 + X22 + u ; 7

Kramer et al. (1996) give the following necessary and sucient condition for the equality of ^2 and ~ 2: ^2 = ~ 2 () R(V X? ) (R(X1) R(X? )) ; where X? is a matrix such that R(X? ) = R(X)?, the orthogonal complement of R(X), and is the direct sum of subspaces. The problem with the above condition is, as in Kruskal's theorem, that the unknown parameter in the matrix V should be given. The following result, which is based on Theorem 3, provides a necessary and sucient condition for the equality of ^2 and 2 ~ under the rst-order spatial error process that works without specifying the value of. Corollary 2 Let W be T T weights matrix and V be a T T matrix of the form V = (I + W )(I + W 0 ) ; (9) where 2. Then the following statements are equivalent: (a) R(V X? ) R(X1) R(X? ) for all 2. (b) (c) R(V X? ) R(X1) R(X? ) for two dierent 1; 2 2. R((W + W 0 )X? ) R(X1) R(X? ) and R(W W 0 X? ) R(X1) R(X? ). Proof: See Theorem 3. 3 Remarks: In order to check the equality of ^2 and ~ 2, statement (iii) can be applied independent of. 8

For the matrix of the form (9) the following holds (see Theorem 3): If R(W X? ) R(X1) R(X? ) and R(W 0 X? ) R(X1) R(X? ) ; then ^2 = ~ 2. Another well known condition for the coincidence of OLS and GLS estimators in the linear regression model (1) is based on the symmetry of the matrix product P X V. That is, in the regression model (1) ^ = ~ () P X V = V P X : (10) For the application of this condition the values of the unknown parameters in the matrix V should again be given. The following sucient condition can be applied under the rst-order spatial error processes, irrespective of the parameters in V. Corollary 3 Assume that the components of the disturbance vector u in model (1) follow a rst-order spatial moving average or autoregressive process. Let W be a T T weights matrix. The estimators ^ and ~ coincide if P X W = W P X : (11) Proof: MA(1) process: Under spatial MA(1) error process the matrix V is given by V = (I + W )(I + W 0 ) : From equation (10) the estimators ^ and ~ coincide if and only if P X V = V P X : The above equation holds if for 6= 0 P X W 0 + P X W + P X W W 0 = W 0 P X + W P X + W W 0 P X : (12) 9

By equation (11), applying the symmetry of P X, we get P X W 0 = W 0 P X and from (12) follows P X V = V P X implying the equality of the estimators ^ and ~. AR(1) process: Under spatial AR(1) error process we have V = ((I? W 0 )(I? W ))?1 and V?1 = (I? W 0 )(I? W ): Furthermore, P X V = V P X if and only if P X V?1 = V?1 P X : (13) Equation (13) holds if P X W 0 W? P X W 0? P X W = W 0 W P X? W 0 P X? W P X (14) with 6= 0. By equation (14), applying the symmetry of P X and equation (11), we obtain P X V?1 = V?1 P X implying the equality of ^ and. ~ 3 Remarks: It can be shown that the condition of Corollary 3 is also necessary if (see Gotu, 1997) { the weights matrix W is symmetric and orthogonal. { the components of the disturbance vector u follow a conditional rst-order spatial process with V given in (5). A counter-example that the condition of Corollary 3 is necessary in general can be obtained by taking W = 0 B @ 0 2=3 1=3 1=3 0 2=3 2=3 1=3 0 1 C A X = 0 B @ 1 0 1 1 1?1 1 C A V = (I + W 0 )(I + W ) and = 3=4. In this case P X V = V P X although P X W 6= W P X. 10

Acknowledgements: The author is grateful to Prof. Dr. S. Schach, Prof. Dr. G. Trenkler and Dr. J. Gro for their useful comments on earlier drafts of the paper. References Baksalary, J. K. (1988): \Two Notes of JKB", unpublished. Cli, A. D. and Ord, J. K. (1981): Spatial processes: Models and applications. Pion, London. Gotu, B. (1997): \Schatz- und Testmethoden im linearen Regressionsmodell bei Vorliegen von raumlicher Korrelation", Doctoral Thesis, University of Dortmund. Horn, R. A. and Johnson, C. R. (1985): Matrix Analysis. Cambridge University Press, Cambridge. Kramer, W., Bartels, R. and Fiebig, D. G. (1996): \ Another Twist on the Equality of OLS and GLS", Statistical Papers 37, 277-281. Kramer, W. and Donninger, C. (1987): \ Spatial Autocorrelation among Errors and the Relative Eciency of OLS in the Linear Regression Model", Journal of the American Statistical Association 82, 577-579. Kruskal, W. (1968): \When are Gauss-Markov and Least Squares Estimator Identical? A Coordinate-Free Approach", The Annals of Mathematical Statistics 39, 70-75. Mathew, T. (1984): \On Inference in a General Linear Model with an Incorrect Disperssion Matrix ", in Calinski, T. and Klonecki, W. (eds.): Linear Statistical Inference, Proceedings of Poznan Conference. Springer Verlag, Berlin, 200-210. 11

Puntanen, S. and Styan, G. P. H. (1989): \The Equality of the Ordinary Least Squares Estimator and the Best Linear Unbiased Estimator", The American Statistician 43, 153-164. Zyskind, G. (1967): \ On Canonical Forms, Non-Negative Covariance Matrices and Best and Simple Least Squares Linear Estimators in Linear Models", The Annals of Mathematical Statistics 38, 1092-1109. 12