New insights into best linear unbiased estimation and the optimality of least-squares

Similar documents
On V-orthogonal projectors associated with a semi-norm

A note on the equality of the BLUPs for new observations under two linear models

Journal of Multivariate Analysis. Sphericity test in a GMANOVA MANOVA model with normal error

The generalized Schur complement in group inverses and (k + 1)-potent matrices

The equalities of ordinary least-squares estimators and best linear unbiased estimators for the restricted linear model

Matrix Inequalities by Means of Block Matrices 1

On Sums of Conjugate Secondary Range k-hermitian Matrices

A new algebraic analysis to linear mixed models

On equality and proportionality of ordinary least squares, weighted least squares and best linear unbiased estimators in the general linear model

A revisit to a reverse-order law for generalized inverses of a matrix product and its variations

Solutions of a constrained Hermitian matrix-valued function optimization problem with applications

Analytical formulas for calculating the extremal ranks and inertias of A + BXB when X is a fixed-rank Hermitian matrix

Analytical formulas for calculating extremal ranks and inertias of quadratic matrix-valued functions and their applications

The symmetric minimal rank solution of the matrix equation AX=B and the optimal approximation

SPECIAL FORMS OF GENERALIZED INVERSES OF ROW BLOCK MATRICES YONGGE TIAN

Yongge Tian. China Economics and Management Academy, Central University of Finance and Economics, Beijing , China

Math Camp Lecture 4: Linear Algebra. Xiao Yu Wang. Aug 2010 MIT. Xiao Yu Wang (MIT) Math Camp /10 1 / 88

The Equality of OLS and GLS Estimators in the

Large Sample Properties of Estimators in the Classical Linear Regression Model

More on generalized inverses of partitioned matrices with Banachiewicz-Schur forms

On a matrix result in comparison of linear experiments

Journal of Inequalities in Pure and Applied Mathematics

AN ELEMENTARY PROOF OF THE SPECTRAL RADIUS FORMULA FOR MATRICES

Absolute value equations

A CLASS OF THE BEST LINEAR UNBIASED ESTIMATORS IN THE GENERAL LINEAR MODEL

ELA THE OPTIMAL PERTURBATION BOUNDS FOR THE WEIGHTED MOORE-PENROSE INVERSE. 1. Introduction. Let C m n be the set of complex m n matrices and C m n

References. Press, New York

Moore Penrose inverses and commuting elements of C -algebras

Optimization problems on the rank and inertia of the Hermitian matrix expression A BX (BX) with applications

On some matrices related to a tree with attached graphs

On some linear combinations of hypergeneralized projectors

An Alternative Proof of the Greville Formula

Inverse of a Square Matrix. For an N N square matrix A, the inverse of A, 1

~ g-inverses are indeed an integral part of linear algebra and should be treated as such even at an elementary level.

Ma 3/103: Lecture 24 Linear Regression I: Estimation

Lecture notes: Applied linear algebra Part 1. Version 2

Two Results About The Matrix Exponential

Re-nnd solutions of the matrix equation AXB = C

Recursive Determination of the Generalized Moore Penrose M-Inverse of a Matrix

Generalized Principal Pivot Transform

Hands-on Matrix Algebra Using R

Rank and inertia optimizations of two Hermitian quadratic matrix functions subject to restrictions with applications

arxiv: v1 [math.ra] 14 Apr 2018

Eigenvalues and Eigenvectors

Vector spaces. DS-GA 1013 / MATH-GA 2824 Optimization-based Data Analysis.

Matrix Differential Calculus with Applications in Statistics and Econometrics

On Hadamard and Kronecker Products Over Matrix of Matrices

Testing Some Covariance Structures under a Growth Curve Model in High Dimension

22.3. Repeated Eigenvalues and Symmetric Matrices. Introduction. Prerequisites. Learning Outcomes

ELA

MATH 423 Linear Algebra II Lecture 33: Diagonalization of normal operators.

Pattern correlation matrices and their properties

PROBLEMS AND SOLUTIONS

Chapter 3. Matrices. 3.1 Matrices

Some inequalities for sum and product of positive semide nite matrices

Rank equalities for idempotent and involutory matrices

Instrumental Variables

Elementary linear algebra

Nonsingularity and group invertibility of linear combinations of two k-potent matrices

Linear Algebra and its Applications

Massachusetts Institute of Technology Department of Economics Statistics. Lecture Notes on Matrix Algebra

Linear estimation in models based on a graph

Rank and inertia of submatrices of the Moore- Penrose inverse of a Hermitian matrix

Ridge Regression. Flachs, Munkholt og Skotte. May 4, 2009

ON VARIANCE COVARIANCE COMPONENTS ESTIMATION IN LINEAR MODELS WITH AR(1) DISTURBANCES. 1. Introduction

Properties of Matrices and Operations on Matrices

Lecture 6: Geometry of OLS Estimation of Linear Regession

304 A^VÇÚO 1n ò while the commonly employed loss function for the precision of estimation is squared error loss function ( β β) ( β β) (1.3) or weight

Notes on Linear Algebra and Matrix Theory

Lecture: Linear algebra. 4. Solutions of linear equation systems The fundamental theorem of linear algebra

Repeated Eigenvalues and Symmetric Matrices

SOLUTIONS TO PROBLEMS

THE PERTURBATION BOUND FOR THE SPECTRAL RADIUS OF A NON-NEGATIVE TENSOR

Recall the convention that, for us, all vectors are column vectors.

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.

Intrinsic products and factorizations of matrices

1 Last time: least-squares problems

The Algebra of the Kronecker Product. Consider the matrix equation Y = AXB where

Lecture Notes 1: Vector spaces

MLES & Multivariate Normal Theory

A note on the product of two skew-hamiltonian matrices

Estimation under a general partitioned linear model

ON EXACT INFERENCE IN LINEAR MODELS WITH TWO VARIANCE-COVARIANCE COMPONENTS

Abstract. In this article, several matrix norm inequalities are proved by making use of the Hiroshima 2003 result on majorization relations.

Chap 3. Linear Algebra

A property of orthogonal projectors

Introduction to Quantitative Techniques for MSc Programmes SCHOOL OF ECONOMICS, MATHEMATICS AND STATISTICS MALET STREET LONDON WC1E 7HX

Some Reviews on Ranks of Upper Triangular Block Matrices over a Skew Field

Inequalities for Modules and Unitary Invariant Norms

Math 443 Differential Geometry Spring Handout 3: Bilinear and Quadratic Forms This handout should be read just before Chapter 4 of the textbook.

(AA-A = A) and A + be the Moore-Penrose inverse of A [4]. A matrix A is called EP r

APPENDIX A. Background Mathematics. A.1 Linear Algebra. Vector algebra. Let x denote the n-dimensional column vector with components x 1 x 2.

MATH 1120 (LINEAR ALGEBRA 1), FINAL EXAM FALL 2011 SOLUTIONS TO PRACTICE VERSION

Zellner s Seemingly Unrelated Regressions Model. James L. Powell Department of Economics University of California, Berkeley

ON WEIGHTED PARTIAL ORDERINGS ON THE SET OF RECTANGULAR COMPLEX MATRICES

2. Linear algebra. matrices and vectors. linear equations. range and nullspace of matrices. function of vectors, gradient and Hessian

Linear algebra. S. Richard

NOTES on LINEAR ALGEBRA 1

Math 350 Fall 2011 Notes about inner product spaces. In this notes we state and prove some important properties of inner product spaces.

ELA THE MINIMUM-NORM LEAST-SQUARES SOLUTION OF A LINEAR SYSTEM AND SYMMETRIC RANK-ONE UPDATES

Transcription:

Journal of Multivariate Analysis 97 (2006) 575 585 www.elsevier.com/locate/jmva New insights into best linear unbiased estimation and the optimality of least-squares Mario Faliva, Maria Grazia Zoia Istituto di Econometria e Matematica, Università Cattolica, Milano, Italy Received 23 July 2002 Available online 30 August 2005 Abstract New results in matrix algebra applied to the fundamental bordered matrix of linear estimation theory pave the way towards obtaining additional and informative closed-form expressions for the best linear unbiased estimator (BLUE). The results prove significant in several respects. Indeed, more light is shed on the BLUE structure and on the working of the OLS estimator under nonsphericalness in (possibly) singular models. 2005 Elsevier Inc. All rights reserved. AMS 1991 subject classification: Primary 62J12; 62F10; secondary 15A09 Keywords: BLUE closed-forms; Nonsphericalness; OLS optimality 1. Introduction A recent result about partitioned inversion [1] together with newly found related results paves the way towards gaining new insights into linear estimation. Indeed, general closedform expressions for the best linear unbiased estimator (BLUE) and its dispersion matrix are obtained, which clear the ground for a thorough analysis of the BLU property for the ordinary least-squares (OLS) under nonsphericalness conditions. The paper is organized as follows: in Section 2 a (possibly) singular linear model is properly specified and the main results both classical and newly found in estimation are Corresponding author. Fax: +39 2 72342324. E-mail address: maria.zoia@unicatt.it (M.G. Zoia). 0047-259X/$ - see front matter 2005 Elsevier Inc. All rights reserved. doi:10.1016/j.jmva.2005.07.001

576 M. Faliva, M.G. Zoia / Journal of Multivariate Analysis 97 (2006) 575 585 reported. Section 3 provides the algebraic tool kit, hinging on partitioned inversion and side results tailored to estimation needs. Section 4 derives new closed-form expressions for the BLUE and its dispersion matrix, with a glance at the parent least-squares problems. Last but not the least interesting, Section 5 sheds new light on the issue of the optimality vs. the efficiency gap of OLS under nonsphericalness in a (possibly) singular model, in the wake of the results established in Section 4. 2. The reference model and the basic results on estimation Consider the general linear model y = Xβ + ε, E{ε} =0, (2.1) (2.2) E{εε }=σ 2 V, tr(v) = r(v), (2.3) r(x) = k<n, r([v, X]) = n, (2.4) (2.5) where y is the n-vector of values of the regressand, X is the n k matrix of values of the k linearly independent regressors, ε is the n-vector of the disturbances, β is a k-vector of parameters to be estimated, σ 2 is an unknown scalar factor, and V is a given nonnegative definite matrix normalized so that the trace and the rank are the same. Under the rank assumptions (2.4) and (2.5) the bordered matrix [ V X X 0 ] is nonsingular and the framework of the unified theory of linear estimation [7,8] holds true. Classical and newly found results regarding the BLU estimator b of β and its (estimated) dispersion matrix D(b) can be summarized as follows: where b OLS denotes the OLS estimator of β, b = C 2 y, (2.6) = b OLS X + VC 1 y, (2.7) b = b OLS if X VX = 0, (2.8) D(b) = s 2 C 3, (2.9) = D(b OLS ) s 2 X + VC 1 V(X + ), (2.10) s 2 = (n k) 1 y C 1 y (2.11) is a quadratic unbiased estimator of σ 2, [ C1 C ] 2 = C 2 C 3 [ ] 1 V X X, (2.12) 0

M. Faliva, M.G. Zoia / Journal of Multivariate Analysis 97 (2006) 575 585 577 X + denotes the Moore Penrose generalized inverse of X and X is a full column-rank n (n k) matrix henceforth referred to as an orthogonal complement of X such that r([x, X]) = n, X X = 0, (2.13) EG = tr{d((x X) 1/2 b OLS ) D((X X) 1/2 b)}, p ( s 2 λi ) 2 λ n i+1, p = min(k, n k), (2.14) i=1 where EG denotes the efficiency gap of b OLS under nonsphericalness and λ 1 λ 2 λ n > 0 are the eigenvalues V + I VV +. While (2.6), (2.8), (2.9) and (2.11) reflect known results of the theory of linear estimation, the other results represent the outcomes of this paper, which provides new and informative representations for the BLUE and an upper bound for the OLS efficiency gap due to nonsphericalness in (possibly) singular models. 3. A useful theorem on partitioned inversion and side results In this section, we shall obtain several expressions for the inverse of a bordered matrix which are of prominent interest in best linear unbiased estimation. Before proving the main theorem we will establish two lemmas. Lemma 3.1. Let A be a nonnegative definite matrix of order n and B be a full column-rank n k matrix, such that r([a, B]) = n. (3.1) Then the following results hold for the product AB : r(ab ) = n k, (3.2) (AB ) = (A + Λ) 1 B, (3.3) where Λ denotes a nonnegative definite matrix of order n satisfying r([a, Λ]) = n, ΛB = 0. (3.4) (3.5) If A is positive definite, then a trivial choice for Λ is the null matrix. Upon noting that M(B ) M(A), (3.6) where M(...) denotes the column space of the argument matrix, a natural choice for Λ turns out to be Λ = I AA +. (3.7)

578 M. Faliva, M.G. Zoia / Journal of Multivariate Analysis 97 (2006) 575 585 Proof. The proof of (3.2) follows from (3.1) by considering that ([ ] ) A r(ab ) = r B B = r(b ) (3.8) and using the elementary properties of the rank. To prove (3.3) observe that, under (3.4) and (3.5), the orthogonality condition (AB ) (A + Λ) 1 B = B (A + Λ)(A + Λ) 1 B = 0 (3.9) holds along with the side rank condition r([ab,(a + Λ) 1 B]) = r(b ) + r(b) = n. (3.10) Relationship (3.6) is implied by the rank assumption (3.1) along with the rank conditions (see formula (2.13)) inherent in the definition of the orthogonal complement. Then, observe that (see e.g. [11, Lemma 2.2.4]) M(B ) M(A) AA + B = B, (3.11) whence it is easy to verify that both conditions (3.4) and (3.5) hold true for Λ = I AA +. Lemma 3.2. Let A and B be defined as in Lemma 3.1 and let C be a full column-rank matrix of order n (n k), such that r([b, C]) = n. (3.12) Then the following identities hold: B(C B) 1 C + C(B C) 1 B = I, (3.13) B[(AB ) B] 1 (AB ) + AB (B AB ) 1 B = I, (3.14) B[B (A + Λ) 1 B] 1 B (A + Λ) 1 + AB (B AB ) 1 B = I. (3.15) Proof. Observe that since the square matrix [B, C] is nonsingular, both C B and B C are also nonsingular. Further observe that [ (C B) 1 C ] [ ] Ik 0 (B C) 1 B [B, C] =. (3.16) 0 I n k [ (C ] This shows that B) 1 C (B C) 1 B is the inverse of [B, C]. Hence identity (3.13) follows from the commutative property of the inverse. Besides, identity (3.14) ensues from (3.13) by taking C = AB (3.17) bearing in mind (3.2). Identity (3.15) follows, in turn, from (3.14) by considering (AB ) = (A + Λ) 1 B, (3.18) where Λ is defined as in Lemma 3.1.

M. Faliva, M.G. Zoia / Journal of Multivariate Analysis 97 (2006) 575 585 579 We now prove the main theorem. Theorem 3.3. Let A and B be defined as in Lemma 3.1. Consider the bordered matrix [ ] A B P = B. (3.19) 0 [ Then the following representations hold for P 1 C1 C = ] 2. C 2 C 3 [ C1 C ] [ 2 D (I DA)(B = + ) ] C 2 C 3 B + (I AD) B + (ADA A)(B + ) (3.20) [ B (C = B ) 1 B C (B C ) 1 ] (C B) 1 C (C B) 1 C AC (B C ) 1 (3.21) [ (A + Λ) = 1 (A + Λ) 1 BF 1 B (A + Λ) 1 (A + Λ) 1 BF 1 ] F 1 B (A + Λ) 1 F 1 + B + Λ(B + ), (3.22) where D = B (B AB ) 1 B, (3.23) C = AB, Λ is defined as in Lemma 3.1 and the matrix F is given by (3.24) F = B (A + Λ) 1 B. (3.25) Proof. The Expression (3.20) is a special case of a result on partitioned inversion recently established by Faliva and Zoia [1]. Expressions (3.21), (3.22) are obtained from (3.20) with simple computations by making use of the identities established in Lemma 3.2. Corollary 3.3.1. If A is positive definite, then the following proves to be true: B + A(B + ) (B A 1 B) 1 = B + AB (B AB ) 1 B A(B+ ). (3.26) Moreover, under the homogeneous constraint B + AB = 0 the reverse order law [5] (3.27) (B A 1 B) 1 = B + A(B + ) (3.28) applies as well. Proof. The proof of (3.26) follows by comparing the lower diagonal blocks of the partitioned inverses (3.20) and (3.22), by noting first of all that C 3 = B + AC 1 A(B + ) B + A(B + ). (3.29)

580 M. Faliva, M.G. Zoia / Journal of Multivariate Analysis 97 (2006) 575 585 On the other hand, if A is nonsingular and Λ is the null matrix, the following holds: C 3 = (B A 1 B) 1. (3.30) Hence, equating the right-hand sides of (3.29) and (3.30) and rearranging the terms yield (3.26). The proof of (3.28) is straightforward once we observe that under (3.27) the right-hand side of (3.26) reduces to a null matrix. 4. Closed-forms of the BLUE and related results We will now establish a theorem which provides new and informative closed-form representations for the BLUE and its dispersion matrix. Theorem 4.1. The following closed-form representations hold for the BLUE of β: where b = b OLS X + VX (X VX ) 1 X y (4.1) =[(VX ) X] 1 (VX ) y (4.2) =[X (V + Λ) 1 X] 1 X (V + Λ) 1 y, (4.3) b OLS = X + y (4.4) and Λ is specified as in Lemma 3.1 with A and B replaced by V and X, respectively. Further, the following closed-form representations hold for the estimated dispersion matrix D(b) of b: where D(b) = D(b OLS ) s 2 X + VX (X VX ) 1 X V(X+ ) (4.5) = D(b OLS ) s 2 {X + (V + Λ)(X + ) [X (V + Λ) 1 X] 1 }, (4.6) s 2 = 1 n k y X (X VX ) 1 X y, (4.7) D(b OLS ) = s 2 X + V(X + ). (4.8) Proof. By (2.6), (2.9), (2.11) and (2.12) it follows that [ ] 1 [ ] V X y b = C 2 y =[0 I] X, (4.9) 0 0 D(b) = s 2 C 3 = s 2 [0 I] [ ] 1 [ ] V X 0 X, (4.10) 0 I s 2 = 1 n k y C 1 y = 1 [ ] 1 [ ] n k [y 0 V X y ] X. (4.11) 0 0

M. Faliva, M.G. Zoia / Journal of Multivariate Analysis 97 (2006) 575 585 581 Then, a straightforward application of the partitioned inversion formulae (3.20) (3.22) of Theorem 3.3, with A and B replaced by V and X, respectively, leads by simple computations to the representations stated in the theorem for the BLUE of β, D(b) and s 2. This theorem is significant in several respects. On the one hand, expressions (4.1) and (4.5) for the estimator b and its estimated dispersion matrix D(b), respectively, shed new light on the structure of the BLUE with respect to the OLS estimator and paves the way towards revisiting (in the next section) the issue of the optimality versus the efficiency gap of the OLS estimator under nonsphericalness of disturbances. On the other hand, result (4.2) reads as an instrumental variable estimator (see e.g. [2]) with (VX ) playing the role of an instrumental variable matrix. This provides a way to look at b as the solution of the least-squares problem min {(y Xβ) (VX ) [(VX ) (VX ) ] 1 (VX ) (y Xβ)}. (4.12) β Finally, expression (4.3) takes us back to Rao s unified theory of least-squares [9] as long as the estimator b turns out to be the solution of the least-squares problem min β {(y Xβ) (V + Λ) 1 (y Xβ)}. (4.13) The following corollaries help in gaining a deeper insight into the relationships among BLU, OLS and generalized least-squares (GLS) estimators. Corollary 4.1.1. If the homogeneous constraint X VX = 0 (4.14) holds true, we have the following: b = b OLS = X + y, (4.15) D(b) = D(b OLS ) = s 2 X + V(X + ). (4.16) Proof. Under (4.14), the second term on the right-hand sides of (4.1) and (4.5) reduces to null matrices; therefore, (4.15) and (4.16) hold trivially true. Corollary 4.1.2. If V is positive definite, then we have the following: b = b GLS = (X V 1 X) 1 X V 1 y, (4.17) D(b) = D(b GLS ) = s 2 (X V 1 X) 1. (4.18) Proof. Under the positive definiteness of V and the equality Λ = 0 as a side-effect, the right-hand sides of (4.3) and (4.6) reduce to (4.17) and (4.18) above.

582 M. Faliva, M.G. Zoia / Journal of Multivariate Analysis 97 (2006) 575 585 Corollary 4.1.3. When both the assumptions of Corollaries 4.1.1 and 4.1.2 hold, then we have the following: b = b GLS, (4.19) = b OLS = X + y, (4.20) D(b) = D(b GLS ) = s 2 (X V 1 X) 1, (4.21) = D(b OLS ) = s 2 X + V(X + ). (4.22) Proof. The proof is straightforward and is omitted. 5. Optimality versus loss of the efficiency for the ordinary least-squares estimator under nonsphericalness specifications In this section, we will first address the issue of optimality for the OLS estimator by specifying when nonsphericalness happens to be immaterial and the BLU property holds accordingly. Indeed, the question turns out to hinge on a homogeneous constraint involving both the dispersion and the regression matrices together with the orthogonal complement of the latter which is instrumental in providing a characterization of the error dispersion matrix. Besides, we will tackle the problem of evaluating by a suitably chosen measure the loss of efficiency for the OLS estimator with respect to the BLUE inasmuch as the aforementioned constraint is violated and the lack of sphericity penalizes OLS. We begin by establishing the following: Theorem 5.1. Let X VX = 0 (5.1) Then the following results hold: (i) b = b OLS = X + y, (5.2) (ii) D(b) = D(b OLS ) = s 2 X + V(X + ), (5.3) (iii) V = XX + ΦXX + + X X + ΦX X +, (5.4) where Φ is an arbitrary nonnegative definite matrix subject to r(φx ) = r(x ). (5.5) Proof. Although result (i) is already known (see the original contributions by Rao [6] and Zyskind [13], the proof becomes considerably simplified because of (4.1) and (4.4).

M. Faliva, M.G. Zoia / Journal of Multivariate Analysis 97 (2006) 575 585 583 Analogously, (ii) follows directly from (4.5). To establish result (iii) observe that (5.1) can be restated as (X + ) X VX X + = 0. (5.6) Hence, Theorem 2.5 of Khatri and Mitra [3] applies and eventually leads to a general nonnegative definite solution which can be written in the form of (5.4) by using the identity (see Lemma 3.2) (X + ) X + X X + = I. (5.7) As far as the rank qualification (5.5) is concerned, the reader should be made aware that the right-hand side of (5.4) should comply with the rank requirement (2.5). In fact, by Theorem 19 of Marsaglia and Styan [4] the following holds: r([v, X]) = r(x) + r((i XX + )V) = r(x) + r(x X + ΦX X + ) = r(x) + r(φx ) = n r(φx ) = r(x ) (5.8) and the other way round. Remark. In view of the foregoing, by replacing Φ with V in (5.4) we are led to conclude that, insofar as the representation V = XX + VXX + + X X + VX X + (5.9) holds, the OLS estimator enjoys the BLU property. Conversely, for an arbitrary dispersion matrix V, the Euclidean norm of the difference between the given matrix and the expression on the right-hand side of (5.9) can be taken as an indicator of the mismatching between the underlying OLS and BLUE estimators. Some computations yield V XX + VXX + X X + VX X + = XX + VX X + + X X + VXX+ ={tr[(xx + VX X + + X X + VXX+ ) (XX + VX X + + X X + VXX+ ) ]} 1/2 ={tr[xx + VX X + VXX+ + X X + VXX+ VX X + ]}1/2 ={tr[ X V X X V X + X V X X V X ]} 1/2 = 2 X V X, (5.10) where X = X(X X) 1/2, X = X (X X ) 1/2. (5.11) As a matter of fact, a number of authors (see [10,12] and the references quoted therein) have investigated the issue of the OLS performances under nonsphericalness; several measures of inefficiency have been suggested and bounds established accordingly. Here, we will handle the problem by giving an ad hoc definition of the efficiency gap for the OLS estimator and by establishing an upper bound of the same.

584 M. Faliva, M.G. Zoia / Journal of Multivariate Analysis 97 (2006) 575 585 The efficiency gap of the OLS estimator with respect to the BLUE is the trace of the nonnegative definite matrix obtained as the difference between the dispersion of (X X) 1/2 b OLS and of (X X) 1/2 b EG = tr{d((x X) 1/2 b OLS ) D((X X) 1/2 b)}. (5.12) In light of (4.6) the following representation of the argument matrix on the right-hand side of (5.12) proves to be true by carrying out simple computations. D((X X) 1/2 b OLS ) D((X X) 1/2 b) = s 2 { X (V + Λ) X [ X (V + Λ) 1 X] 1 }, (5.13) where Λ and X are defined as in Lemma.3.1 and in (5.11) above, respectively. Under these conditions, we will now prove a theorem which sheds more light on the OLS estimator performances under nonsphericalness in a (possibly) singular linear model. Theorem 5.2. The following inequality holds: EG s 2 p i=1 ( λi λ n i+1 ) 2, p = min(k, n k), (5.14) where λ 1 λ 2 λ n > 0 are the eigenvalues of the positive definite matrix Ṽ = V + I VV +. (5.15) Proof. To prove (5.14) reference must be made to (5.13) above, by considering Lemma 3.1 and by taking Λ = I VV +. (5.16) Hence, we obtain the expression for EG EG = s 2 tr[ X Ṽ X ( X Ṽ 1 X)] 1 ]. (5.17) Rao s theorem [10] applies to the right-hand side matrix of (5.17) and (5.14) holds accordingly. Insofar as no positive-definiteness assumption on V is required, inequality (5.14) provides an upper bound for the OLS efficiency gap in singular models as well. References [1] M. Faliva, M.G. Zoia, On a partitioned inversion formula having useful applications in econometrics, Econom. Theory 18 (2002) 525 530. [2] A.S. Goldberger, Econometric Theory, Wiley, New York, NY, 1964. [3] C.G. Khatri, S.K. Mitra, Hermitian and nonnegative definite solutions of linear matrix equations, SIAM J. Appl. Math. 31 (1976) 579 585. [4] G. Marsaglia, G.P.H. Styan, Equalities and inequalities for ranks of matrices, Linear Multilinear Algebra 2 (1974) 269 292.

M. Faliva, M.G. Zoia / Journal of Multivariate Analysis 97 (2006) 575 585 585 [5] R.M. Pringle, A.A. Rayner, Generalized Inverse Matrices with Applications to Statistics, Griffin, London, 1971. [6] C.R. Rao, Least squares theory using an estimated dispersion matrix and its application to measurement of signals, in: L.M. Le Cam, J. Neyman (Eds.), Proceedings of the Fifth Berkeley Symposium on Mathematical Statistics and Probability, vol. 1, University of California Press, Berkeley, 1967, pp. 355 372. [7] C.R. Rao, Unified theory of linear estimation, Sankhyā Ser. A 33 (1971) 371 394. [8] C.R. Rao, Linear Statistical Inference and Its Applications, second ed., Wiley, New York, NY, 1973. [9] C.R. Rao, Unified theory of least squares, Comm. Statist. Part A 1 (1973) 1 8. [10] C.R. Rao, The inefficiency of least squares: extensions of the Kantorovich inequality, Linear Algebra Appl. 70 (1985) 249 255. [11] C.R. Rao, S.K. Mitra, Generalized Inverse of Matrices and Its Applications, Wiley, New York, NY, 1971. [12] C.R. Rao, M.B. Rao, Matrix Algebra and Its Applications to Statistics and Econometrics, World Scientific, Singapore, 1998. [13] G. Zyskind, On canonical forms, non negative covariance matrices and best simple least squares linear estimators in linear models, Ann. Math. Statist. 38 (1967) 1092 1109.