Chapter 3 Best Linear Unbiased Estimation

Similar documents
Chapter 5 Prediction of Random Variables

Chapter 12 REML and ML Estimation

Chapter 11 MIVQUE of Variances and Covariances

Xβ is a linear combination of the columns of X: Copyright c 2010 Dan Nettleton (Iowa State University) Statistics / 25 X =

Estimation of the Response Mean. Copyright c 2012 Dan Nettleton (Iowa State University) Statistics / 27

MIXED MODELS THE GENERAL MIXED MODEL

Properties of Matrices and Operations on Matrices

LECTURE 2 LINEAR REGRESSION MODEL AND OLS

ML and REML Variance Component Estimation. Copyright c 2012 Dan Nettleton (Iowa State University) Statistics / 58

Matrix Operations: Determinant

ANALYTICAL MATHEMATICS FOR APPLICATIONS 2018 LECTURE NOTES 3

Lecture 6: Geometry of OLS Estimation of Linear Regession

Review of Classical Least Squares. James L. Powell Department of Economics University of California, Berkeley

The matrix will only be consistent if the last entry of row three is 0, meaning 2b 3 + b 2 b 1 = 0.

Estimation of Variances and Covariances

1 Determinants. 1.1 Determinant

Cayley-Hamilton Theorem

3.2 Gaussian Elimination (and triangular matrices)

MA 575 Linear Models: Cedric E. Ginestet, Boston University Regularization: Ridge Regression and Lasso Week 14, Lecture 2

Financial Econometrics

Likelihood Methods. 1 Likelihood Functions. The multivariate normal distribution likelihood function is

3.4 Elementary Matrices and Matrix Inverse

ON VARIANCE COVARIANCE COMPONENTS ESTIMATION IN LINEAR MODELS WITH AR(1) DISTURBANCES. 1. Introduction

A note on the equality of the BLUPs for new observations under two linear models

Inverse of a Square Matrix. For an N N square matrix A, the inverse of A, 1

COMP 558 lecture 18 Nov. 15, 2010

Solution Set 3, Fall '12

VARIANCE COMPONENT ESTIMATION & BEST LINEAR UNBIASED PREDICTION (BLUP)

1 Mixed effect models and longitudinal data analysis

Lecture Notes. Introduction

The dual simplex method with bounds

Solving Homogeneous Systems with Sub-matrices

Math 240 Calculus III

identity matrix, shortened I the jth column of I; the jth standard basis vector matrix A with its elements a ij

Advanced Engineering Mathematics Prof. Pratima Panigrahi Department of Mathematics Indian Institute of Technology, Kharagpur

Math 60. Rumbos Spring Solutions to Assignment #17

Chapter 2: Matrix Algebra

ANALYSIS OF VARIANCE AND QUADRATIC FORMS

Chapter 1. Matrix Algebra

2.7 Estimation with linear Restriction

4 Multiple Linear Regression

MATH 2360 REVIEW PROBLEMS

ENGR-1100 Introduction to Engineering Analysis. Lecture 21

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS. + + x 1 x 2. x n 8 (4) 3 4 2

Linear Models Review

Foundations of Matrix Analysis

Topic 7 - Matrix Approach to Simple Linear Regression. Outline. Matrix. Matrix. Review of Matrices. Regression model in matrix form

Chapter 3: Theory Review: Solutions Math 308 F Spring 2015

MATRICES AND MATRIX OPERATIONS

Chapter 5 Matrix Approach to Simple Linear Regression

STAT 100C: Linear models

UNIT 3 MATRICES - II

Problem set 5: SVD, Orthogonal projections, etc.

The Linear Regression Model

Review of Vectors and Matrices

MAT 2037 LINEAR ALGEBRA I web:

ELA THE MINIMUM-NORM LEAST-SQUARES SOLUTION OF A LINEAR SYSTEM AND SYMMETRIC RANK-ONE UPDATES

Linear Algebra Review

Matrix operations Linear Algebra with Computer Science Application

Projection. Ping Yu. School of Economics and Finance The University of Hong Kong. Ping Yu (HKU) Projection 1 / 42

NOTES ON MATRICES OF FULL COLUMN (ROW) RANK. Shayle R. Searle ABSTRACT

Mixed-Models. version 30 October 2011

Lesson 3. Inverse of Matrices by Determinants and Gauss-Jordan Method

3 Multiple Linear Regression

Problem Selected Scores

LECTURE 11: GENERALIZED LEAST SQUARES (GLS) In this lecture, we will consider the model y = Xβ + ε retaining the assumption Ey = Xβ.

Lecture 6: Linear models and Gauss-Markov theorem

Gauss Markov & Predictive Distributions

Solution Set 7, Fall '12

MLES & Multivariate Normal Theory

Linear Algebra. The analysis of many models in the social sciences reduces to the study of systems of equations.

MAT 1332: CALCULUS FOR LIFE SCIENCES. Contents. 1. Review: Linear Algebra II Vectors and matrices Definition. 1.2.

LS.1 Review of Linear Algebra

3 (Maths) Linear Algebra

Linear Regression and Its Applications

Exam 2 Solutions. (a) Is W closed under addition? Why or why not? W is not closed under addition. For example,

Regression. Oscar García

Linear Mixed-Effects Models. Copyright c 2012 Dan Nettleton (Iowa State University) Statistics / 34

Preliminaries. Copyright c 2018 Dan Nettleton (Iowa State University) Statistics / 38

Linear Algebra II. 2 Matrices. Notes 2 21st October Matrix algebra

MA 138 Calculus 2 with Life Science Applications Matrices (Section 9.2)

Econ 620. Matrix Differentiation. Let a and x are (k 1) vectors and A is an (k k) matrix. ) x. (a x) = a. x = a (x Ax) =(A + A (x Ax) x x =(A + A )

Math 2331 Linear Algebra

[y i α βx i ] 2 (2) Q = i=1

This can be accomplished by left matrix multiplication as follows: I

MATH 320: PRACTICE PROBLEMS FOR THE FINAL AND SOLUTIONS

BASIC NOTIONS. x + y = 1 3, 3x 5y + z = A + 3B,C + 2D, DC are not defined. A + C =

APPLICATIONS The eigenvalues are λ = 5, 5. An orthonormal basis of eigenvectors consists of

Chapter 1: Systems of linear equations and matrices. Section 1.1: Introduction to systems of linear equations

Linear models. Linear models are computationally convenient and remain widely used in. applied econometric research

30.3. LU Decomposition. Introduction. Prerequisites. Learning Outcomes

EE731 Lecture Notes: Matrix Computations for Signal Processing

Restricted Maximum Likelihood in Linear Regression and Linear Mixed-Effects Model

Math 215 HW #11 Solutions

Mixed-Model Estimation of genetic variances. Bruce Walsh lecture notes Uppsala EQG 2012 course version 28 Jan 2012

ECE 275A Homework 6 Solutions

Zellner s Seemingly Unrelated Regressions Model. James L. Powell Department of Economics University of California, Berkeley

1 Multiply Eq. E i by λ 0: (λe i ) (E i ) 2 Multiply Eq. E j by λ and add to Eq. E i : (E i + λe j ) (E i )

Solutions for Econometrics I Homework No.1

1 Linear Algebra Problems

Transcription:

Chapter 3 Best Linear Unbiased Estimation C R Henderson 1984 - Guelph In Chapter 2 we discussed linear unbiased estimation of k β, having determined that it is estimable Let the estimate be a y, and if k β is estimable, some a exists such that Ea y k β Assuming that more than one a gives an unbiased estimator, which one should be chosen? The most common criterion for choice is minimum sampling variance Such an estimator is called the best linear unbiased estimator BLUE Thus we find a such that Ea y k β and, in the class of such estimators, has minimum sampling variance Now V ara y a V arya a Va, where V ary V, assumed known, for the moment For unbiasedness we require a X k Consequently we find a that minimizes a Va subject to a X k Using a Lagrange multiplier, θ, and applying differential calculus we need to solve for a in equations V X X 0 a θ 0 k This is a consistent set of equations if and only if k β is estimable In that case the unique solution to a is V 1 XX V 1 X k A solution to θ is X V 1 X k, and this is not unique when X and consequently X V 1 X is not full rank Nevertheless the solution to a is invariant to the choice of a g-inverse of X V 1 X Thus, BLUE of k β is k X V 1 X X V 1 y But let β o X V 1 X X V 1 y, 1

where β o is any solution to X V 1 Xβ o X V 1 y known as generalized least squares GLS equations, Aitken 1935 Superscript 0 is used to denote some solution, not a unique solution Therefore BLUE of k β is k β o and y Let us illustrate with A solution is X 1 1 2 1 2 4 1 1 2 1 3 6, 5 2 4 3 Suppose V ary Iσ 2 e Then the GLS equations are σ 2 e 4 7 14 7 15 30 14 30 60 β o 14 22 44 β o 56 10 0/11 σ 2 e Then BLUE of 0 1 2β, which has been shown to be estimable, is Another solution to β o is 0 1 256 10 0 /11 10/11 56 0 5 /11 Then BLUE of 0 1 2β is 10/11, the same as the other solution to β o 1 Mixed Model Method For BLUE One frequent difficulty with GLS equations, particularly in the mixed model, is that V ZGZ + R is large and non-diagonal Consequently V 1 is difficult or impossible to compute by usual methods It was proved by Henderson et al 1959 that V 1 R 1 R 1 ZZ R 1 Z + G 1 1 Z R 1 Now if R 1 is easier to compute than V 1, as is often true, if G 1 is easy to compute, and Z R 1 Z + G 1 1 is easy to compute, this way of computing V 1 may have important advantages Note that this result can be obtained by writing equations, known as Henderson s mixed model equations 1950 as follows, X R 1 X X R 1 Z β o Z R 1 X Z R 1 Z + G 1 û 2 X R 1 y Z R 1 y

get Note that if we solve for û in the second equation and substitute this in the first we X [R 1 R 1 ZZ R 1 Z + G 1 1 Z R 1 ]Xβ o X [R 1 R 1 ZZ R 1 Z + G 1 1 Z R 1 ]y, or from the result for V 1 X V 1 Xβ o X V 1 y Thus, a solution to β o in the mixed model equations is a GLS solution An interpretation of û is given in Chapter 5 The mixed model equations are often well suited to an iterative solution Let us illustrate the mixed model method for BLUE with and X 1 1 1 2 1 1 1 3, Z 1 0 1 0 1 0 0 1, G R I, y [5 4 3 2] Then the mixed model equations are 4 7 3 1 7 15 4 3 β o 3 4 13 0 û 1 3 0 11 1 0 0 1 The solution is [286 50 2 2] /57 In this case the solution is unique because X has full column rank 14 22 12 2, Now consider a GLS solution V [ZGZ + R] 11 1 1 0 1 11 1 0 1 1 11 0 0 0 0 11 V 1 1 143 Then X V 1 Xβ o X V 1 y becomes The solution is 286 1 143 460 830 830 1852 132 11 11 0 11 132 11 0 11 11 132 0 0 0 0 130 β o 1 1580 143 2540 50/57 as in the mixed model equations 3

2 Variance of BLUE Once having an estimate of k β we should like to know its sampling variance Consider a set of estimators, K β o V ark β o V ar[k X V 1 X X V 1 y] K X V 1 X X V 1 VV 1 XX V 1 X K K X V 1 X K provided K β is estimable The variance is invariant to the choice of a g-inverse provided K β is estimable We can also obtain this result from a g-inverse of the coefficient matrix of the mixed model equations Let a g-inverse of this matrix be C11 C 12 C 21 C 22 Then This result can be proved by noting that V ark β o K C 11 K C 11 X [R 1 R 1 ZZ R 1 Z + G 1 1 Z R 1 ]X X V 1 X Using the mixed model example, let K 1 0 0 1 A g-inverse regular inverse of the coefficient matrix is 1 570 926 415 86 29 415 230 25 25 86 25 56 1 29 25 1 56 Then V ark β o 1 926 415 570 415 230 ] The same result can be obtained from the inverse of the GLS coefficient matrix because 1 460 830 143 1 1 926 415 830 1852 570 415 230 4

3 Generalized Inverses and Mixed Model Equations Earlier in this chapter we found that BLUE of K β, estimable, is K β o, where β o is any solution to either GLS or mixed model equations Also the sampling variance requires a g- inverse of the coefficient matrix of either of these sets of equations We define X V 1 X as a g-inverse of X V 1 X There are various types of generalized inverses, but the one we shall use is defined as follows A is a g-inverse of A provided that A A A A Then if we have a set of consistent equations, A p z, a solution to p is A z We shall be concerned, in this chapter, only with g-inverses of singular, symmetric matrices characteristic of GLS and mixed model equations 31 First type of g-inverse Let W be a symmetric matrix with order, s, and rank, t < s Partition W with possible re-ordering of rows and the same re-ordering of columns as W11 W W 12, W 12 W 22 W where W 11 is a non-singular matrix with order t Then W 1 11 0 0 0 It is of interest that for this type of W it is true that W W W W as well as W W W W This is called a reflexive g-inverse To illustrate, suppose W is a GLS coefficient matrix, 4 7 8 15 7 15 17 32 W 8 17 22 39 15 32 39 71 This matrix has rank 3 and the upper 3 3 is non-singular with inverse 41 18 1 30 1 18 24 12 1 12 11 5

Therefore a g-inverse is Another g-inverse of this type is 30 1 30 1 41 18 1 0 18 24 12 0 1 12 11 0 0 0 0 0 41 17 0 1 17 59 0 23 0 0 0 0 1 23 0 11 This was obtained by inverting the full rank submatrix composed of rows and columns 1, 2, 4 of W This type of g-inverse is described in Searle 1971b In the mixed model equations a comparable g-inverse is obtained as follows Partition X R 1 X with possible re-ordering of rows and columns as X 1R 1 X 1 X 1R 1 X 2 X 2R 1 X 1 X 2R 1 X 2 so that X 1R 1 X 1 has order r and is full rank Compute X 1R 1 X 1 Z R 1 X 1 Then a g-inverse of the coefficient matrix is mixed model coefficient matrix as follows X 1R 1 1 Z C00 C Z R 1 Z + G 1 02 C 02 C 22 C 00 0 C 02 0 0 0 C 02 0 C 22 5 8 8 3 2 8 16 16 4 4 8 16 16 4 4 3 4 4 8 0 2 4 4 0 7 We illustrate with a where X has 3 columns and Z has 2 Therefore X R 1 X is the upper 3 x 3 submatrix It has rank 2 because the 3rd column is the negative of the second Consequently find a g-inverse by inverting the matrix with the 3rd row and column deleted This gives 560 1 656 300 0 96 16 300 185 0 20 20 0 0 0 0 0 96 20 0 96 16 16 20 0 16 96 6

With this type of g-inverse the solution to β o is β o 1 0, where β o 1 has r elements Only the first p rows of the mixed model equations contribute to lack of rank of the mixed model matrix The matrix has order p + q and rank r + q, where r rank of X, p columns in X, and q columns in Z 32 Second type of g-inverse A second type of g-inverse is one which imposes restrictions on the solution to β o Let M β be a set of p r linearly independent, non-estimable functions of β Then a g-inverse X for the GLS matrix is obtained as follows V 1 1 X M C11 C M 12 O C 12 C 22 C 11 is a reflexive g-inverse of X V 1 X This type of solution is described in Kempthorne 1952 Let us illustrate GLS equations as follows 11 5 6 3 8 12 5 5 0 2 3 7 6 0 6 1 5 β o 5 3 2 1 3 0 8 8 3 5 0 8 4 This matrix has order 5 but rank only 3 Two independent non-estimable functions are needed Among others the following qualify 0 1 1 0 0 β 0 0 0 1 1 Therefore we invert which is 244 1 11 5 6 3 8 0 0 5 5 0 2 3 1 0 6 0 6 1 5 1 0 3 2 1 3 0 0 1 8 3 5 0 8 0 1 0 1 1 0 0 0 0 0 0 0 1 1 0 0, 28 1 1 13 13 122 122 1 24 24 7 7 122 0 1 24 24 7 7 122 0 13 7 7 30 30 0 122 13 7 7 30 30 0 122 122 122 122 0 0 0 0 122 0 0 122 122 0 0 7

The upper 5 x 5 submatrix is a g-inverse This gives a solution β o 386 8 8 262 262 /244 A corresponding g-inverse for the mixed model is as follows X R 1 X X R 1 Z M Z R 1 X Z R 1 Z + G 1 0 M 0 0 1 Then C11 C 12 C 12 C 22 C 11 C 12 C 13 C 12 C 22 C 23 C 13 C 23 C 33 is a g-inverse of the mixed model coefficient matrix The property of β o coming from this type of g-inverse is M β o 0 33 Third type of g-inverse A third type of g-inverse uses M of the previous section as follows X V 1 X + MM 1 C Then C is a g-inverse of X V 1 X In this case CX V 1 XC C This is described in Rao and Mitra 1971 We illustrate with the same GLS matrix as before and 0 1 1 0 0 M 0 0 0 1 1 as before with inverse X V 1 X + MM 244 1 11 5 6 3 8 5 6 1 2 3 6 1 7 1 5 3 2 1 4 1 8 3 5 1 9 150 62 60 48 74 62 85 37 7 7 60 37 85 7 7 48 7 7 91 31 74 7 7 31 91 which is a g-inverse of the GLS matrix The resulting solution to β o is the same as the previous section, 8

The corresponding method for finding a g-inverse of the mixed model matrix is X R 1 X + MM X R 1 1 Z C Then C is a g-inverse The property Z R 1 X Z R 1 Z + G 1 of the solution to β o is M β o 0 4 Reparameterization An entirely different method for dealing with the not full rank X problem is reparameterization Let K β be a set of r linearly independent, estimable functions of β Let ˆα be BLUE of K β To find ˆα solve K K 1 K X V 1 XKK K 1 ˆα K K 1 K X V 1 y ˆα has a unique solution, and the regular inverse of the coefficient matrix is V ar ˆα This corresponds to a model Ey X KK K 1 α This method was suggested to me by Gianola 1980 From the immediately preceding example we need 3 estimable functions An independent set is 1 1/2 1/2 1/2 1/2 0 1 1 0 0 0 0 0 1 1 The corresponding GLS equations are The solution is This is identical to 11 50 250 5 275 75 25 75 275 from the previous solution in which was forced to equal 0 ˆα ˆα 193 8 262/122 1 1/2 1/2 1/2 1/2 0 1 1 0 0 0 0 0 1 1 0 1 1 0 0 0 0 0 1 1 β o β o 12 1 2 9

The corresponding set of equations for mixed models is K K 1 K X R 1 XKK K 1 K K 1 K X R 1 Z Z R 1 XKK K 1 ˆα û Z R 1 Z + G 1 K K 1 K X R 1 y Z R 1 y 5 Precautions in Solving Equations Precautions must be observed in the solution to equations, especially if there is some doubt about the rank of the matrix If a supposed g-inverse is calculated, it may be advisable to check that AA A A Another check is to regenerate the right hand sides as follows Let the equations be C ˆα r Having computed ˆα, compute C ˆα and check that it is equal, except for rounding error, to r 10