LARGE SAMPLE VARIANCES OF MAXDUM LJICELIHOOD ESTIMATORS OF VARIANCE COMPONENTS. S. R. Searle. Biometrics Unit, Cornell University, Ithaca, New York

Similar documents
Mauricio A. Elzo, University of Florida, 1996, 2005, 2006, 2010, [8-1] ANIMAL BREEDING NOTES CHAPTER 8 BEST PREDICTION

BUILT-IN RESTRICTIONS ON BEST LINEAR UNBIASED PREDICTORS (BLUP) OF RANDOM EFFECTS IN MIXED MODELSl/

of, denoted Xij - N(~i,af). The usu~l unbiased

Vector spaces. DS-GA 1013 / MATH-GA 2824 Optimization-based Data Analysis.

Lecture Notes. Introduction

5. PROPAGATION OF VARIANCES APPLIED TO LEAST SQUARES ADJUSTMENT OF INDIRECT OBSERVATIONS., the inverse of the normal equation coefficient matrix is Q

C<»iMENTS ON ANOVA CALCUlATIONS FOR MESSY DATA t S. R. SEARLE. Biometrics Unit, Cornell University* and Florida State University.

RECENT DEVELOPMENTS IN VARIANCE COMPONENT ESTIMATION

An Overview of Variance Component Estimation

Inverse of a Square Matrix. For an N N square matrix A, the inverse of A, 1

S. R. Searle:-- Biometrics Unit, Cornell University, Ithaca, N. Y. Introductory Example

Do not copy, quote, or cite without permission LECTURE 4: THE GENERAL LISREL MODEL

Biometrics Unit. 337 Warren Hall. Cornell University. Ithaca, NY 14853

Unit roots in vector time series. Scalar autoregression True model: y t 1 y t1 2 y t2 p y tp t Estimated model: y t c y t1 1 y t1 2 y t2

ESTIMATING VARIANCE CO~WONENTS FOR THE NESTED MODEL, MINQUE OR RESTRICTED MAXIMUM LIKELIHOOD

All variances and covariances appearing in this formula are understood to be defined in the usual manner for finite populations; example

First Year Examination Department of Statistics, University of Florida

Principal Component Analysis

Variance Component Estimation Using Constrained Nonlinear.Maximization. Franz Preitschopf Universitat Augsburg. George Casella Cornell University

11 a 12 a 21 a 11 a 22 a 12 a 21. (C.11) A = The determinant of a product of two matrices is given by AB = A B 1 1 = (C.13) and similarly.

Kriging models with Gaussian processes - covariance function estimation and impact of spatial sampling

Mixed-Models. version 30 October 2011

. a m1 a mn. a 1 a 2 a = a n

Model Estimation Example

Testing Structural Equation Models: The Effect of Kurtosis

Chapter 7 - Section 8, Morris H. DeGroot and Mark J. Schervish, Probability and Statistics, 3 rd

A MODIFICATION OF THE HARTUNG KNAPP CONFIDENCE INTERVAL ON THE VARIANCE COMPONENT IN TWO VARIANCE COMPONENT MODELS

Mauricio A. Elzo, University of Florida, 1996, 2005, 2006, 2010, [10-1] ANIMAL BREEDING NOTES CHAPTER 10 BEST LINEAR UNBIASED PREDICTION

Math 40510, Algebraic Geometry

Marcia Gumpertz and Sastry G. Pantula Department of Statistics North Carolina State University Raleigh, NC

STAT 135 Lab 13 (Review) Linear Regression, Multivariate Random Variables, Prediction, Logistic Regression and the δ-method.

Estimation Tasks. Short Course on Image Quality. Matthew A. Kupinski. Introduction

ECE 275A Homework 6 Solutions

Kevin James. MTHSC 3110 Section 2.2 Inverses of Matrices

Lecture 13: Simple Linear Regression in Matrix Format. 1 Expectations and Variances with Vectors and Matrices

Canonical Correlation Analysis of Longitudinal Data

ON EXACT INFERENCE IN LINEAR MODELS WITH TWO VARIANCE-COVARIANCE COMPONENTS

Testing the homogeneity of variances in a two-way classification

79 1,2 jj8 STATISTICAL PROPERTIES OF ALLOCATION AVERAGES DDC. Research Memorandum C-, Behavioral Science Research Laboratory ,*., U S.

. ffflffluary 7, 1855.

Generalized, Linear, and Mixed Models

MATH 583A REVIEW SESSION #1

MS&E 226: Small Data

The volume growth of complete gradient shrinking Ricci solitons

Use of sparse matrix absorption in animal breeding

MATH 304 Linear Algebra Lecture 8: Vector spaces. Subspaces.

Chapter 11 MIVQUE of Variances and Covariances

Statistical Inference Using Maximum Likelihood Estimation and the Generalized Likelihood Ratio

Lecture Notes 4 Vector Detection and Estimation. Vector Detection Reconstruction Problem Detection for Vector AGN Channel

Sociedad de Estadística e Investigación Operativa

MAXIMUM LIKELIHOOD ESTIMATION IN A MULTINOMIAL MIXTURE MODEL. Charles E. McCulloch Cornell University, Ithaca, N. Y.

Econ 620. Matrix Differentiation. Let a and x are (k 1) vectors and A is an (k k) matrix. ) x. (a x) = a. x = a (x Ax) =(A + A (x Ax) x x =(A + A )

Covariance and Correlation

Empirical Power of Four Statistical Tests in One Way Layout

Lecture 13: Simple Linear Regression in Matrix Format

Optimal estimation for doubly multivariate data in blocked compound symmetric covariance structure

Covariance function estimation in Gaussian process regression

Alternative implementations of Monte Carlo EM algorithms for likelihood inferences

Lecture Notes 1: Vector spaces

7 Planar systems of linear ODE

ON VARIANCE COVARIANCE COMPONENTS ESTIMATION IN LINEAR MODELS WITH AR(1) DISTURBANCES. 1. Introduction

Variations. ECE 6540, Lecture 02 Multivariate Random Variables & Linear Algebra

Large Sample Properties of Estimators in the Classical Linear Regression Model

Efficient Estimation for the Partially Linear Models with Random Effects

An overview of Fay Herriot model with our package smallarea

i=1 h n (ˆθ n ) = 0. (2)

A Simple Procedure for Constructing Experiment Designs with Incomplete Blocks of Sizes 2 and 3

MULTINOMIAL PROBABILITY DISTRIBUTION

Stat 521A Lecture 18 1

MATH 423 Linear Algebra II Lecture 10: Inverse matrix. Change of coordinates.

Variations. ECE 6540, Lecture 10 Maximum Likelihood Estimation

Estimation of Unique Variances Using G-inverse Matrix in Factor Analysis

Minimax design criterion for fractional factorial designs

Chapter 4: Factor Analysis

A note on the equality of the BLUPs for new observations under two linear models

REGRESSION WITH SPATIALLY MISALIGNED DATA. Lisa Madsen Oregon State University David Ruppert Cornell University

Hypothesis Testing for Var-Cov Components

Maximum Likelihood Tests and Quasi-Maximum-Likelihood

New insights into best linear unbiased estimation and the optimality of least-squares

INEQUALITIES CONCERNING THE INVERSES OF POSITIVE DEFINITE MATRICES

STAT 100C: Linear models

Matrix Arithmetic. a 11 a. A + B = + a m1 a mn. + b. a 11 + b 11 a 1n + b 1n = a m1. b m1 b mn. and scalar multiplication for matrices via.

Sometimes the domains X and Z will be the same, so this might be written:

Approximating the Covariance Matrix with Low-rank Perturbations

Chapter 3 ANALYSIS OF RESPONSE PROFILES

HOW IS GENERALIZED LEAST SQUARES RELATED TO WITHIN AND BETWEEN ESTIMATORS IN UNBALANCED PANEL DATA?

Mixed-Model Estimation of genetic variances. Bruce Walsh lecture notes Uppsala EQG 2012 course version 28 Jan 2012

The quantum mechanics approach to uncertainty modeling in structural dynamics

INTER- AND INTRA-BLOCK ESTIMATION OF TREATMENT EFFECTS IN RANDOMIZED BLOCKS. Shayle R. Searle. Biometrics Unit, Cornell University, Ithaca, New York

Projektpartner. Sonderforschungsbereich 386, Paper 163 (1999) Online unter:

ESTIMATION OF THE RELATIVE FISHING POWER OF INDIVIDUAL SHIPS. D. s. Robson*

Regression Graphics. 1 Introduction. 2 The Central Subspace. R. D. Cook Department of Applied Statistics University of Minnesota St.

Introduction to Matrix Algebra

On the convergence of the iterative solution of the likelihood equations

B. L. Raktoe* and W. T. Federer University of Guelph and Cornell University. Abstract

Summer School in Statistics for Astronomers V June 1 - June 6, Regression. Mosuk Chow Statistics Department Penn State University.

Defense Technical Information Center Compilation Part Notice

HENDERSON'S APPROACH TO VARIANCE COMPONENTS ESTIMATION FOR UNBALANCED DATA UDC Vera Djordjević, Vinko Lepojević

COMPETING RISKS WEIBULL MODEL: PARAMETER ESTIMATES AND THEIR ACCURACY

LINEAR MODELS FOR CLASSIFICATION. J. Elder CSE 6390/PSYC 6225 Computational Modeling of Visual Perception

Transcription:

BU-255-M LARGE SAMPLE VARIANCES OF MAXDUM LJICELIHOOD ESTIMATORS OF VARIANCE COMPONENTS. S. R. Searle Biometrics Unit, Cornell University, Ithaca, New York May, 1968 Introduction Asymtotic properties of maximum likelihood (M.L.) estimators are usually considered in terms of sample size tefid:ing 'to infinity. When data are from a linear model involving one or more classifications, then the concept of sample size tending to infinity must be specified in a manner that takes into account the sample sizes in each s~bclass of the model. Hartley and Rao (1967) have discussed this problem, and it.is within their framework that we here consider asymtotic variances. Iterative procedures for deriving ML estimators are presented in that paper; an expr'e;ssion for asymtotic variances of those estimators is now obtained. Model We first note that in the most general terms, all linear models of the customary form z = 1-l_! +!!?. + ~ can be considered as mixed models. For, in the usual fixed model, there is one random term,!; and in the random model there is one fixed effect, lli thus, without loss of generality, any linear model can be considered as a mixed model and expressed as l =!!?. + z u where z is a vector of N observations;!?_ is a p X 1 vector of fixed effects, that include the general mean IJ.j ~ is a :ector of random effects, that include the error terms ~; and X and ~ are known matrices---often, but not always, design matrices. The random effects are further specified as having zero mean and a

- 2 - variance-covariance matrix ~' that involves q variance components cr~, In addition, for purposes of using maximum likelihood we adopt normality assumptions and so have ~ distributed as N(Q, ~). Thus l has variance-covariance matrix V = Z A Z 1 and is distributed as N(:?f~, ~) where elements of V are functions of the q variance components. Likelihood and variances For the above model the likelihood of the data is and, apart from a constant, the logarithm of this is Now the variance-covariance matrix of the large sample M.L. estimators of the p elements of ~ and the q variance components is minus the inverse of the expected value of the Hessian of L with respect to these p + q parameters. The sub-matrices of this Hessian are: ~~ a pxp matrix of terms d2l for h,k = 1, ' ' p; of3h of\ ~~cr2 ' a pxq matrix of terms o~h o2l ocrj for h = 1, p and j = 1, q; and L 2 2, a qxq matrix of terms -cr cr Then the matrix we seek is C\cr~ C\2L l ocr~ J for i,j = 1, q. v = -ML var (~) cov (~~2) cov (~2~) var (~2)

I - 3 - = - E (~13) - E (~cr2) I - E (~02) I - E (L -cr 2 2) a L -1 where ~ is the M.L. estimator of ~ and ~ 2 is the vector of M.L. estimators of the q variance components. For convenience write d for log 1~1. Then L = ~ ~ ( -2d - "'2 l ~~) 'y_-\r!~) and ~13 = -x'v-~ -- - ~02 = {~'(y_- 1 )cr~cr - ~~ )} for j = 1, 2,... ) q J and ~0202 = {-~d 2 2 cr.cr. - ~(r - ~~ ) '(( 1 )cr~o~<r -!. )} ~ J In these expressions for i,j 1, 2, q. dcr~o~ o2d ocrf?jcrj o2v-l and (y_-1)0202 = ocr~ ocr~ { cvr,s I =.. f for r,s = oa~ 1a~ 1, 2, N, where vr,s is the (r,s)'th element in v-1; i.e., (v~ 1 ) 2 2 is the matrix v-1 with - - aicrj every element differentiated with respect to cr~ and o~.

- 4 - \ Taking expectations we now have = = o, for j = 1, 2,, q; and for i,j = 1, 2,, q. Utilizing the fact that a scalar is its own trace, and that under the trace operation matrix products are cyclically commutable, we have = {-~d 2 2 - ttr [y(y-l)o~o~j} 0.0. Hence for i,j = 1, 2, ' q. YML x'v-ix 0 [ycy-l)o~o-:j} 0 ~ { d 2 + tr i,j = 1, q 0~0~ -1 Several points about this result are worthy of note. The first is that covariances between large sample M.L. estimators of fixed effects and variance components are zero. Bearing in mind that under conditions of normality the mean of a sample and its sum of squares are independent, this result is not surprising; but its generality is to be observed. An obvious consequence is, of course, that the variance-covariance matrix of the large sample M.L. estimators of the fixed

- 5 - effects is, fromyml' (_!'~(~)-l; and that for the variance components is the inverse of q. ---(1) With d being log f.yi this matrix is, it will be noted, free of the fixed effects and solely a function of the variance-covariance matrix y of the vector of observations I. Although y = ~A~ 1 and it is only A that involves the variance components, it does not appear to be more usef'ul to write (1) in terms of Z and A instead of v. Application The procedure for using (1) is clear: for any model find y,!yl, d =log IYI -1 and V Then, for every pair of variance components af and a~ (including i = j), derive d0 ~ 0 ~, (y_- 1 ) 0 ~ 0~ and tr [y(y- 1 ) 0~ 0 ~] When there are q components the ~J ~J... ~J values t {d 2 2 + tr [v(v-1) 2 2 lj.. } will, as in (1), constitute a square matrix a.a. -- aiaj. of order q, whose inverse yields the variances (and covariances) of the large sample estimators of the a~'s. As a simple example consider the model y = ~ + ei for i = 1, 2, N, where the e 1 are NID(O, a 2 ). Then V = a 2 I, IYI = a 2 N, d = N log a 2, and v- 1 = (l/a 2 )I. Hence and so, by substitution in (1), with cr 2 being the M.L. estimator of a2,

- 6-.. ( -N 2N )-1 2cr 2cr = -::--!++-::--!+ = 20'4 N as is to be expected. Further application of (1), to the random model, 1-way classification, unbalanced data, yields the results given in Searle (1956). In that case V = zc+jv., - '-"-J_ i the direct sum of matrices v 1 = cr2i + a2 J - e ni a ni Additional application, to the random model, 2-way classification, unbalanced data is currently being pursued. The matrix V is then V = I: G)v., with - i -J. References Hartley, H. 0. and J. N. K. Rao (1967). mixed analysis of variance model. Maximum likelihood estimation for the Biometrika 54:93-loB. Searle, S. R. (1956). Matrix methods in component of variance arid covariance analysis. Ann. Math. Stat. 27:737-748.