HENDERSON'S APPROACH TO VARIANCE COMPONENTS ESTIMATION FOR UNBALANCED DATA UDC Vera Djordjević, Vinko Lepojević

Similar documents
RECENT DEVELOPMENTS IN VARIANCE COMPONENT ESTIMATION

Lecture Notes. Introduction

VARIANCE COMPONENT ANALYSIS

ANOVA Variance Component Estimation. Copyright c 2012 Dan Nettleton (Iowa State University) Statistics / 32

ANOVA Variance Component Estimation. Copyright c 2012 Dan Nettleton (Iowa State University) Statistics / 32

Mixed-Model Estimation of genetic variances. Bruce Walsh lecture notes Uppsala EQG 2012 course version 28 Jan 2012

HOW IS GENERALIZED LEAST SQUARES RELATED TO WITHIN AND BETWEEN ESTIMATORS IN UNBALANCED PANEL DATA?

Measure for measure: exact F tests and the mixed models controversy

Chapter 3 Best Linear Unbiased Estimation

MIXED MODELS THE GENERAL MIXED MODEL

Linear Models in Statistics

ON VARIANCE COVARIANCE COMPONENTS ESTIMATION IN LINEAR MODELS WITH AR(1) DISTURBANCES. 1. Introduction

VARIANCE COMPONENT ESTIMATION & BEST LINEAR UNBIASED PREDICTION (BLUP)

LARGE SAMPLE VARIANCES OF MAXDUM LJICELIHOOD ESTIMATORS OF VARIANCE COMPONENTS. S. R. Searle. Biometrics Unit, Cornell University, Ithaca, New York

Math 423/533: The Main Theoretical Topics

Wiley. Methods and Applications of Linear Models. Regression and the Analysis. of Variance. Third Edition. Ishpeming, Michigan RONALD R.

BUILT-IN RESTRICTIONS ON BEST LINEAR UNBIASED PREDICTORS (BLUP) OF RANDOM EFFECTS IN MIXED MODELSl/

Introduction to Simple Linear Regression

Inverse of a Square Matrix. For an N N square matrix A, the inverse of A, 1

Multiple Regression Analysis. Part III. Multiple Regression Analysis

Topic 7 - Matrix Approach to Simple Linear Regression. Outline. Matrix. Matrix. Review of Matrices. Regression model in matrix form

Least Squares Estimation-Finite-Sample Properties

Chapter 11 MIVQUE of Variances and Covariances

Stochastic Design Criteria in Linear Models

MA 575 Linear Models: Cedric E. Ginestet, Boston University Mixed Effects Estimation, Residuals Diagnostics Week 11, Lecture 1

Parameter Estimation

Charles E. McCulloch Biometrics Unit and Statistics Center Cornell University

ANALYSIS OF INFLUENCE OF PARAMETERS ON TRANSFER FUNCTIONS OF APERIODIC MECHANISMS UDC Života Živković, Miloš Milošević, Ivan Ivanov

Lecture 3: Linear Models. Bruce Walsh lecture notes Uppsala EQG course version 28 Jan 2012

Chapter 14 Stein-Rule Estimation

International Journal of Pure and Applied Mathematics Volume 21 No , THE VARIANCE OF SAMPLE VARIANCE FROM A FINITE POPULATION

Mixed-Models. version 30 October 2011

Regression. Oscar García

An Overview of Variance Component Estimation

Chapter 5 Matrix Approach to Simple Linear Regression

Linear Models in Machine Learning

Domain estimation under design-based models

On Selecting Tests for Equality of Two Normal Mean Vectors

Linear regression with nested errors using probability-linked data

Large Sample Properties of Estimators in the Classical Linear Regression Model

An Introduction to Parameter Estimation

Sociedad de Estadística e Investigación Operativa

Alternative implementations of Monte Carlo EM algorithms for likelihood inferences

Fisher information for generalised linear mixed models

On V-orthogonal projectors associated with a semi-norm

A note on the equality of the BLUPs for new observations under two linear models

Practice Final Exam. December 14, 2009

Introduction to General and Generalized Linear Models

Lecture 3: Just a little more math

SOME THEOREMS ON QUADRATIC FORMS AND NORMAL VARIABLES. (2π) 1 2 (σ 2 ) πσ

Linear Model Under General Variance

Multiple comparisons of slopes of regression lines. Jolanta Wojnar, Wojciech Zieliński

Variance Component Estimation Using Constrained Nonlinear.Maximization. Franz Preitschopf Universitat Augsburg. George Casella Cornell University

ON INTERACTION VARIANCE CCMPONENTS IN MIXED MODELS

Likelihood Methods. 1 Likelihood Functions. The multivariate normal distribution likelihood function is

COMPREHENSIVE WRITTEN EXAMINATION, PAPER III FRIDAY AUGUST 26, 2005, 9:00 A.M. 1:00 P.M. STATISTICS 174 QUESTION

THE 'IMPROVED' BROWN AND FORSYTHE TEST FOR MEAN EQUALITY: SOME THINGS CAN'T BE FIXED

STAT 705 Chapter 19: Two-way ANOVA

Review of Classical Least Squares. James L. Powell Department of Economics University of California, Berkeley

STAT 705 Chapter 19: Two-way ANOVA

EXPLICIT SOLUTION OF THE OPERATOR EQUATION A X + X A = B

MA 575 Linear Models: Cedric E. Ginestet, Boston University Midterm Review Week 7

Lecture 2: Linear Models. Bruce Walsh lecture notes Seattle SISG -Mixed Model Course version 23 June 2011

ECON The Simple Regression Model

On the Behavior of Marginal and Conditional Akaike Information Criteria in Linear Mixed Models

REML Variance-Component Estimation

[y i α βx i ] 2 (2) Q = i=1

Multiple Linear Regression

Simple Regression Model Setup Estimation Inference Prediction. Model Diagnostic. Multiple Regression. Model Setup and Estimation.

Lecture 34: Properties of the LSE

Two-Way Factorial Designs

THE POWER OF THE MATRIX. Shayle R. Searle

Lecture 5: BLUP (Best Linear Unbiased Predictors) of genetic values. Bruce Walsh lecture notes Tucson Winter Institute 9-11 Jan 2013

Chapter 12 REML and ML Estimation

Topic 9: Factorial treatment structures. Introduction. Terminology. Example of a 2x2 factorial

STATISTICS 174: APPLIED STATISTICS FINAL EXAM DECEMBER 10, 2002

Maximum Likelihood Estimation

ON EXACT INFERENCE IN LINEAR MODELS WITH TWO VARIANCE-COVARIANCE COMPONENTS

Multivariate Regression (Chapter 10)

Combining data from two independent surveys: model-assisted approach

CoDa-dendrogram: A new exploratory tool. 2 Dept. Informàtica i Matemàtica Aplicada, Universitat de Girona, Spain;

Unsupervised Learning with Permuted Data

Contextual Effects in Modeling for Small Domains

Linear Models 1. Isfahan University of Technology Fall Semester, 2014

20.1. Balanced One-Way Classification Cell means parametrization: ε 1. ε I. + ˆɛ 2 ij =

SUPPLEMENTARY SIMULATIONS & FIGURES

Chapter 1: Linear Regression with One Predictor Variable also known as: Simple Linear Regression Bivariate Linear Regression

1 Mixed effect models and longitudinal data analysis

Chap The McGraw-Hill Companies, Inc. All rights reserved.

Are Forecast Updates Progressive?

Best linear unbiased prediction when error vector is correlated with other random vectors in the model

Lecture 13: Data Modelling and Distributions. Intelligent Data Analysis and Probabilistic Inference Lecture 13 Slide No 1

Sensitivity of GLS estimators in random effects models

Lecture: Simultaneous Equation Model (Wooldridge s Book Chapter 16)

Unit 6: Fractional Factorial Experiments at Three Levels

79 1,2 jj8 STATISTICAL PROPERTIES OF ALLOCATION AVERAGES DDC. Research Memorandum C-, Behavioral Science Research Laboratory ,*., U S.

Chapter 2 The Simple Linear Regression Model: Specification and Estimation

ENGRG Introduction to GIS

A new algebraic analysis to linear mixed models

Properties of the least squares estimates

Transcription:

FACTA UNIVERSITATIS Series: Economics and Organization Vol. 2, N o 1, 2003, pp. 59-64 HENDERSON'S APPROACH TO VARIANCE COMPONENTS ESTIMATION FOR UNBALANCED DATA UDC 519.233.4 Vera Djordjević, Vinko Lepojević Faculty of Economics, University of Niš, 18000 Niš, Serbia and Montenegro Abstract. The analysis of variance is a frequently used technique of data analysing and probably the most used statistical method in 20th century. However, there are a lot of ambiguities in the interpretation of particular aspects. The aim of this text is an explanation of one of the fundamental parts of this method - the variance components estimation. Namely, by calculating variability in observations we can divide their variance into parts which correspond to the factors we examine, and that means that the subject of study are the variance components. This text is a presentation of Henderson's method for variance components estimation in models of analysis of variance. 1. INTRODUCTION Estimating variance components from unbalanced data is not as straightforward as that obtained from balanced data. This is so for two reasons. First, several methods of estimation are available (most of which are reduced to the analysis of variance method for balanced data), but not one of them has yet been clearly established as superior to the others. Second, all the methods involve relatively cumbersome algebra; discussion of unbalanced data can therefore easily deteriorate into a welter of symbols, a situation we do our best (perhaps not successfully) to minimize here. It is probably safe to describe the Henderson (1953) paper as the foundation paper dealing with variance component estimation from unbalanced data. The methods there described have, accordingly, often been referred to as Henderson's methods 1,2 and 3. As described in Searle (1968, 1971, 1992), Method 1 is simply an analogue of the analysis of variance method used with balanced data; Method 2 is designed to correct a deficiency of Method 1 that arises with mixed models; and Method 3 is based on the method of fitting constants so often used in fixed effects models. Prior to the publication of these methods Windsor and Clark (1940) had utilized the analysis of variance method in studying variation in the catch of plankton nets, Eisenhart (1947) had clearly specified distinctions between fixed, random, and mixed models, and Crump (1946, 1947, 1951) had established Received December 23, 2003

60 V. DJORDJEVIĆ, V. LEPOJEVIĆ sampling variances of the variance component estimators in the one-way classification. Henderson (1953) however, greatly extended the estimation procedures, especially in describing three different methods and in indicating their use in multi-way classifications. Since then, a number of developments have been made. Variances of estimated components have been considered by Tukey (1957a), Searle (1956, 1958, 1961a, 1992), Khuri, A.I., and Sahai, H. (1985) Hocking (1996); defects in Henderson's Method 2 have been demonstrated by Searle (1968,1992), and difficulties with the mixed model have been discussed by Searle and Henderson (1961), Cunningham and Henderson (1968), and Thompson (1969) Christensen, R. (1996); and other methods of estimation have been developed: maximum likelihood by Hartley and Rao (1967) and large sample variances there from by Searle (1967,1992), and Sahai and Ageel (2001). Not all of these developments have been applied to all of even the most straightforward applications and some of them are more specialized than others. 2. GENERAL QUADRATIC FORMS All currently available methods for estimating variance components from unbalanced data use, in one way or another, quadratic forms of the observations. Before describing the methods we therefore outline properties of quadratic forms of observations coming from a general linear model. This is taken as: y = Xβ + e where y is a vector of observations, X is a matrix of known values, β is a vector of parameters (including both fixed and random effects) and e is a vector of the customary error terms. The vector of means and the variance-covariance matrix are taken respectively as: E(y) = µ and V = var(y) = E(y - µ)(y - µ)'. Expected values The expected value under the above- mentioned model of the quadratic form y'qy is: E(y'Qy) = tr (QV) +!'Qµ (1) where 'tr' represents the trace operation on matrix, that of summing its diagonal elements. This result is the basis of most methods of estimating variance components from unbalanced data. The general methodology is to obtain expected values of quadratic forms from (1) and to equate them to their observed values; i.e. to equate E (y'qy) to the observed y'qy. This is exactly what is done with mean squares (which are quadratic forms of the observations) in the analysis of variance method for balanced data. But, whereas with balanced data there is ' obviously' only one set of quadratic forms to use (the analysis of variance mean squares) and they lead to estimators that have some optimal properties, there are many sets of quadratics that can be used for unbalanced data. However, most of such sets lead to estimators that have few optimal properties and no particular set of quadratics has yet been established as more optimal than any other set. Result (1) applies no matter what form of the model y = Xβ + e is used: β always includes all the effects in the model, be they fixed or random or a mixture of both. In most

Henderson's Approach to Variance Components Estimation for Unbalanced Data 61 situations we assume that E(e) = 0, so that var (e) is " (ee') = σe 2 I. In addition, when β is a vector of fixed effects, E(βe') = βe(e') = 0; and when # includes elements that are random effects they are assumed to have zero mean and zero covariance with the elements in e; thus at all times we take E(βe') = E(eβ') = 0. In a fixed effects model β is a vector of fixed effects, E(y) = Xβ and V = var (y) = var (e) = σ e 2 IN, where are N observations, i.e., y is N X 1. Then (1) becomes: "(y'qy) = β'xqxβ + In a mixed model β' can be partitioned as 2 2σ etr (Q) ' ' ' ' β ' = ( β1β AβB... βk ), where β 1 contains all the fixed effects of the model (including the mean) and where the other β's each represent the set of random effects for the factors A, B, C,, K, these random effects having zero means and zero co variances with the effects of any other set. (Although, only single subscripts are used, interaction effects and/or nested-factor effects are not excluded by this notation. They are considered merely as factors, each identified by a single subscript rather than the letters of the appropriate main effects; for example, AB- interaction effects might be in the vector labeled # F ). 3. THE ANALYSIS OF VARIANCE METHOD (HENDERSON'S METHOD 1) The analysis of variance method of estimating variance components is essentially the only method in use for balanced data. It is also the most frequently used method with unbalanced data, although its application is not as straightforward and its deficiencies are more pronounced. Nevertheless, it is likely to continue as an oft-used method and so considerable attention is devoted to it here. With balanced data the method consists of equating mean squares to their expected values. Essentially the same procedure is used with unbalanced data, as is now shown in terms of an example, the two-way crossed classification with interaction. The model for a levels of an A-factor crossed with b levels of a B-factor is ijk = µ + αi + β j + γij eijk (2) y + where y ijk is the k-th observation (for k = 1,2,,n ij ) in the i-th level of the A-factor and the j-th level of the B-factor, where i = 1,2,., a and j = 1,2,.,b. Thus n ij is the number of observations in the (i,j) cell- the i-th level of A and the j-th level of B. Since not all of the cells may contain observations we let s represent the number that do; i.e., s is the number of n ij 's for which n ij > 0. Thus ab-s is the number of cells containing no data (n ij = 0). In (43), µ is a general mean, α i is the effect due to the i-th level of the A-factor, β j is the effect due to the j-th level of the B-factor, γ ij is the interaction effect and e ijk is the customary error term.

62 V. DJORDJEVIĆ, V. LEPOJEVIĆ 4. ADJUSTING FOR BIAS IN MIXED MODELS (HENDERSON METHOD 2) Mixed models involve dual estimation problems - estimating both fixed effects and variance components. For the moment attention is directed just to estimating the variance components. In some situations this is exactly what might be done in practice; with genetic data, for example, time or year effects might be considered fixed and of little interest compared to the genetic variance components. On the other hand, time trends may be of very real interest in some data, in which case their estimation together with that of the variance components would be considered simultaneously. This dual estimation problem is considered subsequently. The analysis of variance method for mixed model leads, with unbalanced data, to biased estimators of variance components. The method known as Method 2 in Henderson (1953) is designed to correct this deficiency. It uses the data first to estimate fixed effects of the model and then, using these estimators to adjust the data, variance components are estimated from the adjusted data, by the analysis of variance method. The whole procedure is designed so that the resulting variance component estimators are not biased by the presence of the fixed effects in the model, as they are with analysis of variance estimators derived from the basic data. So far as the criterion of unbiased ness is concerned, this is certainly achieved by Method 2. But the general method of analyzing data adjusted accordant to some estimator of the fixed effects is open to criticism on other grounds: it cannot be uniquely defined, and a simplified form of it, of which Henderson's Method 2 is a special case, cannot be used whenever the model includes interactions between the fixed effects and the random effects. The general approach of Method 2 can be considered in terms of the model: y = µ 1+ X f β f + X rβr + e where all fixed effects other than! are represented by # f, and all random effects by # r. As usual E(β r ) = 0 and so E(β r β' r ) = V(β r ), the variance-covariance matrix of the random effects. The general effect of correcting the data vector y accordant to an estimator of the fixed effects β f is to suppose that such an estimator is β ~ f = Ly for some matrix L so that ~ the vector of corrected data is z = yx f β f. It can then be shown (Searle 1992) that the model for z contains no terms in β f provided L is a generalized inverse of $ f. Under this condition the analysis of variance method applied to y - X f β ~ f. will yield unbiased estimators of the variance components. However, the fact that L has only to be a generalized inverse of $ f indicates the arbitrariness of the method. This lack of specificity means the method is not uniquely defined and hence is impractical. The model for z = y -X f β ~ f. just described contains no term in β f. An additional restriction would be for the model to have the same term $ r β r as does the model for y, as wall as a mean term µ 1 1 where µ 1 is not necessarily equal to µ.

Henderson's Approach to Variance Components Estimation for Unbalanced Data 63 5. THE FITTING CONSTANTS METHOD (HENDERSON'S METHOD 3) The third method described by Henderson is based on the method of fitting constants traditionally used in fixed effects models. It uses reductions in sums of squares due to fitting different subgroups of factors in the model, using them in exactly the same manner as the S's are used in the analysis of variance method, namely estimating the variance components by equating each computed reduction to its expected value. In an example of the two-way classification random model with interaction, the four quadratics equated to their expected values in the analysis of variance method are S A, S B, S AB and SSE shown as functions of T's in equations. The fitting constants method also uses four quadratics, derived from fitting the model: and three sub-models: y = µ + α + β + γ + e ijk i y ijk = µ + e ijk, y ijk = µ + αi + eijk, y ijk = µ + αi + β j + eijk In fitting these models the total sum of squares is T 0 = ΣΣΣ y 2 ijk. With each model there is a reduction in sum of squares, y'x(x'x) - X'y, due to fitting the model. These reductions can be donated by: R ( µ, A, B, AB), R (µ), R ( µ, A), R ( µ, A, B) respectively, where the letters in parentheses indicate the factors fitted in the respective models. (In this notation AB represents, as usual, A-by-B interaction). By way of example, the last of the above models fits µ, A-, and B- factors, and so the reduction is symbolized as R(µ,A,B). Writing the model as y = $b + e, where b is the vector containing µ, the α's and β's we have R(µ,A,B) = y'x(x'x) - X'y for that $. The fitting constants method of estimating variance components uses T 0 and these R( ) - reductions by equating certain differences among them to their expected values, so setting up equations in the σ 2 's whose solutions are the estimators. j ij ijk 6. CONCLUSION Henderson's approach to variance components estimation for unbalanced data was the foundation paper for other statisticians in attempt to solve this problem. The methods there described have, accordingly, often been referred to as Henderson's methods 1,2 and 3. Method 1 is simply an analogue of the analysis of variance method used with balanced data; Method 2 is designed to correct a deficiency of Method 1 that arises with mixed models; and Method 3 is based on the method of fitting constants so often used in fixed effects models.

64 V. DJORDJEVIĆ, V. LEPOJEVIĆ REFERENCES 1. Christensen, R. 1996. Analysis of Variance, Design and Regression. London: Chapman&Hall. 2. Eisenhart, C. 1947 "The assumptions underling the analysis of variance." Biometrics 3: 1-21. 3. Hartley, H.O. 1967. "Expectations variances and covariance's of ANOVA mean squares by "synthesis." Biometrics 25: 105-114. 4. Hartley, H.O. and Rao, J.N.K. 1967. "Maximum likelihood estimation for the mixed analysis of variance models." Biometrics 23: 93-108. 5. Hartley, H.O. and Searle, R.R. 1969. "A discontinuity in mixed model analysis." Biometrics 25: 573-576 6. Henderson, C.R. 1953 "Estimation of variance and covariance components." Biometrics 9: 226-252. 7. Hoaglin, D.C., Mosteller, F., and Tukey, J.W. 1991. Fundamentals of Exploratory Analysis of Variance. New York; Wiley. 8. Hockihg, R.R., Green, J.W. and Bremer, R.H. 1989. "Variance component estimation with model-based diagnostics." Tehnomestrics 31: 227-239. 9. Hocking, R.R. 1996. Methods and Applications of Linear Models: Regression and the Analysis of Variance. New York: Wiley. 10. Kempthorne, O. 1975. "Fixed and mixed models in the analysis of variance." Biometrics 31:473-486. 11. Khuri, A.I., and Sahai, H. 1985. "Variance components analysis: Selective literature review." International Statistics 11: 279-300. 12. Miller, R.G. 1997. Beyond ANOVA. Chapman&Hall 13. Sahai, H. 1979. "A Bibliography on variance components." International Statistical Reviews 47: 177-222. 14. Sahai H., Ageel M., 2001. "The analysis of variance", Birkhauser, 15. Sheffe, H. 1959. The analysis of Variance. New York: Wiley. 16. Searle, S.R. 1969. "Another look at Hendersons methods of estimating variance components." Biometrics 24:.749-788. 17. Searle, S.R. 1971. "Topics in variance component estimation." Biometrics 27: 1-76. 18. Searle, S.R., Casella, G., and McCulloch, C.E. 1992. Variance Components. New York: Wiley. 19. Thompson, W.A.Jr., and Moore, J.R. 1963. Non-negative estimates of variance components. Tehnometrics 5: 441-450. HENDERSONOV PRISTUP OCENI VARIJANSNIH KOMPONENTI ZA NEURAVNOTEŽENE PODATKE Vera Djordjević, Vinko Lepojević Analiza varijanse je obilato korišćena tehnika analize podataka i verovatno najčešće korišćeni statistički metod u dvadesetom veku. No, i pored toga, još uvek postoje nejasnoće u tumačenju pojedinih njenih aspekata. Ovaj rad ima za cilj osvetljavanje jednog od fundamentalnih delova ovog metoda - ocenu varijansnih komponenti. Naime, u cilju izračunavanja varijabiliteta u posmatranjima možemo razložiti njihovu varijansu na delove koji odgovaraju faktorima koje proučavamo, a to u stvari znači da predmet proučavanja predstavljaju varijansne komponente. U radu je prezentovan kritički pristup Henderson-ov metode ocene varijansnih komponenti u modelima analize varijanse.