Marcia Gumpertz and Sastry G. Pantula Department of Statistics North Carolina State University Raleigh, NC

Similar documents
A Simple Approach to Inference in Random Coefficient Models

Approximating Bayesian Posterior Means Using Multivariate Gaussian Quadrature

Multivariate Time Series: Part 4

Large Sample Properties of Estimators in the Classical Linear Regression Model

Restricted Maximum Likelihood in Linear Regression and Linear Mixed-Effects Model

Some Monte Carlo Evidence for Adaptive Estimation of Unit-Time Varying Heteroscedastic Panel Data Models

Mantel-Haenszel Test Statistics. for Correlated Binary Data. Department of Statistics, North Carolina State University. Raleigh, NC

G. S. Maddala Kajal Lahiri. WILEY A John Wiley and Sons, Ltd., Publication

An Introduction to Multivariate Statistical Analysis

HOW IS GENERALIZED LEAST SQUARES RELATED TO WITHIN AND BETWEEN ESTIMATORS IN UNBALANCED PANEL DATA?

ON EXACT INFERENCE IN LINEAR MODELS WITH TWO VARIANCE-COVARIANCE COMPONENTS

STAT5044: Regression and Anova. Inyoung Kim

Lecture 3: Multiple Regression

2 Regression Analysis

Regression and Statistical Inference

TESTING FOR NORMALITY IN THE LINEAR REGRESSION MODEL: AN EMPIRICAL LIKELIHOOD RATIO TEST

Analysis of variance, multivariate (MANOVA)

Nonstationary Panels

Econ 510 B. Brown Spring 2014 Final Exam Answers

On Selecting Tests for Equality of Two Normal Mean Vectors

Dr. Junchao Xia Center of Biophysics and Computational Biology. Fall /1/2016 1/46

1. The Multivariate Classical Linear Regression Model

A TIME SERIES PARADOX: UNIT ROOT TESTS PERFORM POORLY WHEN DATA ARE COINTEGRATED

Fundamental Probability and Statistics

y it = α i + β 0 ix it + ε it (0.1) The panel data estimators for the linear model are all standard, either the application of OLS or GLS.

Inverse of a Square Matrix. For an N N square matrix A, the inverse of A, 1

Review of Classical Least Squares. James L. Powell Department of Economics University of California, Berkeley

Asymptotic Theory. L. Magee revised January 21, 2013

Stochastic Design Criteria in Linear Models

Testing Some Covariance Structures under a Growth Curve Model in High Dimension

Heteroskedasticity. in the Error Component Model

LECTURE 2 LINEAR REGRESSION MODEL AND OLS

Economics 573 Problem Set 5 Fall 2002 Due: 4 October b. The sample mean converges in probability to the population mean.

Empirical Power of Four Statistical Tests in One Way Layout

Biostatistics Workshop Longitudinal Data Analysis. Session 4 GARRETT FITZMAURICE

Generalized, Linear, and Mixed Models

[y i α βx i ] 2 (2) Q = i=1

Sociedad de Estadística e Investigación Operativa

Sensitivity of GLS estimators in random effects models

Statistics 910, #5 1. Regression Methods

COMPARISON OF FIVE TESTS FOR THE COMMON MEAN OF SEVERAL MULTIVARIATE NORMAL POPULATIONS

Unit roots in vector time series. Scalar autoregression True model: y t 1 y t1 2 y t2 p y tp t Estimated model: y t c y t1 1 y t1 2 y t2

Unit 12: Analysis of Single Factor Experiments

Linear Models 1. Isfahan University of Technology Fall Semester, 2014

Statistics for Engineers Lecture 9 Linear Regression

A MODEL-BASED EVALUATION OF SEVERAL WELL-KNOWN VARIANCE ESTIMATORS FOR THE COMBINED RATIO ESTIMATOR

Canonical Correlation Analysis of Longitudinal Data

Applied Econometrics (QEM)

statistical sense, from the distributions of the xs. The model may now be generalized to the case of k regressors:

Introduction to Eco n o m et rics

Introduction to Estimation Methods for Time Series models. Lecture 1

14 Multiple Linear Regression

Ratio of Linear Function of Parameters and Testing Hypothesis of the Combination Two Split Plot Designs

Linear models. Linear models are computationally convenient and remain widely used in. applied econometric research

PANEL DATA RANDOM AND FIXED EFFECTS MODEL. Professor Menelaos Karanasos. December Panel Data (Institute) PANEL DATA December / 1

Linear Model Under General Variance

Bootstrapping the Grainger Causality Test With Integrated Data

A Multivariate Two-Sample Mean Test for Small Sample Size and Missing Data

Bootstrapping Heteroskedasticity Consistent Covariance Matrix Estimator

An Approximate Test for Homogeneity of Correlated Correlation Coefficients

Sample size calculations for logistic and Poisson regression models

Economics 240A, Section 3: Short and Long Regression (Ch. 17) and the Multivariate Normal Distribution (Ch. 18)

Testing for Unit Roots in Autoregressive Moving Average Models: An Instrumental Variable Approach. Sastry G. Pantula* and Alastair Hall

The LIML Estimator Has Finite Moments! T. W. Anderson. Department of Economics and Department of Statistics. Stanford University, Stanford, CA 94305

Working Paper No Maximum score type estimators

Consistency of test based method for selection of variables in high dimensional two group discriminant analysis

Ordinary Least Squares Regression

MA 575 Linear Models: Cedric E. Ginestet, Boston University Non-parametric Inference, Polynomial Regression Week 9, Lecture 2

the error term could vary over the observations, in ways that are related

Heteroskedasticity. Part VII. Heteroskedasticity

Cointegration Lecture I: Introduction

Inferences on a Normal Covariance Matrix and Generalized Variance with Monotone Missing Data

11. Bootstrap Methods

Vector Auto-Regressive Models

Test Code: STA/STB (Short Answer Type) 2013 Junior Research Fellowship for Research Course in Statistics

Testing Homogeneity Of A Large Data Set By Bootstrapping

HANDBOOK OF APPLICABLE MATHEMATICS

ANALYSIS OF PANEL DATA MODELS WITH GROUPED OBSERVATIONS. 1. Introduction

VAR Models and Applications

Econ 583 Homework 7 Suggested Solutions: Wald, LM and LR based on GMM and MLE

THE EFFECTS OF MULTICOLLINEARITY IN ORDINARY LEAST SQUARES (OLS) ESTIMATION

Mean squared error matrix comparison of least aquares and Stein-rule estimators for regression coefficients under non-normal disturbances

TIME SERIES DATA ANALYSIS USING EVIEWS

GMM estimation of spatial panels

A Practical Guide for Creating Monte Carlo Simulation Studies Using R

ANALYSING BINARY DATA IN A REPEATED MEASUREMENTS SETTING USING SAS

NONLINEAR REGRESSION FOR SPLIT PLOT EXPERIMENTS

Multivariate Regression

Testing Random Effects in Two-Way Spatial Panel Data Models

Missing dependent variables in panel data models

So far our focus has been on estimation of the parameter vector β in the. y = Xβ + u

Applied Multivariate and Longitudinal Data Analysis

Chapter 14 Stein-Rule Estimation

UNIVERSITY OF TORONTO Faculty of Arts and Science

Regression Analysis. y t = β 1 x t1 + β 2 x t2 + β k x tk + ϵ t, t = 1,..., T,

Panel Data Model (January 9, 2018)

Stein-Rule Estimation under an Extended Balanced Loss Function

MATH5745 Multivariate Methods Lecture 07

Bootstrap Approach to Comparison of Alternative Methods of Parameter Estimation of a Simultaneous Equation Model

Keywords: One-Way ANOVA, GLM procedure, MIXED procedure, Kenward-Roger method, Restricted maximum likelihood (REML).

Transcription:

A Simple Approach to Inference in Random Coefficient Models March 8, 1988 Marcia Gumpertz and Sastry G. Pantula Department of Statistics North Carolina State University Raleigh, NC 27695-8203 Key Words and Phrases: Repeated measures regression; Asymptotic inference; Estimated generalized least squares; Growth curve. ABSTRACT Random coefficient regression models have been used to analyze crosssectional and longitudinal data in economics and growth curve data from biological and agricultural experiments. In the literature several estimators, including the ordinary least squares (OLS) and the estimated generalized least squares (EGLS), have been considered for estimating the parameters of the mean model. Based on the asymptotic properties of the EGLS estimators, test statistics have been proposed for testing linear hypotheses involving the parameters of the mean model. An alternative estimator, the simple mean of the individual regression coefficients, provides estimation and hypothesis testing procedures that are simple to compute and simple to teach. The large sample properties of this simple estimator are shown to be similar to that of the EGLS estimator. The performance of the proposed estimator is compared with that of the existing estimators by Monte Carlo simulation.

1. INTRODUCTION Frequently in biological, medical, agricultural and clinical studies several measurements are taken on the same experimental unit over time with the objective of fitting a response curve to the data. Such studies are called growth curve, repeated measure or longitudinal studies. In many bio-medica1 and agricultural experiments the number of experimental units is large and the number of repeated measurements on each unit is small. On the other hand, some economic investigations and meteorological experiments involve a small number of units observed over a long period of time. Several models for analyzing such data exist in the literature; the models usually differ in their covariance structures. See Harville (1977) and Jennrich and Schluchter (1986) for a review of the models and of approaches for estimating parameters. Recently, there seems to be a renewed interest in analyzing repeated measures data using Random Coefficient Regression (RCR) models. In this article we present a brief description of the RCR models and some of the existing results for these models. We present a simple approach that is not difficult to discuss in a course on linear models. We compare our estimation procedure with two of the existing methods. Section 2 contains the assumptions of the model and a brief review of the literature. In section 3 we present the properties of the simple estimator. A Monte Carlo comparison of the estimators is presented in Section 4. Finally, we conclude with a summary which includes some possible extensions.

Page 2 2. RCR MODEL Suppose that the t observations on the ith of n experimental units are described by the model y. = X.f3. + e. i=1,2,..., n (2.1), ",, ",,,, where y. = (y. 1,y. 2,..., y. t )' is a tx1 vector of observations on the response variable, X. is a txk matrix of observations on k explanatory variables, 13. is a kx1 vector of coefficients unique to the ith experimental unit and e. is a tx1 vector of errors. Each experimental unit and its response curve is considered to be selected from a larger population of response curves, thus the regression coefficient vectors 13.,, i=1,2,...,n may be viewed as random drawings from some k-variate population and hence (2.1) is called a RCR model. In this paper we discuss the estimation and testing of such models under the following assumptions: (i) the e., vectors are independent multivariate normal variables with mean zero and covariance matrix a 2 I t i (ii) the f3 i vectors are independent multivariate normal variables with mean 13 and nonsingular covariance matrix Eo ; (iii) the vectors 13. and e. are independent for all i and j; (iv) the X. 1Jf3, J, matrices are fixed and of full row rank for each i; (v) min(n,t) > k and -1 (vi) there exists an M < ~ such that the elements of t(x~x.) are less than M in absolute value for all i and t.,, The assumption (vi) is not very restrictive. It is satisfied for the models that include polynomials in time and stationary exogeneous variables. Several authors, including Rao (1965), Swamy (1971), Hsiao (1975), Harville (1977), Laird and Ware (1982), Jennrich and Schluchter (1986) and Carter and Yang (1986) have considered the estimation and testing for the RCR models. We summarize the results of Carter and Yang (1986) since they consider the large sample distribution of the estimated generalized least squares (EGLS) estimator

Page 3 as nand/or t tend to infinity. For the sake of simplicity, we have assumed that equal number of repeated measurements are taken on all experimental units and that the variance of the error vector e. does not depend on i. 1 However, similar results exist for more general cases and will be discussed in the summary. Consider the least squares estimators b. = (X'.X.)-1 X'.y. 11111 i=1,2,..., n (2.2) of ~ computed for each individual experimental unit. Note that the b. 's are 1 independent and normally distributed with mean ~ and variance -1 2-1 Wi = E~~ + a (XiXi) Therefore, the best (linear) unbiased estimator of ~ is the generalized least squares (GLS) estimator. Swamy (1971) showed that ~GLS n n = ( E w.)-'( E N.b.) i=1 1 i=1 1 1 (2.3) that is, the GLS estimator is the "weighted" least squares (average) estimator of b. where the weights are the inverse variance-covariance matrices of b.. 1 1 Under the normality assumption, ~GLS is also the maximum likelihood estimator 2 2 of ~ (provided E~~ and a are known). The elements of E~~ and a are seldom known and hence we consider the estimated GLS (EGLS) estimator ~EGLS n = ( E W.) (E W.b.) 1. 1 1 1 i=1 1= A 1 n (2.4) where

Page 4 A_1 A2-1 Wi = E~~ + a (X ixi) -1 A 2 n 1 = Sbb - n a E (X~X.)- i =1 ' n = (n-1) -1 E (b.- b) (b. - b) I i =1', -1 n = [n(t-k)] E [y~y. - b~x~y.], i=1 " ", and b -1 n = neb.. 1 ',= Carter and Yang (1986) suggested inference procedures based on the large sample distribution of the estimator ~EGLS. Their results are summarized below. (They suggested a slightly different estimator of E~~ in the case E~~ is not nonnegative definite.) Result 2.1: Consider the model given in (2.1) with the assumptions (i) through (vi). Consider the statistic A 2 A n -1-1 A T = [L ~EGLS - ~O] I [L( i:1 Wi) L I] [L ~EGLS - ~o], (2.5) for testing H o : L~ = ~o ' where L is a qxk matrix of q linearly independent rows. Then, (a) for a fixed nand t tending to infinity: -1-1 2 (n-q)q (n-1) T is (asymptotically) distributed as F(q,n-q), (b) for a fixed t and n tending to infinity: T 2 is (asymptotically) distributed as chi-square with q degrees of freedom, and

Page 5 (c) for the case where nt is large and q=1: T 2 is approximately distributed as F(1,v) where II = L I, and Proof: See Carter and Yang (1986). [ ] Carter and Yang (1986) proved part (b) of the above result by observing that the distribution of ~EGLS is asymptotically (as n ~~) equivalent to that of ~GLS' To prove part (a), they observed that the distribution of ~EGLS is asymptotically (as t ~~) equivalent to that of II 13-1 = n (2.6) which is also asymptotically equivalent to I3 GLS as t ~~. Finally, when nt is large, Satterthwaite's approximation was used to approximate the 2 distribution of T. In the next section we present inference procedures based on the large sample distribution of the simple estimator b. 3. A SIMPLE APPROACH It is well known that the GLS estimator ~GLS is the best (linear) unbiased estimator of ~ and that (under some regularity conditions) the EGLS estimator ~EGLS is asymptotically (as n ~~) equivalent to the GLS estimator. However, in small samples, the distribution of ~EGLS may be far from being

Page 6 normal. It is also argued that the estimator ~EGLS may even be worse than the ordinary least squares (OLS) estimator, ~OLS n n = ( E X'.X.)-'( E X'.X.b.) i=1 1 1 i=1 1 1 1 (3.1 ) because ~EGLS depends on the estimated variance-covariance matrix which may introduce additional variability. It is easy to see that the OLS estimator, ~OLS' is normally distributed with mean ~ and variance n 1[ n 2 n ] n -1 (E. 1X '. X. ) - E. 1X '. X. I:ooX ~ X. + a E. 1X!X. (E. 1X '. X. ) 1 = 1 1 1= 1 1,.,,., 1 1 1 = 1 1 1 = 1 1 Thus, to compute either the EGLS estimate ~EGLS or to compute the variance covariance matrices of ~EGLS and ~OLS' it is necessary to estimate the 2 elements of I:~~ and a We now present the properties of the simple estimator 2 b, which does not require the estimation of I:~~ and a. Note that the GLS, EGLS and OLS estimators are weighted averages of the individual least squares estimators b. 1 The estimator -1 n b =neb. (3.2) i=1 1 is the simple average of the individual least squares estimators. In the special case where the model matrix X. 1 is the same (=A say) for all individuals, then the GLS, EGLS and OLS estimates coincide with the estimator b. The estimator b is normally distributed with mean ~ and variance Var(b) (3.3)

Page 7 Note that n = E[(n-1)-1 E (b.- b)(b.- b)'l i =1', = (n-1 ) -1 E[ ~ (b. - ~)( b. - ~) I - n (b - ~)(b - ~).]. l',,= = (n-1)-1[ ~ var(b.) - n Var(b)). 1 ',= = (n_1)-1[n 2 var(b) - n Var(bG = n var(b) - -1 Therefore, a simple unbiased estimator for var(b) is n Sbb That is, the sample variance (covariance matrix) divided by n is an unbiased estimator for the variance of the sample mean even though the variances (of b.) are not homogeneous. 7 -- Consider the statistic (3.4) for testing H o : L~ = ~o' where L is a qxk matrix of linearly independent rows. A*2 2 Notice that T is the Hotelling's T statistic one would compute if the variances of the b.'s were equal individuals).,, A*2 Before we establish that the statistic T (i.e., if the X.'s were the same for all has similar asymptotic properties as that of the statistict 2, we will make a few remarks. Remark 3.1: Recall that the estimators b., are independently and normally -1 2-1 distributed with mean ~ and variance Wi =E~~ + a (XiXi) Under the assumption (vi), the elements of the matrices t(x~x.)-1 are uniformly (over i),,

bounded. Therefore, the matrices (X~X.)-1,,, i=1,...,n, converge uniformly (over i) to zero as t tends to infinity. Also, note that where Since var(z.) = a2(x~x.)-1 Page 8 b., = ~., + Z., (3.5),,, infinity, the difference between b. and ~. i=1,...,n. converges to zero (uniformly in i) as t tends to,, - -,=, tends to zero in probability. -1 n Therefore, for n fixed and t tending to infinity, b = n E. 1b. and -1 n -0,., = n E., = 1~', are asymptotically equivalent. In fact, since var(z) and where J is a matrix with all elements equal to 1, we have (3.6) Hence b is also asymptotically (as t ~ m) equivalent to ~GLS and ~EGLS. (See also Hsiao (1975) for similar comments.) It is important to note here that the OLS estimator ~OLS' however, is not necessarily asymptotically equivalent to b. For example, suppose X~X.,, = itb where B is a fixed kxk positive definite matrix. satisfied. In this example, the OLS estimator ~OLS is and hence ~OLS is not asymptotically equivalent to b = Then the assumption (vi) is (E n.)-1 En. b i=1' i=1' i -1 n n E. 1b..,=,

Page 9 Remark 3.2: For a fixed t and n tending to infinity, the estimator b may not be asymptotically equivalent to ~GlS and hence may not be an efficient estimator. However, we know that the exact distribution of b is normal and hence the (exact) distribution of is chi-square with q degrees of freedom, where l is a qxk matrix of rank q. A*2 We now present the asymptotic distribution of the T statistic as n and/or t tends to infinity. Result 3.1: Consider the model given in (2.1) with the assumptions (i) through (vi). Consider the test statistic r*2 defined in (3.4) based on the estimator b. Then, (a) (b) for a fixed nand t tending to infinity: _1-1 A*2 (n-q)q (n-1) T is (asymptotically) distributed as F(q,n-q), A*2 for a fixed t and n tending to infinity: T is (asymptotically) distributed as chi-square with q degrees of freedom, and (c) A*2 for the case where nt is large and g=1: T distributed as F(1,v * ) where is approximately * 1-1 2 n 1 2 v = g- [i'i: i + n a E i'(x'.x.)- i] ~~ i=1', g and L = i'. Proof: See Appendix. ()

Page 10 With the exception of v *, the Satterthwaite's approximation for the degrees of freedom, the asymptotic distributions of T 2 and r*2 are identical. The A*2 2 advantage of T over T is that it is simple to compute and is simple to explain. 2 * Note that, as in the case of T, the degrees of freedom v to (n-1) as t tends to infinity and (b) tends to infinity as n tends to (a) tends infinity. Also, the degrees of freedom v* is always greater than or equal to (n-1) and hence the approximation in (c) serves as a compromise between the F and chi-square approximations. To summarize, we have seen that asymptotically (as t - ~), the estimators ~GLS' ~EGLS and b are equivalent and are efficient. Also, asymptotically (as n - ~), the estimators ~EGLS and ~GLS are equivalent and are efficient. However, for a fixed t and n large b may not be as efficient as ~GLS and hence the tests based on b may not be as powerful as the tests based on ~GLS' The distribution of b is exactly normal for all nand t, whereas the exact distribution of ~EGLS is unknown. A small Monte Carlo study was conducted to compare the performance of the test statistics based on b and ~EGLS' (In the study, the test statistics based on ~OLS were also included.) The results of the study are summarized in the next section. Consider the model 4. MONTE CARLO SIMULATION Yij = ~Oi + ~1iXij + e ij, i=1,...,n j=1,..,t where ~i = (~Oi' ~1i)' are NID(O,E~~); xij's are independent N(O,9) random variables if i is even and N(O,4) if i is odd; e.. 's are NID(O,4); (~.}, {x.. }, J,, J and {e..} are independent and 'J

Page 11 4 E~~ = [ 4 :] The values for nand t are taken to be 5, 10 and 50 to represent small, moderate and large samples. A set of 250 x. 0 values were generated once for 'J all and the same values of x. 0' i=1,...,n; j=1,...,t, were used in all of the 'J replications. For each pair of values of nand t, 100 Monte Carlo replications were used. In each replication, independent ~o 's and e.. 's were generated., 'J 2 A*2 2 Test statistics (T, T and T OLS ) based on ~EGLS' b and ~OLS for testing the hypotheses (i) H O : ~1 = 0, (ii) H O : ~O = ~1 = 0; (iii) H O : ~1 = and (iv) H O : ~O = ~1 = 1 were computed. The number of times the test statistics rejected the hypotheses are summarized in Tables 1 and 2. From the asymptotic results in section 3 we would expect that the EGLS estimator ~EGLS and the simple estimator b perform equally well when t is large. However, we do not expect the ordinary least squares estimator to do as well as b when t is large. For t=50 this expectation was borne out. At all values of n the probability of rejecting a true hypothesis (using the F- approximation) was 9% or less for all three statistics, but the power for rejecting either of the false hypotheses was always greater for ~EGLS and b than for ~OLS. Furthermore the rejection rates for ~EGLS and b were identical. A look at the true variances of ~GLS' b and ~OLS revealed that the relative efficiency of b was almost 100% for both the intercept and the slope parameters, whereas for ~OLS it was only 67% for the intercept parameter and 89% for the slope parameter in the case when n=5 and t=50. For smaller t, the efficiency of ~OLS was even worse. However, the efficiency of b was always close to 100%. Similar values for the relative efficiencies of the estimators were observed when n=5 and n=50.

Page 12 Table 1. Comparison of the Levels of Test Criteria: The Number of Times a 0.05 Level Test Criterion Rejects the Hypothesis (out of 100 replications). ( i ) H O : 13 1 = 0 ( i i ) H O : 13 0 = 13 1 = 0 2 Estimator F 2 n t 1, n-1 X 1 F1,v * F 2,n-2 X 2 5 5 EGLS 8 15 11 4 26 BBAR 9 16 10 10 32 OLS 6 14 0 19 10 EGLS 9 18 9 2 24 BBAR 9 18 9 11 30 OLS 8 17 1 14 50 EGLS 5 12 5 8 29 BBAR 5 12 5 9 29 OLS 5 12 5 18 10 5 EGLS 3 5 3 3 10 BBAR 3 5 3 2 11 OLS 2 5 2 6 10 EGLS 7 9 7 9 16 BBAR 7 10 7 10 16 OLS 6 8 3 6 50 EGLS 4 5 4 5 13 BBAR 4 5 4 5 13 OLS 3 3 2 5 50 5 EGLS 8 9 8 6 10 BBAR 7 8 7 8 9 OLS 7 7 8 9 10 EGLS 1 2 1 0 2 BBAR 1 2 1 1 2 OLS 2 3 1 1 50 EGLS 4 4 4 2 2 BBAR 4 4 4 2 2 OLS 4 4 1 4

Page 13 Table 2. Comparison of the Powers of Test Criteria: The Number of Times a 0.05 Level Test Criterion Rejects the Hypothesis (out of 100 repl ications). n Estimator F t 1 I n-1 ( ; ) H O : ~1 = 1 ( i i ) H O : ~O = ~1 = 2 2 X 1 F1 * F,V 2, n-2 X 2 5 5 EGLS 13 24 13 5 34 BBAR 13 25 15 10 43 OLS 13 23 3 22 10 EGLS 12 18 12 8 34 BBAR 12 18 13 16 41 OLS 12 23 4 27 50 EGLS 13 27 13 11 48 BBAR 13 27 13 11 49 OLS 11 26 7 31 10 5 EGLS 13 17 14 15 26 BBAR 12 16 13 17 28 OLS 10 15 5 16 10 EGLS 18 27 18 19 33 BBAR 19 27 19 19 34 OLS 16 23 6 21 50 EGLS 15 20 15 18 35 BBAR 15 20 15 18 35 OLS 14 20 9 24 50 5 EGLS 69 70 70 79 82 BBAR 67 69 67 80 84 OLS 51 54 52 57 10 EGLS 70 70 70 86 87 BBAR 70 71 70 85 88 OLS 56 60 51 53 50 EGLS 67 69 67 83 88 BBAR 67 69 67 83 88 OLS 56 58 72 76

Page 14 As n approaches infinity for fixed t we would expect ~EGLS to be more powerful than b. As it turned out, for n=50 the rejection rates for ~EGLS and b (using the x 2 approximation) were nearly indistinguishable. The rejection rate for ~OLS ranged from 14 to 39 percent lower than that of the other two estimators. For small sample sizes none of the estimators was very powerful. However, contrary to our expectation, the performance of ~EGLS was reasonable. The estimator b may have been more powerful than ~EGLS in rejecting H O : ~o = ~1 =1, but by the same token, b rejected the true hypotheses, H O : ~O = ~1 = 0, more often than ~EGLS' One problem that other authors (e.g., Jennrich & Schluchter (1986), Carter and Yang (1986)) have noted is that, with small sample sizes, E~~ is often not a positive definite matrix. In our simulation this occurred 34% of the time for n=t=5, but for moderate sample sizes, (n=t=10) this was no longer a problem. (If E~~ is not positive definite, the modified estimator suggested by Carter and Yang (1986) was used.), In our simulation even though the X. matrices were different for different, individuals, the weight matrices W. turned out to be close to n- 1 1. This may be one of the reasons why the tests based on b and ~EGLS had very similar power for all sample sizes. 5. SUMMARY In random coefficient regression models several estimators for ~ exist in the literature. Carter and Yang (1986) derived the asymptotic distribution of the estimated generalized least squares estimator as either n, the number of experimental units, tends to infinity and/or as t, the number of repeated measurements on each unit, tends to infinity. They proposed test statistics based on the EGLS estimator. - -1 n The simple average b = n E. 1b. of the,=,

Page 15 regression estimates from each unit has not received much attention in the literature. The main contribution of the paper is to show that inferences can be made, without much difficulty, using the simple estimator b. Asymptotic results for the estimator b, similar to those derived by Carter and Yang (1986) for ~EGLS' are derived. Also, the results of a small Monte Carlo study indicate that it is reasonable to use b for inferences on ~. It is important to emphasize the simplicity of the estimator b, the test statistics based on b and their asymptotic properties. The estimator ~EGLS is not as simple to compute. Also, the estimator E~~ that enters the computation of ~EGLS may need to be adjusted so that E~~ is positive A definite. We are, however, not suggesting that ~EGLS be ignored. The estimator ~EGLS may perform very well for several model matrices (especially when n is large). Our results extend to the case where unequal number (r.,, say) of measurements are made on different individuals. In this case, part (a) of Result 3.1 should be modified to say "for a fixed n and minimum (r.) tending to, infinity." Also, when minimum (r.) is large, the Result 3.1 (a) holds even if 2 o. = variance (e..) is not the same for different experimental units, 'J 2 (provided one uses s., regression mean square error for the regression of ith,, individual, to estimate o~). Wh~n n is large, Result 3.1 (b) holds even if o~ ~ 0 2 for all i, provided we assume that for all i, o~ ~ 0 2 for some finite of. Our results can also be extended to the case where the errors e.. are 'J correlated over time. For example, suppose for each i, {e.. : j=1,...,t} is a 'J stationary time series with variance covariance matrix of e., given by E ee. It -1 - is easy to see that n Sbb is still an unbiased estimator of var(b). Under,

Page 16 some regularity conditions (similar to those given in Section 9.1 of Fuller (1976» on X" L oo and L one can obtain the asymptotic results for the, ~~ ee test statistic based on band Sbb. the sake of brevity. The proofs, however, are not included for ACKNOWLEDGEMENTS The work of S. G. Pantula was partially supported by the National Science Foundation. REFERENCES Carter, R. L. and Yang, M. C. K. (1986). "Large Sample Inference in Random Coefficient Regression Models," Communications in Statistics - Theory and Methods, 15(8), 2507-2525. Fuller, W. A. (1976). Introduction to Statistical Time Series, New York: John Wiley and Sons. Harville, D. A. (1977). "Maximum Likelihood Approaches to Variance Component Estimation and to Related Problems," Journal of the American Statistical Association, 72, 320-340. Hsiao, C. (1975). "Some Estimation Methods for a Random Coefficient Model," Econometrica, 43, 305-325. Jennrich, R. I. and Schluchter, M. D. (1986). "Unbalanced Repeated-Measures Models with Structured Covariance Matrices," Biometrics, 42, 805-820. Laird, N. M. and Ware, J. H. (1982). "Random-Effects Models for Longitudinal Data," Biometrics, 38, 963-974. Rao, C. R. (1965). "The Theory of Least Squares When the Parameters are Stochastic and its Applications to the Analysis of Growth Curves," Biometrika, 52, 447-458. Swamy, P. A. V. B. (1971). Statistical Inference in Random Coefficient Regression Models. Berlin: Springer-Verlag.

Page 17 APPENDIX In the appendix, we outline the proof of Result 3.1. (a) n fixed and t tends to infinity: From Remark 3.1, we know that (A.1 ) and hence the statistic r*2 = n(m~ - ~)LI[LSbbL,]-1L(m~ -~) + 0p(t- 1 / 2 ). A1so, reca 11, -1 n Sbb = (n-1) E (b. -b)(b.- b) I i =1 1 1-1 n = (n-1 ) E [~. + Z. - II - Z][~. + Z. -II-Z]' i=1 1 1 ~ 1 1 ~ = S/3~ + SZZ + S/3Z + S' /3Z (A. 2) where -1 n = (n-1) E (c. - c)(d. - d)', i =1 1 1 and Z. is defined in Remark 3.1. Since~. and Z. are independent normal random 111 variables with means ~ E[S/3Z] = 0 and 0 respectively, and the variance of the (~,m)th element of S~Z is -1 n var[(n-1) E (/3. - m )Z. ] i=1 1,~ /3,~ 1,m -2 n = (n-1) E i=1

Page 18 Therefore, S~z (A. 3) Now, Szz = (n_1)-1[~ Z.Z.'- nil'] = i =1 ', since from Remark 1 we know that Z., = 0 (t- 1 / 2 ) and Z p Therefore, from (A.2) and (A.3), we have = 5 + 0 (t- 1 / 2 ) (A.4) ~~ P Combining (A.1) and (A.4), we get under H O : L~ = ~o' A*2 2 T - T m = 0 (t-1/2) p where -1 n(..~- ~)' L' [L S~~L'] L(m~ -~). Now, the result (a) follows because T 2 has the Hotelling's T 2 distribution m with (n-1) degrees of freedom. (b) t fixed and n tends to infinity: From Remark 3.2, we know that the exact distribution of T*2 is chisquare with q degrees of freedom. *2 A*2 The difference between T and T is that the matrix n var(b) is replaced by it's unbiased estimator Sbb. If we can show that Sbb is consistent (as n ~ m), then the result (b) will follow from Slutsky's Theorem.

Page 19 From (A.2) and (A.3) we have, Now _1 n = (n-1) [E i=1 Z.Z ~ - nzz '] 1 1 _1 n -1-1 = n E Z.Z~ + n (n-1) i =1 1 1 n E i=1 Z.Z~ 1 1 - (n-1)-1 n ii' _1 n 1 1 = n E Z.Z~ + 0 (n- t- ) i=1 ' 1 P (A.5 ) (A. 6) Now, since ~i's are iid N(O, E~~) variables, we have,,, Also, since Z.'s are independent N(O, a2(x~x.)-1) _1 n E[n E Z.Z.'] i=1 ', n = n- 1 0 2 E (X '. X. )- 1,=. 1 ' 1 variables, we have (A.7) (A.S) and _1 n Var[n ~' E Z.Z~~] i =1 " (A.9) for any arbitrary vector~. Therefore, -1 2 n -1-1/2 Sbb =E + n 0 E (X~X.) + 0 (n ) ~/3 i =1 " P = n var(b) + 0 (n- 1 / 2 ) p

Page 20 and the result (b) follows. (c) nt large and g=1: Consider the t-statistic for testing the hypothesis H O : l'~ = A O ' We know that the variable has a standard normal distribution. ~* * T = T [l'n var(b)l] 1/2(l'S lf 1 / 2 bb To show that is (approximately) distributed as Student's t-distribution with v* degrees of * --1 freedom, we need to show that v [l' n var(b) l] l'sbbl is (approximately) a chi-square random variable with v* when nt is large) independent of l'b. degrees of freedom and is (asymptotically, From (A.6), (A.S) and (A.9) we have n = 5 + 0 2 n- 1 E (X~X.)-1 + ~~ i=1 1 1 ~2-1 n 1 = S~~ + 0 n E (X~X.)-. 1 1 1 1= where ~2 is defined in Section 2. Note that (n-1)(l's~~l)(lie~~l)-1 is a x 2 (n-1) random variable and (nt-nk)~2/02 is a x 2 (nt-nk) random variable. Therefore, Sbb is the sum of independent scalar multiples of chi-square random variables. Ignoring the terms of order (nt)-1/2 and using Satterthwaite's * --1 approximation, we have that v [l'n var(b)l] l'sbbl is approximately distributed as chi-square with v* degrees of freedom.

* Now, to show the (asymptotic) independence of T and Sbb' note that b =.~ + Z is independent of S~~ since ~i's are NID(~,E~~) and are independent of {Z.}. Also, for each i, the least squares estimator b. is 1 1 independent of the residual sums of squares y~y. - b~x~y. and hence b 1 1 1 1 1-2 and a are independent. Therefore, for nt large, the distribution of T can be approximated by Student's t-distribution with v* Page 21 -* degrees of freedom.