ON THE CONSEQUENCES OF MISSPECIFING ASSUMPTIONS CONCERNING RESIDUALS DISTRIBUTION IN A REPEATED MEASURES AND NONLINEAR MIXED MODELLING CONTEXT

Similar documents
Supporting Information for Estimating restricted mean. treatment effects with stacked survival models

Confidence Estimation Methods for Neural Networks: A Practical Comparison

A Non-parametric bootstrap for multilevel models

Estimation of Parameters

A Simulation Study on Confidence Interval Procedures of Some Mean Cumulative Function Estimators

Some Theoretical Properties and Parameter Estimation for the Two-Sided Length Biased Inverse Gaussian Distribution

Supplement to A Hierarchical Approach for Fitting Curves to Response Time Measurements

Computationally Efficient Estimation of Multilevel High-Dimensional Latent Variable Models

Impact of serial correlation structures on random effect misspecification with the linear mixed model.

Research Article A Nonparametric Two-Sample Wald Test of Equality of Variances

Plausible Values for Latent Variables Using Mplus

The Nonparametric Bootstrap

11. Bootstrap Methods

A nonparametric two-sample wald test of equality of variances

Inference on reliability in two-parameter exponential stress strength model

Chapter 1 Statistical Inference

Formal Statement of Simple Linear Regression Model

Bayesian spatial quantile regression

Estimation of AUC from 0 to Infinity in Serial Sacrifice Designs

COMPARISON OF THE ESTIMATORS OF THE LOCATION AND SCALE PARAMETERS UNDER THE MIXTURE AND OUTLIER MODELS VIA SIMULATION

INVERTED KUMARASWAMY DISTRIBUTION: PROPERTIES AND ESTIMATION

Categorical Data Analysis Chapter 3

Estimation in an Exponentiated Half Logistic Distribution under Progressively Type-II Censoring

Spatially Smoothed Kernel Density Estimation via Generalized Empirical Likelihood

Glossary. The ISI glossary of statistical terms provides definitions in a number of different languages:

Fractional Imputation in Survey Sampling: A Comparative Review

Nonlinear mixed-effects models using Stata

Confidence Intervals for the Process Capability Index C p Based on Confidence Intervals for Variance under Non-Normality

The formal relationship between analytic and bootstrap approaches to parametric inference

Akaike Information Criterion to Select the Parametric Detection Function for Kernel Estimator Using Line Transect Data

Statistics - Lecture One. Outline. Charlotte Wickham 1. Basic ideas about estimation

When is a nonlinear mixed-effects model identifiable?

STAT 6350 Analysis of Lifetime Data. Probability Plotting

Application of Variance Homogeneity Tests Under Violation of Normality Assumption

Statistics Boot Camp. Dr. Stephanie Lane Institute for Defense Analyses DATAWorks 2018

On Modifications to Linking Variance Estimators in the Fay-Herriot Model that Induce Robustness

Covariance function estimation in Gaussian process regression

Best Linear Unbiased Prediction: an Illustration Based on, but Not Limited to, Shelf Life Estimation

Single Index Quantile Regression for Heteroscedastic Data

Simulating Uniform- and Triangular- Based Double Power Method Distributions

Propensity Score Weighting with Multilevel Data

Generalized Linear Models (GLZ)

Reliability of Acceptance Criteria in Nonlinear Response History Analysis of Tall Buildings

Estimators as Random Variables

ON THE FAILURE RATE ESTIMATION OF THE INVERSE GAUSSIAN DISTRIBUTION

Semiparametric Mixed Effects Models with Flexible Random Effects Distribution

Psychology 282 Lecture #4 Outline Inferences in SLR

A Note on Bayesian Inference After Multiple Imputation

REGRESSION WITH SPATIALLY MISALIGNED DATA. Lisa Madsen Oregon State University David Ruppert Cornell University

Terminology Suppose we have N observations {x(n)} N 1. Estimators as Random Variables. {x(n)} N 1

Measurement error as missing data: the case of epidemiologic assays. Roderick J. Little

Machine Learning CSE546 Carlos Guestrin University of Washington. September 30, What about continuous variables?

The Bootstrap: Theory and Applications. Biing-Shen Kuo National Chengchi University

Bayesian Analysis of Latent Variable Models using Mplus

Machine Learning CSE546 Sham Kakade University of Washington. Oct 4, What about continuous variables?

4. STATISTICAL SAMPLING DESIGNS FOR ISM

Gravity Models, PPML Estimation and the Bias of the Robust Standard Errors

The Problem of Modeling Rare Events in ML-based Logistic Regression s Assessing Potential Remedies via MC Simulations

Statistical Practice

A Monte-Carlo study of asymptotically robust tests for correlation coefficients

On the Application of the Generalized Pareto Distribution for Statistical Extrapolation in the Assessment of Dynamic Stability in Irregular Waves

A NOTE ON ROBUST ESTIMATION IN LOGISTIC REGRESSION MODEL

Gradient types. Gradient Analysis. Gradient Gradient. Community Community. Gradients and landscape. Species responses

Intraclass Correlations in One-Factor Studies

Nonresponse weighting adjustment using estimated response probability

Integrated likelihoods in survival models for highlystratified

Default Priors and Effcient Posterior Computation in Bayesian

401 Review. 6. Power analysis for one/two-sample hypothesis tests and for correlation analysis.

Characterizing Forecast Uncertainty Prediction Intervals. The estimated AR (and VAR) models generate point forecasts of y t+s, y ˆ

Approaches to Modeling Menstrual Cycle Function

Advanced Statistical Methods. Lecture 6

Measurement And Uncertainty

The Bayesian Approach to Multi-equation Econometric Model Estimation

Math 423/533: The Main Theoretical Topics

ON THE PROBLEM OF COMPARING THE MEANS AND MEDIANS OF TWO INDEPENDENT LOGNORMAL DISTRIBUTIONS

Robust Backtesting Tests for Value-at-Risk Models

Estimation of Quantiles

Mean. Pranab K. Mitra and Bimal K. Sinha. Department of Mathematics and Statistics, University Of Maryland, Baltimore County

Econ 582 Nonparametric Regression

Spatial and temporal extremes of wildfire sizes in Portugal ( )

Machine Learning CSE546 Carlos Guestrin University of Washington. September 30, 2013

Confidence Intervals in Ridge Regression using Jackknife and Bootstrap Methods

Overview of Extreme Value Theory. Dr. Sawsan Hilal space

BIAS OF MAXIMUM-LIKELIHOOD ESTIMATES IN LOGISTIC AND COX REGRESSION MODELS: A COMPARATIVE SIMULATION STUDY

Effects of Outliers and Multicollinearity on Some Estimators of Linear Regression Model

Inferring from data. Theory of estimators

Bootstrap and Parametric Inference: Successes and Challenges

Least Absolute Value vs. Least Squares Estimation and Inference Procedures in Regression Models with Asymmetric Error Distributions

Finite Population Sampling and Inference

COMPARISON OF GMM WITH SECOND-ORDER LEAST SQUARES ESTIMATION IN NONLINEAR MODELS. Abstract

Continuous Time Survival in Latent Variable Models

Generalized Linear. Mixed Models. Methods and Applications. Modern Concepts, Walter W. Stroup. Texts in Statistical Science.

Point and Interval Estimation for Gaussian Distribution, Based on Progressively Type-II Censored Samples

Statistical Data Analysis Stat 3: p-values, parameter estimation

Diagnostics and Remedial Measures

Bayesian inference for sample surveys. Roderick Little Module 2: Bayesian models for simple random samples

One-way ANOVA Model Assumptions

Introduction to Rare Event Simulation

1.4 Properties of the autocovariance for stationary time-series

interval forecasting

Transcription:

ON THE CONSEQUENCES OF MISSPECIFING ASSUMPTIONS CONCERNING RESIDUALS DISTRIBUTION IN A REPEATED MEASURES AND NONLINEAR MIXED MODELLING CONTEXT Rachid el Halimi and Jordi Ocaña Departament d Estadística Universitat de Barcelona Abstract In this paper we describe the results of a simulation study performed to elucidate the robustness of the Lindstrom and Bates (1990) approximation method under non-normality of the residuals, under different situations. Concerning the fixed effects, the observed coverage probabilities and the true bias and mean square error values, show that some aspects of this inferential approach are not completely reliable. When the true distribution of the residuals is asymmetrical, the true coverage is markedly lower than the nominal one. The best results are obtained for the skew normal distribution, and not for the normal distribution. On the other hand, the results are partially reversed concerning the random effects. Soybean genotypes data are used to illustrate the methods and to motivate the simulation scenarios. 1. Motivation and Introduction The nonlinear mixed effects model is used to represent data in pharmacokinetics (Davidian and Giltinan, 1995), breast cancer dynamics (El Halimi et al., 2003), growth curves in Soybean genotypes data (Pinheiro and Bates, 2000) and other areas, where the within-individual model is a function of individual-scientific, scientifically meaningful parameters. It is well known that maximum likelihood estimation for nonlinear mixed effects models leads to a cumbersome integration problem, because random parameters appear inside the nonlinear expectation 1

function. To avoid this problem, several approximation have been proposed. The parametric approach to non-linear mixed-effects modeling using the LB-method (Lindstrom and Bates, 1990) is, essentially, based on the standard assumption of normality of the errors and random effects. But these assumptions may not always be realistic or, in any case, difficult to verify as they are not directly observed. In this paper we investigate the impact of the non-normality conditions on estimating fixed and random components parameters, via a Monte-Carlo simulation study by considering the Soybean genotypes model reported in Davidian and Giltinan (1995) and analyzed by Pinheiro and Bates (2000). Typical profiles are displayed in Figure 1, where the response of leaf weight are plotted by subject. The goal of the study was to compare the growth patterns of two soybean genotypes, a commercial variety, Forrest (F) and an experimental strain, Plan Introduction #416937 (P). Data were collected during three years, from 1988 to 1990. At the beginning of the growing season in each year, 16 plots were planted with seeds; 8 plots with each genotype. Each plot was sampled eight to ten times at approximately weekly intervals. At each sampling time, six plants were randomly selected from each plot, leaves from this plant were weighted, and the average leaf weight per plant (in g) was calculated for each plot. Different plots in different sites were used in different years. The logistic model derived from Pinheiro and Bates (2000) is an appropriate characterization of leaf weight response, where the parameter may vary across subjects. δ yij 1+ exp ( δ t )/ δ 2i ij 3i δ = δ + η 1i 1 1i δ = δ + η. 2i 2 2i δ = δ + η 3i 3 3i 1i = + eij (1.1) where y represents the average leaf weight/plant in subject i, i = 1,, 48, at time t. The ij ij random effects ' η = ( η, η, η ) are (0, D) and the eij are (0, σ 2 ) and are independent of the i 1i 2i 3 η i. The association of the fixed effects δ with the random effects vector is represented by the 2

linear function above, where the subject-specific parameters δ are independent across i. But, as some of the analyses presented in El Halimi et al. (2003) and the profile of qq-norm of residuals (under homogeneity assumption) displayed in Figure 2 suggest, this assumption may not always be realistic. These violations of the assumptions of the model pose questions on the validity of the inferences made during the modeling process. As a first approach to answering these questions, we performed a simulation study emulating the conditions of the soybean genotypes studies described above. i 2. Simulation study on the distributional assumptions We carried out several simulation studies in which data were generated according to the soybean genotypes model given in equation (1.1), with known population or true parameter values. For fixed effects, these values were taken as δ = (19.26, 55, 8.4). For random effects, the covariance matrix was D 25 2.50 4.00 = 8.00 2.32. 2.00 The random effects were generated according to an expression equivalent to η = Lh, where h i stands for a standardised version of the vector of random effects (common for all i), generated from a normal distribution with zero mean and unit variance, and L stands for the lower triangular matrix resulting from the Cholesky decomposition of a covariance matrix D. These values were chosen near to the estimated values given by the splus 2000 implementation of LBmethod (nlme function) and according to the maximum likelihood variant of the estimation procedure. The residuals or errors were generated in similar way, first as i.i.d standardised values and subsequently converted to values with standard deviation σ (in this particular cases σ=1) and according to the following marginal distributions: N- Normal distribution, which represent the case where the usual assumption of normality on the errors is valid. 3

SN- Skew normal distribution with location parameter (a typical or central value that best describes the data, its effect is to translate the graph, relative to the standard normal distribution) ζ = -0.05, scale parameter (its effect is to stretch out the graph) ψ = 0.2 and shape parameter (its effect is to allow the distribution to take a variety of shapes, depending on its value) ϕ = 0.02, as described in Azzalini (1986). NPM- Non-parametric gaussian kernel estimated from experimental data with the optimal bandwidth implemented in S-plus for the gaussian kernel by Venables (1997) (width.sj function). E- Exponential distribution. G- Gamma distribution. The last two distributions (G and E) were used to represent a situation where the true distribution of errors is not symmetrically distributed and has heavier tails than expected from a normal distribution. For each possible distribution of the residuals, series of 1000 simulated data sets were generated and processed to fit a nonlinear mixed model like (1.1), via the LB-method, maintaining the same starting values for ML estimation procedure. The resulting set of parameter estimates for each series was used to compute the summaries used to evaluate the performance of the inferential methods, like true coverage of the confidence intervals or the true estimator biases. 3. Results Figures 3 to 5 contain box-plots for the mixed parameter estimates (fixed and random components) from all simulation runs. For each box-plot, the character in the abscises axis represents the simulated distribution residuals, while the simulated distribution of random effects was maintained normal. The horizontal continuous straight line represents the true value of the corresponding parameter. Thus, in this way, the box-plots give a graphical idea of the bias, standard deviation and dispersion of the mixed effects estimators, under different possible distributions. 4

Table 1 contains simulation summary statistics for the fixed-effects parameters estimates. In this table, if δ ˆ stands for a fixed effect or parameter estimate at the k-th simulation replication and k δt for the true value of the parameter (value used to generate the data), the summary measures for the estimators of the fixed effects are defined as: MEAN: denotes the average of each series of 1000 estimates, 1 1000 δˆ k= 1 k, 1000 BIAS: corresponds to the simulation estimate of bias, that is, MEAN δ T, the bias of a statistic indicates, on average, how much the estimator will over- or underestimate the true parameter value. C.Ι Bias : denotes approximate confidence intervals for bias, BIAS ± Ζ α 2 1000 normal critical value at a 95% confidence level and S = δˆ MEAN / 999 ˆ. k MSE: is the mean squared error, into account both, bias and standard error. δ k= 1 ( S 2 δ ˆ, where Ζ α is the 1000 δˆ δ /1000, is a measure of its accuracy that takes 1000 ( ) 2 k = 1 k T COVERAGE-PROBABILITY: denotes the observed coverage of t-based 95% confidence intervals computed using the model-based standard errors and t-distribution critical values based on 362 degrees of freedom ( the intervals provided by the nlme function), PROB. LOW (PROB. UPP): denotes the proportions of δ T lower (upper) than the lower (upper) bound of the confidence intervals (that is, they correspond to non-coverage probabilities). Finally, Average width: denotes the arithmetic average of the 1000 observed lengths of the asymptotic intervals described above. ) 2 For the δ 1 parameter, from Figure 3 and Table 1 we infer that virtually there are not differences between the results for the five distributions under consideration, with respect to coverage, bias and MSE. For this last measure, the results are similar with values ranging between 0.54 and 0.67. None of the confidence interval coverages attain the nominal 95% value, but there is acceptable robustness, with coverages ranging from 94% under the sn residuals distribution to 5

91.5% under the exponential distribution. The NN case has a 92.1% coverage. The last column of Table 1 also shows very homogeneous results with respect to the width of the intervals. Additionally, the intervals are appreciably equitailed. For the δ 2 parameter, the results shown in Table 1 demonstrate a negative bias indicating that, on average, the true parameter value is underestimated. The bias and MSE values are nearly identical for this parameter, except for the sn distribution, in which these measures are considerably smaller. The confidence intervals perform poorly in all cases, except for the sn residuals distribution. The coverage ranges from 79.9% for the exponential distribution (and 84.6% for the gamma distribution) to 93.5% for the sn distribution, with a poor 83.2% value for normal residuals. In correspondence with the coverages, the widths of the intervals show a similar tendency, with the best (shorter) value obtained under the sn distribution. The intervals have appreciably unequal tails, except for the sn case. For the δ parameter, we find that under model (1.1) and using the LB-method, the true 3 parameter value is also systematically underestimated. We remark that the sn residuals distribution gives also the best results, not only from the point of view of the MSE values (0.06), but also from the point of view of coverage probability (90.7%) and equitaildness, but lower than the nominal 95% level. In all cases, the coverage probabilities are poor, ranging between 78.8% (exponential) to 90.7% for (sn), with 80% in the normal case, indicating a low robustness of the LB-method for this parameter. Again, all intervals have similar mean lengths and are not equitailed, except for the sn distribution, which is also associated to the best values. These simulations confirm that the LB-method is not very efficient, even under normal conditions. But, in contrast with other simulation studies, like El Halimi et al. (2004) which is inspired in breast cancer data, a surprising result is that the best results for fixed effects are obtained when the errors are skew normal (sn). On the other hand, this result is not maintained for random components (D and σ). Concretely, the results for these model components are 6

displayed in Figures 4 to 5. Estimation under the nonlinear Soybean genotypes model tells a vastly different story. Here, all simulation conditions underestimate the random components of D, except the estimates under the sn distribution, which provide reasonable unbiased estimates, but perform considerably worse in terms of coverage probabilities, especially for the D 33 parameter (24.5% of coverage probability). In almost all cases the coverage probabilities are lower than their nominal value for D 12, D 13 and D 33 and exceed their nominal value otherwise. Figure 5 displays the results for the parameter σ. Its estimation becomes less precise for the NPM and sn distributions, while the different measures of dispersion are in close agreement for E and G, and the magnitude of the variability is different otherwise. Regarding the coverage probabilities, results ranging from 99.3% for N to 0% for sn are observed. 4. Conclusions and discussion Under the model (1.1) and according to the observed coverages for confidence intervals with a 95% nominal coverage, the results indicate that the LB-method is not robust when the normality assumptions of errors are violated, and even under normal conditions. In fact, the ML procedures seem not adequate even under normal conditions and for a large number of observations per subject and a large number of subjects. The same results are also observed concerning the precision (length) of the intervals, and the bias and MSE of the point estimates. The performance of all these inferential methods is highly variable and dependent on the concrete simulated distributions and the concrete model parameters, in a complex and difficult to forecast way. Acknowledgements The research was supported by Instituto de Salud Carlos III FIS, grant 00/1130; and by Generalitat de Catalunya, grant 2001/SGR/00067. 7

References Azzalini, A., 1986. Further results on a class of distributions which includes the normal ones. Statistica, 46, 199-208. Davidian, M.; Giltinan, D.M. Nonlinear Models for Repeated Measurement Data; Chapman & Hall: London, 1995. El Halimi, R., Ocaña, J, Ruiz De Villa, M.C., Escrich, E., Solanas, M. Modeling Tumor Growth Data Using a Nonlinear Mixed-Effects Model. InterStat 2003, http://jscs.stat.vt.edu/interstat/articles/2003/abstracts/0309002.html-ssi El Halimi, R., Ocaña, J., Ruiz de Villa, M.C. A simulation study on the robustness of parametric inference in a nonlinear mixed modelling context. www.mathpreprints.com, 2004. Lindstrom, M.J.; Bates, D.M. Nonlinear Mixed Effects Models for Repeated Measures Data. Biometrics 1990, 46, 673-687. Pinheiro, J.C.; Bates, D.M. Mixed-Effects Models in S and S-Plus; Springer- Verlag: Berlin, 2000. Venables, et al. Modern Applied Statistics with S-plus, New York: Springer- Verlag, 1997. 8

Table 1: summary results for fixed effect parameters estimates for 1000 simulations Fixed effect parameter : Maximum likelihood (ML) δ 1 RANDOM EFFECTS ERRORS MEAN Bias±C.I Bias MSE COVERAGE- PROBABILITY (ASYMP. INTERVALS) PROB PROB. PROB. LOW UPP Average width N N 18.99208-0.2679243±0.04776559 0.665096 1% 92.1% 6.9% 2.876276 N sn 19.15177-0.1082278±0.04559323 0.552286 1.9% 94% 4.1% 2.798492 N NPM 19.04983-0.2101692±0.04359019 0.5382893 1.2% 93.8% 5% 2.87612 N E 18.95573-0.3042746±0.04281623 0.6537758 0.9% 91.5% 7.6% 2.897845 N G 18.96339-0.2966082±0.04650596 0.6504091 0.5% 91.6% 7.9% 2.899373 Fixed effect parameter : Maximum likelihood (ML) δ 2 DISTRIBUTIONS N N 54.52249-0.4775082±0.03030467 0.4668351 0.1% 83.2% 16.7% 1.95098 N sn 54.96844-0.03156185±0.02606712 0.1776974 2.7% 93.5% 3.8% 1.591223 N NPM 54.56497-0.4350315±0.03086829 0.4370394 0.1% 83.7% 16.2.9% 1.927866 N E 54.50822-0.491783±0.03020465 0.5211324 0.3% 79.9% 19.8% 1.974908 N G 54.53612-0.4638765±0.0309442 0.4641885 0.2% 84.6% 15.2% 1.97972 Fixed effect parameter : Maximum likelihood (ML) δ 3 N N 8.100563-0.2994368±0.01805612 0.1744442 0% 80% 20% 1.117007 N sn 8.290249-0.1097514±0.0133505 0.05839522 0.6% 90.7% 8.7% 0.8053575 N NPM 8.166408-0.2335917±0.0178202 0.1371466 0% 84.5% 15.5% 1.124151 N E 8.079644-0.3203559±0.01745259 0.1958706 0.1% 78.8% 21.1% 1.140723 N G 8.100581-0.2994189±0.01805035 0.1743792 0.1% 80.3% 19.6% 1.137076 Soybean genotypes model: y ij δ 1 + exp ( δ t )/ δ 2i ij 3i = 1i + eij 9

1990F1 1990F8 1990F7 1990F6 1990P8 1990P7 1990P3 1990P1 1990P6 1990P5 1990P2 1990P4 30 25 20 15 10 5 Leaf weight/plant (g) 30 25 20 15 10 5 0 0 1989P7 1989P4 1989P6 1989P5 1989P1 1989P3 1989P2 1989P8 1990F2 1990F3 1990F4 1990F5 1988P7 1988P3 1988P2 1988P6 1989F6 1989F5 1989F4 1989F1 1989F2 1989F7 1989F8 1989F3 30 25 20 15 10 5 30 1988F4 1988F2 1988F1 1988F7 1988F5 1988F8 1988F6 1988F3 1988P1 1988P5 1988P4 1988P8 0 25 20 15 10 5 0 Time since planting (days) Figure1: Growth curves in Soybean genotypes data 10

residulas -6-4 -2 0 2 4 delta_1-10 -5 0 5 10 delta_2-5 0 5 delta_3-2 -1 0 1 2-3 -2-1 0 1 2 3-2 -1 0 1 2-2 -1 0 1 2-2 -1 0 1 2 Quantiles of Standard Normal Quantiles of Standard Normal Quantiles of Standard Normal Quantiles of Standard Normal Figure 2: qq-norm of residuals and random effects for Soybean data 11

ML-method delta_1 16 17 18 19 20 21 ML-method delta_2 53 54 55 56 Distribution ML-method delta_3 7.0 7.5 8.0 8.5 9.0 Figure 3: Box- plot of fixed effects of simulation results for soybean genotypes model 12

ML-method ML-method ML-method D11 10 20 30 40 50 98.6% 98.9% 99% 99.3% 99.1% D12-10 0 5 10 81.1% 61.4% 79.8% 80.6% 81.3% Distribution D13-2 2 4 6 8 86.8% 48.2% 82.7% 85.2% 85.8% ML-method ML-method D22 0 5 10 15 99.9% 98.6% 999% 99.7% 99.7% D23 0 2 4 6 98.4% 92.1% 95.8% 99.3% 98.9% ML-method D33 0 1 2 3 4 5 80% 24.5% 69.2% 70.8%71.5% Figure 4: Box-plots for variance-covariance parameters of Soybean genotypes model 13

ML-method sigma 0.2 0.4 0.6 0.8 1.0 1.2 9 9. 3 % 0 % 8 4. 5 % 88.3% 8 7 4 % Figure 5: Box-plots for sigma parameter of genotype model 14