Hypothesis Testing for Var-Cov Components

Similar documents
Research Design: Topic 18 Hierarchical Linear Modeling (Measures within Persons) 2010 R.C. Gardner, Ph.d.

Model Estimation Example

y response variable x 1, x 2,, x k -- a set of explanatory variables

SuperMix2 features not available in HLM 7 Contents

Computationally Efficient Estimation of Multilevel High-Dimensional Latent Variable Models

AGEC 621 Lecture 16 David Bessler

Institute of Actuaries of India

(a) (3 points) Construct a 95% confidence interval for β 2 in Equation 1.

CHAPTER 9 EXAMPLES: MULTILEVEL MODELING WITH COMPLEX SURVEY DATA

Principal Components. Summary. Sample StatFolio: pca.sgp

CENTERING IN MULTILEVEL MODELS. Consider the situation in which we have m groups of individuals, where

Multiple group models for ordinal variables

STRUCTURAL EQUATION MODELING. Khaled Bedair Statistics Department Virginia Tech LISA, Summer 2013

Package SEMModComp. R topics documented: February 19, Type Package Title Model Comparisons for SEM Version 1.0 Date Author Roy Levy

The Chi-Square Distributions

y ˆ i = ˆ " T u i ( i th fitted value or i th fit)

The Chi-Square Distributions

Multiple Linear Regression

Homoskedasticity. Var (u X) = σ 2. (23)

The Finite Sample Properties of the Least Squares Estimator / Basic Hypothesis Testing

Multilevel Models in Matrix Form. Lecture 7 July 27, 2011 Advanced Multivariate Statistical Methods ICPSR Summer Session #2

Unit 10: Simple Linear Regression and Correlation

Linear Regression. In this problem sheet, we consider the problem of linear regression with p predictors and one intercept,

1. How will an increase in the sample size affect the width of the confidence interval?

Multilevel Modeling: When and Why 1. 1 Why multilevel data need multilevel models

Stats Review Chapter 14. Mary Stangler Center for Academic Success Revised 8/16

STAT 501 EXAM I NAME Spring 1999

2.1 Linear regression with matrices

Subject CS1 Actuarial Statistics 1 Core Principles

Box-Cox Transformations

Newsom Psy 510/610 Multilevel Regression, Spring

Inferences for Regression

Partitioning variation in multilevel models.

Heteroskedasticity. Part VII. Heteroskedasticity

Ronald Heck Week 14 1 EDEP 768E: Seminar in Categorical Data Modeling (F2012) Nov. 17, 2012

Logistic Regression Analysis

ROBUSTNESS OF MULTILEVEL PARAMETER ESTIMATES AGAINST SMALL SAMPLE SIZES

LAB 5 INSTRUCTIONS LINEAR REGRESSION AND CORRELATION

Chapter 15 Panel Data Models. Pooling Time-Series and Cross-Section Data

Testing Indirect Effects for Lower Level Mediation Models in SAS PROC MIXED

Additional Notes: Investigating a Random Slope. When we have fixed level-1 predictors at level 2 we show them like this:

36-309/749 Experimental Design for Behavioral and Social Sciences. Dec 1, 2015 Lecture 11: Mixed Models (HLMs)

Ratio of Polynomials Fit One Variable

Specifying Latent Curve and Other Growth Models Using Mplus. (Revised )

Hypothesis Tests and Estimation for Population Variances. Copyright 2014 Pearson Education, Inc.

Confidence Intervals, Testing and ANOVA Summary

Designing Multilevel Models Using SPSS 11.5 Mixed Model. John Painter, Ph.D.

" M A #M B. Standard deviation of the population (Greek lowercase letter sigma) σ 2

Longitudinal Modeling with Logistic Regression

Concordia University (5+5)Q 1.

Using Tables and Graphing Calculators in Math 11

THE PRINCIPLES AND PRACTICE OF STATISTICS IN BIOLOGICAL RESEARCH. Robert R. SOKAL and F. James ROHLF. State University of New York at Stony Brook

Introduction to Structural Equation Modeling

STK4900/ Lecture 3. Program

Comparing two independent samples

Psychology 282 Lecture #4 Outline Inferences in SLR

Simple Linear Regression

Class Notes: Week 8. Probit versus Logit Link Functions and Count Data

LINEAR REGRESSION ANALYSIS. MODULE XVI Lecture Exercises

Assessing the relation between language comprehension and performance in general chemistry. Appendices

Assumptions, Diagnostics, and Inferences for the Simple Linear Regression Model with Normal Residuals

Econometrics. 4) Statistical inference

2008 Winton. Statistical Testing of RNGs

Determining Sample Sizes for Surveys with Data Analyzed by Hierarchical Linear Models

Investigating Models with Two or Three Categories

Testing Linear Restrictions: cont.

The goodness-of-fit test Having discussed how to make comparisons between two proportions, we now consider comparisons of multiple proportions.

Using SPSS for One Way Analysis of Variance

Binary choice 3.3 Maximum likelihood estimation

Problem Set #6: OLS. Economics 835: Econometrics. Fall 2012

401 Review. 6. Power analysis for one/two-sample hypothesis tests and for correlation analysis.

Technical Appendix C: Methods

Tutorial 2: Power and Sample Size for the Paired Sample t-test

The linear mixed model: introduction and the basic model

PRINCIPAL COMPONENTS ANALYSIS

Stat 5102 Final Exam May 14, 2015

MULTILEVEL MODELS. Multilevel-analysis in SPSS - step by step

Hypothesis testing:power, test statistic CMS:

A Non-Parametric Approach of Heteroskedasticity Robust Estimation of Vector-Autoregressive (VAR) Models

* Tuesday 17 January :30-16:30 (2 hours) Recored on ESSE3 General introduction to the course.

TESTING FOR NORMALITY IN THE LINEAR REGRESSION MODEL: AN EMPIRICAL LIKELIHOOD RATIO TEST

Notation Survey. Paul E. Johnson 1 2 MLM 1 / Department of Political Science

MS&E 226: Small Data

LAB 3 INSTRUCTIONS SIMPLE LINEAR REGRESSION

The use of structural equation modeling to examine consumption of pornographic materials in Chinese adolescents in Hong Kong

Subject-specific observed profiles of log(fev1) vs age First 50 subjects in Six Cities Study

Lecture 8. Using the CLR Model

School of Mathematical Sciences. Question 1

Large Sample Properties of Estimators in the Classical Linear Regression Model

Multivariate Time Series: Part 4

Multivariate Capability Analysis Using Statgraphics. Presented by Dr. Neil W. Polhemus

Lecture 6 Multiple Linear Regression, cont.

Review of Statistics

Correlation and Regression

Economics 471: Econometrics Department of Economics, Finance and Legal Studies University of Alabama

STA 101 Final Review

[y i α βx i ] 2 (2) Q = i=1

Exploiting TIMSS and PIRLS combined data: multivariate multilevel modelling of student achievement

Technical Appendix C: Methods. Multilevel Regression Models

STATISTICS 141 Final Review

Transcription:

Hypothesis Testing for Var-Cov Components When the specification of coefficients as fixed, random or non-randomly varying is considered, a null hypothesis of the form is considered, where Additional output files are obtained by selecting the Output Settings option from the Other Settings menu to bring up the Output Settings HLM3 dialog box. On this dialog box, the Print variance-covariance matrices option should be checked. With the above selection, an additional three variance-covariance files will be generated for a HLM3 model along with the output file, these being gamvc.dat, gamvcr.dat and tauvc.dat. Tauvc.dat contains the tau matrices and the inverse of the associated information matrix. For a model with r random effects at level-1, and r2 random effects at level-2, tauvc.dat contains the following:, in r columns and r rows, in r2 columns and r2 rows the inverse of the information matrix. This dimensions of the matrix are The tauvc.dat file for a HLM3 example is given below: 0.6404879 0.0467574 0.0467574 0.0112249 0.1653112 0.0170460 0.0170460 0.0110198 0.0006327 0.0000285 0.0000014-0.0000281-0.0000014-0.0000001 0.0000285 0.0000249 0.0000020-0.0000014-0.0000011-0.0000001 0.0000014 0.0000020 0.0000039-0.0000001-0.0000001-0.0000002.

-0.0000281-0.0000014-0.0000001 0.0013258 0.0001252 0.0000117-0.0000014-0.0000011-0.0000001 0.0001252 0.0000518 0.0000087-0.0000001-0.0000001-0.0000002 0.0000117 0.0000087 0.0000063 For this model, r = r2 = 2, so that,, and the information matrix is of dimension. The first two lines contain the estimates of the level-2 variance-covariance matrix,, and corresponds with the Tau(pi) matrix as listed in the output file. The elements of this matrix are,,, and, with =0.6404879, = =0.0467574, and = 0.0112249 respectively. The third and fourth lines are the level-3 variance-covariance matrix,. We see that = 0.1653112, =0.0170460, and = 0.0110298. This matrix is also called the asymptotic covariance matrix of the estimated -elements. The matrices are followed by the inverse of the information matrix,. The information matrix can be divided into four blocks. The two blocks on the diagonal (see table below) contain the variances of the elements of the matrices at levels 2 and 3 respectively. The off-diagonal blocks, which are the same due to the symmetry of the information matrix, contains the covariances of the elements of the matrices. 0.0006327 0.0000285 0.0000020-0.0000281-0.0000014-0.0000001 0.0000285 0.0000249 0.0000020-0.0000014-0.0000011-0.0000001 0.0000014 0.0000020 0.0000039-0.0000001-0.0000001-0.0000002-0.0000281-0.0000014-0.0000001 0.0013258 0.0001252 0.000011-0.0000014-0.0000011-0.0000001 0.0001252 0.0000518 0.0000087-0.0000001-0.0000001-0.0000002 0.0000117 0.0000087 0.0000063 The square roots of the diagonal elements of this matrix are the standard errors of the s. For example, if we denote the q-th diagonal element of the matrix by, it follows that is the square of the standard error of the estimated level-2 variance component,

is that of, and is that of. The remaining three diagonal elements are the square of the standard errors of the level-3 variances and covariance, the s. The standard errors of,,,,, are thus 0.025153529, 0.00498999, 0.001974842, 0.036411537, 0.007197222, and 0.00250998 respectively. From the almost zero values of all the covariances shown in the off-diagonal blocks, it can be concluded that the elements of the estimated are uncorrelated with the elements of the estimated. Based on the information matrix, hypothesis tests and confidence intervals for the individual s may be constructed. For example, we can compute the 95% confidence interval for the variance component as shown below: In addition, one can test whether significant random variation exists for random coefficient with a null hypothesis of the form against the alternative hypothesis, where. Rejection of the null hypothesis will lead to the conclusion that there is random variation in. If most or all groups have sufficient data to compute OLS estimates, this can be tested under the model using the statistic

which will be distributed approximately with J Sq - 1 degrees of freedom (Raudenbush & Bryk, 2002). A test for this purpose is based on the standard error of information matrix shown above. The statistic, as obtained from the is approximately normally distributed under the large sample theory of maximum likelihood estimates. The assumption underlying both the confidence interval shown above and the single parameter hypothesis test for random variation in depends on the quality of the normal approximation. For large samples and non-normal data, the use of the ratio shown above is preferred. If this test is used to test the significance of a variance component, it is appropriate to use a one-sided test (Hox, 2002). However, when the normality approximation is rather poor, as will be the case when is near zero and a small sample of groups, the distribution of this statistic is nonnormal as the sampling distribution of the variances is skewed. As pointed out by Raudenbush & Bryk (2002): In general, the smaller the variance component, the more data are needed to justify the large-sample normality approximation on which this test is based. A possible rule of thumb for evaluating the size of is to avoid using this statistic if the contribution of to the total variation in the model is less than 1%. When use of the statistic discussed above is inadvisable, the chi-square test based on the residuals as used by Raudenbush & Bryk (2002), is the appropriate choice. For a twolevel model, this is defined as (also see above) with degrees of freedom equal to, with J the number of level-2 units, and p is the number of predictors in the model. is the OLS estimate of a regression coefficient, computed separately in group j, with its estimated sampling variance in group j denoted

by. This is the approach followed in the calculation of the - statistics reported for individual variance components as reported in the HLM output. For more on this topic, please see: Raudenbush, S.W. & Bryk, A.S. (2002). Hierarchical Linear Models. Applications and Data Analysis Methods. Second Edition. Sage. pp. 61-64; 283-284. Hox, J. (2002). Multilevel Analysis. Techniques and Applications. IEA. pp.42-3.