Testing Structural Equation Models: The Effect of Kurtosis

Similar documents
Testing structural equation models: the effect of kurtosis. Tron Foss BI Norwegian Business School. Karl G. Jøreskog BI Norwegian Business School

Scaled and adjusted restricted tests in. multi-sample analysis of moment structures. Albert Satorra. Universitat Pompeu Fabra.

Uppsala University and Norwegian School of Management, b Uppsala University, Online publication date: 08 July 2010

Evaluating the Sensitivity of Goodness-of-Fit Indices to Data Perturbation: An Integrated MC-SGR Approach

Misspecification in Nonrecursive SEMs 1. Nonrecursive Latent Variable Models under Misspecification

Nesting and Equivalence Testing

1 The Robustness of LISREL Modeling Revisited

A Threshold-Free Approach to the Study of the Structure of Binary Data

Robust Means Modeling vs Traditional Robust Tests 1

Evaluating Small Sample Approaches for Model Test Statistics in Structural Equation Modeling

Global Model Fit Test for Nonlinear SEM

Latent variable interactions

Computationally Efficient Estimation of Multilevel High-Dimensional Latent Variable Models

EVALUATION OF STRUCTURAL EQUATION MODELS

Statistical Inference of Moment Structures

A Cautionary Note on the Use of LISREL s Automatic Start Values in Confirmatory Factor Analysis Studies R. L. Brown University of Wisconsin

Inference with Heywood cases

Can Variances of Latent Variables be Scaled in Such a Way That They Correspond to Eigenvalues?

Ridge Structural Equation Modeling with Correlation Matrices for Ordinal and Continuous Data. Ke-Hai Yuan University of Notre Dame

ABSTRACT. Between-Subjects Design under Variance. Heterogeneity and Nonnormality. Evaluation

STRUCTURAL EQUATION MODELS WITH LATENT VARIABLES

Recovery of weak factor loadings in confirmatory factor analysis under conditions of model misspecification

The comparison of estimation methods on the parameter estimates and fit indices in SEM model under 7-point Likert scale

Jan. 22, An Overview of Structural Equation. Karl Joreskog and LISREL: A personal story

miivfind: A command for identifying model-implied instrumental variables for structural equation models in Stata

Gramian Matrices in Covariance Structure Models

Essays on Estimation Methods for Factor Models and Structural Equation Models

Strati cation in Multivariate Modeling

SEM for Categorical Outcomes

WHAT IS STRUCTURAL EQUATION MODELING (SEM)?

RANDOM INTERCEPT ITEM FACTOR ANALYSIS. IE Working Paper MK8-102-I 02 / 04 / Alberto Maydeu Olivares

Multiple group models for ordinal variables

arxiv: v1 [math.st] 7 Oct 2016

A note on structured means analysis for a single group. André Beauducel 1. October 3 rd, 2015

ANALYSIS OF TWO-LEVEL STRUCTURAL EQUATION MODELS VIA EM TYPE ALGORITHMS

Fit Indices Versus Test Statistics

Title. Description. Remarks and examples. stata.com. stata.com. Variable notation. methods and formulas for sem Methods and formulas for sem

Journal of Multivariate Analysis. Use of prior information in the consistent estimation of regression coefficients in measurement error models

CONSTRUCTION OF COVARIANCE MATRICES WITH A SPECIFIED DISCREPANCY FUNCTION MINIMIZER, WITH APPLICATION TO FACTOR ANALYSIS

Model fit evaluation in multilevel structural equation models

General structural model Part 1: Covariance structure and identification. Psychology 588: Covariance structure and factor models

Extending the Robust Means Modeling Framework. Alyssa Counsell, Phil Chalmers, Matt Sigal, Rob Cribbie

Multivariate Distributions

FACTOR ANALYSIS AS MATRIX DECOMPOSITION 1. INTRODUCTION

UCLA Department of Statistics Papers

Structural Equation Modeling and Confirmatory Factor Analysis. Types of Variables

Large Sample Properties of Estimators in the Classical Linear Regression Model

Asymptotic robustness of standard errors in multilevel structural equation models

Causal Inference Using Nonnormality Yutaka Kano and Shohei Shimizu 1

Methods for Handling Missing Non-Normal Data in Structural Equation Modeling. Fan Jia

Model Estimation Example

An Alternative to Cronbach s Alpha: A L-Moment Based Measure of Internal-consistency Reliability

Assessing Factorial Invariance in Ordered-Categorical Measures

DOCUMENTOS DE TRABAJO BILTOKI. Spanish Customer Satisfaction Indices by Cumulative Panel Data.

sempower Manual Morten Moshagen

New Introduction to Multiple Time Series Analysis

DIAGNOSTICS OF THE ERROR FACTOR COVARIANCES OTTÓ HAJDU 1

Cointegration Lecture I: Introduction

Matrix Algebra Review

MICHAEL SCHREINER and KARL SCHWEIZER

ON NORMAL THEORY AND ASSOCIATED TEST STATISTICS IN COVARIANCE STRUCTURE ANALYSIS UNDER TWO CLASSES OF NONNORMAL DISTRIBUTIONS

Psychology 454: Latent Variable Modeling How do you know if a model works?

Testing Algebraic Hypotheses

TESTING FOR NORMALITY IN THE LINEAR REGRESSION MODEL: AN EMPIRICAL LIKELIHOOD RATIO TEST

POWER AND TYPE I ERROR RATE COMPARISON OF MULTIVARIATE ANALYSIS OF VARIANCE

How to use Stata s sem with small samples? New corrections for the L. R. χ 2 statistics and fit indices

Confirmatory Factor Analysis. Psych 818 DeShon

DATE: 9/ L I S R E L 8.80

Department of Psychology, Beihang University, China 2. Department of Psychology, University of Notre Dame, USA 3

Inference using structural equations with latent variables

Haruhiko Ogasawara. This article gives the first half of an expository supplement to Ogasawara (2015).

A Scaled Diæerence Chi-square Test Statistic. Albert Satorra. Universitat Pompeu Fabra. and. Peter M. Bentler. August 3, 1999

Correlations with Categorical Data

Structural Equation Modeling

More Linear Algebra. Edps/Soc 584, Psych 594. Carolyn J. Anderson

Preliminaries. Copyright c 2018 Dan Nettleton (Iowa State University) Statistics / 38

Streamlining Missing Data Analysis by Aggregating Multiple Imputations at the Data Level

Alma Mater Studiorum - Università di Bologna

Improving Model Selection in Structural Equation Models with New Bayes Factor Approximations

The Effects of Monetary Policy on Stock Market Bubbles: Some Evidence

Analysis of variance, multivariate (MANOVA)

Improper Solutions in Exploratory Factor Analysis: Causes and Treatments

An Introduction to Multivariate Statistical Analysis

INTRODUCTION TO STRUCTURAL EQUATION MODELS

Estimating the validity of administrative and survey variables through structural equation modeling

Effect of Estimation Method on Incremental Fit Indexes for Covariance Structure Models

W-BASED VS LATENT VARIABLES SPATIAL AUTOREGRESSIVE MODELS: EVIDENCE FROM MONTE CARLO SIMULATIONS

MODEL IMPLIED INSTRUMENTAL VARIABLE ESTIMATION FOR MULTILEVEL CONFIRMATORY FACTOR ANALYSIS. Michael L. Giordano

Using Structural Equation Modeling to Conduct Confirmatory Factor Analysis

GARCH Models Estimation and Inference. Eduardo Rossi University of Pavia

Specifying Latent Curve and Other Growth Models Using Mplus. (Revised )

Confirmatory Factor Analysis: Model comparison, respecification, and more. Psychology 588: Covariance structure and factor models

Using Estimating Equations for Spatially Correlated A

Do not copy, quote, or cite without permission LECTURE 4: THE GENERAL LISREL MODEL

Factor Analysis. Qian-Li Xue

Algorithms and Code JMASM24: Numerical Computing for Third-Order Power Method Polynomials (Excel)

Accounting for Population Uncertainty in Covariance Structure Analysis

1 Data Arrays and Decompositions

Simulating Uniform- and Triangular- Based Double Power Method Distributions

Appendix A: The time series behavior of employment growth

Transcription:

Testing Structural Equation Models: The Effect of Kurtosis Tron Foss, Karl G Jöreskog & Ulf H Olsson Norwegian School of Management October 18, 2006

Abstract Various chi-square statistics are used for testing structural equation models. A commonly used chi-square is the Satorra-Bentler (SB) statistic. This is a normal-theory chi-square corrected for non-normality in the observed variables. The relationship between the SB statistic and kurtosis is developed and it is shown that the SB statistic tends to decrease with increasing kurtosis if the model does not hold. It is shown that the population value of the correction factor for the SB statistic decreases toward zero with increasing kurtosis.

Various chi-square statistics are used for testing structural equation models. If the model is fitted by the maximum likelihood (ML) method, one such chi-square statistic is obtained as n times the minimum of the ML fit function, where n is the sample size. An asymptotically equivalent chi-square statistic is n times the minimum value of a quadratic form using a weight matrix evaluated under multivariate normality of the observed variables. Following the notation in Jöreskog, Sörbom, Du Toit, & Du Toit (2003), these chi-square statistics are denoted c 1 and c 2, respectively. They are valid under multivariate normality of the observed variables and if the model holds. If the observed variables are non-normal, Satorra & Bentler (1988) proposed another chi-square statistic c 3 (often called the SB statistic) which is c 1 or c 2 multiplied by a scale factor which is estimated from the sample and involves an estimate of the asymptotic covariance matrix (ACM) of the sample variances and covariances. Satorra & Bentler (1988) show that under multivariate non-normality, c 1 and c 2 have asymptotically a distribution which is a linear combination of chisquares with one degree of freedom. The scale factor is estimated such that c 3 has an asymtotically correct mean even though it does not have an asymptotic chi-square distribution. In practice, c 3 is conceived of as a way of correcting c 1 or c 2 for the effects of non-normality and c 3 is often used as it performs better than the ADF test proposed by Browne (1984), particularly if n is not very large, see e.g., Hu, Bentler, & Kano (1992). In this paper we develop the relationship between c 3 and kurtosis and we show that the c 3 statistic tends to decrease with increasing kurtosis. The practical consequence of this is that models that do not hold tend to be accepted by the chi-square test if kurtosis is large. Thus, this test have low power for detecting misspecified models. Although the results developed here can be demonstrated by simulating and analyzing random samples, we will use a different approach. Curran, West and Finch (1996) presented a simulation study of these chi-square statitics and concluded The most surprising findings are related to the behavior of the SB and ADF test statistics under simultanous conditions of misspecification and multivariate nonnormality (Models 3 and 4). The expected values of these test statistics markedly decreased with increasing nonnormality. That is, all else being equal, the SB and ADF test statistics were less likely to dectect a specification error given increasing departures from a multivariate normal distribution. The more severe the nonormality, the greater the corresponding loss of power. This result was unexpected, and we are not aware of any previous discussions of this finding (Curran, West and Finch, 1996, p.25). This paper provides an explanation for their results. The Asymptotic Covariance Matrix under Non-Normality Browne & Shapiro (1988) considered the following general structure for an observable k 1 random vector z: g z = µ + A i v i, (1) where µ is a constant vector, A i is a constant k m i matrix and the v i are independent m i 1 vector variates for i =1, 2,...,g. Let S be a sample covariance matrix estimated from a random sample of n observations of z, and let s =(s 11,s 21,s 22,...,s kk ) be a vector of order 1 2k(k +1) 1 of the non-duplicated elements of S. Let k = 1 2 k(k + 1). We assume that S converge in probability to Σ 0 as n. i=i 1

It follows from the multivariate central limit theorem that n 1 2 (s σ 0 ) d N(0, Ω), (2) d where denotes convergence in distribution. Browne & Shapiro (1988, Equation 2.7) give Ω as } g Ω=K {2(Σ 0 Σ 0 )+ (A i A i )C i (A i A i) K (3) i=1 where K is the matrix K k of order k 2 k defined in Browne (1974, Section 2) or in Browne (1984, Section 4), and denotes the Kronecker product. The matrix C i is the fourth order cumulant matrix of v i, i =1, 2,...,g. For our purpose it is sufficient to consider a special case of (1), namely when each v i is a scalar random variable. Then A i is a column vector a i and (1) can be written z = Av, (4) where A is a matrix of order k g and v is a g 1 vector of independent random variables having moments up to order four. For convenience, we assume that v i has mean 0 and variance 1. The rationale for (4) is as follows (see also Olsson, Foss & Troye, 2003 and Mattson, 1997). Suppose we are given a structural equation model (SEM) with specified parameter matrices, for example, the factor analysis model z =Λξ + δ, (5) where ξ(q 1) and δ(k 1) are independent with covariance matrices Φ and Θ δ respectively, both assumed to be positive definite. Since Φ is positive definite there exist a q q matrix T 1 such that T 1 T 1 = Φ. Then we can write ξ = T 1v 1, where we assume that the elements of v 1 (q 1) are independent. In the same manner we can write δ = T 2 v 2, where the elements of v 2 of order k 1 are independent and T 2 (k k) is a non-singular matrix such that T 2 T 2 =Θ δ. Assuming that v 1 and v 2 are independent it follows that ξ and δ are independent and that ( ) v1 z =Λξ + δ =(ΛT 1 T 2 ) = Av, (6) as in (4). The Appendix shows that more general SEM models can also be constructed from a set of independent random variables v. Note that v is not a vector of latent variables (not a ξ-vector) but rather a vector of random drawings from a given distribution. We have assumed the v s are independent. But this does not imply that the ξ s (ξ = T 1 v 1 ) are independent; they will in general be correlated. But it does imply that ξ and δ are independent vectors as required. The mean vector of z is 0 and the covariance matrix is v 2 Σ 0 = AA. (7) Let µ 4i = E(vi 4). The matrix C i in (3) is the 1 1 matrix γ 2i = µ 4i 3, the fourth order cumulant or excess kurtosis of v i. Then (3) can be written in the following form } g Ω=K {2(Σ Σ) + (a i a i )(a i a i ) γ 2i K. (8) i=1 2

Let G =[(a 1 a 1 )(a 2 a 2 )...(a g a g )] and let M = diag(γ 21,γ 22,...,γ 2g ). G is of order k 2 g and M is of order g g. Then Ω=K [2(Σ 0 Σ 0 )+GMG ]K. (9) If v i is normally distributed, then µ 4i = 3 and γ 2i = 0. Then the corresponding diagonal element of M is zero and the corresponding column of G can be eliminated. If v i is normally distributed for all i, then v has a multivariate normal distribution and M = 0 so that the second term in (9) vanishes. It is convenient to use the notation Ω nnt for the matrix in (9) and the notation Ω nt for the matrix K [2(Σ 0 Σ 0 )]K. Thus, from (9) it follows that Ω nnt =Ω nt + K GMG K. (10) A special case of (9) is when all elements of v have the same kurtosis so that γ 2i = γ 2, say, which is the same for all i. Then M = γ 2 I. Note that we made no assumption about third order moments of the elements of v. The assumption of an elliptic distribution of z with homogeneous kurtosis parameters was considered by several authors, see, e.g., Browne (1984), and Browne & Shapiro (1987). A general case with heterogenous kurtosis parameters is considered by Kano, Bercane & Bentler (1990). The Effect of Kurtosis Consider a general model Σ(θ), where θ is a parameter vector of order t, and a general fit function of the form F [s,σ(θ)] satisfying the conditions of Browne (1984) and Satorra (1989). For example, F can be F [s,σ(θ)] = [s σ(θ)] V[s σ(θ)] (11) where V is either a fixed positive definite matrix or a random matrix converging in probability to a positive definite matrix V. This covers all fit functions available in computer softwares for estimating structural equation models. F is to be minimized with respect to the model parameters θ. Let ˆθ be the minimizer of F [s,σ(θ)] and let θ 0 be a unique minimizer of F [σ 0,σ(θ)]. We assume that the model does not hold so that F [σ 0,σ(θ 0 )] > 0. Browne & Shapiro (1988, equation 3.9, see also Satorra (1989, p. 135) considered a different case, the model of parametric drift, where σ 0 = σ(θ 0 )+n 1/2 µ, where µ is a constant vector. This implies that the model holds in the population so that F [σ 0,σ(θ 0 )] = 0, contrary to our assumption. Using the notation in Jöreskog & Sörbom (1999), the chi-square statistics c 2 and c 3 discussed initially are defined as c 2 = n(s ˆσ) c ( cw nt c ) 1 c(s ˆσ) (12) where d is the degrees of freedom and c 3 = d h c 2, (13) h = tr[( c W nt c) 1 ( c W nnt c)]. (14) Here ˆσ = σ(ˆθ), W nt and W nnt are consistent estimates of Ω nt and Ω nnt, respectively, and c is an orthogonal complement to the matrix = σ/ θ evaluated at ˆθ. The matrix is of order k t with t<k and the matrix c is of order k d, where d = k t is the degrees of freedom of the model. 3

Since we assume that the model does not hold, the matrices W nt and W nnt are estimated without regard to the model. For example, take the elements of W nt and W nnt as w nt ghij = s gis hj + s gj s hi (15) w nnt ghij = m ghij s gh s ij, (16) where n m ghij =(1/n) (z ag z g )(z ah z h )(z ai z i )(z aj z j ). (17) a=1 Most computer programs for structural equation modeling assume that the model holds and therefore use ˆσ ij = σ ij (ˆθ) instead of s ij in (15). The correction factor in (13) is d/h, where d is a constant and h is a random variable converging in probability to H, say. To obtain H we replace s by σ 0,ˆσ by σ(θ 0 ), and c by 0c, where 0c is evaluated at θ 0. Furthermore, W nt and W nnt are replaced by Ω nt and Ω nnt. We assume that 0c has rank d. Then H = tr[( 0cΩ nt 0c ) 1 ( 0cΩ nnt 0c )]. (18) The influence of kurtosis on H is only via the diagonal matrix M. All other matrices in (10) are independent of kurtosis. From (10) we have 0cΩ nnt 0c = 0cΩ nt 0c + 0cK GMG K 0c. (19) Hence, ( 0cΩ nt 0c ) 1 ( 0cΩ nnt 0c )=I d +( 0cΩ nt 0c ) 1 PMP, (20) where I d is the identity matrix of order d and P = 0c K G. (21) Taking the trace of (20), gives where H = d + tr(qm), (22) Q = P ( 0cΩ nt 0c ) 1 P. (23) Q is symmetric and of order g g. Since M is diagonal, H = d + g q ii γ 2i. (24) i Q is positive semidefinite and if Q = 0,q ii > 0 for at least one i. Thus, if γ 2i for all i, it follows that H.Ifγ 2i = γ 2 for all i, then H = d +(trq)γ 2. (25) increases linearly with γ 2. Thus, the correction factor d/h tends to be small in large samples if kurtosis is large. 4

References Browne, M.W. (1974) Generalized least squares estimators in the analysis of covariance structures. South African Statistical Journal, 8, 1 24. Browne, M.W. (1984) Asymptotically distribution-free methods for the analysis of covariance structures. British Journal of Mathematical and Statistical Psychology, 37, 62 83. Browne, M.W. & Shapiro, A. (1987) Adjustments for kurtosis in factor analysis with elliptically distributed errors. Journal of the Royal Statistical Society. Series B (Methodological), 49, 346 352. Curran, P.J., West S.G. & Finch, J.F. (1996) The robustness of test statistics to nonnormality and specification error in confirmatory factor analysis. Psychological Methods, 1, 16 29. Hu, L., Bentler, P.M., & Kano, Y. (1992) Can test statistics in covariance structure analysis be trusted? Psychological Bulletin, Vol. 112, 351 362. Jöreskog, K.G. & Sörbom, D. (1999) LISREL 8: User s Reference Guide. Second Edition. Lincolnwood, IL: Scientific Software International. Jöreskog, K.G., Sörbom, D., Du Toit, S., & Du Toit, M. (2003) LISREL 8: New Statistical Features. Third printing with revisions. Lincolnwood, IL: Scientific Software International. Kano, Y., Berkane, M., & Bentler, P.M. (1990) Covariance structure analysis with heterogeneous kurtosis parameters. Biometrika, 77, 575 585. Mattsson, S. (1997) How to generate non-normal data for simulation of structural equation models. Multivariate Behavioral Research, 32, 355 373. Olsson, U.H., Foss, T., & Troye, S.V. (2003) Does the ADF fit function decrease when the kurtosis increases. British Journal of Mathematical and Statistical Psychology, 56, 289 303. Satorra, A. (1989) Alternative test criteria in covariance structure analysis: A unified approach. Psychometrika, 54, 131 151. Satorra, A., & Bentler, P.M. (1988) Scaling corrections for chi-square statistics in covariance structure analysis. Proceedings of the Business and Economic Statistics Section of the American Statistical Association, 308 313. 5

Appendix In a general SEM model we need to construct independent random vectors ξ(n 1), ς(m 1), ε(p 1), and δ(q 1) with covariance matrices Φ, Ψ, Θ ε, Θ δ, respectively, satisfying η = Bη +Γξ + ς (26) y =Λ y η + ε (27) x =Λ x ξ + δ (28) where B, Γ, Λ y, Λ x are coefficient matrices with I B non-singular. For further explication of this model, see Jöreskog & Sörbom (1999). Solving (29) for η gives and substituting this into (30) gives η =(I B) 1 Γξ +(I B) 1 ς, (29) Combining (33) and (31) gives ( y x ) = y =Λ y (I B) 1 Γξ +Λ y (I B) 1 ς + ε, (30) ( Λy (I B) 1 Γ Λ y (I B) 1 I 0 Λ x 0 0 I ) Take k=p+qand g = n+m+p + q in (4) and v =(v 1,v 2,v 3,v 4),where the subvectors are of orders n, m, p and q, respectively. Choose square non-singular matrices T 1,T 2,T 3 and T 4, such that Φ = T 1 T 1, Ψ=T 2T 2, Θ ε = T 3 T 3 and Θ δ = T 4 T 4 and choose ξ = T 1v 1, ς = T 2 v 2, ε = T 3 v 3 and δ = T 4 v 4. Then ( y x which is of the form (4). ) ( Λy (I B) = 1 ΓT 1 Λ y (I B) 1 T 2 T 3 0 Λ x T 1 0 0 T 4 ξ ς ε δ ) v 1 v 2 v 3 v 4 6