Latent variable interactions

Similar documents
Global Model Fit Test for Nonlinear SEM

Plausible Values for Latent Variables Using Mplus

Nesting and Equivalence Testing

Computationally Efficient Estimation of Multilevel High-Dimensional Latent Variable Models

Variable-Specific Entropy Contribution

Ruth E. Mathiowetz. Chapel Hill 2010

Using Mplus individual residual plots for. diagnostics and model evaluation in SEM

Moderation 調節 = 交互作用

Estimation of Curvilinear Effects in SEM. Rex B. Kline, September 2009

STT 843 Key to Homework 1 Spring 2018

General structural model Part 1: Covariance structure and identification. Psychology 588: Covariance structure and factor models

SRMR in Mplus. Tihomir Asparouhov and Bengt Muthén. May 2, 2018

Latent Variable Centering of Predictors and Mediators in Multilevel and Time-Series Models

Centering Predictor and Mediator Variables in Multilevel and Time-Series Models

Do not copy, quote, or cite without permission LECTURE 4: THE GENERAL LISREL MODEL

Misspecification in Nonrecursive SEMs 1. Nonrecursive Latent Variable Models under Misspecification

Testing Structural Equation Models: The Effect of Kurtosis

Title. Description. Remarks and examples. stata.com. stata.com. Variable notation. methods and formulas for sem Methods and formulas for sem

Inference using structural equations with latent variables

STAT 730 Chapter 9: Factor analysis

CHAPTER 9 EXAMPLES: MULTILEVEL MODELING WITH COMPLEX SURVEY DATA

Multilevel Structural Equation Modeling

SEM 2: Structural Equation Modeling

WHAT IS STRUCTURAL EQUATION MODELING (SEM)?

SEM 2: Structural Equation Modeling

Bayesian Analysis of Latent Variable Models using Mplus

Strati cation in Multivariate Modeling

State-space Model. Eduardo Rossi University of Pavia. November Rossi State-space Model Financial Econometrics / 49

Assignment 1. SEM 2: Structural Equation Modeling

Consistent Bivariate Distribution

Specifying Latent Curve and Other Growth Models Using Mplus. (Revised )

What is in the Book: Outline

Introduction to Confirmatory Factor Analysis

Structural Equation Modeling and Confirmatory Factor Analysis. Types of Variables

Overview. 1. Terms and Definitions. 2. Model Identification. 3. Path Coefficients

Advanced Econometrics I

Problem Set #6: OLS. Economics 835: Econometrics. Fall 2012

State-space Model. Eduardo Rossi University of Pavia. November Rossi State-space Model Fin. Econometrics / 53

Advances in Mixture Modeling And More

SEM for Categorical Outcomes

Structural Equation Modeling An Econometrician s Introduction

Supplementary materials for: Exploratory Structural Equation Modeling

An Introduction to Path Analysis

Mplus Short Courses Topic 3. Growth Modeling With Latent Variables Using Mplus: Introductory And Intermediate Growth Models

Estimating Three-way Latent Interaction Effects in Structural Equation Modeling

Confidence Intervals for a Semiparametric Approach to Modeling Nonlinear Relations among Latent Variables

Bayesian SEM: A more flexible representation of substantive theory

An Introduction to Mplus and Path Analysis

Confirmatory Factor Analysis. Psych 818 DeShon

Latent Factor Regression Models for Grouped Outcomes

4. Path Analysis. In the diagram: The technique of path analysis is originated by (American) geneticist Sewell Wright in early 1920.

Consequences of measurement error. Psychology 588: Covariance structure and factor models

An Efficient State Space Approach to Estimate Univariate and Multivariate Multilevel Regression Models

Continuous Time Survival in Latent Variable Models

Mplus Short Courses Day 2. Growth Modeling With Latent Variables Using Mplus

Finite Mixture EFA in Mplus

Introduction to Structural Equation Modeling Dominique Zephyr Applied Statistics Lab

Outline. Overview of Issues. Spatial Regression. Luc Anselin

Citation for published version (APA): Jak, S. (2013). Cluster bias: Testing measurement invariance in multilevel data

Course Introduction and Overview Descriptive Statistics Conceptualizations of Variance Review of the General Linear Model

Estimating the validity of administrative and survey variables through structural equation modeling

Structural Equation Modeling

Latent class analysis for intensive longitudinal data, Hidden Markov processes, Regime switching models and Dynamic Structural Equations in Mplus

Time Dependence of Growth Parameters in Latent Growth Curve Models with Time Invariant Covariates

Covariance and Correlation

Scaled and adjusted restricted tests in. multi-sample analysis of moment structures. Albert Satorra. Universitat Pompeu Fabra.

Old and new approaches for the analysis of categorical data in a SEM framework

Introduction to Matrix Algebra and the Multivariate Normal Distribution

SEM for those who think they don t care

Principle Components Analysis (PCA) Relationship Between a Linear Combination of Variables and Axes Rotation for PCA

2. Matrix Algebra and Random Vectors

Supplemental material for Autoregressive Latent Trajectory 1

Estimation of nonlinear latent structural equation models using the extended unconstrained approach

W-BASED VS LATENT VARIABLES SPATIAL AUTOREGRESSIVE MODELS: EVIDENCE FROM MONTE CARLO SIMULATIONS

ABSTRACT. Chair, Dr. Gregory R. Hancock, Department of. interactions as a function of the size of the interaction effect, sample size, the loadings of

TAMS39 Lecture 10 Principal Component Analysis Factor Analysis

INTRODUCTION TO STRUCTURAL EQUATION MODELS

Journal of Multivariate Analysis. Use of prior information in the consistent estimation of regression coefficients in measurement error models

Time-Invariant Predictors in Longitudinal Models

I L L I N O I S UNIVERSITY OF ILLINOIS AT URBANA-CHAMPAIGN

Using Bayesian Priors for More Flexible Latent Class Analysis

Dynamic Structural Equation Modeling of Intensive Longitudinal Data Using Mplus Version 8

Machine Learning and Pattern Recognition, Tutorial Sheet Number 3 Answers

STRUCTURAL EQUATION MODELS WITH LATENT VARIABLES

Lecture 20: Linear model, the LSE, and UMVUE

Gaussian Processes. Le Song. Machine Learning II: Advanced Topics CSE 8803ML, Spring 2012

SEM Day 1 Lab Exercises SPIDA 2007 Dave Flora

Model Assumptions; Predicting Heterogeneity of Variance

RANDOM INTERCEPT ITEM FACTOR ANALYSIS. IE Working Paper MK8-102-I 02 / 04 / Alberto Maydeu Olivares

Random Vectors 1. STA442/2101 Fall See last slide for copyright information. 1 / 30

Dimensionality Assessment: Additional Methods

7. The Multivariate Normal Distribution

Psychology 454: Latent Variable Modeling How do you know if a model works?

Next is material on matrix rank. Please see the handout

Application of Plausible Values of Latent Variables to Analyzing BSI-18 Factors. Jichuan Wang, Ph.D

The comparison of estimation methods on the parameter estimates and fit indices in SEM model under 7-point Likert scale

Path Analysis. PRE 906: Structural Equation Modeling Lecture #5 February 18, PRE 906, SEM: Lecture 5 - Path Analysis

Maximum Likelihood Tests and Quasi-Maximum-Likelihood

STA 2201/442 Assignment 2

ECON 4160, Autumn term Lecture 1

Transcription:

Latent variable interactions Bengt Muthén & Tihomir Asparouhov Mplus www.statmodel.com November 2, 2015 1

1 Latent variable interactions Structural equation modeling with latent variable interactions has been discussed with respect to maximum-likelihood estimation in Klein and Moosbrugger (2000). Multivariate normality is assumed for the latent variables. The ML computations are heavier than for models without latent variable interactions because numerical integration is needed. For an overview of the ML approach and various estimators suggested in earlier work, see Marsh et al. (2004). Arminger and Muthén (1998), Klein and Muthén (2007), Cudeck et al. (2009), and Mooijaart and Bentler (2010) discuss alternative estimators and algorithms. This section discusses interpretation, model testing, explained variance, standardization, and plotting of effects for models with latent variable interactions. 1.1 Model interpretation As an example, consider the latent variable interaction model of Figure 1. The figure specifies that the factor η 3 is regressed on η 1 and η 2 as well as the interaction between η 1 and η 2, as shown by the structural equation η 3 = β 1 η 1 + β 2 η 2 + β 3 η 1 η 2 + ζ 3. (1) The interaction variable η 1 η 2 involves only one parameter, the slope β 3. The interaction variable does not have a mean or a variance parameter. It does not have parameters for covariances with other variables. It can also not be a dependent variable. As is seen in Figure 1, the model also contains a second structural equation where η 4 is linearly regressed on η 3, so that there is no direct effect on 2

Figure 1: Structural equation model with interaction between latent variables η 4 from η 1 and η 2, or their interaction. For ease of interpretation the (1) regression can be re-written in the equivalent form η 3 = (β 1 + β 3 η 2 ) η 1 + β 2 η 2 + ζ 3, (2) where (β 1 + β 3 η 2 ) is a moderator function (Klein & Moosbrugger, 2000) so that the β 1 strength of influence of η 1 on η 3 is moderated by β 3 η 2. The choice of moderator when translating (1) to (2) is arbitrary from an algebraic point of view, and is purely a choice based on ease of substantive interpretation. As an example, Cudeck et al. (2009) considers school achievement (η 3 ) influenced by general reasoning (η 1 ), quantitative ability (η 2 ), and their interaction. In line with (2) the interaction is expressed as quantitative ability moderating the influence of general reasoning on school achievement. Plotting of interactions further aids the interpretation as discussed in Section 1.5. 3

1.2 Model testing As pointed out in Mooijaart and Satorra (2009), the likelihood-ratio χ 2 obtained by ML for models without latent variable interactions is not sensitive to incorrectly leaving out latent variable interactions. For example, the model of Figure 1 without the interaction term β 3 η 1 η 2 fits data generated as in (1) perfectly. This is due to general maximum-likelihood results on robustness to non-normality (Satorra, 1992, 2002). Misfit can be detected only by considering higher-order moments than the second-order variances and covariances of the outcomes. Without involving higher-order moments, a reasonable modeling strategy is to first fit a model without interactions and obtain a good fit in terms of the ML likelihood-ratio χ 2. An interaction term can then be added and the β 3 significance of the interaction significance tested by either a z-test or a likelihood-ratio χ 2 difference test (Klein & Moosbrugger, 2000). Likelihood-ratio or Wald tests can be used to test the joint significance of several interaction terms. 1.3 Mean, variance, and R 2 To compute a dependent variable mean, variance, and R 2 for models with latent variable interactions, the following results are needed. As discussed in Chapter??, the covariance between two variables x j and x k is defined as Cov(x j, x k ) = E(x j x k ) E(x j ) E(x k ), (3) so that the variance is obtained as Cov(x j, x j ) = V (x j ) = E(x 2 j) [E(x j )] 2. (4) 4

With E(x j ) = 0 or E(x k ) = 0, (3) gives the mean of a product E(x j x k ) = Cov(x j, x k ). (5) Assuming multivariate normality for four random variables x i, x j, x k, x l any third-order moment about the mean (µ) is zero (see, e.g., Anderson, 1984), E(x i µ i )(x j µ j )(x k µ k ) = 0, (6) while the fourth-order moment about the mean is a function of covariances, E(x i µ i )(x j µ j )(x k µ k )(x l µ l ) = σ ij σ kl + σ ik σ jl + σ il σ jk, (7) where for example σ jk = Cov(x j, x k ) and σ kk = V ar(x k ). This gives E(x j x k x j x k ) = V (x j ) V (x k ) + 2 [Cov(x j, x k )] 2, (8) so that the variance of a product is obtained as V (x j x k ) = E(x j x k x j x k ) [E(x j x k )] 2 (9) = V (x j ) V (x k ) + 2 [Cov(x j, x k )] 2 [Cov(x j, x k ] 2 (10) = V (x j ) V (x k ) + [Cov(x j, x k )] 2, (11) Consider the application of these results to the mean and variance of the factor η 3 in (1) of Figure 1. Because of zero factor means, using (5) the mean of η 3 in 5

(1) is obtained as E(η 3 ) = β 1 0 + β 2 0 + β 3 E(η 1 η 2 ) + 0 (12) = β 3 Cov(η 1, η 2 ). (13) Using (4), the variance of η 3 is V (η 3 ) = E(η 3 η 3 ) [E(η 3 )] 2, (14) where the second term has already been determined. As for the first term, multiplying the right-hand side of (1) by itself results in products of two, three, and four factors. Expectations for three- and four-factor terms are simplified by the following two results, assuming bivariate normality and zero means for η 1 and η 2. All third-order moments E(η i η j η k ) are zero by (6). The formula (8) is used to obtain the result E(η 1 η 2 η 1 η 2 ) = V (η 1 ) V (η 2 ) + 2 [Cov(η 1, η 2 )] 2. (15) Collecting terms, it follows that the variance of η 3 is obtained as V (η 3 ) = β 2 1 V (η 1 ) + β 2 2 V (η 2 ) + 2 β 1 β 2 Cov(η 1, η 2 ) + β 2 3 V (η 1 η 2 ) + V (ζ 3 ), (16) where by (11) V (η 1 η 2 ) = V (η 1 ) V (η 2 ) + [Cov(η 1, η 2 )] 2, (17) 6

R-square for η 3 can be expressed as usual as [V (η 3 ) V (ζ 3 )]/V (η 3 ). (18) Using (16), the proportion of V (η 3 ) contributed by the interaction term can be quantified as (cf. Mooijaart & Satorra, 2009; p. 445) β 2 3 [V (η 1 ) V (η 2 ) + [Cov(η 1, η 2 )] 2 ]/V (η 3 ). (19) Consider as a hypothetical example the latent variable interaction model of Figure 2. Here, the latent variable interaction is between an exogenous and an endogenous latent variable. This example is useful to study the details of how to portray the model. The structural equations are η 1 = β η 2 + ζ 1, (20) η 3 = β 1 η 1 + β 2 η 2 + β 3 η 1 η 2 + ζ 3. (21) Let β = 1, β 1 = 0.5, β 2 = 0.7, β 3 = 0.4, V (η 2 ) = 1, V (ζ 1 ) = 1, and V (ζ 3 ) = 1. This implies that V (η 1 ) = β 2 V (η 2 ) + V (ζ 1 ) = 1 2 1 + 1 = 2 and Cov(η 1, η 2 ) = β V (η 2 ) = 1 1 = 1. Using (16), V (η 3 ) = 3.17. The η 3 R-square is 0.68 and the variance percentage due to the interaction is 15%. 1.4 Standardization Because latent variables have arbitrary metrics, it is useful to also present interaction effects in terms of standardized latent variables. Noting that (21) 7

Figure 2: Structural equation model with interaction between an exogenous and an endogenous latent variable is identical to (1), the model interpretation is aided by considering the moderator function (β 1 + β 3 η 2 ) η 1 of (2), so that η 2 moderates the η 1 influence on η 3. As usual, standardization is obtained by dividing by the standard deviation of the dependent variable and multiplying by the standard deviation of the independent variable. The standardized β 1 and β 3 coefficients in the term (β 1 + β 3 η 2 ) η 1 are obtained by dividing both by V (η 3 ) = 3.17, multiplying β 1 by V (η 1 ) = 2, and multiplying β 3 by V (η 1 ) V (η 2 ) = 2. This gives a standardized β 1 = 0.397 and a standardized β 3 = 0.318. The standardization of β 3 is in line with Wen, Marsh, and Hau (2010; equation 10). These authors discuss why standardization of β 3 using V (η 1 ) V (η 2 ) is preferred over using V (η1 η 2 ). 8

The standard deviation change in η 3 as a function of a one standard deviation change in η 1 can now be evaluated at different values of η 2 using the moderator function. At the zero mean of η 2, a standard deviation increase in η 1 leads to a 0.397 standard deviation increase in η 3. At one standard deviation above the mean of η 2, a standard deviation increase in η 1 leads to a 0.397+0.318 1 = 0.715 standard deviation increase in η 3. At one standard deviation below the mean of η 2, a standard deviation increase in η 1 leads to a 0.397 0.318 1 = 0.079 standard deviation increase in η 3. In other words, the biggest effect of η 1 on η 3 occurs for subjects with high values on η 2. 1.5 Plotting of interactions The interaction can be plotted as in Figure 3. Using asterisks to denote standardization, consider the rearranged (21), η 3 = (β 1 + β 3 η 2) η 1 + β 2 η 2 + ζ 3. (22) Using (22), the three lines in the figure are expressed as follows in terms of the conditional expectation function for η 3 at the three levels of η 2, E(η3 η 1, η2 = 0) = β1 η1, (23) E(η3 η 1, η2 = 1) = (β1 + β3) η1 + β2, (24) E(η3 η 1, η2 = 1) = (β1 β3) η1 β2. (25) (26) Here, the standardized value β 2 = β 2 V (η 2 )/ V (η 3 ) = 0.7 1/ 3.17 = 0.393. 9

Figure 3: Interaction plot for structural equation model with interaction between an exogenous and an endogenous latent variable η 3 1.0 0.5 0.0 0.5 1.0 1.5 2.0 Mean of eta2 One SD above the mean of eta2 One SD below the mean of eta2 3 2 1 0 1 2 3 η 1 2 Standardization in matrix terms This section describes standardization of the general model in Mplus when latent variable interactions are present. Suppose that Y is a vector of dependent variables, X is a vector of covariates, η is a vector of latent variables. All residuals are assumed normal. Suppose that the variables (Y,η,X) are split into two disjoint set of variables V 1 and V 2. V 1 set of dependent variables that are not part of interaction terms and V 2 is a set of variables that are a part of interaction terms. Suppose that V 1 is a vector of size p 1 and V 2 is a vector of size p 2. The SEM equation is given by these two equations V 1 = α 1 + B 1 V 1 + C 1 V 2 + k D i (V 2,f(i) V 2,g(i) ) + ε 1 i=1 10

V 2 = α 2 + B 2 V 2 + ε 2 where α 1, B 1, C 1, D i, α 2, B 2 are model parameters. The vectors α 1, D i are of length p 1 while the vector α 2 is of length p 2. The matrices B 1, C 1 and B 2 are of size p 1 p 1, p 1 p 2 and p 2 p 2 respectively. The residual variable ε 1 has zero mean and variance covariance Θ and ε 2 has zero mean and variance covariance Ψ. The residuals ε 1 are not considered independent of the residuals ε 2. Let s call the covariance F = Cov(ε 1, ε 2 ). The functions f(i) and g(i) simply define the interaction terms, i.e., f(i) and g(i) are integers between 1 and p 2 and k is the number of interaction terms in the model. We can assume that all covariates X are in the V 2 vector and the V 1 vector consists only of η and Y variables that are regressed on interaction terms, while the remaining η and Y variables are in vector V 2. We can compute the model implied mean and variance for these variables as follows. For the variables V 2 we get E(V 2 ) = µ 2 = (1 B 2 ) 1 α 2 V ar(v 2 ) = Σ 2 = (1 B 2 ) 1 Ψ((1 B 2 ) 1 ) T For V 1 we get E(V 1 ) = (1 B 1 ) 1 α 1 +(1 B 1 ) 1 C 1 µ 2 +(1 B 1 ) 1 k D i (µ 2,f(i) µ 2,g(i) +Σ 2,f(i),g(i) ). i=1 Denote by V 20 = V 2 µ 2 k µ 10 = (1 B 1 ) 1 α 1 + (1 B 1 ) 1 C 1 µ 2 + (1 B 1 ) 1 D i (µ 2,f(i) µ 2,g(i) ) 11 i=1

k V 10 = (1 B 1 ) 1 C 1 V 20 + (1 B 1 ) 1 ε 1 + (1 B 1 ) 1 D i (µ 2,f(i) V 20,g(i) )+ k (1 B 1 ) 1 D i (V 20,f(i) µ 2,g(i) ). i=1 Then V 1 = µ 10 + V 10 + (1 B 1 ) 1 Another representation for V 10 is i=1 k D i (V 20,f(i) V 20,g(i) ). i=1 V 10 = QV 20 + (1 B 1 ) 1 ε 1 where the matrix Q combines all the coefficients from the terms involving V 20. The above equation is essentially the definition of Q. Note now that Cov(ε 1, V 20 ) = F ((1 B 2 ) 1 ) T and thus V ar(v 10 ) = QΣ 2 Q T + (1 B 1 ) 1 Θ((1 B 1 ) 1 ) T + Q(1 B 2 ) 1 F T ((1 B 1 ) 1 ) T + (1 B 1 ) 1 F ((1 B 2 ) 1 ) T Q T. Using the fact that the covariance between V 20,f(i) V 20,g(i) and V 20 and the covariance between V 20,f(i) V 20,g(i) and ε 1 are zero we get that V ar(v 1 ) = V ar(v 10 ) + i,j D i Cov(V 20,f(i) V 20,g(i), V 20,f(j) V 20,g(j) )D T j = 12

V ar(v 10 ) + i,j D i D T j (Σ 2,f(i),f(j) Σ 2,g(i),g(j) + Σ 2,f(i),g(j) Σ 2,g(i),f(j) ). Note also that Cov(V 1, V 2 ) = Cov(V 10, V 20 ) = QΣ 2 + (1 B 1 ) 1 F ((1 B 2 ) 1 ) T. 13