STRUCTURAL EQUATION MODELING. Khaled Bedair Statistics Department Virginia Tech LISA, Summer 2013

Similar documents
Chapter 8. Models with Structural and Measurement Components. Overview. Characteristics of SR models. Analysis of SR models. Estimation of SR models

Introduction to Structural Equation Modeling Dominique Zephyr Applied Statistics Lab

Factor analysis. George Balabanis

Compiled by: Assoc. Prof. Dr Bahaman Abu Samah Department of Professional Developmentand Continuing Education Faculty of Educational Studies

Hypothesis Testing for Var-Cov Components

Structural Equation Modelling

An Introduction to Path Analysis

Introduction to Structural Equation Modeling

1 Motivation for Instrumental Variable (IV) Regression

Chapter 5. Introduction to Path Analysis. Overview. Correlation and causation. Specification of path models. Types of path models

STAT 730 Chapter 9: Factor analysis

Path Analysis. PRE 906: Structural Equation Modeling Lecture #5 February 18, PRE 906, SEM: Lecture 5 - Path Analysis

An Introduction to Mplus and Path Analysis

SC705: Advanced Statistics Instructor: Natasha Sarkisian Class notes: Introduction to Structural Equation Modeling (SEM)

Confirmatory Factor Analysis: Model comparison, respecification, and more. Psychology 588: Covariance structure and factor models

C:\Users\Rex Kline\AppData\Local\Temp\AmosTemp\AmosScratch.amw. Date: Friday, December 5, 2014 Time: 11:20:30 AM

Multiple group models for ordinal variables

Factor Analysis. Qian-Li Xue

Introduction to Factor Analysis

Confirmatory Factor Analysis. Psych 818 DeShon

4. Path Analysis. In the diagram: The technique of path analysis is originated by (American) geneticist Sewell Wright in early 1920.

Introduction to Confirmatory Factor Analysis

SHOPPING FOR EFFICIENT CONFIDENCE INTERVALS IN STRUCTURAL EQUATION MODELS. Donna Mohr and Yong Xu. University of North Florida

y response variable x 1, x 2,, x k -- a set of explanatory variables

STA 431s17 Assignment Eight 1

A Study of Statistical Power and Type I Errors in Testing a Factor Analytic. Model for Group Differences in Regression Intercepts

Belowisasamplingofstructuralequationmodelsthatcanbefitby sem.

Chapter 3: Testing alternative models of data

LECTURE 4 PRINCIPAL COMPONENTS ANALYSIS / EXPLORATORY FACTOR ANALYSIS

Preface. List of examples

Structural Equation Modeling and Confirmatory Factor Analysis. Types of Variables

Longitudinal Invariance CFA (using MLR) Example in Mplus v. 7.4 (N = 151; 6 items over 3 occasions)

Nesting and Equivalence Testing

Model Invariance Testing Under Different Levels of Invariance. W. Holmes Finch Brian F. French

A study on new customer satisfaction index model of smart grid

Statistics Handbook. All statistical tables were computed by the author.

Related Concepts: Lecture 9 SEM, Statistical Modeling, AI, and Data Mining. I. Terminology of SEM

Systematic error, of course, can produce either an upward or downward bias.

Psychology 830. Introduction to Covariance Structure Modeling

9 Correlation and Regression

SRMR in Mplus. Tihomir Asparouhov and Bengt Muthén. May 2, 2018

BTRY 4830/6830: Quantitative Genomics and Genetics Fall 2014

Introduction to Matrix Algebra and the Multivariate Normal Distribution

2/26/2017. PSY 512: Advanced Statistics for Psychological and Behavioral Research 2

CHAPTER 9 EXAMPLES: MULTILEVEL MODELING WITH COMPLEX SURVEY DATA

Intermediate Social Statistics

Factor Analysis & Structural Equation Models. CS185 Human Computer Interaction

Wooldridge, Introductory Econometrics, 4th ed. Chapter 15: Instrumental variables and two stage least squares

Review of Classical Least Squares. James L. Powell Department of Economics University of California, Berkeley

Multiple Group CFA Invariance Example (data from Brown Chapter 7) using MLR Mplus 7.4: Major Depression Criteria across Men and Women (n = 345 each)

A Cautionary Note on the Use of LISREL s Automatic Start Values in Confirmatory Factor Analysis Studies R. L. Brown University of Wisconsin

How to run the RI CLPM with Mplus By Ellen Hamaker March 21, 2018

Dimensionality Reduction Techniques (DRT)

1 Correlation and Inference from Regression

6. Let C and D be matrices conformable to multiplication. Then (CD) =

Just Enough Likelihood

SEM for those who think they don t care

Chapter 4: Factor Analysis

Short Answer Questions: Answer on your separate blank paper. Points are given in parentheses.

Do not copy, post, or distribute

Glossary. The ISI glossary of statistical terms provides definitions in a number of different languages:

Assessing Factorial Invariance in Ordered-Categorical Measures

Mixture Modeling in Mplus

sempower Manual Morten Moshagen

STAT 501 EXAM I NAME Spring 1999

Structural Equation Modeling

Specifying Latent Curve and Other Growth Models Using Mplus. (Revised )

EVALUATION OF STRUCTURAL EQUATION MODELS

Latent Variable Analysis

Path Diagrams. James H. Steiger. Department of Psychology and Human Development Vanderbilt University

Unless provided with information to the contrary, assume for each question below that the Classical Linear Model assumptions hold.

Analysis of Variance and Co-variance. By Manza Ramesh

A guide to multiple-sample structural equation modeling

What is Structural Equation Modelling?

General structural model Part 1: Covariance structure and identification. Psychology 588: Covariance structure and factor models

SC705: Advanced Statistics Instructor: Natasha Sarkisian Class notes: Model Building Strategies

Package SEMModComp. R topics documented: February 19, Type Package Title Model Comparisons for SEM Version 1.0 Date Author Roy Levy

CONFIRMATORY FACTOR ANALYSIS

Can Variances of Latent Variables be Scaled in Such a Way That They Correspond to Eigenvalues?

Confirmatory Factor Analysis

1 Least Squares Estimation - multiple regression.

Model Estimation Example

PLEASE SCROLL DOWN FOR ARTICLE

Equation Number 1 Dependent Variable.. Y W's Childbearing expectations

Econ 836 Final Exam. 2 w N 2 u N 2. 2 v N

Assignment 1. SEM 2: Structural Equation Modeling

Evaluation of structural equation models. Hans Baumgartner Penn State University

STRUCTURAL EQUATION MODEL (SEM)

LISREL Workshop at M3 Conference 2015

The Common Factor Model. Measurement Methods Lecture 15 Chapter 9

Relating Latent Class Analysis Results to Variables not Included in the Analysis

Factor Analysis: An Introduction. What is Factor Analysis? 100+ years of Factor Analysis FACTOR ANALYSIS AN INTRODUCTION NILAM RAM

Condition 9 and 10 Tests of Model Confirmation with SEM Techniques

Econometric Analysis of Cross Section and Panel Data

7 Principal Components and Factor Analysis

Using Structural Equation Modeling to Conduct Confirmatory Factor Analysis

Causal Inference Using Nonnormality Yutaka Kano and Shohei Shimizu 1

SEM Day 3 Lab Exercises SPIDA 2007 Dave Flora

Test 3 Practice Test A. NOTE: Ignore Q10 (not covered)

The 3 Indeterminacies of Common Factor Analysis

Transcription:

STRUCTURAL EQUATION MODELING Khaled Bedair Statistics Department Virginia Tech LISA, Summer 2013

Introduction: Path analysis Path Analysis is used to estimate a system of equations in which all of the variables are observed. This type of model is often used when one or more variables is thought to mediate the relationship between two others (mediation models). Models where the relationship between variables is thought to vary across groups (multiple group models). Similar models setups can be used to estimate models where the errors (residuals) of two otherwise unrelated dependent variables are allowed to correlated (seemingly unrelated regression)

Introduction: Confirmatory factor analysis (CFA) Confirmatory factor analysis (CFA) is a statistical technique used to verify the factor structure of a set of observed variables. CFA allows the researcher to test the hypothesis that a relationship between observed variables and their underlying latent constructs exists. The researcher uses knowledge of the theory, empirical research, or both, postulates the relationship pattern a priori and then tests the hypothesis statistically.

Introduction:Structural Equation Models (SEM) SEM is used to test complex 'relationships between observed (measured) and unobserved (latent) variables and also relationships between two or more latent variables.

Outline We introduce an example to demonstrate structural modeling with time-related latent variables. The use of modification indices and critical ratios in exploratory analyses. How to compare multiple models in a single analysis, and computation of implied moments, factor score weights, total effects, and indirect effects.

About the Data Wheaton et al. (1977) reported a longitudinal study of 932 persons over the period from 1966 to 1971. Six of Wheaton's measures will be used for this example. Measure Explanation Anomia67: 1967 score on the anomia scale Anomia71: 1971 anomia score Powles67: 1967 score on the powerlessness scale Powles71: 1971 powerlessness score Education: Years of schooling recorded in 1966 SEI Duncan's Socioeconomic: Index administered in 1966

About the Data Basically all SEM computer tools accept either a raw data file or a matrix summary of the data. There are two basic types of summaries of raw data correlation matrices with standard deviations and covariance matrices.

Specifying the Model To specify Model A, draw the path diagram shown

Model A The model asserts that all of the observed variables depend on underlying, unobserved variables. For example, anomia67 and powles67 both depend on the unobserved variable 67_alienation. The eight unique variables (delta1, delta2, zeta1, zeta2, and eps1 through eps4) are uncorrelated among themselves and with the three latent variables: ses, 67_alienation, and 71_alienation. In this case, eps1 and eps2 should be thought of as representing not only errors of measurement in anomia67 and powles67 but in every other variable that might affect scores on the two tests.

Identification There is just not enough information. We can solve this identification problem by fixing either The regression weight applied to errors, or the variance of the errors variable itself, at an arbitrary, nonzero value. Let s fix the regression weight for the errors at 1. In that path diagram, three weights are fixed at 1, which is one fixed regression weight for each unobserved variable. These constraints are sufficient to make the model identified.

Assumptions In maximum likelihood or generalized least-squares estimation depend on certain assumptions. First, observations must be independent. Second, the observed variables must meet some distributional requirements. Multivariate normality of all observed variables is a standard distribution assumption. If some exogenous variables are fixed (that is, they are either known before hand or measured without error), their distributions may have any shape,.

Result for Model A

Results of the Analysis The model has 15 parameters to be estimated (6 regression weights and 9 variances). There are 21 sample moments (6 sample variances and 15 covariances). This leaves 6 degrees of freedom The data depart significantly from Model A. Chi-square = 71.544 Degrees of freedom = 6 Probability level = 0.000

Dealing with Rejection We have several options when a proposed model has to be rejected on statistical grounds: We can point out that statistical hypothesis testing can be a poor tool for choosing a model. It is a widely accepted view that a model can be only an approximation at best, and that, fortunately, a model can be useful. You can start from scratch to devise another model to substitute for the rejected one. You can try to modify the rejected model in small ways so that it fits the data better. It is the last tactic that will be demonstrated in this example

Dealing with Rejection The most natural way of modifying a model to make it fit better is to relax some of its assumptions. For example, Model A assumes that eps1 and eps3 are uncorrelated. You could relax this restriction by connecting eps1 and eps3 with a double-headed arrow. The model also specifies that anomia67 does not depend directly on ses. You could remove this assumption by drawing a single-headed arrow from ses to anomia67.

Dealing with Rejection Model A does not happen to constrain any parameters to be equal to other parameters, But if such constraints were present, you might consider removing them in hopes of getting a better fit. Of course, you have to be careful when relaxing the assumptions of a model that you do not turn an identified model into an unidentified one.

Modification Indices You can test various modifications of a model by carrying out a separate analysis for each potential modification, but this approach is time-consuming. Modification indices allow you to evaluate many potential modifications in a single analysis. They provide suggestions for model modifications that are likely to pay off in smaller chi square values.

Using Modification Indices The following are the modification indices for Model A

Using Modification Indices The first modification index listed (5.905) is a conservative estimate of the decrease in chi-square that will occur if eps2 and delta1 are allowed to be correlated. The new chi-square statistic would have 5 degrees of freedom and would be no greater than 65.639. The actual decrease of the chisquare statistic might be much larger than 5.905. The column labeled Par Change gives approximate estimates of how much each parameter would change if it were estimated rather than fixed at 0. it would be hard to justify this particular modification on theoretical grounds even if it did produce an acceptable fit.

Using Modification Indices The largest modification index in Model A is 40.911. It indicates that allowing eps1 and eps3 to be correlated will decrease the chi-square statistic by at least 40.911. In fact, you would expect the correlation to be positive, which is consistent with the fact that the number in the Par Change column is positive.

Using Modification Indices The theoretical reasons for suspecting that eps1 and eps3 might be correlated apply to eps2 and eps4 as well. The modification indices also suggest allowing eps2 and eps4 to be correlated. However, we will ignore this potential modification and proceed immediately to look at the results of modifying Model A by allowing eps1 and eps3 to be correlated. The new model is Jöreskog and Sörbom s Model B.

Model B The path diagram for Model B

Model B The added covariance between eps1 and eps3 decreases the degrees of freedom by 1.The chi-square statistic is reduced by substantially more than the promised 40.911. Model A: Chi-square = 71.544 Degrees of freedom = 6 Probability level = 0.000 Model B: Chi-square = 6.383 Degrees of freedom = 5 Probability level = 0.271 Model B cannot be rejected. Since the fit of Model B is so good.

Text Output The large critical ratio associated with the new covariance path. The covariance between eps1 and eps3 is clearly different from 0. This explains the poor fit of Model A, in which that covariance was fixed at 0.

Misuse of Modification Indices A modification should be considered only if it makes theoretical or common sense. A slavish reliance on modification indices without such a limitation amounts to sorting through a very large number of potential modifications in search of one that provides a big improvement in fit. Such a strategy is prone, through capitalization on chance, to producing an incorrect (and absurd) model that has an acceptable chi-square value.

Critical Ratios The Critical Ratio R.C. technique introduces additional constraints in such a way as to produce a relatively large increase in degrees of freedom, coupled with a relatively small increase in the chi-square statistic. The square of the critical ratio C.R. of a parameter is, approximately, the amount by which the chi-square statistic will increase if the analysis is repeated with that parameter fixed at 0.

Critical Ratio We might be able to improve the chi-square test of fit by postulating a new model where those two parameters are specified to be exactly equal. we need to provide a critical ratio for every pair of parameters.

Critical Ratio The square of the critical ratio is approximately the amount by which the chi-square statistic would increase if the two parameters were set equal to each other. The smallest critical ratio in the table is 0.077, for the parameters labeled par_14 and par_15. These two parameters are the variances will result in a chi-square value that exceeds 6.383 by about 0.006, but with 6 degrees of freedom instead of 5. The associated probability level would be about 0.381. The only problem with this modification is that there does not appear to be any justification for it.

Critical Ratio The variances of eps1 and eps3 (par_11 and par_13) differ with a critical ratio of 0.51. The variances of eps2 and eps4 (par_12 and par_14) differ with a critical ratio of 1.00. The weights for the regression of powerlessness on alienation (par_1 and par_2) differ with a critical ratio of 0.88. None of these differences, taken individually, is significant at any conventional significance level. This suggests that it may be worthwhile to investigate more carefully a model in which all three differences are constrained to be 0. We will call this new model Model C.

Model C

Parameter estimate Model C

Model C Model A: Sample moments = 21 parameters = 15 Chi-square = 71.544 Degrees of freedom = 6 Probability level = 0.000 Model B: Sample moments = 21 parameters =16 Chi-square = 6.383 Degrees of freedom = 5 Probability level = 0.271 Model C: Sample moments = 21 parameters = 13 Chi-square = 7.501 Degrees of freedom = 8 Probability level = 0.484 We can test Model C against Model B by examining the difference in chi-square values ( 7.501-6.383) and the difference in degrees of freedom (8-5 ). A chi-square value of 1.118 with 3 degrees of freedom is not significant.

Output form multiple Models Model D incorporates the constraints of both Model A and Model C. The following is the portion of the output that shows the chisquare statistic: The CMIN column contains the minimum discrepancy for each model. In the case of maximum likelihood estimation (the default), the CMIN column contains the chi-square statistic. The p column contains the corresponding upper-tail probability for testing each model.

This table shows, for example, that Model C does not fit significantly worse than Model B (p = 0.773 ). In other words, assuming that Model B is correct, you would accept the hypothesis of time invariance. On the other hand, the table shows that Model A fits significantly worse than Model B ( P=00). In other words, assuming that Model B is correct, you would reject the hypothesis that eps1 and eps3 are uncorrelated. Model Comparisons

Implied Covariances i The implied variances and covariances for the observed variables are not the same as The sample variances and covariances. As estimates of the corresponding population values. The implied variances and covariances are superior to the sample variances and covariances (assuming that Model C is correct). i

Factor Score Weights The table of factor score weights has a separate row for each unobserved variable. A separate column for each observed variable. Suppose you wanted to estimate the ses score of an individual. You would compute a weighted sum of the individual s six observed scores using the six weights in the ses row of the table.

Total effect The coefficients associated with the single-headed arrows in a path diagram are sometimes called direct effects. In Model C, for example, ses has a direct effect on 71_alienation. In turn, 71_alienation has a direct effect on powles71. Ses is then said to have an indirect effect (through the intermediary of 71_alienation) on powles71.

Total Effect The first row of the table indicates that 67_alienation depends, directly or indirectly, on ses only. The total effect of ses on 67_alienation is 0.56. Looking in the fifth row of the table, powles71 depends, directly or indirectly, on ses, 67_alienation, and 71_alienation. Low scores on ses, high scores on 67_alienation, and high scores on71_alienation are associated with high scores on powles71.

Thanks Comments and Questions???