Introduction to Confirmatory Factor Analysis

Similar documents
Instrumental variables regression on the Poverty data

Regression without measurement error using proc calis

SAS Example 3: Deliberately create numerical problems

CONFIRMATORY FACTOR ANALYSIS

EVALUATION OF STRUCTURAL EQUATION MODELS

UNIVERSITY OF TORONTO MISSISSAUGA April 2009 Examinations STA431H5S Professor Jerry Brunner Duration: 3 hours

C:\Users\Rex Kline\AppData\Local\Temp\AmosTemp\AmosScratch.amw. Date: Friday, December 5, 2014 Time: 11:20:30 AM

Confirmatory Factor Analysis. Psych 818 DeShon

* IVEware Analysis Examples Replication for ASDA 2nd Edition * Berglund April 2017 * Chapter 13 ;

Introduction to Structural Equation Modeling with Latent Variables

THE GENERAL STRUCTURAL EQUATION MODEL WITH LATENT VARIATES

The Role of Leader Motivating Language in Employee Absenteeism (Mayfield: 2009)

Chapter 13 Introduction to Structural Equation Modeling

An Introduction to Path Analysis

Factor analysis. George Balabanis

An Introduction to Mplus and Path Analysis

Introduction to Confirmatory Factor Analysis

The use of structural equation modeling to examine consumption of pornographic materials in Chinese adolescents in Hong Kong

Psychology 454: Latent Variable Modeling How do you know if a model works?

Inference using structural equations with latent variables

Structural Equation Modeling and Confirmatory Factor Analysis. Types of Variables

SC705: Advanced Statistics Instructor: Natasha Sarkisian Class notes: Model Building Strategies

Model Estimation Example

DATE: 9/ L I S R E L 8.80

Introduction to Structural Equation Modeling Dominique Zephyr Applied Statistics Lab

Chapter 3: Testing alternative models of data

Evaluation of structural equation models. Hans Baumgartner Penn State University

sempower Manual Morten Moshagen

Multi-sample structural equation models with mean structures, with special emphasis on assessing measurement invariance in cross-national research

Introduction to Structural Equation Modeling

Condition 9 and 10 Tests of Model Confirmation with SEM Techniques

Confirmatory Factor Analysis: Model comparison, respecification, and more. Psychology 588: Covariance structure and factor models

2/26/2017. PSY 512: Advanced Statistics for Psychological and Behavioral Research 2

SEM Day 1 Lab Exercises SPIDA 2007 Dave Flora

STRUCTURAL EQUATION MODELING. Khaled Bedair Statistics Department Virginia Tech LISA, Summer 2013

Multiple Group CFA Invariance Example (data from Brown Chapter 7) using MLR Mplus 7.4: Major Depression Criteria across Men and Women (n = 345 each)

The Common Factor Model. Measurement Methods Lecture 15 Chapter 9

Chapter 8. Models with Structural and Measurement Components. Overview. Characteristics of SR models. Analysis of SR models. Estimation of SR models

Longitudinal Invariance CFA (using MLR) Example in Mplus v. 7.4 (N = 151; 6 items over 3 occasions)

Structural Equation Modeling

Structural Equation Modeling

2.2 Classical Regression in the Time Series Context

CONFIRMATORY FACTOR ANALYSIS

Structural Equation Modelling

Psychology 454: Latent Variable Modeling How do you know if a model works?

THE GENERAL STRUCTURAL EQUATION MODEL WITH LATENT VARIATES

SRMR in Mplus. Tihomir Asparouhov and Bengt Muthén. May 2, 2018

Multi-group analyses for measurement invariance parameter estimates and model fit (ML)

Related Concepts: Lecture 9 SEM, Statistical Modeling, AI, and Data Mining. I. Terminology of SEM

The comparison of estimation methods on the parameter estimates and fit indices in SEM model under 7-point Likert scale

Using Structural Equation Modeling to Conduct Confirmatory Factor Analysis

Confirmatory Factor Models (CFA: Confirmatory Factor Analysis)

Factor Analysis & Structural Equation Models. CS185 Human Computer Interaction

Maximum Likelihood Estimation; Robust Maximum Likelihood; Missing Data with Maximum Likelihood

Recovery of weak factor loadings in confirmatory factor analysis under conditions of model misspecification

Introduction to Matrix Algebra and the Multivariate Normal Distribution

Exploring Cultural Differences with Structural Equation Modelling

This course. Tutors. Jon Heron, PhD (Bristol) Anna Brown, PhD (Cambridge)

Case of single exogenous (iv) variable (with single or multiple mediators) iv à med à dv. = β 0. iv i. med i + α 1

AN INTRODUCTION TO STRUCTURAL EQUATION MODELING WITH AN APPLICATION TO THE BLOGOSPHERE

Title. Description. Remarks and examples. stata.com. stata.com. Variable notation. methods and formulas for sem Methods and formulas for sem

Compiled by: Assoc. Prof. Dr Bahaman Abu Samah Department of Professional Developmentand Continuing Education Faculty of Educational Studies

SHOPPING FOR EFFICIENT CONFIDENCE INTERVALS IN STRUCTURAL EQUATION MODELS. Donna Mohr and Yong Xu. University of North Florida

DIAGNOSTICS OF THE ERROR FACTOR COVARIANCES OTTÓ HAJDU 1

Chapter 1 Statistical Inference

LECTURE 4 PRINCIPAL COMPONENTS ANALYSIS / EXPLORATORY FACTOR ANALYSIS

Introduction to Within-Person Analysis and RM ANOVA

Path Analysis. PRE 906: Structural Equation Modeling Lecture #5 February 18, PRE 906, SEM: Lecture 5 - Path Analysis

Hypothesis Testing for Var-Cov Components

Factor Analysis. Qian-Li Xue

Dimensionality Reduction Techniques (DRT)

SAS/STAT 14.1 User s Guide. Introduction to Structural Equation Modeling with Latent Variables

A Re-Introduction to General Linear Models (GLM)

Manual Of The Program FACTOR. v Windows XP/Vista/W7/W8. Dr. Urbano Lorezo-Seva & Dr. Pere Joan Ferrando

Evaluating the Sensitivity of Goodness-of-Fit Indices to Data Perturbation: An Integrated MC-SGR Approach

Mixture Modeling. Identifying the Correct Number of Classes in a Growth Mixture Model. Davood Tofighi Craig Enders Arizona State University

UNIVERSITY OF CALGARY. The Influence of Model Components and Misspecification Type on the Performance of the

Applied Multivariate Analysis

Nesting and Equivalence Testing

Model fit evaluation in multilevel structural equation models

Improper Solutions in Exploratory Factor Analysis: Causes and Treatments

Spatial inference. Spatial inference. Accounting for spatial correlation. Multivariate normal distributions

36-309/749 Experimental Design for Behavioral and Social Sciences. Dec 1, 2015 Lecture 11: Mixed Models (HLMs)

Step 2: Select Analyze, Mixed Models, and Linear.

PLEASE SCROLL DOWN FOR ARTICLE. Full terms and conditions of use:

Factor Analysis: An Introduction. What is Factor Analysis? 100+ years of Factor Analysis FACTOR ANALYSIS AN INTRODUCTION NILAM RAM

SEM for Categorical Outcomes

Analysis of Longitudinal Data: Comparison Between PROC GLM and PROC MIXED. Maribeth Johnson Medical College of Georgia Augusta, GA

Effect of Estimation Method on Incremental Fit Indexes for Covariance Structure Models

Exploratory Factor Analysis and Principal Component Analysis

Statistical Distribution Assumptions of General Linear Models

Univariate Normal Distribution; GLM with the Univariate Normal; Least Squares Estimation

E X P L O R E R. R e l e a s e A Program for Common Factor Analysis and Related Models for Data Analysis

dm'log;clear;output;clear'; options ps=512 ls=99 nocenter nodate nonumber nolabel FORMCHAR=" = -/\<>*"; ODS LISTING;

Measuring relationships among multiple responses

A Cautionary Note on the Use of LISREL s Automatic Start Values in Confirmatory Factor Analysis Studies R. L. Brown University of Wisconsin

SEM Analysis of Epigenetic Data

Measurement Invariance Testing with Many Groups: A Comparison of Five Approaches (Online Supplements)

Introduction to Bayesian Statistics and Markov Chain Monte Carlo Estimation. EPSY 905: Multivariate Analysis Spring 2016 Lecture #10: April 6, 2016

Exploratory Factor Analysis and Principal Component Analysis

Transcription:

Introduction to Confirmatory Factor Analysis In Exploratory FA, the analyst has no strong prior notion of the structure of the factor solution the goal is to infer factor structure from the patterns of correlation in the data Exploratory FA suggests hypotheses, but does not justify knowledge. Confirmatory FA Confirmatory FA is a theory-testing model as opposed to a theory-generating method In confirmatory FA, the analyst tests if the prior notion regarding the factor solution is consistent with the patterns in the data. The model, or hypothesis, that is tested specifies which variables will be correlated with which factors and which factors are correlated. The hypothesis is usually based on a strong theoretical and/or empirical foundation. With confirmatory FA, it is possible to measure the goodness-of-fit of a particular model and to statistically test the adequacy of the model fit. Confirmatory FA also offers the analyst a more viable method for evaluating construct validity. A construct is said to be reliably measured (have construct validity) when multiple measurements of the same construct using different methods are highly correlated. Ex. Menezes and Elbert (1979), as cited in Lattin, Carroll and Green (2003) used multiple measurement methods to examine multiple characteristics of the stores in a major local grocery chain. They selected two traits for their study: store appearance (A) and product assortment (P). Each trait was measured using three different measurement methods: the Likert scale, the semantic differential scale, and the Stapel scale: 1. Likert scale Selection is wide. SA GA MA MD GD SD Where SA = strongly agree, GA =generally agree, MA = moderately agree and SD, GD, MD correspond to disagree. 2. Semantic differential scale Extremely Quite Slight Slight Quite Extremely Wide Selection Limited Selection 36

3. Stapel scale +3 +2 +1 Wide Selection -1-2 -3 Menezes and Elbert tested whether any one scale is any more reliable and concluded that that there were no marked differences across the 3 methods. Confirmatory FA can also be used to test for discriminant validity, which in the context of the Menezres and Elbert example, holds when the two traits store appearance and product assortment can be separated into distinct factors. Confirmatory FA Procedure 1. The analyst begins with a correlation matrix or a variance/covariance matrix. 2. The analyst then proposes competing models, based on theory or existing data, that are hypothesized to fit the data. The models may specify the degree of correlation, if any, between each pair of common factors; the degree of correlation between individual variables and one or more factors, which particular pairs of unique factors are correlated. The different models are determined by "fixing" or "freeing" specific parameters such as: the factor loadings or coefficients, the factor correlation coefficients, and the variance/covariance of the error of measurement. These parameters are set according to the theoretical expectation of the analyst. One fixes a parameter by setting the parameter at a specific value based on one's expectations. One frees a parameter by allowing the parameter to be estimated during the analysis. 3. The competing models or hypotheses about the structure of the data are then tested against one another with the use of goodness-of-fit statistics. The fit statistics determine how well the competing models fit the data, or explain the covariation among the variables. Examples of these statistics include the chi square/degrees of freedom ratio, the Bentler comparative fit index (CFI), the parsimony ratio, and the Goodness-of-fit Index (GFI). 37

Example. Holzinger and Swineford s (1939) study on psychological testing of 7 th and 8 th grade children: X 1 = paragraph comprehension X 2 = sentence completion X 3 = word meaning X 4 = addition X 5 = counting dots The correlation matrix based on a sample of 145 children is given by 1.000.722.714.203.095 1.000.685.246.181 R = 1.000.170.113 1.000.585 1.000 An exploratory FA with two factors in a sense hypothesizes that all five variables load on both factors: X 1 = γ 11 F 1 + γ 12 F 2 + ε 1 X 2 = γ 21 F 1 + γ 22 F 2 + ε 2 X 3 = γ 31 F 1 + γ 32 F 2 + ε 3 X 4 = γ 41 F 1 + γ 42 F 2 + ε 4 X 5 = γ 51 F 1 + γ 52 F 2 + ε 5 But we may believe that two factors underlie test performance: verbal ability and quantitative ability. Moreover, we may believe that the two factors are distinct, but are positively correlated. The following FA model may thus be hypothesized: X 1 = γ 11 F 1 + ε 1 X 2 = γ 21 F 1 + ε 2 X 3 = γ 31 F 1 + ε 3 X 4 γ 42 F 2 + ε 4 X 5 = γ 52 F 2 + ε 5 Corr(F 1,F 2 ) = ϕ 12. We thus fixed γ 12 = γ 22 = γ 32 = γ 41 = γ 51 = 0. We can further specify that var(f 1 )=var(f 2 )=1, while allowing Corr(F 1,F 2 ) = ϕ 12 to be freely estimated. Note that with this model, we have 15 pieces of information (the five 1 s on the main diagonal and the 10 off-diagonal elements of R) with which to estimate 11 model parameters: the five γ loadings, ϕ 12 and the five specific variances (variances of the ε s). 38

Because there are more observations (15) than unknown parameters (11), this model is overdetermined. In general, with an overdetermined model, it will not be possible to fit the observed values exactly. Having more observations than model parameters is only a necessary, but not sufficient, condition for each model parameter to be identified. However, if the model exhibits a maximally simple structure (i.e., each variable loads on only one factor), error terms are independent, and the number of factors is small relative to the number of measures, then the model is identified. Estimation Methods and Fitting Functions There are various estimation methods available for CFA. These methods all obtain parameter estimates collected in a vector θhat by iteratively minimizing a certain function F[S, Σ(θhat)] so that S and Σ(θhat) are close or, equivalently, the elements in the residual matrix S - Σ(θhat) are small. F is called the fitting function and its form depends on the estimation method used. Five estimation methods are available under PROC CALIS of SAS. These are Unweighted least squares (ULS) Generalized Least Squares (GLS) Normal theory Maximum Likelihood (ML) Weighted Least Squares (WLS, ADF) Diagonally Weighted Least Squares (DWLS). The default is GLS. Let S denote the sample correlation or covariance matrix for a sample size N and let C be the predicted moment matrix [i.e., C =Σ(θhat) ]. The fitting functions are as follows F ULS = 0.5 Tr(S C) 2 F GLS = 0.5 Tr(S -1 (S C)) 2 F ML = Tr(SC -1 )- n +log(det C) log(det S), where n = number of manifest variables F WLS = Vec(s ij - c ij )' W -1 Vec(s ij - c ij ), where Vec(s ij - c ij ) denotes the vector of the n(n+1)/2 elements of the lower triangle of the symmetric matrix S - C, and W = (w ij,kl ) is a positive definite symmetric matrix with n(n+1)/2 rows and columns If the moment matrix S is a correlation (rather than a covariance) matrix, the default setting of W = (w ij,kl ) is the estimators of the asymptotic covariances correlations S = (s ij ). of the 39

F DWLS = Vec(s ij - c ij )' diag(w -1 )Vec(s ij - c ij ) Using only the diagonal elements of W -1 eases computational burden. But statistical properties of DWLS estimates are still not known. The Fit Statistics Chi square statistic and Chi square/degrees of freedom ratio The chi square test tests the hypothesis that the model is consistent with the pattern of covariation among the observed variables. Specifically, it tests the null hypothesis that there is no statistically significant difference in the observed covariance matrix and the theoretical covariance structure matrix implied by the model. Under Ho: Σ = Σ(θ) and certain distributional assumptions (e.g., multivariate normality if the ML method is used), χ 2 = (n -1) F[S, Σ(θhat)] is asymptotically chi-square with degrees of freedom equal to the difference (c p) where c is the number of nonredundant variances and covariances of the observed variables and p the number of parameters to be estimated. Here n is the sample size. In the case of the chi-square statistic, smaller rather than larger values indicate a good fit. Suppose an SEM model with 5 degrees of freedom yields a discrepancy function of χ 2 = 16.37 with a p-value of 0.01. This implies that the model with 5 degrees of freedom is significantly worse than if it had estimated 5 more parameters and had 0 degrees of freedom. The chi-square statistic is very sensitive to sample size. When the sample size is large, even the smallest discrepancies between the fitted model matrix and the observed data matrix will be judged significant. In such situations, it is unclear whether the statistical significance of the chi square statistic is due to poor fit of the model or to the size of the sample. Thus, alternative fit statistics were developed. Because of its limitations, some researchers advocate that the chi-square not be used as a formal test statistic. They instead suggest that divide the Chi-square value be divided by its numbers of degrees of freedom to obtain the Chi square ratio. rule of thumb well-fitted: ratio < 2 acceptable: ratio < 3 not acceptable: ratio > 5 40

Goodness-of-fit index (GFI) and adjusted goodness-of-fit index (AGFI) The good of fit index "is a measure of the relative amount of variances and covariances jointly accounted for by the model". It can be thought of as being roughly analogous to the multiple R squared in multiple regression. The closer the GFI is to 1.00, the better is the fit of the model to the data. GFI = 1 - F[S, Σ(θhat)] F[S, Σ(θ)] Here the numerator is the minimum value of the fitting function F for the hypothesized model, while the denominator is the minimum value of F when there is no hypothesized model. Thus GFI measures how much better the model fits compared to no model at all. Rules of thumb: good fit: GFI > 0.95 acceptable fit: GFI > 0.90. The adjusted goodness of fit statistic (AGFI) is based on a correction for the number of degrees of freedom in a less restricted model obtained by freeing more parameters. It penalizes more complex models (i.e., those with more parameters to be estimated) with a downward adjustment in the fit index. c AGFI = 1 - df h F[S, Σ(θhat)] F[S, Σ(θ)] Here c is as before and df h = c p, the df for the hypothesized model. Rules of thumb: good fit: AGFI > 0.90 acceptable fit: AGFI > 0.80. Both the GFI and the AGFI are less sensitive to sample size than the chi square statistic. The Normed and Nonnormed Fit Index and Their Extensions These measures fit by comparing a hypothesized model to some reasonable but more restrictive (fewer parameters to be estimated) baseline model. The independence model is usually taken to be the baseline model. Under the independence model (1) no latent variables are hypothesized to underlie the observed variables; (2) observed variables are measured without error; and the observed variables are independent. Let F h denote the minimum value of the fitting function F for the hypothesized model with associated degrees of freedom df h. Similarly, let F i denote the minimum value of F for the hypothesized model with associated degrees of freedom df i. 41

χ 2 2 i - χ h F h NFI = = 1 -. 2 χ i F i The Normed Fit Index (NFI) is normed to a 0 to 1 range with large values ( >.90) indicating a better fit. However, the NFI is affected by sample size and might not equal 1 even if the correct model is hypothesized. Thus the NFI model often underestimates model fit especially for small samples. The Nonnormed Fit Index (NNFI) penalizes more complex (or less restrictive) models by a downward adjustment and rewards more parsimonious (more restrictive models) by an upward adjustment. As its name suggests, the NNFI is not normed and may take values outside the 0 to 1 range. χ 2 2 i df i. - χ h df h NNFI =. χ 2 i df i. - 1 The normed Comparative Fit Index (CFI) was introduced to address the deficiency of the NFI, while the nonnormed Fit Index (FI) was introduced to address the deficiency of the NNFI. Let l h = [(n 1 ) F h - df h ] and l i = [(n 1 ) F i df i ]. Then CFI = 1 - l 1 l 2 where l 1 = max ( l h, 0) and l 2 = max ( l h, l i, 0). l h FI = 1 -. l i CFI is restricted to a 0 to 1 range while FI is not. Comparisons of CFI and FI with NFI and NNFI show that underestimation of fit happens less with CFI than with NFI and that FI behaves better than NNFI at the endpoints of the [0,1] interval in that it is not as frequently negative as NNFI and when it exceeds 1, it does so by a smaller amount. CFI is also less affected by sample size than NFI. Values above 0.90 are considered to be indicative of good overall fit for NFI, NNFI CFI and FI. 42

The Parsimonious Fit Indices and the Parsimony Ratio The parsimony fit indices (Parsimonious Goodness of Fit Index or PGFI and Parsimony Normed Fit Index or PNFI) and the parsimony ratio take into consideration the number of parameters estimated in the model. The fewer number of parameters necessary to specify the model, the more parsimonious is the model. By multiplying the parsimony ratio by a fit statistic, an index of both the overall efficacy of the model explaining the covariance among the variables and the parsimony of the proposed model is obtained. df h PGFI = GFI df n PNFI = df h df i NFI where df h is the degrees of freedom associated with the hypothesized model, df n = c, the degrees of freedom when no model is hypothesized (the number of nonredundant variances and covariances of the observed variables) and df i is the degrees of freedom for the independence baseline model. Both the AGFI and PGFI adjust the GFI downward for model complexity, but the PGFI provides a harsher penalty than the AGFI for hypothesizing a less restrictive (therefore less parsimonious) model. Muelller (1996) provides an example where both the GFI and NFI are very high (0.999 and 0.997, respectively) but PGFI and PNFI are 0.2 and 0.332, respectively. Root Mean Square Error of Approximation (RMS or RMSEA) This is a standardized summary of the average covariance residuals. Covariance residuals are the differences between the observed and model-implied covariances. Perfect model fit: RMSEA=0 Close fit : RMSEA < 0.05 Akaike's Information Criterion (AIC) Bozdogan's Consistent version of the AIC (CAIC) Browne-Cudeck Criterion (BCC) Bayes Information Criterion (BIC) AIC and CAIC address the issue of parsimony in the assessment of model fit are used to compare two or more models, with smaller values representing a better fit of the hypothesized model AIC : a modification of the standard goodness of fit Chi-square that includes a penalty for complexity provides a quantitative method for model selection, whether or not the models are hierarchical 43

is defined as: AIC= log e (L) + 2q where log e (L) = log-likelihood function q = number of parameters in the model rationale: Adding additional parameters (and increasing the complexity of a model) will always improve the fit of a model; however, improvement in fit may not be enough to justify the added complexity To use: AIC is computed for each candidate model Choose model with lowest AIC. The AIC has the best literature supporting its use is based on strong underlying statistical reasoning has been subjected to numerous Monte Carlo simulations and is found to be well behaved reflects the extent to which parameter estimates from the original sample will cross-validate in future samples CAIC: a modification of the AIC that takes into account the sample size assigns greater penalty to model complexity than either the AIC or BCC, but not as great as the BIC The AIC, CAIC and BCC assume that no true model exists, but try to find the best one among those being considered. BCC: was developed specifically for the analysis of structural eqns. There is some evidence that the BCC may be superior to the other measures. It imposes a slightly greater penalty for model complexity than does the AIC. BIC: assigns more penalty for complexity than the AIC, BCC and CAIC and hence has a greater tendency to pick parsimonious models. has been found in Monte Carlo simulations to perform comparably to the AIC assumes that the true model is in the set of candidate models and that the goal of model selection is to find the true model. requires that the sample size to be very large. In most outputs, these measures are applied to three models: (1) the "default" model, aka the model that you have specified (2) the independence model (one in which it is assumed that there are no correlations between any of the variables), and (3) the saturated model, one in which there are as many parameters as there are observations. 44

In a sense, one is looking to see how bad your model is compared with a "worst-case" scenario. The closer the AIC, CAID, BCC or BIC are to the saturated value and the further from the independence model, the better the fit. The more common use is when one develops two reasonably fit models; these measures can be used to help decide which model to use. Hoelter's critical N. This fit statistic focuses on the adequacy of sample size, rather than on model fit. Its purpose is to estimate a sample size that would be sufficient to yield an adequate model fit for a χ 2 test. A model is considered adequately fitting if the Hoelter's N is greater than 200. Guiding Principles in Assessing Model Fit 1. Use a strong underlying theory as the primary guide to assessing model adequacy since the objective of structural equation modeling in general, and CFA in particular, is to understand a substantive area. 2. If possible, formulate several alternative models prior to analysis rather than considering a single model. 3. Whenever possible, compare fit results to that obtained in previous studies for the same or similar models. 4. In addition assessing the fit of the whole model, consider assessing the fit of components of the model. 5. Use different measures of fit from different families of measures. 6. Use overall fit indices that take the degrees of freedom of the model into account and that depend as little as possible on the sample size. Interpreting Confirmatory Factor Analyses It is important to remember when interpreting the findings from a confirmatory factor analysis that more than one model can be determined that will adequately fit the data. Therefore, finding a model with good fit does not mean that the model is the only, or optimal model for that data. 45

46

Model Modification When a confirmatory analysis fails to fit the observed factor structure with the theoretical structure, the analyst can 1. Re-specify another a priori model, based on theory and your knowledge; or 2. Modify the original model to improve the fit. Improving the fit can be done by: (a) (b) examining the fitted model for regression weights or covariances that are not significant and pruning back these covariances or paths. Rerunning the model and checking if the fit is now adequate. However, post-hoc model modifications should make substantive sense and should be theoretically justifiable. Two types of model specification errors 1. External omission of important variables from the model Can be identified based on knowledge of the substantive area 2. Internal omission of important relations within the model Indicators of possible internal specification errors: 1. Modification indices (MI) The MI is an estimate of the decrease in the χ 2 statistic (which is actually a badness of fit measure) if a fixed parameter is freed. 2. Expected parameter change statistics (EPC) The EPC is an estimate of the value of the freed parameter. Suggestions: Large MI, large EPC : free the parameter Large MI, small EPC : keep parameter fixed Small MI, small EPC : keep parameter fixed Other suggestion: free those with large MI indices Consequences of post hoc model modification: 1. Model modification based on these measures might lead to several competing models that should be compared on substantive as well as statistical merits. 2. Fit statistics of a modified structure that was applied to the same data set that yielded the MIs and EPCs will be biased upward. These do not necessarily indicate that a structure that better reflects the true structure in the population is obtained; they 47

could simply indicate that a model that is particularly suited to the sampled data at hand has been obtained. Solution to overfitting to the sample data: cross validate the modified structure with a new and independent sample. Split the sample randomly into calibration and validation subsamples and compute the cross validation index (CVI) due to Cudeck and Brown. The CVI measures the distance between the unrestricted variance-covariance matrix from the validation sample S v and the model-implied variance-covariance matrix from the calibration sample Σ(θ c hat). That is CVI = F[S v, Σ(θ c hat)]. The smaller the value of CVI, the better the estimated predictive validity of the model. One can compare the competing models based on CVI, or compare the CVI obtained for the restructured model to that obtained from the saturated model (an unrestricted model with df=0) or the independence model. If the sample data cannot be split, use the expected value of the cross validation index (ECVI) given by ECVI = F[S, Σ(θ c hat)] + ( 2p / (n 1)) where F[S, Σ(θ c hat)] is the minimum value of the fitting function for the hypothesized structure, p is the number of parameters to be estimated based on the model and n is the sample size. Comparing the ECVI value for the hypothesized model to the values obtained for the independence and saturated models shows that ECVI s < ECVI h < ECVI i implying that technically the saturated model has better predictive validity than the hypothesized model. But if the differences between the more parsimonious hypothesized model and the saturated model is small, the former may be taken to be an adequate approximation to reality. Example 1 from SAS manual: Second-Order Confirmatory Factor Analysis A second-order confirmatory factor analysis model is applied to a correlation matrix of Thurstone reported by McDonald (1985). Using the LINEQS statement, the three-term secondorder factor analysis model is specified in equations notation. The first-order loadings for the three factors, F1, F2, and F3, each refer to three variables, X1-X3, X4-X6, and X7-X9. One second-order factor, F4, reflects the correlations among the three first-order factors. The second-order factor correlation matrix P is defined as a 1 1 identity matrix. Choosing the second-order uniqueness matrix U2 as a diagonal matrix with parameters U21-U23 gives an unidentified model. To compute identified maximum likelihood estimates, the matrix U2 is defined as a 3 3 identity matrix. The following code generates results that are partially displayed in the output that follows. data Thurst(TYPE=CORR); Title "Example of THURSTONE resp. McDONALD (1985, p.57, p.105)"; _TYPE_ = 'CORR'; Input _NAME_ $ Obs1-Obs9; Label Obs1='Sentences' Obs2='Vocabulary' Obs3='Sentence Completion' Obs4='First Letters' Obs5='Four-letter Words' Obs6='Suffices' Obs7='Letter series' Obs8='Pedigrees' Obs9='Letter Grouping'; 48

datalines; Obs1 1......... Obs2.828 1........ Obs3.776.779 1....... Obs4.439.493.460 1...... Obs5.432.464.425.674 1..... Obs6.447.489.443.590.541 1.... Obs7.447.432.401.381.402.288 1... Obs8.541.537.534.350.367.320.555 1.. Obs9.380.358.359.424.446.325.598.452 1. ; proc calis data=thurst method=max edf=212 pestim se; Title2 "Identified Second Order Confirmatory Factor Analysis"; Title3 "C = F1 * F2 * P * F2' * F1' + F1 * U2 * F1' + U1, With P=U2=Ide"; Lineqs Obs1 = X1 F1 + E1, Obs2 = X2 F1 + E2, Obs3 = X3 F1 + E3, Obs4 = X4 F2 + E4, Obs5 = X5 F2 + E5, Obs6 = X6 F2 + E6, Obs7 = X7 F3 + E7, Obs8 = X8 F3 + E8, Obs9 = X9 F3 + E9, F1 = X10 F4 + E10, F2 = X11 F4 + E11, F3 = X12 F4 + E12; Std F4 = 1., E1-E9 = U11-U19, E10-E12 = 3 * 1.; Bounds 0. <= U11-U19; run; Output: Second-Order Confirmatory Factor Analysis Example of THURSTONE resp. McDONALD (1985, p.57, p.105) Identified Second Order Confirmatory Factor Analysis C = F1 * F2 * P * F2' * F1' + F1 * U2 * F1' + U1, With P=U2=Ide The CALIS Procedure Covariance Structure Analysis: Maximum Likelihood Estimation Parameter Estimates 21 49

Functions (Observations) 45 Lower Bounds 9 Upper Bounds 0 GCONV convergence criterion satisfied. Fit Function 0.1802 Goodness of Fit Index (GFI) 0.9596 GFI Adjusted for Degrees of Freedom (AGFI) 0.9242 Root Mean Square Residual (RMR) 0.0436 Parsimonious GFI (Mulaik, 1989) 0.6397 Chi-Square 38.1963 Chi-Square DF 24 Pr > Chi-Square 0.0331 Independence Model Chi-Square 1101.9 Independence Model Chi-Square DF 36 RMSEA Estimate 0.0528 RMSEA 90% Lower Confidence Limit 0.0153 RMSEA 90% Upper Confidence Limit 0.0831 ECVI Estimate 0.3881 ECVI 90% Lower Confidence Limit. ECVI 90% Upper Confidence Limit 0.4888 Probability of Close Fit 0.4088 Bentler's Comparative Fit Index 0.9867 Normal Theory Reweighted LS Chi-Square 40.1947 Akaike's Information Criterion -9.8037 Bozdogan's (1987) CAIC -114.4747 Schwarz's Bayesian Criterion -90.4747 McDonald's (1989) Centrality 0.9672 Bentler & Bonett's (1980) Non-normed Index 0.9800 Bentler & Bonett's (1980) NFI 0.9653 James, Mulaik, & Brett (1982) Parsimonious NFI 0.6436 Z-Test of Wilson & Hilferty (1931) 1.8373 Bollen (1986) Normed Index Rho1 0.9480 Bollen (1988) Non-normed Index Delta2 0.9868 Hoelter's (1983) Critical N 204 Manifest Variable Equations Obs1 = 0.5151* F1 + 1.0000 E1 Std Err 0.0629 X1 t Value 8.1868 Obs2 = 0.5203* F1 + 1.0000 E2 Std Err 0.0634 X2 50

t Value 8.2090 Obs3 = 0.4874* F1 + 1.0000 E3 Std Err 0.0608 X3 t Value 8.0151 Obs4 = 0.5211* F2 + 1.0000 E4 Std Err 0.0611 X4 t Value 8.5342 Obs5 = 0.4971* F2 + 1.0000 E5 Std Err 0.0590 X5 t Value 8.4213 Obs6 = 0.4381* F2 + 1.0000 E6 Std Err 0.0560 X6 t Value 7.8283 Obs7 = 0.4524* F3 + 1.0000 E7 Std Err 0.0660 X7 t Value 6.8584 Obs8 = 0.4173* F3 + 1.0000 E8 Std Err 0.0622 X8 t Value 6.7135 Obs9 = 0.4076* F3 + 1.0000 E9 Std Err 0.0613 X9 t Value 6.6484 Latent Variable Equations F1 = 1.4438* F4 + 1.0000 E10 Std Err 0.2565 X10 t Value 5.6282 F2 = 1.2538* F4 + 1.0000 E11 Std Err 0.2114 X11 t Value 5.9320 F3 = 1.4065* F4 + 1.0000 E12 Std Err 0.2689 X12 t Value 5.2307 The CALIS Procedure Covariance Structure Analysis: Maximum Likelihood Estimation Variances of Exogenous Variables Variable Paramete r Estimat e Standar d Error t Value F4 1.00000 E1 U11 0.18150 0.02848 6.37 E2 U12 0.16493 0.02777 5.94 E3 U13 0.26713 0.03336 8.01 51

E4 U14 0.30150 0.05102 5.91 E5 U15 0.36450 0.05264 6.93 E6 U16 0.50642 0.05963 8.49 E7 U17 0.39032 0.05934 6.58 E8 U18 0.48138 0.06225 7.73 E9 U19 0.50509 0.06333 7.98 E10 1.00000 E11 1.00000 E12 1.00000 The CALIS Procedure Covariance Structure Analysis: Maximum Likelihood Estimation Squared Multiple Correlations Variable Error Variance Total Variance R- Square 1 Obs1 0.18150 1.00000 0.8185 2 Obs2 0.16493 1.00000 0.8351 3 Obs3 0.26713 1.00000 0.7329 4 Obs4 0.30150 1.00000 0.6985 5 Obs5 0.36450 1.00000 0.6355 6 Obs6 0.50642 1.00000 0.4936 7 Obs7 0.39032 1.00000 0.6097 8 Obs8 0.48138 1.00000 0.5186 9 Obs9 0.50509 1.00000 0.4949 10 F1 1.00000 3.08452 0.6758 11 F2 1.00000 2.57213 0.6112 12 F3 1.00000 2.97832 0.6642 52

Example 2 : CFA on needs data SAS Program ods html file='e:\mvar\0708\calisout_html'; proc calis data=e.need tech=nr nobs=285 all; lineqs imp1 = lam1 F1 + E1, imp2 = lam2 F1 + E2, imp3 = F1 + E3, imp4 = lam4 F1 + E4, imp5 = lam5 F1 + E5, imp6 = lam6 F2 + E6, imp7 = lam7 F2 + E7, imp8 = F2 + E8, imp9 = lam9 F2 + E9, F1 = gam1 F3 + D1, F2 = gam2 F3 + D2; STD E1-E9= the1-the9, D1-D2=psi1-psi2, F3=1; COV D1 D2 = phi; run; ods html close; run; Note: the default estimation method is ML. tech=nr specifies that the Newton Raphson algorithm be used in optimization. The following output shows that the data do not follow a multivariate normal distribution. None of the marginal distributions has zero kurtosis. Mardia s multivariate kurtosis is positive, indicating leptokursis. Violating the multivariate normality assumption in Generalized Least Squares and Maximum Likelihood usually leads to the wrong approximate standard errors and incorrect fit statistics based on the χ 2 value. In general the parameter estimates are more stable against violation of the assumption of multivariate normality. 53

The CALIS Procedure Covariance Structure Analysis: Maximum Likelihood Estimation Observations 285 Model Terms 1 Variables 9 Model Matrices 4 Informations 45 Parameters 21 Variable Mean Std Dev Skewness Kurtosis IMP1 4.63860 0.50272-0.99713 0.81035 IMP2 4.60702 0.50346-0.60675-1.23090 IMP3 4.63158 0.49758-0.72046-1.06096 IMP4 4.61053 0.53658-1.48328 5.44282 IMP5 4.63860 0.48851-0.67111-1.31892 IMP6 4.58596 0.50051-0.43565-1.60273 IMP7 4.61404 0.48768-0.47097-1.79080 IMP8 4.57193 0.50968-0.45265-1.41305 IMP9 4.55088 0.53901-0.74738 0.25909 Mardia's Multivariate Kurtosis 144.3764 Relative Multivariate Kurtosis 2.4583 Normalized Multivariate Kurtosis 86.6076 Mardia Based Kappa (Browne, 1982) 1.4583 Mean Scaled Univariate Kurtosis -0.0706 Adjusted Mean Scaled Univariate Kurtosis 0.1200 54

The CALIS Procedure Covariance Structure Analysis: Maximum Likelihood Estimation Fit Function 0.5899 Goodness of Fit Index (GFI) 0.8774 GFI Adjusted for Degrees of Freedom (AGFI) 0.7701 Root Mean Square Residual (RMR) 0.0390 Parsimonious GFI (Mulaik, 1989) 0.5849 Chi-Square 167.5285 Chi-Square DF 24 Pr > Chi-Square <.0001 Independence Model Chi-Square 2517.5 Independence Model Chi-Square DF 36 Specifying that weighted least squares estimation, which is equivalent to Browne's asymptotically distribution-free estimation, yields much improved fit results. WLS is invoked by typing proc calis data=e.need method=wls nobs=285 all; The CALIS Procedure Covariance Structure Analysis: Weighted Least-Squares Estimation Fit Function 0.1997 Goodness of Fit Index (GFI) 0.9998 GFI Adjusted for Degrees of Freedom (AGFI) 0.9996 Root Mean Square Residual (RMR) 0.1411 Parsimonious GFI (Mulaik, 1989) 0.6665 Chi-Square 56.7052 Chi-Square DF 24 Pr > Chi-Square 0.0002 Independence Model Chi-Square 9113.5 Independence Model Chi-Square DF 36 55

RMSEA Estimate 0.0693 RMSEA 90% Lower Confidence Limit 0.0461 RMSEA 90% Upper Confidence Limit 0.0928 ECVI Estimate 0.3476 ECVI 90% Lower Confidence Limit 0.2833 ECVI 90% Upper Confidence Limit 0.4389 Probability of Close Fit 0.0824 Bentler's Comparative Fit Index 0.9964 Akaike's Information Criterion 8.7052 Bozdogan's (1987) CAIC -102.9545 Schwarz's Bayesian Criterion -78.9545 McDonald's (1989) Centrality 0.9442 Bentler & Bonett's (1980) Non-normed Index 0.9946 Bentler & Bonett's (1980) NFI 0.9938 James, Mulaik, & Brett (1982) Parsimonious NFI 0.6625 Z-Test of Wilson & Hilferty (1931) 3.5454 Bollen (1986) Normed Index Rho1 0.9907 Bollen (1988) Non-normed Index Delta2 0.9964 Hoelter's (1983) Critical N 184 56

Manifest Variable Equations with Estimates IMP1 = 0.9368 * F1 + 1.0000 E1 Std Err 0.0216 lam1 t Value 43.3864 IMP2 = 0.9515 * F1 + 1.0000 E2 Std Err 0.0169 lam2 t Value 56.4608 IMP3 = 1.0000 F1 + 1.0000 E3 IMP4 = 0.9631 * F1 + 1.0000 E4 Std Err 0.0299 lam4 t Value 32.1829 IMP5 = 0.9631 * F1 + 1.0000 E5 Std Err 0.0179 lam5 t Value 53.6875 IMP6 = 0.9720 * F2 + 1.0000 E6 Std Err 0.0290 lam6 t Value 33.4809 IMP7 = 1.0435 * F2 + 1.0000 E7 Std Err 0.0262 lam7 t Value 39.8476 IMP8 = 1.0000 F2 + 1.0000 E8 IMP9 = 0.9451 * F2 + 1.0000 E9 Std Err 0.0281 lam9 t Value 33.5879 57

The CALIS Procedure Covariance Structure Analysis: Weighted Least-Squares Estimation Latent Variable Equations with Estimates F1 = 0.9422 * F3 + 1.0000 D1 Std Err 0.00653 gam1 t Value 144.3 F2 = 0.8910 * F3 + 1.0000 D2 Std Err 0.0153 gam2 t Value 58.0883 Other Differences Between EFA and CFA 1. Factor rotation does not apply to CFA. Identification restrictions associated with CFA are achieved in part by fixing most or all indicator cross-loadings to zero. Thus rotation is not necessary in CFA because simple structure is obtained by specifying indicators to load on just one factor. 2. CFA models are typically more parsimonious than EFA solutions. Only primary loadings and factor correlations are typically freely estimated. 3. Unlike EFA, CFA allows for ability to specify the nature of measurement errors (unique variances) of the indicators. In EFA, factor models are specified under the assumption that measurement error is random. CFA allows for the specification of correlation between the error terms of indicator (manifest) variables. This specification allows for the possibility that some of the covariation between the indicators is due to the shared influence of the underlying latent factor and some may be due to sources other than the common factor, e.g., common assessment method like reversed wording of questions, may have resulted in additional indicator covariation. Example. Self-Esteem Scale Positively worded items, e.g., I feel good about myself. Negatively worded items, e.g., At times I think I am no good at all. Application of EFA results in a positive self-evaluation dimension and a negative self-evaluation dimension Allowing for correlated errors results in a global self-esteem solution. 58

R E F E R E N C E S http://v8doc.sas.com/sashtml/stat/chap19/sect49.htm Jackson, J. L. et. al. (2005). Introduction to Structural Equation Modeling (Path Analysis). (From www) Lattin, J., Carroll, J. D., and Green, P.(2003). Analyzing Multivariate Data. Ca: Brooks/ Cole Thomson Learning. Mueller, Ralph O. (1996). Basic Principles of Structural Equation Modeling. New York: Springer Verlag. Stapleton, Connie D. (1997). Basic Concepts and Procedures of Confirmatory Factor Analysis in http://searcheric.org/ericdb/ed407416.htm. 59