Introduction to Confirmatory Factor Analysis

Size: px
Start display at page:

Download "Introduction to Confirmatory Factor Analysis"

Transcription

1 Introduction to Confirmatory Factor Analysis In Exploratory FA, the analyst has no strong prior notion of the structure of the factor solution the goal is to infer factor structure from the patterns of correlation in the data Exploratory FA suggests hypotheses, but does not justify knowledge. Confirmatory FA Confirmatory FA is a theory-testing model as opposed to a theory-generating method In confirmatory FA, the analyst tests if the prior notion regarding the factor solution is consistent with the patterns in the data. The model, or hypothesis, that is tested specifies which variables will be correlated with which factors and which factors are correlated. The hypothesis is usually based on a strong theoretical and/or empirical foundation. With confirmatory FA, it is possible to measure the goodness-of-fit of a particular model and to statistically test the adequacy of the model fit. Confirmatory FA also offers the analyst a more viable method for evaluating construct validity. A construct is said to be reliably measured (have construct validity) when multiple measurements of the same construct using different methods are highly correlated. Ex. Menezes and Elbert (1979), as cited in Lattin, Carroll and Green (2003) used multiple measurement methods to examine multiple characteristics of the stores in a major local grocery chain. They selected two traits for their study: store appearance (A) and product assortment (P). Each trait was measured using three different measurement methods: the Likert scale, the semantic differential scale, and the Stapel scale: 1. Likert scale Selection is wide. SA GA MA MD GD SD Where SA = strongly agree, GA =generally agree, MA = moderately agree and SD, GD, MD correspond to disagree. 2. Semantic differential scale Extremely Quite Slight Slight Quite Extremely Wide Selection Limited Selection 36

2 3. Stapel scale Wide Selection Menezes and Elbert tested whether any one scale is any more reliable and concluded that that there were no marked differences across the 3 methods. Confirmatory FA can also be used to test for discriminant validity, which in the context of the Menezres and Elbert example, holds when the two traits store appearance and product assortment can be separated into distinct factors. Confirmatory FA Procedure 1. The analyst begins with a correlation matrix or a variance/covariance matrix. 2. The analyst then proposes competing models, based on theory or existing data, that are hypothesized to fit the data. The models may specify the degree of correlation, if any, between each pair of common factors; the degree of correlation between individual variables and one or more factors, which particular pairs of unique factors are correlated. The different models are determined by "fixing" or "freeing" specific parameters such as: the factor loadings or coefficients, the factor correlation coefficients, and the variance/covariance of the error of measurement. These parameters are set according to the theoretical expectation of the analyst. One fixes a parameter by setting the parameter at a specific value based on one's expectations. One frees a parameter by allowing the parameter to be estimated during the analysis. 3. The competing models or hypotheses about the structure of the data are then tested against one another with the use of goodness-of-fit statistics. The fit statistics determine how well the competing models fit the data, or explain the covariation among the variables. Examples of these statistics include the chi square/degrees of freedom ratio, the Bentler comparative fit index (CFI), the parsimony ratio, and the Goodness-of-fit Index (GFI). 37

3 Example. Holzinger and Swineford s (1939) study on psychological testing of 7 th and 8 th grade children: X 1 = paragraph comprehension X 2 = sentence completion X 3 = word meaning X 4 = addition X 5 = counting dots The correlation matrix based on a sample of 145 children is given by R = An exploratory FA with two factors in a sense hypothesizes that all five variables load on both factors: X 1 = γ 11 F 1 + γ 12 F 2 + ε 1 X 2 = γ 21 F 1 + γ 22 F 2 + ε 2 X 3 = γ 31 F 1 + γ 32 F 2 + ε 3 X 4 = γ 41 F 1 + γ 42 F 2 + ε 4 X 5 = γ 51 F 1 + γ 52 F 2 + ε 5 But we may believe that two factors underlie test performance: verbal ability and quantitative ability. Moreover, we may believe that the two factors are distinct, but are positively correlated. The following FA model may thus be hypothesized: X 1 = γ 11 F 1 + ε 1 X 2 = γ 21 F 1 + ε 2 X 3 = γ 31 F 1 + ε 3 X 4 γ 42 F 2 + ε 4 X 5 = γ 52 F 2 + ε 5 Corr(F 1,F 2 ) = ϕ 12. We thus fixed γ 12 = γ 22 = γ 32 = γ 41 = γ 51 = 0. We can further specify that var(f 1 )=var(f 2 )=1, while allowing Corr(F 1,F 2 ) = ϕ 12 to be freely estimated. Note that with this model, we have 15 pieces of information (the five 1 s on the main diagonal and the 10 off-diagonal elements of R) with which to estimate 11 model parameters: the five γ loadings, ϕ 12 and the five specific variances (variances of the ε s). 38

4 Because there are more observations (15) than unknown parameters (11), this model is overdetermined. In general, with an overdetermined model, it will not be possible to fit the observed values exactly. Having more observations than model parameters is only a necessary, but not sufficient, condition for each model parameter to be identified. However, if the model exhibits a maximally simple structure (i.e., each variable loads on only one factor), error terms are independent, and the number of factors is small relative to the number of measures, then the model is identified. Estimation Methods and Fitting Functions There are various estimation methods available for CFA. These methods all obtain parameter estimates collected in a vector θhat by iteratively minimizing a certain function F[S, Σ(θhat)] so that S and Σ(θhat) are close or, equivalently, the elements in the residual matrix S - Σ(θhat) are small. F is called the fitting function and its form depends on the estimation method used. Five estimation methods are available under PROC CALIS of SAS. These are Unweighted least squares (ULS) Generalized Least Squares (GLS) Normal theory Maximum Likelihood (ML) Weighted Least Squares (WLS, ADF) Diagonally Weighted Least Squares (DWLS). The default is GLS. Let S denote the sample correlation or covariance matrix for a sample size N and let C be the predicted moment matrix [i.e., C =Σ(θhat) ]. The fitting functions are as follows F ULS = 0.5 Tr(S C) 2 F GLS = 0.5 Tr(S -1 (S C)) 2 F ML = Tr(SC -1 )- n +log(det C) log(det S), where n = number of manifest variables F WLS = Vec(s ij - c ij )' W -1 Vec(s ij - c ij ), where Vec(s ij - c ij ) denotes the vector of the n(n+1)/2 elements of the lower triangle of the symmetric matrix S - C, and W = (w ij,kl ) is a positive definite symmetric matrix with n(n+1)/2 rows and columns If the moment matrix S is a correlation (rather than a covariance) matrix, the default setting of W = (w ij,kl ) is the estimators of the asymptotic covariances correlations S = (s ij ). of the 39

5 F DWLS = Vec(s ij - c ij )' diag(w -1 )Vec(s ij - c ij ) Using only the diagonal elements of W -1 eases computational burden. But statistical properties of DWLS estimates are still not known. The Fit Statistics Chi square statistic and Chi square/degrees of freedom ratio The chi square test tests the hypothesis that the model is consistent with the pattern of covariation among the observed variables. Specifically, it tests the null hypothesis that there is no statistically significant difference in the observed covariance matrix and the theoretical covariance structure matrix implied by the model. Under Ho: Σ = Σ(θ) and certain distributional assumptions (e.g., multivariate normality if the ML method is used), χ 2 = (n -1) F[S, Σ(θhat)] is asymptotically chi-square with degrees of freedom equal to the difference (c p) where c is the number of nonredundant variances and covariances of the observed variables and p the number of parameters to be estimated. Here n is the sample size. In the case of the chi-square statistic, smaller rather than larger values indicate a good fit. Suppose an SEM model with 5 degrees of freedom yields a discrepancy function of χ 2 = with a p-value of This implies that the model with 5 degrees of freedom is significantly worse than if it had estimated 5 more parameters and had 0 degrees of freedom. The chi-square statistic is very sensitive to sample size. When the sample size is large, even the smallest discrepancies between the fitted model matrix and the observed data matrix will be judged significant. In such situations, it is unclear whether the statistical significance of the chi square statistic is due to poor fit of the model or to the size of the sample. Thus, alternative fit statistics were developed. Because of its limitations, some researchers advocate that the chi-square not be used as a formal test statistic. They instead suggest that divide the Chi-square value be divided by its numbers of degrees of freedom to obtain the Chi square ratio. rule of thumb well-fitted: ratio < 2 acceptable: ratio < 3 not acceptable: ratio > 5 40

6 Goodness-of-fit index (GFI) and adjusted goodness-of-fit index (AGFI) The good of fit index "is a measure of the relative amount of variances and covariances jointly accounted for by the model". It can be thought of as being roughly analogous to the multiple R squared in multiple regression. The closer the GFI is to 1.00, the better is the fit of the model to the data. GFI = 1 - F[S, Σ(θhat)] F[S, Σ(θ)] Here the numerator is the minimum value of the fitting function F for the hypothesized model, while the denominator is the minimum value of F when there is no hypothesized model. Thus GFI measures how much better the model fits compared to no model at all. Rules of thumb: good fit: GFI > 0.95 acceptable fit: GFI > The adjusted goodness of fit statistic (AGFI) is based on a correction for the number of degrees of freedom in a less restricted model obtained by freeing more parameters. It penalizes more complex models (i.e., those with more parameters to be estimated) with a downward adjustment in the fit index. c AGFI = 1 - df h F[S, Σ(θhat)] F[S, Σ(θ)] Here c is as before and df h = c p, the df for the hypothesized model. Rules of thumb: good fit: AGFI > 0.90 acceptable fit: AGFI > Both the GFI and the AGFI are less sensitive to sample size than the chi square statistic. The Normed and Nonnormed Fit Index and Their Extensions These measures fit by comparing a hypothesized model to some reasonable but more restrictive (fewer parameters to be estimated) baseline model. The independence model is usually taken to be the baseline model. Under the independence model (1) no latent variables are hypothesized to underlie the observed variables; (2) observed variables are measured without error; and the observed variables are independent. Let F h denote the minimum value of the fitting function F for the hypothesized model with associated degrees of freedom df h. Similarly, let F i denote the minimum value of F for the hypothesized model with associated degrees of freedom df i. 41

7 χ 2 2 i - χ h F h NFI = = χ i F i The Normed Fit Index (NFI) is normed to a 0 to 1 range with large values ( >.90) indicating a better fit. However, the NFI is affected by sample size and might not equal 1 even if the correct model is hypothesized. Thus the NFI model often underestimates model fit especially for small samples. The Nonnormed Fit Index (NNFI) penalizes more complex (or less restrictive) models by a downward adjustment and rewards more parsimonious (more restrictive models) by an upward adjustment. As its name suggests, the NNFI is not normed and may take values outside the 0 to 1 range. χ 2 2 i df i. - χ h df h NNFI =. χ 2 i df i. - 1 The normed Comparative Fit Index (CFI) was introduced to address the deficiency of the NFI, while the nonnormed Fit Index (FI) was introduced to address the deficiency of the NNFI. Let l h = [(n 1 ) F h - df h ] and l i = [(n 1 ) F i df i ]. Then CFI = 1 - l 1 l 2 where l 1 = max ( l h, 0) and l 2 = max ( l h, l i, 0). l h FI = 1 -. l i CFI is restricted to a 0 to 1 range while FI is not. Comparisons of CFI and FI with NFI and NNFI show that underestimation of fit happens less with CFI than with NFI and that FI behaves better than NNFI at the endpoints of the [0,1] interval in that it is not as frequently negative as NNFI and when it exceeds 1, it does so by a smaller amount. CFI is also less affected by sample size than NFI. Values above 0.90 are considered to be indicative of good overall fit for NFI, NNFI CFI and FI. 42

8 The Parsimonious Fit Indices and the Parsimony Ratio The parsimony fit indices (Parsimonious Goodness of Fit Index or PGFI and Parsimony Normed Fit Index or PNFI) and the parsimony ratio take into consideration the number of parameters estimated in the model. The fewer number of parameters necessary to specify the model, the more parsimonious is the model. By multiplying the parsimony ratio by a fit statistic, an index of both the overall efficacy of the model explaining the covariance among the variables and the parsimony of the proposed model is obtained. df h PGFI = GFI df n PNFI = df h df i NFI where df h is the degrees of freedom associated with the hypothesized model, df n = c, the degrees of freedom when no model is hypothesized (the number of nonredundant variances and covariances of the observed variables) and df i is the degrees of freedom for the independence baseline model. Both the AGFI and PGFI adjust the GFI downward for model complexity, but the PGFI provides a harsher penalty than the AGFI for hypothesizing a less restrictive (therefore less parsimonious) model. Muelller (1996) provides an example where both the GFI and NFI are very high (0.999 and 0.997, respectively) but PGFI and PNFI are 0.2 and 0.332, respectively. Root Mean Square Error of Approximation (RMS or RMSEA) This is a standardized summary of the average covariance residuals. Covariance residuals are the differences between the observed and model-implied covariances. Perfect model fit: RMSEA=0 Close fit : RMSEA < 0.05 Akaike's Information Criterion (AIC) Bozdogan's Consistent version of the AIC (CAIC) Browne-Cudeck Criterion (BCC) Bayes Information Criterion (BIC) AIC and CAIC address the issue of parsimony in the assessment of model fit are used to compare two or more models, with smaller values representing a better fit of the hypothesized model AIC : a modification of the standard goodness of fit Chi-square that includes a penalty for complexity provides a quantitative method for model selection, whether or not the models are hierarchical 43

9 is defined as: AIC= log e (L) + 2q where log e (L) = log-likelihood function q = number of parameters in the model rationale: Adding additional parameters (and increasing the complexity of a model) will always improve the fit of a model; however, improvement in fit may not be enough to justify the added complexity To use: AIC is computed for each candidate model Choose model with lowest AIC. The AIC has the best literature supporting its use is based on strong underlying statistical reasoning has been subjected to numerous Monte Carlo simulations and is found to be well behaved reflects the extent to which parameter estimates from the original sample will cross-validate in future samples CAIC: a modification of the AIC that takes into account the sample size assigns greater penalty to model complexity than either the AIC or BCC, but not as great as the BIC The AIC, CAIC and BCC assume that no true model exists, but try to find the best one among those being considered. BCC: was developed specifically for the analysis of structural eqns. There is some evidence that the BCC may be superior to the other measures. It imposes a slightly greater penalty for model complexity than does the AIC. BIC: assigns more penalty for complexity than the AIC, BCC and CAIC and hence has a greater tendency to pick parsimonious models. has been found in Monte Carlo simulations to perform comparably to the AIC assumes that the true model is in the set of candidate models and that the goal of model selection is to find the true model. requires that the sample size to be very large. In most outputs, these measures are applied to three models: (1) the "default" model, aka the model that you have specified (2) the independence model (one in which it is assumed that there are no correlations between any of the variables), and (3) the saturated model, one in which there are as many parameters as there are observations. 44

10 In a sense, one is looking to see how bad your model is compared with a "worst-case" scenario. The closer the AIC, CAID, BCC or BIC are to the saturated value and the further from the independence model, the better the fit. The more common use is when one develops two reasonably fit models; these measures can be used to help decide which model to use. Hoelter's critical N. This fit statistic focuses on the adequacy of sample size, rather than on model fit. Its purpose is to estimate a sample size that would be sufficient to yield an adequate model fit for a χ 2 test. A model is considered adequately fitting if the Hoelter's N is greater than 200. Guiding Principles in Assessing Model Fit 1. Use a strong underlying theory as the primary guide to assessing model adequacy since the objective of structural equation modeling in general, and CFA in particular, is to understand a substantive area. 2. If possible, formulate several alternative models prior to analysis rather than considering a single model. 3. Whenever possible, compare fit results to that obtained in previous studies for the same or similar models. 4. In addition assessing the fit of the whole model, consider assessing the fit of components of the model. 5. Use different measures of fit from different families of measures. 6. Use overall fit indices that take the degrees of freedom of the model into account and that depend as little as possible on the sample size. Interpreting Confirmatory Factor Analyses It is important to remember when interpreting the findings from a confirmatory factor analysis that more than one model can be determined that will adequately fit the data. Therefore, finding a model with good fit does not mean that the model is the only, or optimal model for that data. 45

11 46

12 Model Modification When a confirmatory analysis fails to fit the observed factor structure with the theoretical structure, the analyst can 1. Re-specify another a priori model, based on theory and your knowledge; or 2. Modify the original model to improve the fit. Improving the fit can be done by: (a) (b) examining the fitted model for regression weights or covariances that are not significant and pruning back these covariances or paths. Rerunning the model and checking if the fit is now adequate. However, post-hoc model modifications should make substantive sense and should be theoretically justifiable. Two types of model specification errors 1. External omission of important variables from the model Can be identified based on knowledge of the substantive area 2. Internal omission of important relations within the model Indicators of possible internal specification errors: 1. Modification indices (MI) The MI is an estimate of the decrease in the χ 2 statistic (which is actually a badness of fit measure) if a fixed parameter is freed. 2. Expected parameter change statistics (EPC) The EPC is an estimate of the value of the freed parameter. Suggestions: Large MI, large EPC : free the parameter Large MI, small EPC : keep parameter fixed Small MI, small EPC : keep parameter fixed Other suggestion: free those with large MI indices Consequences of post hoc model modification: 1. Model modification based on these measures might lead to several competing models that should be compared on substantive as well as statistical merits. 2. Fit statistics of a modified structure that was applied to the same data set that yielded the MIs and EPCs will be biased upward. These do not necessarily indicate that a structure that better reflects the true structure in the population is obtained; they 47

13 could simply indicate that a model that is particularly suited to the sampled data at hand has been obtained. Solution to overfitting to the sample data: cross validate the modified structure with a new and independent sample. Split the sample randomly into calibration and validation subsamples and compute the cross validation index (CVI) due to Cudeck and Brown. The CVI measures the distance between the unrestricted variance-covariance matrix from the validation sample S v and the model-implied variance-covariance matrix from the calibration sample Σ(θ c hat). That is CVI = F[S v, Σ(θ c hat)]. The smaller the value of CVI, the better the estimated predictive validity of the model. One can compare the competing models based on CVI, or compare the CVI obtained for the restructured model to that obtained from the saturated model (an unrestricted model with df=0) or the independence model. If the sample data cannot be split, use the expected value of the cross validation index (ECVI) given by ECVI = F[S, Σ(θ c hat)] + ( 2p / (n 1)) where F[S, Σ(θ c hat)] is the minimum value of the fitting function for the hypothesized structure, p is the number of parameters to be estimated based on the model and n is the sample size. Comparing the ECVI value for the hypothesized model to the values obtained for the independence and saturated models shows that ECVI s < ECVI h < ECVI i implying that technically the saturated model has better predictive validity than the hypothesized model. But if the differences between the more parsimonious hypothesized model and the saturated model is small, the former may be taken to be an adequate approximation to reality. Example 1 from SAS manual: Second-Order Confirmatory Factor Analysis A second-order confirmatory factor analysis model is applied to a correlation matrix of Thurstone reported by McDonald (1985). Using the LINEQS statement, the three-term secondorder factor analysis model is specified in equations notation. The first-order loadings for the three factors, F1, F2, and F3, each refer to three variables, X1-X3, X4-X6, and X7-X9. One second-order factor, F4, reflects the correlations among the three first-order factors. The second-order factor correlation matrix P is defined as a 1 1 identity matrix. Choosing the second-order uniqueness matrix U2 as a diagonal matrix with parameters U21-U23 gives an unidentified model. To compute identified maximum likelihood estimates, the matrix U2 is defined as a 3 3 identity matrix. The following code generates results that are partially displayed in the output that follows. data Thurst(TYPE=CORR); Title "Example of THURSTONE resp. McDONALD (1985, p.57, p.105)"; _TYPE_ = 'CORR'; Input _NAME_ $ Obs1-Obs9; Label Obs1='Sentences' Obs2='Vocabulary' Obs3='Sentence Completion' Obs4='First Letters' Obs5='Four-letter Words' Obs6='Suffices' Obs7='Letter series' Obs8='Pedigrees' Obs9='Letter Grouping'; 48

14 datalines; Obs Obs Obs Obs Obs Obs Obs Obs Obs ; proc calis data=thurst method=max edf=212 pestim se; Title2 "Identified Second Order Confirmatory Factor Analysis"; Title3 "C = F1 * F2 * P * F2' * F1' + F1 * U2 * F1' + U1, With P=U2=Ide"; Lineqs Obs1 = X1 F1 + E1, Obs2 = X2 F1 + E2, Obs3 = X3 F1 + E3, Obs4 = X4 F2 + E4, Obs5 = X5 F2 + E5, Obs6 = X6 F2 + E6, Obs7 = X7 F3 + E7, Obs8 = X8 F3 + E8, Obs9 = X9 F3 + E9, F1 = X10 F4 + E10, F2 = X11 F4 + E11, F3 = X12 F4 + E12; Std F4 = 1., E1-E9 = U11-U19, E10-E12 = 3 * 1.; Bounds 0. <= U11-U19; run; Output: Second-Order Confirmatory Factor Analysis Example of THURSTONE resp. McDONALD (1985, p.57, p.105) Identified Second Order Confirmatory Factor Analysis C = F1 * F2 * P * F2' * F1' + F1 * U2 * F1' + U1, With P=U2=Ide The CALIS Procedure Covariance Structure Analysis: Maximum Likelihood Estimation Parameter Estimates 21 49

15 Functions (Observations) 45 Lower Bounds 9 Upper Bounds 0 GCONV convergence criterion satisfied. Fit Function Goodness of Fit Index (GFI) GFI Adjusted for Degrees of Freedom (AGFI) Root Mean Square Residual (RMR) Parsimonious GFI (Mulaik, 1989) Chi-Square Chi-Square DF 24 Pr > Chi-Square Independence Model Chi-Square Independence Model Chi-Square DF 36 RMSEA Estimate RMSEA 90% Lower Confidence Limit RMSEA 90% Upper Confidence Limit ECVI Estimate ECVI 90% Lower Confidence Limit. ECVI 90% Upper Confidence Limit Probability of Close Fit Bentler's Comparative Fit Index Normal Theory Reweighted LS Chi-Square Akaike's Information Criterion Bozdogan's (1987) CAIC Schwarz's Bayesian Criterion McDonald's (1989) Centrality Bentler & Bonett's (1980) Non-normed Index Bentler & Bonett's (1980) NFI James, Mulaik, & Brett (1982) Parsimonious NFI Z-Test of Wilson & Hilferty (1931) Bollen (1986) Normed Index Rho Bollen (1988) Non-normed Index Delta Hoelter's (1983) Critical N 204 Manifest Variable Equations Obs1 = * F E1 Std Err X1 t Value Obs2 = * F E2 Std Err X2 50

16 t Value Obs3 = * F E3 Std Err X3 t Value Obs4 = * F E4 Std Err X4 t Value Obs5 = * F E5 Std Err X5 t Value Obs6 = * F E6 Std Err X6 t Value Obs7 = * F E7 Std Err X7 t Value Obs8 = * F E8 Std Err X8 t Value Obs9 = * F E9 Std Err X9 t Value Latent Variable Equations F1 = * F E10 Std Err X10 t Value F2 = * F E11 Std Err X11 t Value F3 = * F E12 Std Err X12 t Value The CALIS Procedure Covariance Structure Analysis: Maximum Likelihood Estimation Variances of Exogenous Variables Variable Paramete r Estimat e Standar d Error t Value F E1 U E2 U E3 U

17 E4 U E5 U E6 U E7 U E8 U E9 U E E E The CALIS Procedure Covariance Structure Analysis: Maximum Likelihood Estimation Squared Multiple Correlations Variable Error Variance Total Variance R- Square 1 Obs Obs Obs Obs Obs Obs Obs Obs Obs F F F

18 Example 2 : CFA on needs data SAS Program ods html file='e:\mvar\0708\calisout_html'; proc calis data=e.need tech=nr nobs=285 all; lineqs imp1 = lam1 F1 + E1, imp2 = lam2 F1 + E2, imp3 = F1 + E3, imp4 = lam4 F1 + E4, imp5 = lam5 F1 + E5, imp6 = lam6 F2 + E6, imp7 = lam7 F2 + E7, imp8 = F2 + E8, imp9 = lam9 F2 + E9, F1 = gam1 F3 + D1, F2 = gam2 F3 + D2; STD E1-E9= the1-the9, D1-D2=psi1-psi2, F3=1; COV D1 D2 = phi; run; ods html close; run; Note: the default estimation method is ML. tech=nr specifies that the Newton Raphson algorithm be used in optimization. The following output shows that the data do not follow a multivariate normal distribution. None of the marginal distributions has zero kurtosis. Mardia s multivariate kurtosis is positive, indicating leptokursis. Violating the multivariate normality assumption in Generalized Least Squares and Maximum Likelihood usually leads to the wrong approximate standard errors and incorrect fit statistics based on the χ 2 value. In general the parameter estimates are more stable against violation of the assumption of multivariate normality. 53

19 The CALIS Procedure Covariance Structure Analysis: Maximum Likelihood Estimation Observations 285 Model Terms 1 Variables 9 Model Matrices 4 Informations 45 Parameters 21 Variable Mean Std Dev Skewness Kurtosis IMP IMP IMP IMP IMP IMP IMP IMP IMP Mardia's Multivariate Kurtosis Relative Multivariate Kurtosis Normalized Multivariate Kurtosis Mardia Based Kappa (Browne, 1982) Mean Scaled Univariate Kurtosis Adjusted Mean Scaled Univariate Kurtosis

20 The CALIS Procedure Covariance Structure Analysis: Maximum Likelihood Estimation Fit Function Goodness of Fit Index (GFI) GFI Adjusted for Degrees of Freedom (AGFI) Root Mean Square Residual (RMR) Parsimonious GFI (Mulaik, 1989) Chi-Square Chi-Square DF 24 Pr > Chi-Square <.0001 Independence Model Chi-Square Independence Model Chi-Square DF 36 Specifying that weighted least squares estimation, which is equivalent to Browne's asymptotically distribution-free estimation, yields much improved fit results. WLS is invoked by typing proc calis data=e.need method=wls nobs=285 all; The CALIS Procedure Covariance Structure Analysis: Weighted Least-Squares Estimation Fit Function Goodness of Fit Index (GFI) GFI Adjusted for Degrees of Freedom (AGFI) Root Mean Square Residual (RMR) Parsimonious GFI (Mulaik, 1989) Chi-Square Chi-Square DF 24 Pr > Chi-Square Independence Model Chi-Square Independence Model Chi-Square DF 36 55

21 RMSEA Estimate RMSEA 90% Lower Confidence Limit RMSEA 90% Upper Confidence Limit ECVI Estimate ECVI 90% Lower Confidence Limit ECVI 90% Upper Confidence Limit Probability of Close Fit Bentler's Comparative Fit Index Akaike's Information Criterion Bozdogan's (1987) CAIC Schwarz's Bayesian Criterion McDonald's (1989) Centrality Bentler & Bonett's (1980) Non-normed Index Bentler & Bonett's (1980) NFI James, Mulaik, & Brett (1982) Parsimonious NFI Z-Test of Wilson & Hilferty (1931) Bollen (1986) Normed Index Rho Bollen (1988) Non-normed Index Delta Hoelter's (1983) Critical N

22 Manifest Variable Equations with Estimates IMP1 = * F E1 Std Err lam1 t Value IMP2 = * F E2 Std Err lam2 t Value IMP3 = F E3 IMP4 = * F E4 Std Err lam4 t Value IMP5 = * F E5 Std Err lam5 t Value IMP6 = * F E6 Std Err lam6 t Value IMP7 = * F E7 Std Err lam7 t Value IMP8 = F E8 IMP9 = * F E9 Std Err lam9 t Value

23 The CALIS Procedure Covariance Structure Analysis: Weighted Least-Squares Estimation Latent Variable Equations with Estimates F1 = * F D1 Std Err gam1 t Value F2 = * F D2 Std Err gam2 t Value Other Differences Between EFA and CFA 1. Factor rotation does not apply to CFA. Identification restrictions associated with CFA are achieved in part by fixing most or all indicator cross-loadings to zero. Thus rotation is not necessary in CFA because simple structure is obtained by specifying indicators to load on just one factor. 2. CFA models are typically more parsimonious than EFA solutions. Only primary loadings and factor correlations are typically freely estimated. 3. Unlike EFA, CFA allows for ability to specify the nature of measurement errors (unique variances) of the indicators. In EFA, factor models are specified under the assumption that measurement error is random. CFA allows for the specification of correlation between the error terms of indicator (manifest) variables. This specification allows for the possibility that some of the covariation between the indicators is due to the shared influence of the underlying latent factor and some may be due to sources other than the common factor, e.g., common assessment method like reversed wording of questions, may have resulted in additional indicator covariation. Example. Self-Esteem Scale Positively worded items, e.g., I feel good about myself. Negatively worded items, e.g., At times I think I am no good at all. Application of EFA results in a positive self-evaluation dimension and a negative self-evaluation dimension Allowing for correlated errors results in a global self-esteem solution. 58

24 R E F E R E N C E S Jackson, J. L. et. al. (2005). Introduction to Structural Equation Modeling (Path Analysis). (From www) Lattin, J., Carroll, J. D., and Green, P.(2003). Analyzing Multivariate Data. Ca: Brooks/ Cole Thomson Learning. Mueller, Ralph O. (1996). Basic Principles of Structural Equation Modeling. New York: Springer Verlag. Stapleton, Connie D. (1997). Basic Concepts and Procedures of Confirmatory Factor Analysis in 59

Instrumental variables regression on the Poverty data

Instrumental variables regression on the Poverty data Instrumental variables regression on the Poverty data /********************** poverty2.sas **************************/ options linesize=79 noovp formdlim='-' nodate; title 'UN Poverty Data: Instrumental

More information

Regression without measurement error using proc calis

Regression without measurement error using proc calis Regression without measurement error using proc calis /* calculus2.sas */ options linesize=79 pagesize=500 noovp formdlim='_'; title 'Calculus 2: Regression with no measurement error'; title2 ''; data

More information

SAS Example 3: Deliberately create numerical problems

SAS Example 3: Deliberately create numerical problems SAS Example 3: Deliberately create numerical problems Four experiments 1. Try to fit this model, failing the parameter count rule. 2. Set φ 12 =0 to pass the parameter count rule, but still not identifiable.

More information

CONFIRMATORY FACTOR ANALYSIS

CONFIRMATORY FACTOR ANALYSIS 1 CONFIRMATORY FACTOR ANALYSIS The purpose of confirmatory factor analysis (CFA) is to explain the pattern of associations among a set of observed variables in terms of a smaller number of underlying latent

More information

EVALUATION OF STRUCTURAL EQUATION MODELS

EVALUATION OF STRUCTURAL EQUATION MODELS 1 EVALUATION OF STRUCTURAL EQUATION MODELS I. Issues related to the initial specification of theoretical models of interest 1. Model specification: a. Measurement model: (i) EFA vs. CFA (ii) reflective

More information

UNIVERSITY OF TORONTO MISSISSAUGA April 2009 Examinations STA431H5S Professor Jerry Brunner Duration: 3 hours

UNIVERSITY OF TORONTO MISSISSAUGA April 2009 Examinations STA431H5S Professor Jerry Brunner Duration: 3 hours Name (Print): Student Number: Signature: Last/Surname First /Given Name UNIVERSITY OF TORONTO MISSISSAUGA April 2009 Examinations STA431H5S Professor Jerry Brunner Duration: 3 hours Aids allowed: Calculator

More information

C:\Users\Rex Kline\AppData\Local\Temp\AmosTemp\AmosScratch.amw. Date: Friday, December 5, 2014 Time: 11:20:30 AM

C:\Users\Rex Kline\AppData\Local\Temp\AmosTemp\AmosScratch.amw. Date: Friday, December 5, 2014 Time: 11:20:30 AM Page 1 of 7 C:\Users\Rex Kline\AppData\Local\Temp\AmosTemp\AmosScratch.amw Analysis Summary Date and Time Date: Friday, December 5, 2014 Time: 11:20:30 AM Title Groups Group number 1 (Group number 1) Notes

More information

Confirmatory Factor Analysis. Psych 818 DeShon

Confirmatory Factor Analysis. Psych 818 DeShon Confirmatory Factor Analysis Psych 818 DeShon Purpose Takes factor analysis a few steps further. Impose theoretically interesting constraints on the model and examine the resulting fit of the model with

More information

* IVEware Analysis Examples Replication for ASDA 2nd Edition * Berglund April 2017 * Chapter 13 ;

* IVEware Analysis Examples Replication for ASDA 2nd Edition * Berglund April 2017 * Chapter 13 ; IVEware Analysis Example Replication C13 * IVEware Analysis Examples Replication for ASDA 2nd Edition * Berglund April 2017 * Chapter 13 ; * 13.3.3 Alternative Approaches to Fitting GLMMs to Survey Data:

More information

Introduction to Structural Equation Modeling with Latent Variables

Introduction to Structural Equation Modeling with Latent Variables SAS/STAT 9.2 User s Guide Introduction to Structural Equation Modeling with Latent Variables (Book Excerpt) SAS Documentation This document is an individual chapter from SAS/STAT 9.2 User s Guide. The

More information

THE GENERAL STRUCTURAL EQUATION MODEL WITH LATENT VARIATES

THE GENERAL STRUCTURAL EQUATION MODEL WITH LATENT VARIATES THE GENERAL STRUCTURAL EQUATION MODEL WITH LATENT VARIATES I. Specification: A full structural equation model with latent variables consists of two parts: a latent variable model (which specifies the relations

More information

The Role of Leader Motivating Language in Employee Absenteeism (Mayfield: 2009)

The Role of Leader Motivating Language in Employee Absenteeism (Mayfield: 2009) DATE: 12/15/2009 TIME: 5:50 Page 1 LISREL 8.80 (STUDENT EDITION) BY Karl G. J reskog & Dag S rbom This program is published exclusively by Scientific Software International, Inc. 7383 N. Lincoln Avenue,

More information

Chapter 13 Introduction to Structural Equation Modeling

Chapter 13 Introduction to Structural Equation Modeling Chapter 13 Introduction to Structural Equation Modeling Chapter Contents OVERVIEW................................... 203 COMPARISON OF THE CALIS AND SYSLIN PROCEDURES...... 203 MODEL SPECIFICATION...........................

More information

An Introduction to Path Analysis

An Introduction to Path Analysis An Introduction to Path Analysis PRE 905: Multivariate Analysis Lecture 10: April 15, 2014 PRE 905: Lecture 10 Path Analysis Today s Lecture Path analysis starting with multivariate regression then arriving

More information

Factor analysis. George Balabanis

Factor analysis. George Balabanis Factor analysis George Balabanis Key Concepts and Terms Deviation. A deviation is a value minus its mean: x - mean x Variance is a measure of how spread out a distribution is. It is computed as the average

More information

An Introduction to Mplus and Path Analysis

An Introduction to Mplus and Path Analysis An Introduction to Mplus and Path Analysis PSYC 943: Fundamentals of Multivariate Modeling Lecture 10: October 30, 2013 PSYC 943: Lecture 10 Today s Lecture Path analysis starting with multivariate regression

More information

Introduction to Confirmatory Factor Analysis

Introduction to Confirmatory Factor Analysis Introduction to Confirmatory Factor Analysis Multivariate Methods in Education ERSH 8350 Lecture #12 November 16, 2011 ERSH 8350: Lecture 12 Today s Class An Introduction to: Confirmatory Factor Analysis

More information

The use of structural equation modeling to examine consumption of pornographic materials in Chinese adolescents in Hong Kong

The use of structural equation modeling to examine consumption of pornographic materials in Chinese adolescents in Hong Kong The use of structural equation modeling to examine consumption of pornographic materials in Chinese adolescents in Hong Kong Appendix 1 Creating matrices and checking for normality!prelis SYNTAX: Can be

More information

Psychology 454: Latent Variable Modeling How do you know if a model works?

Psychology 454: Latent Variable Modeling How do you know if a model works? Psychology 454: Latent Variable Modeling How do you know if a model works? William Revelle Department of Psychology Northwestern University Evanston, Illinois USA November, 2012 1 / 18 Outline 1 Goodness

More information

Inference using structural equations with latent variables

Inference using structural equations with latent variables This work is licensed under a Creative Commons Attribution-NonCommercial-ShareAlike License. Your use of this material constitutes acceptance of that license and the conditions of use of materials on this

More information

Structural Equation Modeling and Confirmatory Factor Analysis. Types of Variables

Structural Equation Modeling and Confirmatory Factor Analysis. Types of Variables /4/04 Structural Equation Modeling and Confirmatory Factor Analysis Advanced Statistics for Researchers Session 3 Dr. Chris Rakes Website: http://csrakes.yolasite.com Email: Rakes@umbc.edu Twitter: @RakesChris

More information

SC705: Advanced Statistics Instructor: Natasha Sarkisian Class notes: Model Building Strategies

SC705: Advanced Statistics Instructor: Natasha Sarkisian Class notes: Model Building Strategies SC705: Advanced Statistics Instructor: Natasha Sarkisian Class notes: Model Building Strategies Model Diagnostics The model diagnostics and improvement strategies discussed here apply to both measurement

More information

Model Estimation Example

Model Estimation Example Ronald H. Heck 1 EDEP 606: Multivariate Methods (S2013) April 7, 2013 Model Estimation Example As we have moved through the course this semester, we have encountered the concept of model estimation. Discussions

More information

DATE: 9/ L I S R E L 8.80

DATE: 9/ L I S R E L 8.80 98 LAMPIRAN 3 STRUCTURAL EQUATION MODEL ONE CONGINERIC Use of this program is subject to the terms specified in the Convention. Universal Copyright 9/2017 DATE: 9/ TIME: 20:22 Website: www.ssicentral.com

More information

Introduction to Structural Equation Modeling Dominique Zephyr Applied Statistics Lab

Introduction to Structural Equation Modeling Dominique Zephyr Applied Statistics Lab Applied Statistics Lab Introduction to Structural Equation Modeling Dominique Zephyr Applied Statistics Lab SEM Model 3.64 7.32 Education 2.6 Income 2.1.6.83 Charac. of Individuals 1 5.2e-06 -.62 2.62

More information

Chapter 3: Testing alternative models of data

Chapter 3: Testing alternative models of data Chapter 3: Testing alternative models of data William Revelle Northwestern University Prepared as part of course on latent variable analysis (Psychology 454) and as a supplement to the Short Guide to R

More information

Evaluation of structural equation models. Hans Baumgartner Penn State University

Evaluation of structural equation models. Hans Baumgartner Penn State University Evaluation of structural equation models Hans Baumgartner Penn State University Issues related to the initial specification of theoretical models of interest Model specification: Measurement model: EFA

More information

sempower Manual Morten Moshagen

sempower Manual Morten Moshagen sempower Manual Morten Moshagen 2018-03-22 Power Analysis for Structural Equation Models Contact: morten.moshagen@uni-ulm.de Introduction sempower provides a collection of functions to perform power analyses

More information

Multi-sample structural equation models with mean structures, with special emphasis on assessing measurement invariance in cross-national research

Multi-sample structural equation models with mean structures, with special emphasis on assessing measurement invariance in cross-national research 1 Multi-sample structural equation models with mean structures, with special emphasis on assessin measurement invariance in cross-national research Measurement invariance measurement invariance: whether

More information

Introduction to Structural Equation Modeling

Introduction to Structural Equation Modeling Introduction to Structural Equation Modeling Notes Prepared by: Lisa Lix, PhD Manitoba Centre for Health Policy Topics Section I: Introduction Section II: Review of Statistical Concepts and Regression

More information

Condition 9 and 10 Tests of Model Confirmation with SEM Techniques

Condition 9 and 10 Tests of Model Confirmation with SEM Techniques Condition 9 and 10 Tests of Model Confirmation with SEM Techniques Dr. Larry J. Williams CARMA Director Donald and Shirley Clifton Chair of Survey Science Professor of Management University of Nebraska

More information

Confirmatory Factor Analysis: Model comparison, respecification, and more. Psychology 588: Covariance structure and factor models

Confirmatory Factor Analysis: Model comparison, respecification, and more. Psychology 588: Covariance structure and factor models Confirmatory Factor Analysis: Model comparison, respecification, and more Psychology 588: Covariance structure and factor models Model comparison 2 Essentially all goodness of fit indices are descriptive,

More information

2/26/2017. PSY 512: Advanced Statistics for Psychological and Behavioral Research 2

2/26/2017. PSY 512: Advanced Statistics for Psychological and Behavioral Research 2 PSY 512: Advanced Statistics for Psychological and Behavioral Research 2 What is SEM? When should we use SEM? What can SEM tell us? SEM Terminology and Jargon Technical Issues Types of SEM Models Limitations

More information

SEM Day 1 Lab Exercises SPIDA 2007 Dave Flora

SEM Day 1 Lab Exercises SPIDA 2007 Dave Flora SEM Day 1 Lab Exercises SPIDA 2007 Dave Flora 1 Today we will see how to estimate CFA models and interpret output using both SAS and LISREL. In SAS, commands for specifying SEMs are given using linear

More information

STRUCTURAL EQUATION MODELING. Khaled Bedair Statistics Department Virginia Tech LISA, Summer 2013

STRUCTURAL EQUATION MODELING. Khaled Bedair Statistics Department Virginia Tech LISA, Summer 2013 STRUCTURAL EQUATION MODELING Khaled Bedair Statistics Department Virginia Tech LISA, Summer 2013 Introduction: Path analysis Path Analysis is used to estimate a system of equations in which all of the

More information

Multiple Group CFA Invariance Example (data from Brown Chapter 7) using MLR Mplus 7.4: Major Depression Criteria across Men and Women (n = 345 each)

Multiple Group CFA Invariance Example (data from Brown Chapter 7) using MLR Mplus 7.4: Major Depression Criteria across Men and Women (n = 345 each) Multiple Group CFA Invariance Example (data from Brown Chapter 7) using MLR Mplus 7.4: Major Depression Criteria across Men and Women (n = 345 each) 9 items rated by clinicians on a scale of 0 to 8 (0

More information

The Common Factor Model. Measurement Methods Lecture 15 Chapter 9

The Common Factor Model. Measurement Methods Lecture 15 Chapter 9 The Common Factor Model Measurement Methods Lecture 15 Chapter 9 Today s Class Common Factor Model Multiple factors with a single test ML Estimation Methods New fit indices because of ML Estimation method

More information

Chapter 8. Models with Structural and Measurement Components. Overview. Characteristics of SR models. Analysis of SR models. Estimation of SR models

Chapter 8. Models with Structural and Measurement Components. Overview. Characteristics of SR models. Analysis of SR models. Estimation of SR models Chapter 8 Models with Structural and Measurement Components Good people are good because they've come to wisdom through failure. Overview William Saroyan Characteristics of SR models Estimation of SR models

More information

Longitudinal Invariance CFA (using MLR) Example in Mplus v. 7.4 (N = 151; 6 items over 3 occasions)

Longitudinal Invariance CFA (using MLR) Example in Mplus v. 7.4 (N = 151; 6 items over 3 occasions) Longitudinal Invariance CFA (using MLR) Example in Mplus v. 7.4 (N = 151; 6 items over 3 occasions) CLP 948 Example 7b page 1 These data measuring a latent trait of social functioning were collected at

More information

Structural Equation Modeling

Structural Equation Modeling CHAPTER 23 Structural Equation Modeling JODIE B. ULLMAN AND PETER M. BENTLER A FOUR-STAGE GENERAL PROCESS OF MODELING 663 MODEL ESTIMATION TECHNIQUES AND TEST STATISTICS 667 MODEL EVALUATION 671 MODEL

More information

Structural Equation Modeling

Structural Equation Modeling Chapter 11 Structural Equation Modeling Hans Baumgartner and Bert Weijters Hans Baumgartner, Smeal College of Business, The Pennsylvania State University, University Park, PA 16802, USA, E-mail: jxb14@psu.edu.

More information

2.2 Classical Regression in the Time Series Context

2.2 Classical Regression in the Time Series Context 48 2 Time Series Regression and Exploratory Data Analysis context, and therefore we include some material on transformations and other techniques useful in exploratory data analysis. 2.2 Classical Regression

More information

CONFIRMATORY FACTOR ANALYSIS

CONFIRMATORY FACTOR ANALYSIS 1 CONFIRMATORY FACTOR ANALYSIS The purpose of confirmatory factor analysis (CFA) is to explain the pattern of associations among a set of observe variables in terms of a smaller number of unerlying latent

More information

Structural Equation Modelling

Structural Equation Modelling Slide Email: jkanglim@unimelb.edu.au Office: Room 0 Redmond Barry Building Website: http://jeromyanglim.googlepages.com/ Appointments: For appointments regarding course or with the application of statistics

More information

Psychology 454: Latent Variable Modeling How do you know if a model works?

Psychology 454: Latent Variable Modeling How do you know if a model works? Psychology 454: Latent Variable Modeling How do you know if a model works? William Revelle Department of Psychology Northwestern University Evanston, Illinois USA October, 2017 1 / 33 Outline Goodness

More information

THE GENERAL STRUCTURAL EQUATION MODEL WITH LATENT VARIATES

THE GENERAL STRUCTURAL EQUATION MODEL WITH LATENT VARIATES THE GENERAL STRUCTURAL EQUATION MODEL WITH LATENT VARIATES I. Specification: A full structural equation model with latent variables consists of two parts: a latent variable model (which specifies the relations

More information

SRMR in Mplus. Tihomir Asparouhov and Bengt Muthén. May 2, 2018

SRMR in Mplus. Tihomir Asparouhov and Bengt Muthén. May 2, 2018 SRMR in Mplus Tihomir Asparouhov and Bengt Muthén May 2, 2018 1 Introduction In this note we describe the Mplus implementation of the SRMR standardized root mean squared residual) fit index for the models

More information

Multi-group analyses for measurement invariance parameter estimates and model fit (ML)

Multi-group analyses for measurement invariance parameter estimates and model fit (ML) LBP-TBQ: Supplementary digital content 8 Multi-group analyses for measurement invariance parameter estimates and model fit (ML) Medication data Multi-group CFA analyses were performed with the 16-item

More information

Related Concepts: Lecture 9 SEM, Statistical Modeling, AI, and Data Mining. I. Terminology of SEM

Related Concepts: Lecture 9 SEM, Statistical Modeling, AI, and Data Mining. I. Terminology of SEM Lecture 9 SEM, Statistical Modeling, AI, and Data Mining I. Terminology of SEM Related Concepts: Causal Modeling Path Analysis Structural Equation Modeling Latent variables (Factors measurable, but thru

More information

The comparison of estimation methods on the parameter estimates and fit indices in SEM model under 7-point Likert scale

The comparison of estimation methods on the parameter estimates and fit indices in SEM model under 7-point Likert scale The comparison of estimation methods on the parameter estimates and fit indices in SEM model under 7-point Likert scale Piotr Tarka Abstract In this article, the author discusses the issues and problems

More information

Using Structural Equation Modeling to Conduct Confirmatory Factor Analysis

Using Structural Equation Modeling to Conduct Confirmatory Factor Analysis Using Structural Equation Modeling to Conduct Confirmatory Factor Analysis Advanced Statistics for Researchers Session 3 Dr. Chris Rakes Website: http://csrakes.yolasite.com Email: Rakes@umbc.edu Twitter:

More information

Confirmatory Factor Models (CFA: Confirmatory Factor Analysis)

Confirmatory Factor Models (CFA: Confirmatory Factor Analysis) Confirmatory Factor Models (CFA: Confirmatory Factor Analysis) Today s topics: Comparison of EFA and CFA CFA model parameters and identification CFA model estimation CFA model fit evaluation CLP 948: Lecture

More information

Factor Analysis & Structural Equation Models. CS185 Human Computer Interaction

Factor Analysis & Structural Equation Models. CS185 Human Computer Interaction Factor Analysis & Structural Equation Models CS185 Human Computer Interaction MoodPlay Recommender (Andjelkovic et al, UMAP 2016) Online system available here: http://ugallery.pythonanywhere.com/ 2 3 Structural

More information

Maximum Likelihood Estimation; Robust Maximum Likelihood; Missing Data with Maximum Likelihood

Maximum Likelihood Estimation; Robust Maximum Likelihood; Missing Data with Maximum Likelihood Maximum Likelihood Estimation; Robust Maximum Likelihood; Missing Data with Maximum Likelihood PRE 906: Structural Equation Modeling Lecture #3 February 4, 2015 PRE 906, SEM: Estimation Today s Class An

More information

Recovery of weak factor loadings in confirmatory factor analysis under conditions of model misspecification

Recovery of weak factor loadings in confirmatory factor analysis under conditions of model misspecification Behavior Research Methods 29, 41 (4), 138-152 doi:1.3758/brm.41.4.138 Recovery of weak factor loadings in confirmatory factor analysis under conditions of model misspecification CARMEN XIMÉNEZ Autonoma

More information

Introduction to Matrix Algebra and the Multivariate Normal Distribution

Introduction to Matrix Algebra and the Multivariate Normal Distribution Introduction to Matrix Algebra and the Multivariate Normal Distribution Introduction to Structural Equation Modeling Lecture #2 January 18, 2012 ERSH 8750: Lecture 2 Motivation for Learning the Multivariate

More information

Exploring Cultural Differences with Structural Equation Modelling

Exploring Cultural Differences with Structural Equation Modelling Exploring Cultural Differences with Structural Equation Modelling Wynne W. Chin University of Calgary and City University of Hong Kong 1996 IS Cross Cultural Workshop slide 1 The objectives for this presentation

More information

This course. Tutors. Jon Heron, PhD (Bristol) Anna Brown, PhD (Cambridge)

This course. Tutors. Jon Heron, PhD (Bristol) Anna Brown, PhD (Cambridge) This course The course is funded by the ESRC RDI and hosted by The Psychometrics Centre Tutors Jon Heron, PhD (Bristol) jon.heron@bristol.ac.uk Anna Brown, PhD (Cambridge) ab936@medschl.cam.ac.uk Tim Croudace,

More information

Case of single exogenous (iv) variable (with single or multiple mediators) iv à med à dv. = β 0. iv i. med i + α 1

Case of single exogenous (iv) variable (with single or multiple mediators) iv à med à dv. = β 0. iv i. med i + α 1 Mediation Analysis: OLS vs. SUR vs. ISUR vs. 3SLS vs. SEM Note by Hubert Gatignon July 7, 2013, updated November 15, 2013, April 11, 2014, May 21, 2016 and August 10, 2016 In Chap. 11 of Statistical Analysis

More information

AN INTRODUCTION TO STRUCTURAL EQUATION MODELING WITH AN APPLICATION TO THE BLOGOSPHERE

AN INTRODUCTION TO STRUCTURAL EQUATION MODELING WITH AN APPLICATION TO THE BLOGOSPHERE AN INTRODUCTION TO STRUCTURAL EQUATION MODELING WITH AN APPLICATION TO THE BLOGOSPHERE Dr. James (Jim) D. Doyle March 19, 2014 Structural equation modeling or SEM 1971-1980: 27 1981-1990: 118 1991-2000:

More information

Title. Description. Remarks and examples. stata.com. stata.com. Variable notation. methods and formulas for sem Methods and formulas for sem

Title. Description. Remarks and examples. stata.com. stata.com. Variable notation. methods and formulas for sem Methods and formulas for sem Title stata.com methods and formulas for sem Methods and formulas for sem Description Remarks and examples References Also see Description The methods and formulas for the sem commands are presented below.

More information

Compiled by: Assoc. Prof. Dr Bahaman Abu Samah Department of Professional Developmentand Continuing Education Faculty of Educational Studies

Compiled by: Assoc. Prof. Dr Bahaman Abu Samah Department of Professional Developmentand Continuing Education Faculty of Educational Studies Compiled by: Assoc. Prof. Dr Bahaman Abu Samah Department of Professional Developmentand Continuing Education Faculty of Educational Studies Universiti Putra Malaysia Serdang Structural Equation Modeling

More information

SHOPPING FOR EFFICIENT CONFIDENCE INTERVALS IN STRUCTURAL EQUATION MODELS. Donna Mohr and Yong Xu. University of North Florida

SHOPPING FOR EFFICIENT CONFIDENCE INTERVALS IN STRUCTURAL EQUATION MODELS. Donna Mohr and Yong Xu. University of North Florida SHOPPING FOR EFFICIENT CONFIDENCE INTERVALS IN STRUCTURAL EQUATION MODELS Donna Mohr and Yong Xu University of North Florida Authors Note Parts of this work were incorporated in Yong Xu s Masters Thesis

More information

DIAGNOSTICS OF THE ERROR FACTOR COVARIANCES OTTÓ HAJDU 1

DIAGNOSTICS OF THE ERROR FACTOR COVARIANCES OTTÓ HAJDU 1 DIAGNOSTICS OF THE ERROR FACTOR COVARIANCES OTTÓ HAJDU In this paper we explore initial simple factor structure by the means of the so-called EPIC factor extraction method and the orthosim orthogonal rotational

More information

Chapter 1 Statistical Inference

Chapter 1 Statistical Inference Chapter 1 Statistical Inference causal inference To infer causality, you need a randomized experiment (or a huge observational study and lots of outside information). inference to populations Generalizations

More information

LECTURE 4 PRINCIPAL COMPONENTS ANALYSIS / EXPLORATORY FACTOR ANALYSIS

LECTURE 4 PRINCIPAL COMPONENTS ANALYSIS / EXPLORATORY FACTOR ANALYSIS LECTURE 4 PRINCIPAL COMPONENTS ANALYSIS / EXPLORATORY FACTOR ANALYSIS NOTES FROM PRE- LECTURE RECORDING ON PCA PCA and EFA have similar goals. They are substantially different in important ways. The goal

More information

Introduction to Within-Person Analysis and RM ANOVA

Introduction to Within-Person Analysis and RM ANOVA Introduction to Within-Person Analysis and RM ANOVA Today s Class: From between-person to within-person ANOVAs for longitudinal data Variance model comparisons using 2 LL CLP 944: Lecture 3 1 The Two Sides

More information

Path Analysis. PRE 906: Structural Equation Modeling Lecture #5 February 18, PRE 906, SEM: Lecture 5 - Path Analysis

Path Analysis. PRE 906: Structural Equation Modeling Lecture #5 February 18, PRE 906, SEM: Lecture 5 - Path Analysis Path Analysis PRE 906: Structural Equation Modeling Lecture #5 February 18, 2015 PRE 906, SEM: Lecture 5 - Path Analysis Key Questions for Today s Lecture What distinguishes path models from multivariate

More information

Hypothesis Testing for Var-Cov Components

Hypothesis Testing for Var-Cov Components Hypothesis Testing for Var-Cov Components When the specification of coefficients as fixed, random or non-randomly varying is considered, a null hypothesis of the form is considered, where Additional output

More information

Factor Analysis. Qian-Li Xue

Factor Analysis. Qian-Li Xue Factor Analysis Qian-Li Xue Biostatistics Program Harvard Catalyst The Harvard Clinical & Translational Science Center Short course, October 7, 06 Well-used latent variable models Latent variable scale

More information

Dimensionality Reduction Techniques (DRT)

Dimensionality Reduction Techniques (DRT) Dimensionality Reduction Techniques (DRT) Introduction: Sometimes we have lot of variables in the data for analysis which create multidimensional matrix. To simplify calculation and to get appropriate,

More information

SAS/STAT 14.1 User s Guide. Introduction to Structural Equation Modeling with Latent Variables

SAS/STAT 14.1 User s Guide. Introduction to Structural Equation Modeling with Latent Variables SAS/STAT 14.1 User s Guide Introduction to Structural Equation Modeling with Latent Variables This document is an individual chapter from SAS/STAT 14.1 User s Guide. The correct bibliographic citation

More information

A Re-Introduction to General Linear Models (GLM)

A Re-Introduction to General Linear Models (GLM) A Re-Introduction to General Linear Models (GLM) Today s Class: You do know the GLM Estimation (where the numbers in the output come from): From least squares to restricted maximum likelihood (REML) Reviewing

More information

Manual Of The Program FACTOR. v Windows XP/Vista/W7/W8. Dr. Urbano Lorezo-Seva & Dr. Pere Joan Ferrando

Manual Of The Program FACTOR. v Windows XP/Vista/W7/W8. Dr. Urbano Lorezo-Seva & Dr. Pere Joan Ferrando Manual Of The Program FACTOR v.9.20 Windows XP/Vista/W7/W8 Dr. Urbano Lorezo-Seva & Dr. Pere Joan Ferrando urbano.lorenzo@urv.cat perejoan.ferrando@urv.cat Departament de Psicologia Universitat Rovira

More information

Evaluating the Sensitivity of Goodness-of-Fit Indices to Data Perturbation: An Integrated MC-SGR Approach

Evaluating the Sensitivity of Goodness-of-Fit Indices to Data Perturbation: An Integrated MC-SGR Approach Evaluating the Sensitivity of Goodness-of-Fit Indices to Data Perturbation: An Integrated MC-SGR Approach Massimiliano Pastore 1 and Luigi Lombardi 2 1 Department of Psychology University of Cagliari Via

More information

Mixture Modeling. Identifying the Correct Number of Classes in a Growth Mixture Model. Davood Tofighi Craig Enders Arizona State University

Mixture Modeling. Identifying the Correct Number of Classes in a Growth Mixture Model. Davood Tofighi Craig Enders Arizona State University Identifying the Correct Number of Classes in a Growth Mixture Model Davood Tofighi Craig Enders Arizona State University Mixture Modeling Heterogeneity exists such that the data are comprised of two or

More information

UNIVERSITY OF CALGARY. The Influence of Model Components and Misspecification Type on the Performance of the

UNIVERSITY OF CALGARY. The Influence of Model Components and Misspecification Type on the Performance of the UNIVERSITY OF CALGARY The Influence of Model Components and Misspecification Type on the Performance of the Comparative Fit Index (CFI) and the Root Mean Square Error of Approximation (RMSEA) in Structural

More information

Applied Multivariate Analysis

Applied Multivariate Analysis Department of Mathematics and Statistics, University of Vaasa, Finland Spring 2017 Dimension reduction Exploratory (EFA) Background While the motivation in PCA is to replace the original (correlated) variables

More information

Nesting and Equivalence Testing

Nesting and Equivalence Testing Nesting and Equivalence Testing Tihomir Asparouhov and Bengt Muthén August 13, 2018 Abstract In this note, we discuss the nesting and equivalence testing (NET) methodology developed in Bentler and Satorra

More information

Model fit evaluation in multilevel structural equation models

Model fit evaluation in multilevel structural equation models Model fit evaluation in multilevel structural equation models Ehri Ryu Journal Name: Frontiers in Psychology ISSN: 1664-1078 Article type: Review Article Received on: 0 Sep 013 Accepted on: 1 Jan 014 Provisional

More information

Improper Solutions in Exploratory Factor Analysis: Causes and Treatments

Improper Solutions in Exploratory Factor Analysis: Causes and Treatments Improper Solutions in Exploratory Factor Analysis: Causes and Treatments Yutaka Kano Faculty of Human Sciences, Osaka University Suita, Osaka 565, Japan. email: kano@hus.osaka-u.ac.jp Abstract: There are

More information

Spatial inference. Spatial inference. Accounting for spatial correlation. Multivariate normal distributions

Spatial inference. Spatial inference. Accounting for spatial correlation. Multivariate normal distributions Spatial inference I will start with a simple model, using species diversity data Strong spatial dependence, Î = 0.79 what is the mean diversity? How precise is our estimate? Sampling discussion: The 64

More information

36-309/749 Experimental Design for Behavioral and Social Sciences. Dec 1, 2015 Lecture 11: Mixed Models (HLMs)

36-309/749 Experimental Design for Behavioral and Social Sciences. Dec 1, 2015 Lecture 11: Mixed Models (HLMs) 36-309/749 Experimental Design for Behavioral and Social Sciences Dec 1, 2015 Lecture 11: Mixed Models (HLMs) Independent Errors Assumption An error is the deviation of an individual observed outcome (DV)

More information

Step 2: Select Analyze, Mixed Models, and Linear.

Step 2: Select Analyze, Mixed Models, and Linear. Example 1a. 20 employees were given a mood questionnaire on Monday, Wednesday and again on Friday. The data will be first be analyzed using a Covariance Pattern model. Step 1: Copy Example1.sav data file

More information

PLEASE SCROLL DOWN FOR ARTICLE. Full terms and conditions of use:

PLEASE SCROLL DOWN FOR ARTICLE. Full terms and conditions of use: This article was downloaded by: [Howell, Roy][Texas Tech University] On: 15 December 2009 Access details: Access Details: [subscription number 907003254] Publisher Psychology Press Informa Ltd Registered

More information

Factor Analysis: An Introduction. What is Factor Analysis? 100+ years of Factor Analysis FACTOR ANALYSIS AN INTRODUCTION NILAM RAM

Factor Analysis: An Introduction. What is Factor Analysis? 100+ years of Factor Analysis FACTOR ANALYSIS AN INTRODUCTION NILAM RAM NILAM RAM 2018 PSYCHOLOGY R BOOTCAMP PENNSYLVANIA STATE UNIVERSITY AUGUST 16, 2018 FACTOR ANALYSIS https://psu-psychology.github.io/r-bootcamp-2018/index.html WITH ADDITIONAL MATERIALS AT https://quantdev.ssri.psu.edu/tutorials

More information

SEM for Categorical Outcomes

SEM for Categorical Outcomes This work is licensed under a Creative Commons Attribution-NonCommercial-ShareAlike License. Your use of this material constitutes acceptance of that license and the conditions of use of materials on this

More information

Analysis of Longitudinal Data: Comparison Between PROC GLM and PROC MIXED. Maribeth Johnson Medical College of Georgia Augusta, GA

Analysis of Longitudinal Data: Comparison Between PROC GLM and PROC MIXED. Maribeth Johnson Medical College of Georgia Augusta, GA Analysis of Longitudinal Data: Comparison Between PROC GLM and PROC MIXED Maribeth Johnson Medical College of Georgia Augusta, GA Overview Introduction to longitudinal data Describe the data for examples

More information

Effect of Estimation Method on Incremental Fit Indexes for Covariance Structure Models

Effect of Estimation Method on Incremental Fit Indexes for Covariance Structure Models Effect of Estimation Method on Incremental Fit Indexes for Covariance Structure Models Hazuki M. Sugawara and Robert C. MacCallum The Ohio State University In a typical study involving covariance structure

More information

Exploratory Factor Analysis and Principal Component Analysis

Exploratory Factor Analysis and Principal Component Analysis Exploratory Factor Analysis and Principal Component Analysis Today s Topics: What are EFA and PCA for? Planning a factor analytic study Analysis steps: Extraction methods How many factors Rotation and

More information

Statistical Distribution Assumptions of General Linear Models

Statistical Distribution Assumptions of General Linear Models Statistical Distribution Assumptions of General Linear Models Applied Multilevel Models for Cross Sectional Data Lecture 4 ICPSR Summer Workshop University of Colorado Boulder Lecture 4: Statistical Distributions

More information

Univariate Normal Distribution; GLM with the Univariate Normal; Least Squares Estimation

Univariate Normal Distribution; GLM with the Univariate Normal; Least Squares Estimation Univariate Normal Distribution; GLM with the Univariate Normal; Least Squares Estimation PRE 905: Multivariate Analysis Spring 2014 Lecture 4 Today s Class The building blocks: The basics of mathematical

More information

E X P L O R E R. R e l e a s e A Program for Common Factor Analysis and Related Models for Data Analysis

E X P L O R E R. R e l e a s e A Program for Common Factor Analysis and Related Models for Data Analysis E X P L O R E R R e l e a s e 3. 2 A Program for Common Factor Analysis and Related Models for Data Analysis Copyright (c) 1990-2011, J. S. Fleming, PhD Date and Time : 23-Sep-2011, 13:59:56 Number of

More information

dm'log;clear;output;clear'; options ps=512 ls=99 nocenter nodate nonumber nolabel FORMCHAR=" = -/\<>*"; ODS LISTING;

dm'log;clear;output;clear'; options ps=512 ls=99 nocenter nodate nonumber nolabel FORMCHAR= = -/\<>*; ODS LISTING; dm'log;clear;output;clear'; options ps=512 ls=99 nocenter nodate nonumber nolabel FORMCHAR=" ---- + ---+= -/\*"; ODS LISTING; *** Table 23.2 ********************************************; *** Moore, David

More information

Measuring relationships among multiple responses

Measuring relationships among multiple responses Measuring relationships among multiple responses Linear association (correlation, relatedness, shared information) between pair-wise responses is an important property used in almost all multivariate analyses.

More information

A Cautionary Note on the Use of LISREL s Automatic Start Values in Confirmatory Factor Analysis Studies R. L. Brown University of Wisconsin

A Cautionary Note on the Use of LISREL s Automatic Start Values in Confirmatory Factor Analysis Studies R. L. Brown University of Wisconsin A Cautionary Note on the Use of LISREL s Automatic Start Values in Confirmatory Factor Analysis Studies R. L. Brown University of Wisconsin The accuracy of parameter estimates provided by the major computer

More information

SEM Analysis of Epigenetic Data

SEM Analysis of Epigenetic Data SEM Analysis of Epigenetic Data By Azadeh Chizarifard Department of Statistics Uppsala University Supervisors: Åsa Johansson, Rolf Larsson 2014 Abstract DNA methylation as well as glucosylceramide has

More information

Measurement Invariance Testing with Many Groups: A Comparison of Five Approaches (Online Supplements)

Measurement Invariance Testing with Many Groups: A Comparison of Five Approaches (Online Supplements) University of South Florida Scholar Commons Educational and Psychological Studies Faculty Publications Educational and Psychological Studies 2017 Measurement Invariance Testing with Many Groups: A Comparison

More information

Introduction to Bayesian Statistics and Markov Chain Monte Carlo Estimation. EPSY 905: Multivariate Analysis Spring 2016 Lecture #10: April 6, 2016

Introduction to Bayesian Statistics and Markov Chain Monte Carlo Estimation. EPSY 905: Multivariate Analysis Spring 2016 Lecture #10: April 6, 2016 Introduction to Bayesian Statistics and Markov Chain Monte Carlo Estimation EPSY 905: Multivariate Analysis Spring 2016 Lecture #10: April 6, 2016 EPSY 905: Intro to Bayesian and MCMC Today s Class An

More information

Exploratory Factor Analysis and Principal Component Analysis

Exploratory Factor Analysis and Principal Component Analysis Exploratory Factor Analysis and Principal Component Analysis Today s Topics: What are EFA and PCA for? Planning a factor analytic study Analysis steps: Extraction methods How many factors Rotation and

More information