Factor Analysis. Qian-Li Xue
|
|
- Jane Holmes
- 6 years ago
- Views:
Transcription
1 Factor Analysis Qian-Li Xue Biostatistics Program Harvard Catalyst The Harvard Clinical & Translational Science Center Short course, October 7, 06
2 Well-used latent variable models Latent variable scale Observed variable scale Continuous Discrete Continuous Factor analysis Discrete LISREL Latent profile Growth miture Discrete FA IRT (item response) Latent class analysis, regression General software: MPlus, Latent Gold, WinBugs (Bayesian), NLMIXED (SAS)
3 Objectives What is factor analysis? What do we need factor analysis for? What are the modeling assumptions? How to specify, fit, and interpret factor models? What is the difference between eploratory and confirmatory factor analysis? What is and how to assess model identifiability? 3
4 What is factor analysis Factor analysis is a theory driven statistical data reduction technique used to eplain covariance among observed random variables in terms of fewer unobserved random variables named factors 4
5 An Eample: General Intelligence (Charles Spearman, 904) Y δ Y δ General Intelligence Y 3 δ 3 F Y 4 δ 4 Y 5 δ 5 5
6 Why Factor Analysis?. Testing of theory Eplain covariation among multiple observed variables by Mapping variables to latent constructs (called factors ). Understanding the structure underlying a set of measures Gain insight to dimensions Construct validation (e.g., convergent validity) 3. Scale development Eploit redundancy to improve scale s validity and reliability 6
7 Part I. Eploratory Factor Analysis (EFA) 7
8 One Common Factor Model: Model Specification Y δ Y = F + δ F 3 Y δ Y = F + δ Y 3 δ 3 Y 3 = 3 F + δ 3 The factor F is not observed; only Y, Y, Y 3 are observed δ i represent variability in the Y i NOT eplained by F Y i is a linear function of F and δ i 8
9 One Common Factor Model: Model Assumptions Y δ Y = F + δ F Y δ Y = F + δ 3 Y 3 δ 3 Y 3 = 3 F + δ 3 Factorial causation F is independent of δ j, i.e. cov(f,δ j )=0 δ i and δ j are independent for i j, i.e. cov(δ i,δ j )=0 Conditional independence: Given the factor, observed variables are independent of one another, i.e. cov( Y i,y j F ) = 0 9
10 One Common Factor Model: Model Interpretation F Y Y Y 3 = = = 3 Y Y δ δ Y 3 δ 3 3 F F F δ δ δ 3 Given all variables in standardized form, i.e. var(y i )=var(f)= Factor loadings: i i = corr(y i,f) Communality of Y i : h i h i = i = [corr(y i,f)] =% variance of Y i eplained by F Uniqueness of Y i : -h i = residual variance of Y i Degree of factorial determination: =Σ i /n, where n=# observed variables Y 0
11 Two-Common Factor Model (Orthogonal): Model Specification F F Y Y Y 3 Y 4 Y 5 δ δ δ 3 δ 4 δ 5 Y Y Y Y Y Y = F = = F = F F = F = F + F F F F F F + δ + δ + δ + δ + δ + δ Y 6 δ 6 F and F are common factors because they are shared by variables!
12 Matri Notation with n variables and m factors Y n = Λ nm F m + δ n + = n n m m m n nm n m n F F Y Y δ δ
13 Factor Pattern Matri Columns represent derived factors Rows represent input variables Loadings represent degree to which each of the variables correlates with each of the factors Loadings range from - to Inspection of factor loadings reveals etent to which each of the variables contributes to the meaning of each of the factors. High loadings provide meaning and interpretation of factors (~ regression coefficients) n m nm n m 3
14 Two-Common Factor Model (Orthogonal): Model Assumptions F F Y Y Y 3 Y 4 Y 5 δ δ δ 3 δ 4 δ 5 Factorial causation F and F are independent of δ j, i.e. cov(f,δ j )= cov(f,δ j )= 0 δ i and δ j are independent for i j, i.e. cov(δ i,δ j )=0 Conditional independence: Given factors F and F, observed variables are independent of one another, i.e. cov( Y i,y j F, F ) = 0 for i j Orthogonal (=independent): cov(f,f )=0 Y 6 δ 6 4
15 Two-Common Factor Model (Orthogonal): Model Interpretation F Y Y Y 3 Y 4 δ δ δ 3 δ 4 Given all variables in standardized form, i.e. var(y i )=var(f i )=; AND orthogonal factors, i.e. cov(f,f )=0 Factor loadings: ij ij = corr(y i,f j ) Communality of Y i : h i h i = i + i =% variance of Y i eplained by F AND F F 6 5 Y 5 δ 5 Uniqueness of Y i : -h i Y 6 δ 6 Degree of factorial determination: =Σ ij /n, n=# observed variables Y 5
16 Two-Common Factor Model : The Oblique Case F Y Y Y 3 δ δ δ 3 Given all variables in standardized form, i.e. var(y i )=var(f i )=; AND oblique factors (i.e. cov(f,f ) 0) The interpretation of factor loadings: ij is no longer correlation between Y and F; it is direct effect of F on Y 3 4 Y 4 δ 4 The calculation of communality of Yi (h i ) is more comple F 5 Y 5 δ 5 6 Y 6 δ 6 6
17 Etracting initial factors Least-squares method (e.g. principal ais factoring with iterated communalities) Maimum likelihood method 7
18 d) Repeat b) and c) until no improvement can be made 8 Model Fitting: Etracting initial factors Least-squares method (LS) (e.g. principal ais factoring with iterated communalities) v Goal: minimize the sum of squared differences between observed and estimated corr. matrices v Fitting steps: a) Obtain initial estimates of communalities (h ) e.g. squared correlation between a variable and the remaining variables b) Solve objective function: det(r LS -ηi)=0, where R LS is the corr matri with h in the main diag. (also termed adjusted corr matri), η is an eigenvalue c) Re-estimate h
19 Model Fitting: Etracting initial factors Maimum likelihood method (MLE) v v v v v v Goal: maimize the likelihood of producing the observed corr matri Assumption: distribution of variables (Y and F) is multivariate normal Objective function: det(r MLE - ηi)=0, where R MLE =U - (R-U )U - =U - R LS U -, and U is diag(-h ) Iterative fitting algorithm similar to LS approach Eception: adjust R by giving greater weights to correlations with smaller unique variance, i.e. - h Advantage: availability of a large sample χ significant test for goodness-of-fit (but tends to select more factors for large n!) 9
20 Choosing among Different Methods Between MLE and LS LS is preferred with v few indicators per factor v Equeal loadings within factors v No large cross-loadings v No factor correlations v Recovering factors with low loadings (overetraction) MLE if preferred with v Multivariate normality v unequal loadings within factors Both MLE and LS may have convergence problems 0
21 Factor Rotation Goal is simple structure Make factors more easily interpretable While keeping the number of factors and communalities of Ys fied!!! Rotation does NOT improve fit!
22 Factor Rotation To do this we rotate factors: redefine factors such that loadings (or pattern matri coefficients) on various factors tend to be very high (- or ) or very low (0) intuitively, it makes sharper distinctions in the meanings of the factors
23 Factor Rotation (Intuitively) F,, F 5 4 F 4 F Factor Factor Factor Factor
24 Factor Rotation Uses ambiguity or non-uniqueness of solution to make interpretation more simple Where does ambiguity come in? Unrotated solution is based on the idea that each factor tries to maimize variance eplained, conditional on previous factors What if we take that away? Then, there is not one best solution 4
25 Factor Rotation: Orthogonal vs. Oblique Rotation Orthogonal: Factors are independent varima: maimize variance of squared loadings across variables (sum over factors) v Goal: the simplicity of interpretation of factors quartima: maimize variance of squared loadings across factors (sum over variables) v Goal: the simplicity of interpretation of variables Intuition: from previous picture, there is a right angle between aes Note: Uniquenesses remain the same! 5
26 Factor Rotation: Orthogonal vs. Oblique Rotation Oblique: Factors are NOT independent. Change in angle. oblimin: minimize covariance of squared loadings between factors. proma: simplify orthogonal rotation by making small loadings even closer to zero. Target matri: choose simple structure a priori. Intuition: from previous picture, angle between aes is not necessarily a right angle. Note: Uniquenesses remain the same! 6
27 Pattern versus Structure Matri In oblique rotation, one typically presents both a pattern matri and a structure matri Also need to report correlation between the factors The pattern matri presents the usual factor loadings The structure matri presents correlations between the variables and the factors For orthogonal factors, pattern matri=structure matri The pattern matri is used to interpret the factors 7
28 Factor Rotation: Which to use? Choice is generally not critical Interpretation with orthogonal (varima) is simple because factors are independent: Loadings are correlations. Configuration may appear more simple in oblique (proma), but correlation of factors can be difficult to reconcile. Theory? Are the conceptual meanings of the factors associated? 8
29 Factor Rotation: Unique Solution? The factor analysis solution is NOT unique! More than one solution will yield the same result. 9
30 Derivation of Factor Scores Each object (e.g. each person) gets a factor score for each factor: The factors themselves are variables Object s score is weighted combination of scores on input variables Fˆ = WY ˆ, where Wˆ is the weight matri. These weights are NOT the factor loadings! Different approaches eist for estimating Wˆ (e.g. regression method) Factor scores are not unique Using factors scores instead of factor indicators can reduce measurement error, but does NOT remove it. Therefore, using factor scores as predictors in conventional regressions leads to inconsistent coefficient estimators! 30
31 Factor Analysis with Categorical Observed Variables Factor analysis hinges on the correlation matri As long as you can get an interpretable correlation matri, you can perform factor analysis Binary/ordinal items? Pearson corrlation: Epect attenuation! Tetrachoric correlation (binary) Polychoric correlation (ordinal) To obtain polychoric correlation in STATA: polychoric var var var3 var4 var5 To run princial component analysis: pcamat r(r), n(38) To run factor analysis: factormat r(r), fa() ipf n(38) 3
32 Criticisms of Factor Analysis Labels of factors can be arbitrary or lack scientific basis Derived factors often very obvious defense: but we get a quantification Garbage in, garbage out really a criticism of input variables factor analysis reorganizes input matri Correlation matri is often poor measure of association of input variables. 3
33 Major steps in EFA. Data collection and preparation. Choose number of factors to etract 3. Etracting initial factors 4. Rotation to a final solution 5. Model diagnosis/refinement 6. Derivation of factor scales to be used in further analysis 33
34 Part II. Confirmatory Factor Analysis (CFA) 34
35 Eploratory vs. Confirmatory Eploratory: summarize data Factor Analysis describe correlation structure between variables generate hypotheses Confirmatory Testing correlated measurement errors Redundancy test of one-factor vs. multi-factor models Measurement invariance test comparing a model across groups Orthogonality tests 35
36 Confirmatory Factor Analysis (CFA) Takes factor analysis a step further. We can test or confirm or implement a highly constrained a priori structure that meets conditions of model identification But be careful, a model can never be confirmed!! CFA model is constructed in advance number of latent variables ( factors ) is pre-set by analyst (not part of the modeling usually) Whether latent variable influences observed is specified Measurement errors may correlate Difference between CFA and the usual SEM: SEM assumes causally interrelated latent variables CFA assumes interrelated latent variables (i.e. eogenous) 36
37 Eploratory Factor Analysis Two factor model: = Λξ + δ = ξ ξ + δ δ δ 3 δ 4 δ 5 δ 6 ξ ξ δ δ δ 3 δ 4 δ 5 δ 6 37
38 CFA Notation + = δ δ δ δ δ δ ξ ξ Two factor model: = + Λξ δ δ δ δ 3 δ 4 δ 5 δ ξ ξ 38
39 Difference between CFA and EFA CFA = ξ + δ = ξ + δ = ξ + δ = ξ + δ = ξ + δ = ξ + δ cov( ξ, ξ ) = ϕ = ξ + ξ + δ = ξ + ξ + δ = ξ + ξ + δ = ξ + ξ + δ = ξ + ξ + δ = ξ + ξ + δ ξ ξ = 0 cov(, ) EFA 39
40 Model Constraints Hallmark of CFA Purposes for setting constraints: Test a priori theory Ensure identifiability Test reliability of measures 40
41 Identifiability Let θ be a t vector containing all unknown and unconstrained parameters in a model. The parameters θ are identified if Σ(θ )= Σ(θ ) θ =θ Estimability Identifiability!! Identifiability attribute of the model Estimability attribute of the data 4
42 Model Constraints: Identifiability Latent variables (LVs) need some constraints Because factors are unmeasured, their variances can take different values Recall EFA where we constrained factors: F ~ N(0,) Otherwise, model is not identifiable. Here we have two options: Fi variance of latent variables (LV) to be (or another constant) Fi one path between LV and indicator 4
43 Necessary Constraints Fi variances: Fi path: δ δ ξ δ ξ δ ξ 3 4 δ 3 δ 4 ξ 3 4 δ 3 δ 4 5 δ 5 5 δ 5 6 δ 6 6 δ 6 43
44 Model Parametrization Fi variances: = ξ + δ = ξ + δ = ξ + δ = ξ + δ = ξ + δ = ξ + δ cov( ξ, ξ ) var( ξ ) = var( ξ ) = = ϕ Fi path: = ξ + δ = ξ + δ = ξ + δ = ξ + δ 4 4 = ξ + δ = ξ + δ cov( ξ, ξ ) var( ξ ) var( ξ ) = = = ϕ ϕ ϕ 44
45 Identifiability Rules for CFA () Two-indicator rule (sufficient, not necessary) ) At least two factors ) At least two indicators per factor 3) Eactly one non-zero element per row of Λ (translation: each only is pointed at by one LV) 4) Non-correlated errors (Θ is diagonal) (translation: no double-header arrows between the δ s) 5) Factors are correlated (Φ has no zero elements)* (translation: there are double-header arrows between all of the LVs) * Alternative less strict criteria: each factor is correlated with at least one other factor. (see page 47 on Bollen) 45
46 = ξ ξ + δ δ δ 3 δ 4 δ 5 δ 6 ξ ξ δ δ δ 3 δ 4 δ 5 δ 6 Θ = var( δ) = θ θ θ θ θ θ66 Φ = = ϕ var( ξ) ϕ 46
47 Eample: Two-Indicator Rule δ ξ δ ξ 3 δ 3 4 δ 4 ξ 3 5 δ 5 6 δ 6 47
48 Eample: Two-Indicator Rule δ ξ δ ξ 3 δ 3 4 δ 4 ξ 3 5 δ 5 6 δ 6 48
49 Eample: Two-Indicator Rule δ ξ δ ξ 3 δ 3 4 δ 4 ξ 3 5 δ 5 6 δ 6 ξ 4 7 δ 7 8 δ 8 49
50 Identifiability Rules for CFA () Three-indicator rule (sufficient, not necessary) ) at least one factor ) at least three indicators per factor 3) one non-zero element per row of Λ (translation: each only is pointed at by one LV) 4) non-correlated errors (Θ is diagonal) (translation: no double-headed arrows between the δ s) [Note: no condition about correlation of factors (no restrictions on Φ).] δ ξ δ 3 δ 3 50
STAT 730 Chapter 9: Factor analysis
STAT 730 Chapter 9: Factor analysis Timothy Hanson Department of Statistics, University of South Carolina Stat 730: Multivariate Data Analysis 1 / 15 Basic idea Factor analysis attempts to explain the
More informationInference using structural equations with latent variables
This work is licensed under a Creative Commons Attribution-NonCommercial-ShareAlike License. Your use of this material constitutes acceptance of that license and the conditions of use of materials on this
More informationStructural Equation Modeling and Confirmatory Factor Analysis. Types of Variables
/4/04 Structural Equation Modeling and Confirmatory Factor Analysis Advanced Statistics for Researchers Session 3 Dr. Chris Rakes Website: http://csrakes.yolasite.com Email: Rakes@umbc.edu Twitter: @RakesChris
More informationComparing IRT with Other Models
Comparing IRT with Other Models Lecture #14 ICPSR Item Response Theory Workshop Lecture #14: 1of 45 Lecture Overview The final set of slides will describe a parallel between IRT and another commonly used
More informationIntroduction to Confirmatory Factor Analysis
Introduction to Confirmatory Factor Analysis Multivariate Methods in Education ERSH 8350 Lecture #12 November 16, 2011 ERSH 8350: Lecture 12 Today s Class An Introduction to: Confirmatory Factor Analysis
More informationSEM for Categorical Outcomes
This work is licensed under a Creative Commons Attribution-NonCommercial-ShareAlike License. Your use of this material constitutes acceptance of that license and the conditions of use of materials on this
More informationFactor analysis. George Balabanis
Factor analysis George Balabanis Key Concepts and Terms Deviation. A deviation is a value minus its mean: x - mean x Variance is a measure of how spread out a distribution is. It is computed as the average
More informationFactor Analysis Edpsy/Soc 584 & Psych 594
Factor Analysis Edpsy/Soc 584 & Psych 594 Carolyn J. Anderson University of Illinois, Urbana-Champaign April 29, 2009 1 / 52 Rotation Assessing Fit to Data (one common factor model) common factors Assessment
More informationMeasurement Invariance (MI) in CFA and Differential Item Functioning (DIF) in IRT/IFA
Topics: Measurement Invariance (MI) in CFA and Differential Item Functioning (DIF) in IRT/IFA What are MI and DIF? Testing measurement invariance in CFA Testing differential item functioning in IRT/IFA
More informationExploratory Factor Analysis and Principal Component Analysis
Exploratory Factor Analysis and Principal Component Analysis Today s Topics: What are EFA and PCA for? Planning a factor analytic study Analysis steps: Extraction methods How many factors Rotation and
More informationExploratory Factor Analysis and Principal Component Analysis
Exploratory Factor Analysis and Principal Component Analysis Today s Topics: What are EFA and PCA for? Planning a factor analytic study Analysis steps: Extraction methods How many factors Rotation and
More informationIntroduction to Factor Analysis
to Factor Analysis Lecture 10 August 2, 2011 Advanced Multivariate Statistical Methods ICPSR Summer Session #2 Lecture #10-8/3/2011 Slide 1 of 55 Today s Lecture Factor Analysis Today s Lecture Exploratory
More information2/26/2017. PSY 512: Advanced Statistics for Psychological and Behavioral Research 2
PSY 512: Advanced Statistics for Psychological and Behavioral Research 2 What is SEM? When should we use SEM? What can SEM tell us? SEM Terminology and Jargon Technical Issues Types of SEM Models Limitations
More informationLecture 3: Latent Variables Models and Learning with the EM Algorithm. Sam Roweis. Tuesday July25, 2006 Machine Learning Summer School, Taiwan
Lecture 3: Latent Variables Models and Learning with the EM Algorithm Sam Roweis Tuesday July25, 2006 Machine Learning Summer School, Taiwan Latent Variable Models What to do when a variable z is always
More informationIntroduction to Random Effects of Time and Model Estimation
Introduction to Random Effects of Time and Model Estimation Today s Class: The Big Picture Multilevel model notation Fixed vs. random effects of time Random intercept vs. random slope models How MLM =
More informationNesting and Equivalence Testing
Nesting and Equivalence Testing Tihomir Asparouhov and Bengt Muthén August 13, 2018 Abstract In this note, we discuss the nesting and equivalence testing (NET) methodology developed in Bentler and Satorra
More informationSTRUCTURAL EQUATION MODELING. Khaled Bedair Statistics Department Virginia Tech LISA, Summer 2013
STRUCTURAL EQUATION MODELING Khaled Bedair Statistics Department Virginia Tech LISA, Summer 2013 Introduction: Path analysis Path Analysis is used to estimate a system of equations in which all of the
More informationA Study of Statistical Power and Type I Errors in Testing a Factor Analytic. Model for Group Differences in Regression Intercepts
A Study of Statistical Power and Type I Errors in Testing a Factor Analytic Model for Group Differences in Regression Intercepts by Margarita Olivera Aguilar A Thesis Presented in Partial Fulfillment of
More informationMultivariate and Multivariable Regression. Stella Babalola Johns Hopkins University
Multivariate and Multivariable Regression Stella Babalola Johns Hopkins University Session Objectives At the end of the session, participants will be able to: Explain the difference between multivariable
More informationIntroduction to Factor Analysis
to Factor Analysis Lecture 11 November 2, 2005 Multivariate Analysis Lecture #11-11/2/2005 Slide 1 of 58 Today s Lecture Factor Analysis. Today s Lecture Exploratory factor analysis (EFA). Confirmatory
More informationGeneral structural model Part 2: Categorical variables and beyond. Psychology 588: Covariance structure and factor models
General structural model Part 2: Categorical variables and beyond Psychology 588: Covariance structure and factor models Categorical variables 2 Conventional (linear) SEM assumes continuous observed variables
More informationIntroduction to Structural Equation Modeling
Introduction to Structural Equation Modeling Notes Prepared by: Lisa Lix, PhD Manitoba Centre for Health Policy Topics Section I: Introduction Section II: Review of Statistical Concepts and Regression
More informationFactor Analysis: An Introduction. What is Factor Analysis? 100+ years of Factor Analysis FACTOR ANALYSIS AN INTRODUCTION NILAM RAM
NILAM RAM 2018 PSYCHOLOGY R BOOTCAMP PENNSYLVANIA STATE UNIVERSITY AUGUST 16, 2018 FACTOR ANALYSIS https://psu-psychology.github.io/r-bootcamp-2018/index.html WITH ADDITIONAL MATERIALS AT https://quantdev.ssri.psu.edu/tutorials
More informationApplied Multivariate Analysis
Department of Mathematics and Statistics, University of Vaasa, Finland Spring 2017 Dimension reduction Exploratory (EFA) Background While the motivation in PCA is to replace the original (correlated) variables
More informationDimensionality Reduction Techniques (DRT)
Dimensionality Reduction Techniques (DRT) Introduction: Sometimes we have lot of variables in the data for analysis which create multidimensional matrix. To simplify calculation and to get appropriate,
More informationExploratory Factor Analysis and Canonical Correlation
Exploratory Factor Analysis and Canonical Correlation 3 Dec 2010 CPSY 501 Dr. Sean Ho Trinity Western University Please download: SAQ.sav Outline for today Factor analysis Latent variables Correlation
More information1 A factor can be considered to be an underlying latent variable: (a) on which people differ. (b) that is explained by unknown variables
1 A factor can be considered to be an underlying latent variable: (a) on which people differ (b) that is explained by unknown variables (c) that cannot be defined (d) that is influenced by observed variables
More informationConsequences of measurement error. Psychology 588: Covariance structure and factor models
Consequences of measurement error Psychology 588: Covariance structure and factor models Scaling indeterminacy of latent variables Scale of a latent variable is arbitrary and determined by a convention
More informationMACHINE LEARNING ADVANCED MACHINE LEARNING
MACHINE LEARNING ADVANCED MACHINE LEARNING Recap of Important Notions on Estimation of Probability Density Functions 2 2 MACHINE LEARNING Overview Definition pdf Definition joint, condition, marginal,
More informationPrincipal Component Analysis & Factor Analysis. Psych 818 DeShon
Principal Component Analysis & Factor Analysis Psych 818 DeShon Purpose Both are used to reduce the dimensionality of correlated measurements Can be used in a purely exploratory fashion to investigate
More informationAdvanced Structural Equations Models I
This work is licensed under a Creative Commons Attribution-NonCommercial-ShareAlike License. Your use of this material constitutes acceptance of that license and the conditions of use of materials on this
More informationLatent Class Analysis
Latent Class Analysis Karen Bandeen-Roche October 27, 2016 Objectives For you to leave here knowing When is latent class analysis (LCA) model useful? What is the LCA model its underlying assumptions? How
More informationChapter 8. Models with Structural and Measurement Components. Overview. Characteristics of SR models. Analysis of SR models. Estimation of SR models
Chapter 8 Models with Structural and Measurement Components Good people are good because they've come to wisdom through failure. Overview William Saroyan Characteristics of SR models Estimation of SR models
More informationGeneral structural model Part 1: Covariance structure and identification. Psychology 588: Covariance structure and factor models
General structural model Part 1: Covariance structure and identification Psychology 588: Covariance structure and factor models Latent variables 2 Interchangeably used: constructs --- substantively defined
More informationGrowth Mixture Model
Growth Mixture Model Latent Variable Modeling and Measurement Biostatistics Program Harvard Catalyst The Harvard Clinical & Translational Science Center Short course, October 28, 2016 Slides contributed
More informationWhat is in the Book: Outline
Estimating and Testing Latent Interactions: Advancements in Theories and Practical Applications Herbert W Marsh Oford University Zhonglin Wen South China Normal University Hong Kong Eaminations Authority
More informationA Practitioner s Guide to Generalized Linear Models
A Practitioners Guide to Generalized Linear Models Background The classical linear models and most of the minimum bias procedures are special cases of generalized linear models (GLMs). GLMs are more technically
More informationPrinciple Components Analysis (PCA) Relationship Between a Linear Combination of Variables and Axes Rotation for PCA
Principle Components Analysis (PCA) Relationship Between a Linear Combination of Variables and Axes Rotation for PCA Principle Components Analysis: Uses one group of variables (we will call this X) In
More informationMeasurement Theory. Reliability. Error Sources. = XY r XX. r XY. r YY
Y -3 - -1 0 1 3 X Y -10-5 0 5 10 X Measurement Theory t & X 1 X X 3 X k Reliability e 1 e e 3 e k 1 The Big Picture Measurement error makes it difficult to identify the true patterns of relationships between
More informationResearch Design - - Topic 15a Introduction to Multivariate Analyses 2009 R.C. Gardner, Ph.D.
Research Design - - Topic 15a Introduction to Multivariate Analses 009 R.C. Gardner, Ph.D. Major Characteristics of Multivariate Procedures Overview of Multivariate Techniques Bivariate Regression and
More informationFACTOR ANALYSIS AND MULTIDIMENSIONAL SCALING
FACTOR ANALYSIS AND MULTIDIMENSIONAL SCALING Vishwanath Mantha Department for Electrical and Computer Engineering Mississippi State University, Mississippi State, MS 39762 mantha@isip.msstate.edu ABSTRACT
More information6. Vector Random Variables
6. Vector Random Variables In the previous chapter we presented methods for dealing with two random variables. In this chapter we etend these methods to the case of n random variables in the following
More informationRelated Concepts: Lecture 9 SEM, Statistical Modeling, AI, and Data Mining. I. Terminology of SEM
Lecture 9 SEM, Statistical Modeling, AI, and Data Mining I. Terminology of SEM Related Concepts: Causal Modeling Path Analysis Structural Equation Modeling Latent variables (Factors measurable, but thru
More informationMACHINE LEARNING ADVANCED MACHINE LEARNING
MACHINE LEARNING ADVANCED MACHINE LEARNING Recap of Important Notions on Estimation of Probability Density Functions 22 MACHINE LEARNING Discrete Probabilities Consider two variables and y taking discrete
More information2 Statistical Estimation: Basic Concepts
Technion Israel Institute of Technology, Department of Electrical Engineering Estimation and Identification in Dynamical Systems (048825) Lecture Notes, Fall 2009, Prof. N. Shimkin 2 Statistical Estimation:
More informationLecture 3: Pattern Classification. Pattern classification
EE E68: Speech & Audio Processing & Recognition Lecture 3: Pattern Classification 3 4 5 The problem of classification Linear and nonlinear classifiers Probabilistic classification Gaussians, mitures and
More informationIntermediate Social Statistics
Intermediate Social Statistics Lecture 5. Factor Analysis Tom A.B. Snijders University of Oxford January, 2008 c Tom A.B. Snijders (University of Oxford) Intermediate Social Statistics January, 2008 1
More informationMultivariate Fundamentals: Rotation. Exploratory Factor Analysis
Multivariate Fundamentals: Rotation Exploratory Factor Analysis PCA Analysis A Review Precipitation Temperature Ecosystems PCA Analysis with Spatial Data Proportion of variance explained Comp.1 + Comp.2
More informationUnconstrained Ordination
Unconstrained Ordination Sites Species A Species B Species C Species D Species E 1 0 (1) 5 (1) 1 (1) 10 (4) 10 (4) 2 2 (3) 8 (3) 4 (3) 12 (6) 20 (6) 3 8 (6) 20 (6) 10 (6) 1 (2) 3 (2) 4 4 (5) 11 (5) 8 (5)
More informationChapter 4: Factor Analysis
Chapter 4: Factor Analysis In many studies, we may not be able to measure directly the variables of interest. We can merely collect data on other variables which may be related to the variables of interest.
More informationCorrelation and simple linear regression S5
Basic medical statistics for clinical and eperimental research Correlation and simple linear regression S5 Katarzyna Jóźwiak k.jozwiak@nki.nl November 15, 2017 1/41 Introduction Eample: Brain size and
More informationGeneralized Linear Models for Non-Normal Data
Generalized Linear Models for Non-Normal Data Today s Class: 3 parts of a generalized model Models for binary outcomes Complications for generalized multivariate or multilevel models SPLH 861: Lecture
More informationCan Variances of Latent Variables be Scaled in Such a Way That They Correspond to Eigenvalues?
International Journal of Statistics and Probability; Vol. 6, No. 6; November 07 ISSN 97-703 E-ISSN 97-7040 Published by Canadian Center of Science and Education Can Variances of Latent Variables be Scaled
More informationPrinciples of factor analysis. Roger Watson
Principles of factor analysis Roger Watson Factor analysis Factor analysis Factor analysis Factor analysis is a multivariate statistical method for reducing large numbers of variables to fewer underlying
More informationIntroduction to Structural Equation Modeling Dominique Zephyr Applied Statistics Lab
Applied Statistics Lab Introduction to Structural Equation Modeling Dominique Zephyr Applied Statistics Lab SEM Model 3.64 7.32 Education 2.6 Income 2.1.6.83 Charac. of Individuals 1 5.2e-06 -.62 2.62
More informationAssessing Factorial Invariance in Ordered-Categorical Measures
Multivariate Behavioral Research, 39 (3), 479-515 Copyright 2004, Lawrence Erlbaum Associates, Inc. Assessing Factorial Invariance in Ordered-Categorical Measures Roger E. Millsap and Jenn Yun-Tein Arizona
More informationADVANCED C. MEASUREMENT INVARIANCE SEM REX B KLINE CONCORDIA
ADVANCED SEM C. MEASUREMENT INVARIANCE REX B KLINE CONCORDIA C C2 multiple model 2 data sets simultaneous C3 multiple 2 populations 2 occasions 2 methods C4 multiple unstandardized constrain to equal fit
More informationSTA 216, GLM, Lecture 16. October 29, 2007
STA 216, GLM, Lecture 16 October 29, 2007 Efficient Posterior Computation in Factor Models Underlying Normal Models Generalized Latent Trait Models Formulation Genetic Epidemiology Illustration Structural
More informationPreface. List of examples
Contents Preface List of examples i xix 1 LISREL models and methods 1 1.1 The general LISREL model 1 Assumptions 2 The covariance matrix of the observations as implied by the LISREL model 3 Fixed, free,
More informationKey Algebraic Results in Linear Regression
Key Algebraic Results in Linear Regression James H. Steiger Department of Psychology and Human Development Vanderbilt University James H. Steiger (Vanderbilt University) 1 / 30 Key Algebraic Results in
More informationSC705: Advanced Statistics Instructor: Natasha Sarkisian Class notes: Introduction to Structural Equation Modeling (SEM)
SC705: Advanced Statistics Instructor: Natasha Sarkisian Class notes: Introduction to Structural Equation Modeling (SEM) SEM is a family of statistical techniques which builds upon multiple regression,
More informationSEM with observed variables: parameterization and identification. Psychology 588: Covariance structure and factor models
SEM with observed variables: parameterization and identification Psychology 588: Covariance structure and factor models Limitations of SEM as a causal modeling 2 If an SEM model reflects the reality, the
More informationMultivariate Statistics
Multivariate Statistics Chapter 4: Factor analysis Pedro Galeano Departamento de Estadística Universidad Carlos III de Madrid pedro.galeano@uc3m.es Course 2017/2018 Master in Mathematical Engineering Pedro
More informationCS281 Section 4: Factor Analysis and PCA
CS81 Section 4: Factor Analysis and PCA Scott Linderman At this point we have seen a variety of machine learning models, with a particular emphasis on models for supervised learning. In particular, we
More informationPrincipal Component Analysis
Principal Component Analysis 1 Principal Component Analysis Principal component analysis is a technique used to construct composite variable(s) such that composite variable(s) are weighted combination
More informationSEM 2: Structural Equation Modeling
SEM 2: Structural Equation Modeling Week 1 - Causal modeling and SEM Sacha Epskamp 18-04-2017 Course Overview Mondays: Lecture Wednesdays: Unstructured practicals Three assignments First two 20% of final
More informationTHE GENERAL STRUCTURAL EQUATION MODEL WITH LATENT VARIATES
THE GENERAL STRUCTURAL EQUATION MODEL WITH LATENT VARIATES I. Specification: A full structural equation model with latent variables consists of two parts: a latent variable model (which specifies the relations
More informationConfirmatory Factor Analysis: Model comparison, respecification, and more. Psychology 588: Covariance structure and factor models
Confirmatory Factor Analysis: Model comparison, respecification, and more Psychology 588: Covariance structure and factor models Model comparison 2 Essentially all goodness of fit indices are descriptive,
More informationFactor Analysis. Robert L. Wolpert Department of Statistical Science Duke University, Durham, NC, USA
Factor Analysis Robert L. Wolpert Department of Statistical Science Duke University, Durham, NC, USA 1 Factor Models The multivariate regression model Y = XB +U expresses each row Y i R p as a linear combination
More informationCONFIRMATORY FACTOR ANALYSIS
1 CONFIRMATORY FACTOR ANALYSIS The purpose of confirmatory factor analysis (CFA) is to explain the pattern of associations among a set of observed variables in terms of a smaller number of underlying latent
More informationExperimental Design and Data Analysis for Biologists
Experimental Design and Data Analysis for Biologists Gerry P. Quinn Monash University Michael J. Keough University of Melbourne CAMBRIDGE UNIVERSITY PRESS Contents Preface page xv I I Introduction 1 1.1
More informationBelowisasamplingofstructuralequationmodelsthatcanbefitby sem.
Title intro4 Tourofmodels Description Belowisasamplingofstructuralequationmodelsthatcanbefitby sem. Remarks Ifyouhavenotread [SEM]intro 2, please doso.youneed to speak thelanguage.wealso recommend reading[sem]
More informationAn Introduction to Mplus and Path Analysis
An Introduction to Mplus and Path Analysis PSYC 943: Fundamentals of Multivariate Modeling Lecture 10: October 30, 2013 PSYC 943: Lecture 10 Today s Lecture Path analysis starting with multivariate regression
More informationSpecifying Latent Curve and Other Growth Models Using Mplus. (Revised )
Ronald H. Heck 1 University of Hawai i at Mānoa Handout #20 Specifying Latent Curve and Other Growth Models Using Mplus (Revised 12-1-2014) The SEM approach offers a contrasting framework for use in analyzing
More informationFactor Analysis (10/2/13)
STA561: Probabilistic machine learning Factor Analysis (10/2/13) Lecturer: Barbara Engelhardt Scribes: Li Zhu, Fan Li, Ni Guan Factor Analysis Factor analysis is related to the mixture models we have studied.
More informationStat 5101 Lecture Notes
Stat 5101 Lecture Notes Charles J. Geyer Copyright 1998, 1999, 2000, 2001 by Charles J. Geyer May 7, 2001 ii Stat 5101 (Geyer) Course Notes Contents 1 Random Variables and Change of Variables 1 1.1 Random
More informationFactor Analysis Continued. Psy 524 Ainsworth
Factor Analysis Continued Psy 524 Ainsworth Equations Extraction Principal Axis Factoring Variables Skiers Cost Lift Depth Powder S1 32 64 65 67 S2 61 37 62 65 S3 59 40 45 43 S4 36 62 34 35 S5 62 46 43
More informationMetric-based classifiers. Nuno Vasconcelos UCSD
Metric-based classifiers Nuno Vasconcelos UCSD Statistical learning goal: given a function f. y f and a collection of eample data-points, learn what the function f. is. this is called training. two major
More informationFREQUENTIST BEHAVIOR OF FORMAL BAYESIAN INFERENCE
FREQUENTIST BEHAVIOR OF FORMAL BAYESIAN INFERENCE Donald A. Pierce Oregon State Univ (Emeritus), RERF Hiroshima (Retired), Oregon Health Sciences Univ (Adjunct) Ruggero Bellio Univ of Udine For Perugia
More informationModeration 調節 = 交互作用
Moderation 調節 = 交互作用 Kit-Tai Hau 侯傑泰 JianFang Chang 常建芳 The Chinese University of Hong Kong Based on Marsh, H. W., Hau, K. T., Wen, Z., Nagengast, B., & Morin, A. J. S. (in press). Moderation. In Little,
More informationGeneralized Models: Part 1
Generalized Models: Part 1 Topics: Introduction to generalized models Introduction to maximum likelihood estimation Models for binary outcomes Models for proportion outcomes Models for categorical outcomes
More informationImproper Solutions in Exploratory Factor Analysis: Causes and Treatments
Improper Solutions in Exploratory Factor Analysis: Causes and Treatments Yutaka Kano Faculty of Human Sciences, Osaka University Suita, Osaka 565, Japan. email: kano@hus.osaka-u.ac.jp Abstract: There are
More informationSupplementary materials for: Exploratory Structural Equation Modeling
Supplementary materials for: Exploratory Structural Equation Modeling To appear in Hancock, G. R., & Mueller, R. O. (Eds.). (2013). Structural equation modeling: A second course (2nd ed.). Charlotte, NC:
More informationPRINCIPAL COMPONENTS ANALYSIS
121 CHAPTER 11 PRINCIPAL COMPONENTS ANALYSIS We now have the tools necessary to discuss one of the most important concepts in mathematical statistics: Principal Components Analysis (PCA). PCA involves
More informationBayesian Analysis of Latent Variable Models using Mplus
Bayesian Analysis of Latent Variable Models using Mplus Tihomir Asparouhov and Bengt Muthén Version 2 June 29, 2010 1 1 Introduction In this paper we describe some of the modeling possibilities that are
More informationA note on structured means analysis for a single group. André Beauducel 1. October 3 rd, 2015
Structured means analysis for a single group 1 A note on structured means analysis for a single group André Beauducel 1 October 3 rd, 2015 Abstract The calculation of common factor means in structured
More informationComputation fundamentals of discrete GMRF representations of continuous domain spatial models
Computation fundamentals of discrete GMRF representations of continuous domain spatial models Finn Lindgren September 23 2015 v0.2.2 Abstract The fundamental formulas and algorithms for Bayesian spatial
More information6.867 Machine learning
6.867 Machine learning Mid-term eam October 8, 6 ( points) Your name and MIT ID: .5.5 y.5 y.5 a).5.5 b).5.5.5.5 y.5 y.5 c).5.5 d).5.5 Figure : Plots of linear regression results with different types of
More informationMultivariate Statistical Analysis
Multivariate Statistical Analysis Fall 2011 C. L. Williams, Ph.D. Lecture 4 for Applied Multivariate Analysis Outline 1 Eigen values and eigen vectors Characteristic equation Some properties of eigendecompositions
More informationSEM 2: Structural Equation Modeling
SEM 2: Structural Equation Modeling Week 2 - Causality and equivalent models Sacha Epskamp 15-05-2018 Covariance Algebra Let Var(x) indicate the variance of x and Cov(x, y) indicate the covariance between
More information6. Let C and D be matrices conformable to multiplication. Then (CD) =
Quiz 1. Name: 10 points per correct answer. (20 points for attendance). 1. Let A = 3 and B = [3 yy]. When is A equal to B? xx A. When x = 3 B. When y = 3 C. When x = y D. Never 2. See 1. What is the dimension
More informationPackage paramap. R topics documented: September 20, 2017
Package paramap September 20, 2017 Type Package Title paramap Version 1.4 Date 2017-09-20 Author Brian P. O'Connor Maintainer Brian P. O'Connor Depends R(>= 1.9.0), psych, polycor
More informationA Threshold-Free Approach to the Study of the Structure of Binary Data
International Journal of Statistics and Probability; Vol. 2, No. 2; 2013 ISSN 1927-7032 E-ISSN 1927-7040 Published by Canadian Center of Science and Education A Threshold-Free Approach to the Study of
More informationIntroduction to Generalized Models
Introduction to Generalized Models Today s topics: The big picture of generalized models Review of maximum likelihood estimation Models for binary outcomes Models for proportion outcomes Models for categorical
More informationPLEASE SCROLL DOWN FOR ARTICLE
This article was downloaded by: [CDL Journals Account] On: 13 August 2009 Access details: Access Details: [subscription number 794379768] Publisher Psychology Press Informa Ltd Registered in England and
More informationLECTURE 4 PRINCIPAL COMPONENTS ANALYSIS / EXPLORATORY FACTOR ANALYSIS
LECTURE 4 PRINCIPAL COMPONENTS ANALYSIS / EXPLORATORY FACTOR ANALYSIS NOTES FROM PRE- LECTURE RECORDING ON PCA PCA and EFA have similar goals. They are substantially different in important ways. The goal
More informationCHAPTER 9 EXAMPLES: MULTILEVEL MODELING WITH COMPLEX SURVEY DATA
Examples: Multilevel Modeling With Complex Survey Data CHAPTER 9 EXAMPLES: MULTILEVEL MODELING WITH COMPLEX SURVEY DATA Complex survey data refers to data obtained by stratification, cluster sampling and/or
More informationAssessment, analysis and interpretation of Patient Reported Outcomes (PROs)
Assessment, analysis and interpretation of Patient Reported Outcomes (PROs) Day 2 Summer school in Applied Psychometrics Peterhouse College, Cambridge 12 th to 16 th September 2011 This course is prepared
More informationA Re-Introduction to General Linear Models
A Re-Introduction to General Linear Models Today s Class: Big picture overview Why we are using restricted maximum likelihood within MIXED instead of least squares within GLM Linear model interpretation
More information9.1 Orthogonal factor model.
36 Chapter 9 Factor Analysis Factor analysis may be viewed as a refinement of the principal component analysis The objective is, like the PC analysis, to describe the relevant variables in study in terms
More information