Factor Analysis Edpsy/Soc 584 & Psych 594

Size: px
Start display at page:

Download "Factor Analysis Edpsy/Soc 584 & Psych 594"

Transcription

1 Factor Analysis Edpsy/Soc 584 & Psych 594 Carolyn J. Anderson University of Illinois, Urbana-Champaign April 29, / 52

2 Rotation Assessing Fit to Data (one common factor model) common factors Assessment of Fit to Data Rotation Confirmatory Factor Analysis 2 / 52

3 : Rotation Assessing Fit to Data Explore underlying structure data driven. Confirm underlying structure theory driven. method (Factor extraction) Eigen-decomposition based Eigen-decomposition of S which is the principal components solution. Iterative eigen-decompositions of S Ψ which is the Principal factor solution. Maximum likelihood estimation We now must assume that F and ɛ are multivariate normal. Tends to fit data better & yields scale invariance (ie., use either S or R). 3 / 52

4 Rotation Assessing Fit to Data Data from Espelage, D.L., Holt, M.K., & Henkel, R.R. (2003). Examination of Peer-Group contextual effects on aggression during early adolescence. Child Development, 74, Items have a 5 point response scale: Never, 1 to 2 times, 3 to 4 times, 5 to 6 times, 7 or more times Q36: I upset other students for the fun of it. Q37: In a group I teased other students. Q38: I fought students I could easily beat. Q39: Other students picked on me. Q40: Other students made fun of me. Q41: Other students called me names. Q42: I got hit and pushed by other students. Q43: I helped harass other students. 4 / 52

5 Rotation Assessing Fit to Data Items have a 5 point response scale: Never, 1 to 2 times,..., 7 or more times Q44: I teased other students. Q45: I got in a physical fight. Q46: I threatened to hurt or hit another student. Q47: I got into a physical fight because I was angry. Q48: I hit back when someone hit me first. Q49: I was mean to someone when I was angry. Q50: I spread rumors about other students. Q51: I started (instigated) arguments or conflicts. Q52: I encouraged people to fight. Q53: I excluded other students from my clique (group) of friends 5 / 52

6 Rotation Assessing Fit to Data What are the factors and which items load on which factors? Item F 1 F 2 F 3 Item F 1 F 2 F / 52

7 Rotation Assessing Fit to Data α here is Chronbach s alpha for measuring reliability. F 1 = Fighting (α =.88), F 2 = Bullying (α =.88), F 3 = Victimization (α =.87) Item F 1 F 2 F 3 Item F 1 F 2 F 3 36 X 45 X 37 X 46 X 38 X 47 X 39 X 48 X 40 X 49 X 41 X 50 X 42 X 51 X 43 X 52 X 44 X 53 X 7 / 52

8 Factor Analytic The One Common Factor Analytic Unique Variables Unique Variables & Common Variables Implications for Data Implications for Data (continued) The Data is Covariance Matrix Victim Statistics One Factor Solution Rotation 8 / 52

9 Factor Analytic Factor Analytic The One Common Factor Analytic Unique Variables Unique Variables & Common Variables Implications for Data Implications for Data (continued) The Data is Covariance Matrix Victim Statistics One Factor Solution Factor analysis (FA) is a latent variable model. ɛ 1 X 1 ɛ 2 X 2 ɛ 3 X 3 ɛ 4 X 4 l 21 l 11 l 31 l 41 F 1 The One Common Factor X 1 = µ 1 + l 11 F 1 + ɛ 1 X 2 = µ 2 + l 21 F 1 + ɛ 2 X 3 = µ 3 + l 31 F 1 + ɛ 3 X 4 = µ 4 + l 41 F 1 + ɛ 4 Rotation 9 / 52

10 The One Common Factor Analytic Factor Analytic The One Common Factor Analytic Unique Variables Unique Variables & Common Variables Implications for Data Implications for Data (continued) The Data is Covariance Matrix Victim Statistics One Factor Solution X i = µ i + l i1 F 1 + ɛ i The µ i s are the means of the X i s. Common Factor: F 1 is an unobserved random variable with mean 0 and variance φ. The l i1 s are the Factor Loadings. Specificities or Uniquenesses: ɛ i are independent over i, unobserved random variables with means equal to 0 and variances equal to ψ i. F 1 is independent of the ɛ i s. Rotation 10 / 52

11 Factor Analytic The One Common Factor Analytic Unique Variables Unique Variables & Common Variables Implications for Data Implications for Data (continued) The Data is Covariance Matrix Victim Statistics One Factor Solution Unique Variables A little Classical test theory: true score = observed score pure measurement error t i = X i e i If a factor model holds for the observed scores X i, then it should also hold for the true scores t i. The uniqueness in the factor model for the true scores contains specific errors due to the particular variables (items) selected; that is, Solve for X i : t i = l i1 F 1 + s i }{{} specif ic t i = l i1 F 1 + s i = X i e i = X i = l i1 F i + (s i + e i ) }{{} ɛ i Rotation 11 / 52

12 Unique Variables & Common Variables The unique variables ɛ i contain Pure measurement errors Specific errors Factor Analytic The One Common Factor Analytic Unique Variables Unique Variables & Common Variables Implications for Data Implications for Data (continued) The Data is Covariance Matrix Victim Statistics One Factor Solution The observed variables will be correlated because they all depend (in part) on F 1. The common factor model makes implications for covariance (and correlation) matrices; therefore, the data are the covariance (or correlation) matrix. Rotation 12 / 52

13 Factor Analytic The One Common Factor Analytic Unique Variables Unique Variables & Common Variables Implications for Data Implications for Data (continued) The Data is Covariance Matrix Victim Statistics One Factor Solution Implications for Data Let s turn it into a linear algebra problem: X 1 µ 1 X 2 µ 2 X =. µ =... Σ = X p L = l 11 l 21. l p1 Rotation 13 / 52 µ p ɛ = Ψ = So X µ = LF 1 + ɛ ɛ 1 ɛ 2.. ɛ p ψ ψ ψ p σ 11 σ σ 1p σ 12 σ σ 2p σ 1p σ 2p... σ pp var(f) = Φ = φ 11

14 Implications for Data (continued) Factor Analytic The One Common Factor Analytic Unique Variables Unique Variables & Common Variables Implications for Data Implications for Data (continued) The Data is Covariance Matrix Victim Statistics One Factor Solution X µ = LF 1 + ɛ Mean of X µ = LE(F 1 ) + E(ɛ) = 0 Covariance matrix for X or equivalently X µ: Σ = E[(X µ)(x µ) ] = E[(LF 1 + ɛ)(lf 1 ɛ) ] = E[(LF 1 + ɛ)(f 1 L ɛ )] = L E[F1 2 ] L + L E[F 1 ɛ ] +E[ɛF 1 ] L + E[ɛɛ ] }{{}}{{}}{{}}{{} φ 11 =1 0 0 Ψ = LL + Ψ Rotation 14 / 52

15 Factor Analytic The One Common Factor Analytic Unique Variables Unique Variables & Common Variables Implications for Data Implications for Data (continued) The Data is Covariance Matrix Victim Statistics One Factor Solution The Data is Covariance Matrix One common factor model implies : Σ = l 2 i1 = h2 i (l ψ 1) l 11 l l 11 l p1 l 11 l 21 (l ψ 2)... l 21 l p l 11 l I1 l 21 l I1... (l 2 I1 + ψ p) is the Communality of item i. Foreshadowing estimation methods: If Ψ were known, or Σ = (e 1 Σ Ψ = LL = (e 1 λ1 )( λ 1 e 1) λ 1 ) } {{ } L ( λ 1 e 1 ) } {{ } L Rotation 15 / 52 and Ψ = diag(σ L L )

16 Victim Statistics Factor Analytic The One Common Factor Analytic Unique Variables Unique Variables & Common Variables Implications for Data Implications for Data (continued) The Data is Covariance Matrix Victim Statistics One Factor Solution Descriptives Correlations Item Mean Std Dev q39 q40 q41 q42 q q q q Cronbach Coefficient Alpha with Deleted Variable Raw Variables Std Variables Correlation Correlation Deleted Variable with Total Alpha with Total Al q39: other students picked on me q40: students made fun of me q41: students called me names q42: got hit and pushed Rotation 16 / 52

17 One Factor Solution Factor Analytic The One Common Factor Analytic Unique Variables Unique Variables & Common Variables Implications for Data Implications for Data (continued) The Data is Covariance Matrix Victim Statistics One Factor Solution Factor Pattern Item Factor1 Communalities Specific l i1 h 2 i var: ψ i q39 other students picked on me q40 students made fun of me q41 students called me names q42 got hit and pushed (maximum likelihood estimation using the correlation matrix) Residual Correlation with Uniqueness ( ˆψ i ) on the diagonal: q39 q40 q41 q42 q q q q Root Mean Square Off-Diagonal Residuals: Overall = Rotation 17 / 52

18 A Closer Look at Implied Covariance Matrix Summary of Terminology and Components & Victim s Plot of Factor 18 / 52

19 Picture of this for p = 4 and m = 2. ɛ 1 X 1 ɛ 2 X 2 F 1 A Closer Look at Implied Covariance Matrix Summary of Terminology and Components & Victim s Plot of Factor ɛ 3 ɛ 4 X 3 X 4 F 2 19 / 52

20 A Closer Look at Implied Covariance Matrix Summary of Terminology and Components & Victim s Plot of Factor X 1 = µ 1 + l 11 F 1 + l 12 F l 1m F m + ɛ 1 X 2 = µ 2 + l 21 F 1 + l 22 F l 2m F m + ɛ 2.. X p = µ p + l p1 F 1 + l p2 F l pm F m + ɛ p X = µ + LF + ɛ E(F) = 0 and cov(f) = Φ = I... for now E(ɛ) = 0 and cov(ɛ) = Ψ = diag(ψ i ) cov(f,ɛ) = 0 L is matrix of Factor Loadings m q=1 l 2 iq = h2 i =Communality of item i. ψ i is Specific variance of item i. 20 / 52

21 A Closer Look at Implied Covariance Matrix Summary of Terminology and Components & Victim s Plot of Factor Using the model for i = 1,..., I observed variables and m common factors, the covariance matrix is X µ = LF + ɛ Mean of X µ= E(LF + ɛ) = LE(F) + E(ɛ) = 0 Covariance matrix for X or equivalently X µ: Σ = E[(X µ)(x µ) ] = E[(LF + ɛ)(lf ɛ) ] = E[(LF + ɛ)(f L ɛ )] = L E[FF ] }{{} I = LL + Ψ L + L E[F1ɛ ] +E[ɛF ] }{{}}{{} 0 I L + E[ɛɛ ] }{{} Ψ 21 / 52

22 The Common factor model implies : Σ = ( q l2 1q + ψ 1) q l 1ql 2q... q l 1ql pq q l 1ql 2q ( q l2 2q + ψ 2)... q l 2ql pq q l 1q l pq q l 2q l pq... ( q l 2 pq + ψ I ) A Closer Look at Implied Covariance Matrix Summary of Terminology and Components & Victim s Plot of Factor Variance for X i : σ ii = m q=1 l 2 im }{{} h 2 i +ψ i Covariance between X i and X k : m σ ik = l iq l kq q=1 22 / 52

23 A Closer Look at Implied Covariance Matrix A Closer Look at Implied Covariance Matrix Summary of Terminology and Components & Victim s Plot of Factor Covariance between X and F: cov(x, F) = cov((x µ), F) = E[(X µ)(f 0) ] = E[(LF + ɛ)(f) ] = LE(FF ) + E(ɛF ) }{{}}{{} I 0 = L = {l iq } p q The correlation between observed variables and the common factors equal cov(x i, F q ) l = iq h 2 i + ψ i 1 h 2 i + ψ i The correlations ρ(x i, F q ) are called Structure Coefficients. 23 / 52

24 Summary of Terminology and Components A Closer Look at Implied Covariance Matrix Summary of Terminology and Components & Victim s Plot of Factor F q are Common factors (latent variables). l iq are Factor Loadings. ɛ i are latent variables specific or Unique to item i. ψ i are the unique variance or Specifcities. h 2 i = m q=1 l iq are the common variance or Communalities. Correlation between X i and F q are Structure Coefficients: ρ(x i, F q ) = l iq mq l 2 iq + ψ i = l iq h 2 i + ψ i 24 / 52

25 A Closer Look at Implied Covariance Matrix Summary of Terminology and Components & Victim s Plot of Factor Basic Descriptive Statistics with a little item analysis: Corr Variable Mean Std Dev w/ Total Alph q36 upset students for fun q37 in group teased students q43 helped harass students q44 teased other students q49 mean to someone when angry q50 spread rumors q51 started arguments or conflicts q52 encouraged people to fight q53 excluded students from clique / 52

26 A Closer Look at Implied Covariance Matrix Summary of Terminology and Components & Victim s Plot of Factor Correlations: q36 q37 q43 q44 q49 q50 q51 q52 q53 q q q q q q q q q / 52

27 & Victim s A Closer Look at Implied Covariance Matrix Summary of Terminology and Components & Victim s Plot of Factor Correlations: q36 q37 q43 q44 q49 q50 q51 q52 q53 q39 q40 q What do you notice and what does this imply? 27 / 52

28 Plot of Factor Loadings A Closer Look at Implied Covariance Matrix Summary of Terminology and Components & Victim s Plot of Factor 28 / 52

29 A Closer Look at Implied Covariance Matrix Summary of Terminology and Components & Victim s Plot of Factor Factor Loadings Item Factor 1 Factor 2 Bully items q36 upset students for fun q37 in group teased students q43 helped harass students q44 teased other students q49 mean to someone when angry q50 spread rumors q51 started arguments or conflicts q52 encouraged people to fight q53 excluded students from clique Victim items q39 other students picked on me q40 students made fun of me q41 students called me names q42 got hit and pushed / 52

30 Residual Correlation Matrix w/ Uniqueness on Diagonal A Closer Look at Implied Covariance Matrix Summary of Terminology and Components & Victim s Plot of Factor q36 q37 q43 q44 q49 q50 q51 q52 q53 q q41 q42 q q Root Mean Square Off-Diagonal Residuals: Overall = / 52

31 Example Two: Fight A Closer Look at Implied Covariance Matrix Summary of Terminology and Components & Victim s Plot of Factor Deleted Correlation Variable with Total Alpha Label q fought students could beat q got in physical fight q threatened to hurt or hit student q physical fight because angry q hit back when hit first q38 q45 q46 q47 q38 fought students could beat q45 got in physical fight q46 threatened to hurt or hit student q47 physical fight because angry q48 hit back when hit first / 52

32 Correlations between Fight and Bully A Closer Look at Implied Covariance Matrix Summary of Terminology and Components & Victim s Plot of Factor Fight Bully q38 q45 q46 q47 q48 q q q q q q q q q Test of H o : Σ bully,fight = 0 versus H o : Σ bully,fight 0 F(45, 1367) = 8.72, p <.01, canonical correlation = / 52

33 Bully/Fight Factor Structure A Closer Look at Implied Covariance Matrix Summary of Terminology and Components & Victim s Plot of Factor 33 / 52

34 A Closer Look at Implied Covariance Matrix Summary of Terminology and Components & Victim s Plot of Factor Bully/Fight Factor Pattern Item Factor 1 Factor 2 Bully q Bully q Bully q Bully q Bully q Bully q Bully q Bully q Bully q Fight q Fight q Fight q Fight q Fight q Root Mean Square Off-Diagonal Residuals: Overall = / 52

35 Rotation Rotation Factor Pattern not Unique Desirable Structure Orthogonal Rotations Varimax for Victim & Bully Items Varimax Bully/Fight Factor Pattern Desirable (Oblique) Structure with Correlated Factors Oblique Rotation: Bully/Fight Factor Pattern Assessing Fit 35 / 52

36 Factor Pattern not Unique Two kinds: Rotation Factor Pattern not Unique Desirable Structure Orthogonal Rotations Varimax for Victim & Bully Items Varimax Bully/Fight Factor Pattern Desirable (Oblique) Structure with Correlated Factors Oblique Rotation: Bully/Fight Factor Pattern Assessing Fit Orthogonal Oblique Orthogonal: Let T be an orthogonal matrix such that Therefore, TT = T T = I Σ = LL + Ψ = LTT }{{} L + Ψ I = L L + Ψ Still keeping cov(f) = I and the model fit to data remains the same. 36 / 52

37 Desirable Structure for easier interpretation: ɛ 1 X 1 ɛ 2 X 2 F 1 Rotation Factor Pattern not Unique Desirable Structure Orthogonal Rotations Varimax for Victim & Bully Items Varimax Bully/Fight Factor Pattern Desirable (Oblique) Structure with Correlated Factors Oblique Rotation: Bully/Fight Factor Pattern ɛ 3 ɛ 4 X 3 X 4 F 2 Assessing Fit 37 / 52

38 Orthogonal Rotations Rotation Factor Pattern not Unique Desirable Structure Orthogonal Rotations Varimax for Victim & Bully Items Varimax Bully/Fight Factor Pattern Desirable (Oblique) Structure with Correlated Factors Oblique Rotation: Bully/Fight Factor Pattern Assessing Fit Change in coordinates corresponding to a rigid rotation of the axses. For interpretation, it s easier if there is a Simple Structure or Simple Solution such that each factor has either large or small (near zero) loadings. There are 10 methods for orthogonal rotation in SAS PROC FACTOR. The most commonly used one is varimax that is due to Henry Kaiser (1958). Maximize the sum of the variances of the vectors of loadings. Formally V = p (l 2 iq l 2 iq) 2 i=1 If you have a target matrix, then use procrustes. 38 / 52

39 Varimax for Victim & Bully Items Rotation Factor Pattern not Unique Desirable Structure Orthogonal Rotations Varimax for Victim & Bully Items Varimax Bully/Fight Factor Pattern Desirable (Oblique) Structure with Correlated Factors Oblique Rotation: Bully/Fight Factor Pattern Assessing Fit Initial Solution varimax Item Fac 1 Fac 2 Fac 1 Fac 2 Bully items q36 upset students for fun q37 in group teased students q43 helped harass students q44 teased other students q49 mean to someone when angry q50 spread rumors q51 started arguments or conflicts q52 encouraged people to fight q53 excluded students from clique Victim items q39 other students picked on me q40 students made fun of me q41 students called me names q42 got hit and pushed About the same. 39 / 52

40 Varimax Bully/Fight Factor Pattern Rotation Factor Pattern not Unique Desirable Structure Orthogonal Rotations Varimax for Victim & Bully Items Varimax Bully/Fight Factor Pattern Desirable (Oblique) Structure with Correlated Factors Oblique Rotation: Bully/Fight Factor Pattern Initial varimax Item Fac 1 Fac 2 Fac 1 Fac 2 Bully q Bully q Bully q Bully q Bully q Bully q Bully q Bully q Bully q Fight q Fight q Fight q Fight q Fight q Assessing Fit 40 / 52

41 Desirable (Oblique) Structure with Correlated Factors for easier interpretation: ɛ 1 X 1 ɛ 2 X 2 F 1 Rotation Factor Pattern not Unique Desirable Structure Orthogonal Rotations Varimax for Victim & Bully Items Varimax Bully/Fight Factor Pattern Desirable (Oblique) Structure with Correlated Factors Oblique Rotation: Bully/Fight Factor Pattern ɛ 3 ɛ 4 X 3 X 4 F 2 Assessing Fit 41 / 52

42 Oblique Rotation: Bully/Fight Factor Pattern Rotation Factor Pattern not Unique Desirable Structure Orthogonal Rotations Varimax for Victim & Bully Items Varimax Bully/Fight Factor Pattern Desirable (Oblique) Structure with Correlated Factors Oblique Rotation: Bully/Fight Factor Pattern Assessing Fit Initial varimax oblimin Item Fac 1 Fac 2 Fac 1 Fac 2 Fac 1 Fac 2 Bully q Bully q Bully q Bully q Bully q Bully q Bully q Bully q Bully q Fight q Fight q Fight q Fight q Fight q From oblimin, ˆρ(F 1, F 2 ) = / 52

43 Rotation Methods Principal Components Solution of Factor Principal Factor or Axis Solution A Modern Solution: Maximum Likelihood Maximum Likelihood MLE continued Rotation & Indeterminacy Assessing Fit 43 / 52

44 Methods Rotation Methods Principal Components Solution of Factor Principal Factor or Axis Solution A Modern Solution: Maximum Likelihood Maximum Likelihood MLE continued Rotation & Indeterminacy Assessing Fit method (Factor extraction) Eigen-decomposition based Eigen-decomposition of S or the principal components solution of the factor model. Iterative eigen-decompositions of S Ψ or the Principal factor or Principal Axis solution. Maximum likelihood estimation We now must assume that F and ɛ are multivariate normal. Tends to fit data better & yields scale invariance (ie., use either S or R). 44 / 52

45 Principal Components Solution of Factor Rotation Methods Principal Components Solution of Factor Principal Factor or Axis Solution A Modern Solution: Maximum Likelihood Maximum Likelihood MLE continued Rotation & Indeterminacy Assessing Fit Let ˆλ 1 ˆλ 2... ˆλ p be the eigenvalues of the sample covariance matrix S and ê 1, ê 2,...,ê p be corresponding eigenvectors. Let m < p (ie, the number of factors be less than the number of X variables), then the matrix of factor loadings equals ( ) L = ˆλ 1 ê 1, ˆλ 2 ê 2,..., ˆλ m ê m The estimate specific variances are provided by the diagonal elements of the matrix S L L, So Ψ = ψ ψ ψm where ψi = s ii m q=1 And estimates of communalities are h 2 i = m q=1 l2 iq Alternatively, use R instead of S (will get a different result). l 2 iq 45 / 52

46 Principal Factor or Axis Solution Rotation Methods Principal Components Solution of Factor Principal Factor or Axis Solution A Modern Solution: Maximum Likelihood Maximum Likelihood MLE continued Rotation & Indeterminacy Assessing Fit 1. Get an initial estimate of Ψ. Most common choice is the square of the multiple correlation (SMC) coefficient between X i and the other p 1 variables. The SMC is the diagonal element of R 1 and is our initial estimate of the ψ i s; that is, ψ i = the i th diagonal element of R 1 = SMC 2. Find the eigenvalues λ q and eigenvectors e i of R diag(ψ i ) and set L = ( λ 1e 1, λ 2e 2,..., λ me m). 3. New estimates of ψ ψ i = 1 m q=1 l 2 iq = 1 h 2 i. 4. Repeat steps 2 through 3 until convergence. 46 / 52

47 A Modern Solution: Maximum Likelihood Rotation Methods Principal Components Solution of Factor Principal Factor or Axis Solution A Modern Solution: Maximum Likelihood Maximum Likelihood MLE continued Rotation & Indeterminacy Assessing Fit With principal component solution, extraction of an additional factor does not effect the values already extracted; this is not the case with the MLE or the principal factor method. Both MLE and principal factor methods can run into problems (Heywood cases, improper solutions) where the ψs want to be negative. You get different results if you use S or R (ie, not scale invariant); however, this is not the case for MLE: If l iq is a loading from a (population) correlation matrix, then l iq σ ii is the corresponding loading from the (population) covariance matrix. You can get a better fit via MLE. Statistical tests become possible. With MLE it becomes possible to set values for l iq (and/or ψ i )...a confirmatory analysis / 52

48 Rotation Methods Principal Components Solution of Factor Principal Factor or Axis Solution A Modern Solution: Maximum Likelihood Maximum Likelihood MLE continued Rotation & Indeterminacy Assessing Fit Maximum Likelihood For MLE, we need to add the assumption that (or X µ). Note that X N(0,Σ) i.i.d where Σ = LL + Ψ. Σ = LΦL + Ψ Ψ} 1/2 ΣΨ {{ 1/2 } = Ψ} 1/2 {{ LΦ 1/2 } Φ} 1/2 L {{ Ψ 1/2 } +I Σ L L Σ Given Ψ, we can get the L s: = L L + I Σ I = Ψ 1/2 ΣΨ 1/2 I = PΛP = PΛ }{{ 1/2 } Λ} 1/2 {{ P } L L 48 / 52

49 MLE continued Starting with a initial guess for Ψ this gives us an initial estimate of L. Using an optimization algorithm, iteratively up-date estimates by maximizing L(L,Ψ) = n 2 (ln( Σ ) + Tr(SΣ 1 )) + constant Rotation Methods Principal Components Solution of Factor Principal Factor or Axis Solution A Modern Solution: Maximum Likelihood Maximum Likelihood MLE continued Rotation & Indeterminacy Assessing Fit or equivalently minimize F(L,Ψ) = ln( σ ) + Tr(SΣ 1 )) ln( S ) p 49 / 52

50 Rotation & Indeterminacy Rotation Methods Principal Components Solution of Factor Principal Factor or Axis Solution A Modern Solution: Maximum Likelihood Maximum Likelihood MLE continued Rotation & Indeterminacy Assessing Fit To improve interpretation, we can rotate factor loadings using either an orthogonal or oblique method. With oblique, Σ = LΦL The m m matrix Φ is the Factor Pattern Matrix that contains correlations between the factors. The p m matrix LΦ is the Factor Structure Matrix that contains the covariances (or correlations) between the X p 1 observed variables and the F m 1 latent variables. Due to the indeterminacy, computing factor scores is not a reasonable thing to do event though many computer programs will give them to you if asked. With MLE, you can fix parameters (e.g., set some loadings equal to 0) and this leads to confirmatory factor analysis. 50 / 52

51 Assessing Fit to Data Rotation Assessing Fit to Data Or How Many Factors Needed? 51 / 52

52 Or How Many Factors Needed? Rotation Assessing Fit to Data Or How Many Factors Needed? Data points: There are p variances and p(p 1)/2 covariances. Parameters to be estimated: There are pm Factor loadings and p unique variances. Residual correlation matrix Root mean squared errors Statistical tests for number of factors Proportion of total sample variance due to the q th factor Proportion of total sample variance accounted for by the m factors. Others. 52 / 52

9.1 Orthogonal factor model.

9.1 Orthogonal factor model. 36 Chapter 9 Factor Analysis Factor analysis may be viewed as a refinement of the principal component analysis The objective is, like the PC analysis, to describe the relevant variables in study in terms

More information

Chapter 4: Factor Analysis

Chapter 4: Factor Analysis Chapter 4: Factor Analysis In many studies, we may not be able to measure directly the variables of interest. We can merely collect data on other variables which may be related to the variables of interest.

More information

Introduction to Factor Analysis

Introduction to Factor Analysis to Factor Analysis Lecture 10 August 2, 2011 Advanced Multivariate Statistical Methods ICPSR Summer Session #2 Lecture #10-8/3/2011 Slide 1 of 55 Today s Lecture Factor Analysis Today s Lecture Exploratory

More information

2/26/2017. This is similar to canonical correlation in some ways. PSY 512: Advanced Statistics for Psychological and Behavioral Research 2

2/26/2017. This is similar to canonical correlation in some ways. PSY 512: Advanced Statistics for Psychological and Behavioral Research 2 PSY 512: Advanced Statistics for Psychological and Behavioral Research 2 What is factor analysis? What are factors? Representing factors Graphs and equations Extracting factors Methods and criteria Interpreting

More information

Introduction to Factor Analysis

Introduction to Factor Analysis to Factor Analysis Lecture 11 November 2, 2005 Multivariate Analysis Lecture #11-11/2/2005 Slide 1 of 58 Today s Lecture Factor Analysis. Today s Lecture Exploratory factor analysis (EFA). Confirmatory

More information

Factor Analysis. Robert L. Wolpert Department of Statistical Science Duke University, Durham, NC, USA

Factor Analysis. Robert L. Wolpert Department of Statistical Science Duke University, Durham, NC, USA Factor Analysis Robert L. Wolpert Department of Statistical Science Duke University, Durham, NC, USA 1 Factor Models The multivariate regression model Y = XB +U expresses each row Y i R p as a linear combination

More information

Or, in terms of basic measurement theory, we could model it as:

Or, in terms of basic measurement theory, we could model it as: 1 Neuendorf Factor Analysis Assumptions: 1. Metric (interval/ratio) data 2. Linearity (in relationships among the variables--factors are linear constructions of the set of variables; the critical source

More information

3.1. The probabilistic view of the principal component analysis.

3.1. The probabilistic view of the principal component analysis. 301 Chapter 3 Principal Components and Statistical Factor Models This chapter of introduces the principal component analysis (PCA), briefly reviews statistical factor models PCA is among the most popular

More information

Factor Analysis. Qian-Li Xue

Factor Analysis. Qian-Li Xue Factor Analysis Qian-Li Xue Biostatistics Program Harvard Catalyst The Harvard Clinical & Translational Science Center Short course, October 7, 06 Well-used latent variable models Latent variable scale

More information

Principal Component Analysis & Factor Analysis. Psych 818 DeShon

Principal Component Analysis & Factor Analysis. Psych 818 DeShon Principal Component Analysis & Factor Analysis Psych 818 DeShon Purpose Both are used to reduce the dimensionality of correlated measurements Can be used in a purely exploratory fashion to investigate

More information

Applied Multivariate Analysis

Applied Multivariate Analysis Department of Mathematics and Statistics, University of Vaasa, Finland Spring 2017 Dimension reduction Exploratory (EFA) Background While the motivation in PCA is to replace the original (correlated) variables

More information

1 A factor can be considered to be an underlying latent variable: (a) on which people differ. (b) that is explained by unknown variables

1 A factor can be considered to be an underlying latent variable: (a) on which people differ. (b) that is explained by unknown variables 1 A factor can be considered to be an underlying latent variable: (a) on which people differ (b) that is explained by unknown variables (c) that cannot be defined (d) that is influenced by observed variables

More information

More Linear Algebra. Edps/Soc 584, Psych 594. Carolyn J. Anderson

More Linear Algebra. Edps/Soc 584, Psych 594. Carolyn J. Anderson More Linear Algebra Edps/Soc 584, Psych 594 Carolyn J. Anderson Department of Educational Psychology I L L I N O I S university of illinois at urbana-champaign c Board of Trustees, University of Illinois

More information

TAMS39 Lecture 10 Principal Component Analysis Factor Analysis

TAMS39 Lecture 10 Principal Component Analysis Factor Analysis TAMS39 Lecture 10 Principal Component Analysis Factor Analysis Martin Singull Department of Mathematics Mathematical Statistics Linköping University, Sweden Content - Lecture Principal component analysis

More information

STAT 730 Chapter 9: Factor analysis

STAT 730 Chapter 9: Factor analysis STAT 730 Chapter 9: Factor analysis Timothy Hanson Department of Statistics, University of South Carolina Stat 730: Multivariate Data Analysis 1 / 15 Basic idea Factor analysis attempts to explain the

More information

FACTOR ANALYSIS AND MULTIDIMENSIONAL SCALING

FACTOR ANALYSIS AND MULTIDIMENSIONAL SCALING FACTOR ANALYSIS AND MULTIDIMENSIONAL SCALING Vishwanath Mantha Department for Electrical and Computer Engineering Mississippi State University, Mississippi State, MS 39762 mantha@isip.msstate.edu ABSTRACT

More information

Linear Dimensionality Reduction

Linear Dimensionality Reduction Outline Hong Chang Institute of Computing Technology, Chinese Academy of Sciences Machine Learning Methods (Fall 2012) Outline Outline I 1 Introduction 2 Principal Component Analysis 3 Factor Analysis

More information

VAR2 VAR3 VAR4 VAR5. Or, in terms of basic measurement theory, we could model it as:

VAR2 VAR3 VAR4 VAR5. Or, in terms of basic measurement theory, we could model it as: 1 Neuendorf Factor Analysis Assumptions: 1. Metric (interval/ratio) data 2. Linearity (in the relationships among the variables) -Factors are linear constructions of the set of variables (see #8 under

More information

Inter Item Correlation Matrix (R )

Inter Item Correlation Matrix (R ) 7 1. I have the ability to influence my child s well-being. 2. Whether my child avoids injury is just a matter of luck. 3. Luck plays a big part in determining how healthy my child is. 4. I can do a lot

More information

Exploratory Factor Analysis and Principal Component Analysis

Exploratory Factor Analysis and Principal Component Analysis Exploratory Factor Analysis and Principal Component Analysis Today s Topics: What are EFA and PCA for? Planning a factor analytic study Analysis steps: Extraction methods How many factors Rotation and

More information

Exploratory Factor Analysis and Principal Component Analysis

Exploratory Factor Analysis and Principal Component Analysis Exploratory Factor Analysis and Principal Component Analysis Today s Topics: What are EFA and PCA for? Planning a factor analytic study Analysis steps: Extraction methods How many factors Rotation and

More information

Factor Analysis (10/2/13)

Factor Analysis (10/2/13) STA561: Probabilistic machine learning Factor Analysis (10/2/13) Lecturer: Barbara Engelhardt Scribes: Li Zhu, Fan Li, Ni Guan Factor Analysis Factor analysis is related to the mixture models we have studied.

More information

FE670 Algorithmic Trading Strategies. Stevens Institute of Technology

FE670 Algorithmic Trading Strategies. Stevens Institute of Technology FE670 Algorithmic Trading Strategies Lecture 3. Factor Models and Their Estimation Steve Yang Stevens Institute of Technology 09/12/2012 Outline 1 The Notion of Factors 2 Factor Analysis via Maximum Likelihood

More information

I L L I N O I S UNIVERSITY OF ILLINOIS AT URBANA-CHAMPAIGN

I L L I N O I S UNIVERSITY OF ILLINOIS AT URBANA-CHAMPAIGN Canonical Edps/Soc 584 and Psych 594 Applied Multivariate Statistics Carolyn J. Anderson Department of Educational Psychology I L L I N O I S UNIVERSITY OF ILLINOIS AT URBANA-CHAMPAIGN Canonical Slide

More information

The Reliability of a Homogeneous Test. Measurement Methods Lecture 12

The Reliability of a Homogeneous Test. Measurement Methods Lecture 12 The Reliability of a Homogeneous Test Measurement Methods Lecture 12 Today s Class Estimating the reliability of a homogeneous test. McDonald s Omega. Guttman-Chronbach Alpha. Spearman-Brown. What to do

More information

Factor Analysis Continued. Psy 524 Ainsworth

Factor Analysis Continued. Psy 524 Ainsworth Factor Analysis Continued Psy 524 Ainsworth Equations Extraction Principal Axis Factoring Variables Skiers Cost Lift Depth Powder S1 32 64 65 67 S2 61 37 62 65 S3 59 40 45 43 S4 36 62 34 35 S5 62 46 43

More information

Multivariate Statistics

Multivariate Statistics Multivariate Statistics Chapter 4: Factor analysis Pedro Galeano Departamento de Estadística Universidad Carlos III de Madrid pedro.galeano@uc3m.es Course 2017/2018 Master in Mathematical Engineering Pedro

More information

Factor Analysis. -Applied Multivariate Analysis- Lecturer: Darren Homrighausen, PhD

Factor Analysis. -Applied Multivariate Analysis- Lecturer: Darren Homrighausen, PhD Factor Analysis -Applied Multivariate Analysis- Lecturer: Darren Homrighausen, PhD 1 From PCA to factor analysis Remember: PCA tries to estimate a transformation of the data such that: 1. The maximum amount

More information

Factor analysis. George Balabanis

Factor analysis. George Balabanis Factor analysis George Balabanis Key Concepts and Terms Deviation. A deviation is a value minus its mean: x - mean x Variance is a measure of how spread out a distribution is. It is computed as the average

More information

Inferences about a Mean Vector

Inferences about a Mean Vector Inferences about a Mean Vector Edps/Soc 584, Psych 594 Carolyn J. Anderson Department of Educational Psychology I L L I N O I S university of illinois at urbana-champaign c Board of Trustees, University

More information

E X P L O R E R. R e l e a s e A Program for Common Factor Analysis and Related Models for Data Analysis

E X P L O R E R. R e l e a s e A Program for Common Factor Analysis and Related Models for Data Analysis E X P L O R E R R e l e a s e 3. 2 A Program for Common Factor Analysis and Related Models for Data Analysis Copyright (c) 1990-2011, J. S. Fleming, PhD Date and Time : 23-Sep-2011, 13:59:56 Number of

More information

Intermediate Social Statistics

Intermediate Social Statistics Intermediate Social Statistics Lecture 5. Factor Analysis Tom A.B. Snijders University of Oxford January, 2008 c Tom A.B. Snijders (University of Oxford) Intermediate Social Statistics January, 2008 1

More information

I L L I N O I S UNIVERSITY OF ILLINOIS AT URBANA-CHAMPAIGN

I L L I N O I S UNIVERSITY OF ILLINOIS AT URBANA-CHAMPAIGN Principal Analysis Edps/Soc 584 and Psych 594 Applied Multivariate Statistics Carolyn J. Anderson Department of Educational Psychology I L L I N O I S UNIVERSITY OF ILLINOIS AT URBANA-CHAMPAIGN c Board

More information

PCA and admixture models

PCA and admixture models PCA and admixture models CM226: Machine Learning for Bioinformatics. Fall 2016 Sriram Sankararaman Acknowledgments: Fei Sha, Ameet Talwalkar, Alkes Price PCA and admixture models 1 / 57 Announcements HW1

More information

Principle Components Analysis (PCA) Relationship Between a Linear Combination of Variables and Axes Rotation for PCA

Principle Components Analysis (PCA) Relationship Between a Linear Combination of Variables and Axes Rotation for PCA Principle Components Analysis (PCA) Relationship Between a Linear Combination of Variables and Axes Rotation for PCA Principle Components Analysis: Uses one group of variables (we will call this X) In

More information

Multilevel Models in Matrix Form. Lecture 7 July 27, 2011 Advanced Multivariate Statistical Methods ICPSR Summer Session #2

Multilevel Models in Matrix Form. Lecture 7 July 27, 2011 Advanced Multivariate Statistical Methods ICPSR Summer Session #2 Multilevel Models in Matrix Form Lecture 7 July 27, 2011 Advanced Multivariate Statistical Methods ICPSR Summer Session #2 Today s Lecture Linear models from a matrix perspective An example of how to do

More information

I L L I N O I S UNIVERSITY OF ILLINOIS AT URBANA-CHAMPAIGN

I L L I N O I S UNIVERSITY OF ILLINOIS AT URBANA-CHAMPAIGN Introduction Edps/Psych/Stat/ 584 Applied Multivariate Statistics Carolyn J Anderson Department of Educational Psychology I L L I N O I S UNIVERSITY OF ILLINOIS AT URBANA-CHAMPAIGN c Board of Trustees,

More information

18.S096 Problem Set 7 Fall 2013 Factor Models Due Date: 11/14/2013. [ ] variance: E[X] =, and Cov[X] = Σ = =

18.S096 Problem Set 7 Fall 2013 Factor Models Due Date: 11/14/2013. [ ] variance: E[X] =, and Cov[X] = Σ = = 18.S096 Problem Set 7 Fall 2013 Factor Models Due Date: 11/14/2013 1. Consider a bivariate random variable: [ ] X X = 1 X 2 with mean and co [ ] variance: [ ] [ α1 Σ 1,1 Σ 1,2 σ 2 ρσ 1 σ E[X] =, and Cov[X]

More information

LECTURE 4 PRINCIPAL COMPONENTS ANALYSIS / EXPLORATORY FACTOR ANALYSIS

LECTURE 4 PRINCIPAL COMPONENTS ANALYSIS / EXPLORATORY FACTOR ANALYSIS LECTURE 4 PRINCIPAL COMPONENTS ANALYSIS / EXPLORATORY FACTOR ANALYSIS NOTES FROM PRE- LECTURE RECORDING ON PCA PCA and EFA have similar goals. They are substantially different in important ways. The goal

More information

Multivariate and Multivariable Regression. Stella Babalola Johns Hopkins University

Multivariate and Multivariable Regression. Stella Babalola Johns Hopkins University Multivariate and Multivariable Regression Stella Babalola Johns Hopkins University Session Objectives At the end of the session, participants will be able to: Explain the difference between multivariable

More information

Sample Geometry. Edps/Soc 584, Psych 594. Carolyn J. Anderson

Sample Geometry. Edps/Soc 584, Psych 594. Carolyn J. Anderson Sample Geometry Edps/Soc 584, Psych 594 Carolyn J. Anderson Department of Educational Psychology I L L I N O I S university of illinois at urbana-champaign c Board of Trustees, University of Illinois Spring

More information

Vectors and Matrices Statistics with Vectors and Matrices

Vectors and Matrices Statistics with Vectors and Matrices Vectors and Matrices Statistics with Vectors and Matrices Lecture 3 September 7, 005 Analysis Lecture #3-9/7/005 Slide 1 of 55 Today s Lecture Vectors and Matrices (Supplement A - augmented with SAS proc

More information

6348 Final, Fall 14. Closed book, closed notes, no electronic devices. Points (out of 200) in parentheses.

6348 Final, Fall 14. Closed book, closed notes, no electronic devices. Points (out of 200) in parentheses. 6348 Final, Fall 14. Closed book, closed notes, no electronic devices. Points (out of 200) in parentheses. 0 11 1 1.(5) Give the result of the following matrix multiplication: 1 10 1 Solution: 0 1 1 2

More information

Factor Analysis (1) Factor Analysis

Factor Analysis (1) Factor Analysis Factor Analysis (1) Outlines: 1. Introduction of factor analysis 2. Principle component analysis 4. Factor rotation 5. Case Shan-Yu Chou 1 Factor Analysis Combines questions or variables to create new

More information

THE UNIVERSITY OF CHICAGO Graduate School of Business Business 41912, Spring Quarter 2008, Mr. Ruey S. Tsay. Solutions to Final Exam

THE UNIVERSITY OF CHICAGO Graduate School of Business Business 41912, Spring Quarter 2008, Mr. Ruey S. Tsay. Solutions to Final Exam THE UNIVERSITY OF CHICAGO Graduate School of Business Business 41912, Spring Quarter 2008, Mr. Ruey S. Tsay Solutions to Final Exam 1. (13 pts) Consider the monthly log returns, in percentages, of five

More information

Conflict Tactics Scales Kindergarten/Year 1 Fast Track Project Technical Report Cynthia Rains November 15, 2004

Conflict Tactics Scales Kindergarten/Year 1 Fast Track Project Technical Report Cynthia Rains November 15, 2004 Conflict Tactics Scales Kindergarten/Year 1 Fast Track Project Technical Report Cynthia Rains November 15, 2004 Table of Contents I. Scale Description II. Report Sample III. Scaling IV. Differences Between

More information

UCLA STAT 233 Statistical Methods in Biomedical Imaging

UCLA STAT 233 Statistical Methods in Biomedical Imaging UCLA STAT 233 Statistical Methods in Biomedical Imaging Instructor: Ivo Dinov, Asst. Prof. In Statistics and Neurology University of California, Los Angeles, Spring 2004 http://www.stat.ucla.edu/~dinov/

More information

CS281 Section 4: Factor Analysis and PCA

CS281 Section 4: Factor Analysis and PCA CS81 Section 4: Factor Analysis and PCA Scott Linderman At this point we have seen a variety of machine learning models, with a particular emphasis on models for supervised learning. In particular, we

More information

Short Answer Questions: Answer on your separate blank paper. Points are given in parentheses.

Short Answer Questions: Answer on your separate blank paper. Points are given in parentheses. ISQS 6348 Final exam solutions. Name: Open book and notes, but no electronic devices. Answer short answer questions on separate blank paper. Answer multiple choice on this exam sheet. Put your name on

More information

Data Mining. Dimensionality reduction. Hamid Beigy. Sharif University of Technology. Fall 1395

Data Mining. Dimensionality reduction. Hamid Beigy. Sharif University of Technology. Fall 1395 Data Mining Dimensionality reduction Hamid Beigy Sharif University of Technology Fall 1395 Hamid Beigy (Sharif University of Technology) Data Mining Fall 1395 1 / 42 Outline 1 Introduction 2 Feature selection

More information

Factor Analysis: An Introduction. What is Factor Analysis? 100+ years of Factor Analysis FACTOR ANALYSIS AN INTRODUCTION NILAM RAM

Factor Analysis: An Introduction. What is Factor Analysis? 100+ years of Factor Analysis FACTOR ANALYSIS AN INTRODUCTION NILAM RAM NILAM RAM 2018 PSYCHOLOGY R BOOTCAMP PENNSYLVANIA STATE UNIVERSITY AUGUST 16, 2018 FACTOR ANALYSIS https://psu-psychology.github.io/r-bootcamp-2018/index.html WITH ADDITIONAL MATERIALS AT https://quantdev.ssri.psu.edu/tutorials

More information

Estimation: Problems & Solutions

Estimation: Problems & Solutions Estimation: Problems & Solutions Edps/Psych/Stat 587 Carolyn J. Anderson Department of Educational Psychology c Board of Trustees, University of Illinois Fall 2017 Outline 1. Introduction: Estimation of

More information

Capítulo 12 FACTOR ANALYSIS 12.1 INTRODUCTION

Capítulo 12 FACTOR ANALYSIS 12.1 INTRODUCTION Capítulo 12 FACTOR ANALYSIS Charles Spearman (1863-1945) British psychologist. Spearman was an officer in the British Army in India and upon his return at the age of 40, and influenced by the work of Galton,

More information

Canonical Correlations

Canonical Correlations Canonical Correlations Like Principal Components Analysis, Canonical Correlation Analysis looks for interesting linear combinations of multivariate observations. In Canonical Correlation Analysis, a multivariate

More information

Principal Component Analysis (PCA) Principal Component Analysis (PCA)

Principal Component Analysis (PCA) Principal Component Analysis (PCA) Recall: Eigenvectors of the Covariance Matrix Covariance matrices are symmetric. Eigenvectors are orthogonal Eigenvectors are ordered by the magnitude of eigenvalues: λ 1 λ 2 λ p {v 1, v 2,..., v n } Recall:

More information

VAR Model. (k-variate) VAR(p) model (in the Reduced Form): Y t-2. Y t-1 = A + B 1. Y t + B 2. Y t-p. + ε t. + + B p. where:

VAR Model. (k-variate) VAR(p) model (in the Reduced Form): Y t-2. Y t-1 = A + B 1. Y t + B 2. Y t-p. + ε t. + + B p. where: VAR Model (k-variate VAR(p model (in the Reduced Form: where: Y t = A + B 1 Y t-1 + B 2 Y t-2 + + B p Y t-p + ε t Y t = (y 1t, y 2t,, y kt : a (k x 1 vector of time series variables A: a (k x 1 vector

More information

Random Vectors 1. STA442/2101 Fall See last slide for copyright information. 1 / 30

Random Vectors 1. STA442/2101 Fall See last slide for copyright information. 1 / 30 Random Vectors 1 STA442/2101 Fall 2017 1 See last slide for copyright information. 1 / 30 Background Reading: Renscher and Schaalje s Linear models in statistics Chapter 3 on Random Vectors and Matrices

More information

Vector Space Models. wine_spectral.r

Vector Space Models. wine_spectral.r Vector Space Models 137 wine_spectral.r Latent Semantic Analysis Problem with words Even a small vocabulary as in wine example is challenging LSA Reduce number of columns of DTM by principal components

More information

simple if it completely specifies the density of x

simple if it completely specifies the density of x 3. Hypothesis Testing Pure significance tests Data x = (x 1,..., x n ) from f(x, θ) Hypothesis H 0 : restricts f(x, θ) Are the data consistent with H 0? H 0 is called the null hypothesis simple if it completely

More information

Stat 206, Week 6: Factor Analysis

Stat 206, Week 6: Factor Analysis Stat 206, Week 6: Factor Analysis James Johndrow 2016-09-24 Introduction Factor analysis aims to explain correlation between a large set (p) of variables in terms of a smaller number (m) of underlying

More information

Factor Analysis of Data Matrices

Factor Analysis of Data Matrices Factor Analysis of Data Matrices PAUL HORST University of Washington HOLT, RINEHART AND WINSTON, INC. New York Chicago San Francisco Toronto London Contents Preface PART I. Introductory Background 1. The

More information

Statistics for Applications. Chapter 9: Principal Component Analysis (PCA) 1/16

Statistics for Applications. Chapter 9: Principal Component Analysis (PCA) 1/16 Statistics for Applications Chapter 9: Principal Component Analysis (PCA) 1/16 Multivariate statistics and review of linear algebra (1) Let X be a d-dimensional random vector and X 1,..., X n be n independent

More information

UNIVERSITY OF TORONTO Faculty of Arts and Science

UNIVERSITY OF TORONTO Faculty of Arts and Science UNIVERSITY OF TORONTO Faculty of Arts and Science December 2013 Final Examination STA442H1F/2101HF Methods of Applied Statistics Jerry Brunner Duration - 3 hours Aids: Calculator Model(s): Any calculator

More information

Lecture 7 Spectral methods

Lecture 7 Spectral methods CSE 291: Unsupervised learning Spring 2008 Lecture 7 Spectral methods 7.1 Linear algebra review 7.1.1 Eigenvalues and eigenvectors Definition 1. A d d matrix M has eigenvalue λ if there is a d-dimensional

More information

Principal Component Analysis. Applied Multivariate Statistics Spring 2012

Principal Component Analysis. Applied Multivariate Statistics Spring 2012 Principal Component Analysis Applied Multivariate Statistics Spring 2012 Overview Intuition Four definitions Practical examples Mathematical example Case study 2 PCA: Goals Goal 1: Dimension reduction

More information

Regularized Common Factor Analysis

Regularized Common Factor Analysis New Trends in Psychometrics 1 Regularized Common Factor Analysis Sunho Jung 1 and Yoshio Takane 1 (1) Department of Psychology, McGill University, 1205 Dr. Penfield Avenue, Montreal, QC, H3A 1B1, Canada

More information

Psychology 405: Psychometric Theory Constructs, Components, and Factor Models

Psychology 405: Psychometric Theory Constructs, Components, and Factor Models Psychology 405: Psychometric Theory Constructs, Components, and Factor Models William Revelle Department of Psychology Northwestern University Evanston, Illinois USA April, 2012 1 / 58 Outline 1 Preliminaries

More information

Principal Components Theory Notes

Principal Components Theory Notes Principal Components Theory Notes Charles J. Geyer August 29, 2007 1 Introduction These are class notes for Stat 5601 (nonparametrics) taught at the University of Minnesota, Spring 2006. This not a theory

More information

A Study of Statistical Power and Type I Errors in Testing a Factor Analytic. Model for Group Differences in Regression Intercepts

A Study of Statistical Power and Type I Errors in Testing a Factor Analytic. Model for Group Differences in Regression Intercepts A Study of Statistical Power and Type I Errors in Testing a Factor Analytic Model for Group Differences in Regression Intercepts by Margarita Olivera Aguilar A Thesis Presented in Partial Fulfillment of

More information

Exploratory Factor Analysis and Canonical Correlation

Exploratory Factor Analysis and Canonical Correlation Exploratory Factor Analysis and Canonical Correlation 3 Dec 2010 CPSY 501 Dr. Sean Ho Trinity Western University Please download: SAQ.sav Outline for today Factor analysis Latent variables Correlation

More information

[y i α βx i ] 2 (2) Q = i=1

[y i α βx i ] 2 (2) Q = i=1 Least squares fits This section has no probability in it. There are no random variables. We are given n points (x i, y i ) and want to find the equation of the line that best fits them. We take the equation

More information

Dimensionality Reduction Techniques (DRT)

Dimensionality Reduction Techniques (DRT) Dimensionality Reduction Techniques (DRT) Introduction: Sometimes we have lot of variables in the data for analysis which create multidimensional matrix. To simplify calculation and to get appropriate,

More information

MULTIVARIATE TIME SERIES ANALYSIS AN ADAPTATION OF BOX-JENKINS METHODOLOGY Joseph N Ladalla University of Illinois at Springfield, Springfield, IL

MULTIVARIATE TIME SERIES ANALYSIS AN ADAPTATION OF BOX-JENKINS METHODOLOGY Joseph N Ladalla University of Illinois at Springfield, Springfield, IL MULTIVARIATE TIME SERIES ANALYSIS AN ADAPTATION OF BOX-JENKINS METHODOLOGY Joseph N Ladalla University of Illinois at Springfield, Springfield, IL KEYWORDS: Multivariate time series, Box-Jenkins ARIMA

More information

Comparing IRT with Other Models

Comparing IRT with Other Models Comparing IRT with Other Models Lecture #14 ICPSR Item Response Theory Workshop Lecture #14: 1of 45 Lecture Overview The final set of slides will describe a parallel between IRT and another commonly used

More information

CS534 Machine Learning - Spring Final Exam

CS534 Machine Learning - Spring Final Exam CS534 Machine Learning - Spring 2013 Final Exam Name: You have 110 minutes. There are 6 questions (8 pages including cover page). If you get stuck on one question, move on to others and come back to the

More information

2. Matrix Algebra and Random Vectors

2. Matrix Algebra and Random Vectors 2. Matrix Algebra and Random Vectors 2.1 Introduction Multivariate data can be conveniently display as array of numbers. In general, a rectangular array of numbers with, for instance, n rows and p columns

More information

Structural Equation Modeling and Confirmatory Factor Analysis. Types of Variables

Structural Equation Modeling and Confirmatory Factor Analysis. Types of Variables /4/04 Structural Equation Modeling and Confirmatory Factor Analysis Advanced Statistics for Researchers Session 3 Dr. Chris Rakes Website: http://csrakes.yolasite.com Email: Rakes@umbc.edu Twitter: @RakesChris

More information

Dimensionality Reduction

Dimensionality Reduction Dimensionality Reduction Le Song Machine Learning I CSE 674, Fall 23 Unsupervised learning Learning from raw (unlabeled, unannotated, etc) data, as opposed to supervised data where a classification of

More information

Linear Methods for Prediction

Linear Methods for Prediction Chapter 5 Linear Methods for Prediction 5.1 Introduction We now revisit the classification problem and focus on linear methods. Since our prediction Ĝ(x) will always take values in the discrete set G we

More information

STATISTICAL LEARNING SYSTEMS

STATISTICAL LEARNING SYSTEMS STATISTICAL LEARNING SYSTEMS LECTURE 8: UNSUPERVISED LEARNING: FINDING STRUCTURE IN DATA Institute of Computer Science, Polish Academy of Sciences Ph. D. Program 2013/2014 Principal Component Analysis

More information

LOG-MULTIPLICATIVE ASSOCIATION MODELS AS LATENT VARIABLE MODELS FOR NOMINAL AND0OR ORDINAL DATA. Carolyn J. Anderson* Jeroen K.

LOG-MULTIPLICATIVE ASSOCIATION MODELS AS LATENT VARIABLE MODELS FOR NOMINAL AND0OR ORDINAL DATA. Carolyn J. Anderson* Jeroen K. 3 LOG-MULTIPLICATIVE ASSOCIATION MODELS AS LATENT VARIABLE MODELS FOR NOMINAL AND0OR ORDINAL DATA Carolyn J. Anderson* Jeroen K. Vermunt Associations between multiple discrete measures are often due to

More information

ICS 6N Computational Linear Algebra Symmetric Matrices and Orthogonal Diagonalization

ICS 6N Computational Linear Algebra Symmetric Matrices and Orthogonal Diagonalization ICS 6N Computational Linear Algebra Symmetric Matrices and Orthogonal Diagonalization Xiaohui Xie University of California, Irvine xhx@uci.edu Xiaohui Xie (UCI) ICS 6N 1 / 21 Symmetric matrices An n n

More information

Models for Clustered Data

Models for Clustered Data Models for Clustered Data Edps/Psych/Soc 589 Carolyn J Anderson Department of Educational Psychology c Board of Trustees, University of Illinois Spring 2019 Outline Notation NELS88 data Fixed Effects ANOVA

More information

Lecture 13. Principal Component Analysis. Brett Bernstein. April 25, CDS at NYU. Brett Bernstein (CDS at NYU) Lecture 13 April 25, / 26

Lecture 13. Principal Component Analysis. Brett Bernstein. April 25, CDS at NYU. Brett Bernstein (CDS at NYU) Lecture 13 April 25, / 26 Principal Component Analysis Brett Bernstein CDS at NYU April 25, 2017 Brett Bernstein (CDS at NYU) Lecture 13 April 25, 2017 1 / 26 Initial Question Intro Question Question Let S R n n be symmetric. 1

More information

Introduction to Machine Learning

Introduction to Machine Learning 10-701 Introduction to Machine Learning PCA Slides based on 18-661 Fall 2018 PCA Raw data can be Complex, High-dimensional To understand a phenomenon we measure various related quantities If we knew what

More information

Models for Clustered Data

Models for Clustered Data Models for Clustered Data Edps/Psych/Stat 587 Carolyn J Anderson Department of Educational Psychology c Board of Trustees, University of Illinois Fall 2017 Outline Notation NELS88 data Fixed Effects ANOVA

More information

A Factor Analysis of Key Decision Factors for Implementing SOA-based Enterprise level Business Intelligence Systems

A Factor Analysis of Key Decision Factors for Implementing SOA-based Enterprise level Business Intelligence Systems icbst 2014 International Conference on Business, Science and Technology which will be held at Hatyai, Thailand on the 25th and 26th of April 2014. AENSI Journals Australian Journal of Basic and Applied

More information

Multivariate Statistical Analysis

Multivariate Statistical Analysis Multivariate Statistical Analysis Fall 2011 C. L. Williams, Ph.D. Lecture 4 for Applied Multivariate Analysis Outline 1 Eigen values and eigen vectors Characteristic equation Some properties of eigendecompositions

More information

Data Analysis and Manifold Learning Lecture 6: Probabilistic PCA and Factor Analysis

Data Analysis and Manifold Learning Lecture 6: Probabilistic PCA and Factor Analysis Data Analysis and Manifold Learning Lecture 6: Probabilistic PCA and Factor Analysis Radu Horaud INRIA Grenoble Rhone-Alpes, France Radu.Horaud@inrialpes.fr http://perception.inrialpes.fr/ Outline of Lecture

More information

Multivariate Fundamentals: Rotation. Exploratory Factor Analysis

Multivariate Fundamentals: Rotation. Exploratory Factor Analysis Multivariate Fundamentals: Rotation Exploratory Factor Analysis PCA Analysis A Review Precipitation Temperature Ecosystems PCA Analysis with Spatial Data Proportion of variance explained Comp.1 + Comp.2

More information

Factor Models for Asset Returns. Prof. Daniel P. Palomar

Factor Models for Asset Returns. Prof. Daniel P. Palomar Factor Models for Asset Returns Prof. Daniel P. Palomar The Hong Kong University of Science and Technology (HKUST) MAFS6010R- Portfolio Optimization with R MSc in Financial Mathematics Fall 2018-19, HKUST,

More information

A Primer on Statistical Inference using Maximum Likelihood

A Primer on Statistical Inference using Maximum Likelihood A Primer on Statistical Inference using Maximum Likelihood November 3, 2017 1 Inference via Maximum Likelihood Statistical inference is the process of using observed data to estimate features of the population.

More information

Intelligent Data Analysis. Principal Component Analysis. School of Computer Science University of Birmingham

Intelligent Data Analysis. Principal Component Analysis. School of Computer Science University of Birmingham Intelligent Data Analysis Principal Component Analysis Peter Tiňo School of Computer Science University of Birmingham Discovering low-dimensional spatial layout in higher dimensional spaces - 1-D/3-D example

More information

Principal Component Analysis

Principal Component Analysis Principal Component Analysis Giorgos Korfiatis Alfa-Informatica University of Groningen Seminar in Statistics and Methodology, 2007 What Is PCA? Dimensionality reduction technique Aim: Extract relevant

More information

Stat 5101 Lecture Notes

Stat 5101 Lecture Notes Stat 5101 Lecture Notes Charles J. Geyer Copyright 1998, 1999, 2000, 2001 by Charles J. Geyer May 7, 2001 ii Stat 5101 (Geyer) Course Notes Contents 1 Random Variables and Change of Variables 1 1.1 Random

More information

1 Data Arrays and Decompositions

1 Data Arrays and Decompositions 1 Data Arrays and Decompositions 1.1 Variance Matrices and Eigenstructure Consider a p p positive definite and symmetric matrix V - a model parameter or a sample variance matrix. The eigenstructure is

More information

Factor Analysis and Kalman Filtering (11/2/04)

Factor Analysis and Kalman Filtering (11/2/04) CS281A/Stat241A: Statistical Learning Theory Factor Analysis and Kalman Filtering (11/2/04) Lecturer: Michael I. Jordan Scribes: Byung-Gon Chun and Sunghoon Kim 1 Factor Analysis Factor analysis is used

More information

Probability reminders

Probability reminders CS246 Winter 204 Mining Massive Data Sets Probability reminders Sammy El Ghazzal selghazz@stanfordedu Disclaimer These notes may contain typos, mistakes or confusing points Please contact the author so

More information

Improper Solutions in Exploratory Factor Analysis: Causes and Treatments

Improper Solutions in Exploratory Factor Analysis: Causes and Treatments Improper Solutions in Exploratory Factor Analysis: Causes and Treatments Yutaka Kano Faculty of Human Sciences, Osaka University Suita, Osaka 565, Japan. email: kano@hus.osaka-u.ac.jp Abstract: There are

More information

For more information about how to cite these materials visit

For more information about how to cite these materials visit Author(s): Kerby Shedden, Ph.D., 2010 License: Unless otherwise noted, this material is made available under the terms of the Creative Commons Attribution Share Alike 3.0 License: http://creativecommons.org/licenses/by-sa/3.0/

More information