Factor Analysis Edpsy/Soc 584 & Psych 594
|
|
- Matthew Harvey
- 5 years ago
- Views:
Transcription
1 Factor Analysis Edpsy/Soc 584 & Psych 594 Carolyn J. Anderson University of Illinois, Urbana-Champaign April 29, / 52
2 Rotation Assessing Fit to Data (one common factor model) common factors Assessment of Fit to Data Rotation Confirmatory Factor Analysis 2 / 52
3 : Rotation Assessing Fit to Data Explore underlying structure data driven. Confirm underlying structure theory driven. method (Factor extraction) Eigen-decomposition based Eigen-decomposition of S which is the principal components solution. Iterative eigen-decompositions of S Ψ which is the Principal factor solution. Maximum likelihood estimation We now must assume that F and ɛ are multivariate normal. Tends to fit data better & yields scale invariance (ie., use either S or R). 3 / 52
4 Rotation Assessing Fit to Data Data from Espelage, D.L., Holt, M.K., & Henkel, R.R. (2003). Examination of Peer-Group contextual effects on aggression during early adolescence. Child Development, 74, Items have a 5 point response scale: Never, 1 to 2 times, 3 to 4 times, 5 to 6 times, 7 or more times Q36: I upset other students for the fun of it. Q37: In a group I teased other students. Q38: I fought students I could easily beat. Q39: Other students picked on me. Q40: Other students made fun of me. Q41: Other students called me names. Q42: I got hit and pushed by other students. Q43: I helped harass other students. 4 / 52
5 Rotation Assessing Fit to Data Items have a 5 point response scale: Never, 1 to 2 times,..., 7 or more times Q44: I teased other students. Q45: I got in a physical fight. Q46: I threatened to hurt or hit another student. Q47: I got into a physical fight because I was angry. Q48: I hit back when someone hit me first. Q49: I was mean to someone when I was angry. Q50: I spread rumors about other students. Q51: I started (instigated) arguments or conflicts. Q52: I encouraged people to fight. Q53: I excluded other students from my clique (group) of friends 5 / 52
6 Rotation Assessing Fit to Data What are the factors and which items load on which factors? Item F 1 F 2 F 3 Item F 1 F 2 F / 52
7 Rotation Assessing Fit to Data α here is Chronbach s alpha for measuring reliability. F 1 = Fighting (α =.88), F 2 = Bullying (α =.88), F 3 = Victimization (α =.87) Item F 1 F 2 F 3 Item F 1 F 2 F 3 36 X 45 X 37 X 46 X 38 X 47 X 39 X 48 X 40 X 49 X 41 X 50 X 42 X 51 X 43 X 52 X 44 X 53 X 7 / 52
8 Factor Analytic The One Common Factor Analytic Unique Variables Unique Variables & Common Variables Implications for Data Implications for Data (continued) The Data is Covariance Matrix Victim Statistics One Factor Solution Rotation 8 / 52
9 Factor Analytic Factor Analytic The One Common Factor Analytic Unique Variables Unique Variables & Common Variables Implications for Data Implications for Data (continued) The Data is Covariance Matrix Victim Statistics One Factor Solution Factor analysis (FA) is a latent variable model. ɛ 1 X 1 ɛ 2 X 2 ɛ 3 X 3 ɛ 4 X 4 l 21 l 11 l 31 l 41 F 1 The One Common Factor X 1 = µ 1 + l 11 F 1 + ɛ 1 X 2 = µ 2 + l 21 F 1 + ɛ 2 X 3 = µ 3 + l 31 F 1 + ɛ 3 X 4 = µ 4 + l 41 F 1 + ɛ 4 Rotation 9 / 52
10 The One Common Factor Analytic Factor Analytic The One Common Factor Analytic Unique Variables Unique Variables & Common Variables Implications for Data Implications for Data (continued) The Data is Covariance Matrix Victim Statistics One Factor Solution X i = µ i + l i1 F 1 + ɛ i The µ i s are the means of the X i s. Common Factor: F 1 is an unobserved random variable with mean 0 and variance φ. The l i1 s are the Factor Loadings. Specificities or Uniquenesses: ɛ i are independent over i, unobserved random variables with means equal to 0 and variances equal to ψ i. F 1 is independent of the ɛ i s. Rotation 10 / 52
11 Factor Analytic The One Common Factor Analytic Unique Variables Unique Variables & Common Variables Implications for Data Implications for Data (continued) The Data is Covariance Matrix Victim Statistics One Factor Solution Unique Variables A little Classical test theory: true score = observed score pure measurement error t i = X i e i If a factor model holds for the observed scores X i, then it should also hold for the true scores t i. The uniqueness in the factor model for the true scores contains specific errors due to the particular variables (items) selected; that is, Solve for X i : t i = l i1 F 1 + s i }{{} specif ic t i = l i1 F 1 + s i = X i e i = X i = l i1 F i + (s i + e i ) }{{} ɛ i Rotation 11 / 52
12 Unique Variables & Common Variables The unique variables ɛ i contain Pure measurement errors Specific errors Factor Analytic The One Common Factor Analytic Unique Variables Unique Variables & Common Variables Implications for Data Implications for Data (continued) The Data is Covariance Matrix Victim Statistics One Factor Solution The observed variables will be correlated because they all depend (in part) on F 1. The common factor model makes implications for covariance (and correlation) matrices; therefore, the data are the covariance (or correlation) matrix. Rotation 12 / 52
13 Factor Analytic The One Common Factor Analytic Unique Variables Unique Variables & Common Variables Implications for Data Implications for Data (continued) The Data is Covariance Matrix Victim Statistics One Factor Solution Implications for Data Let s turn it into a linear algebra problem: X 1 µ 1 X 2 µ 2 X =. µ =... Σ = X p L = l 11 l 21. l p1 Rotation 13 / 52 µ p ɛ = Ψ = So X µ = LF 1 + ɛ ɛ 1 ɛ 2.. ɛ p ψ ψ ψ p σ 11 σ σ 1p σ 12 σ σ 2p σ 1p σ 2p... σ pp var(f) = Φ = φ 11
14 Implications for Data (continued) Factor Analytic The One Common Factor Analytic Unique Variables Unique Variables & Common Variables Implications for Data Implications for Data (continued) The Data is Covariance Matrix Victim Statistics One Factor Solution X µ = LF 1 + ɛ Mean of X µ = LE(F 1 ) + E(ɛ) = 0 Covariance matrix for X or equivalently X µ: Σ = E[(X µ)(x µ) ] = E[(LF 1 + ɛ)(lf 1 ɛ) ] = E[(LF 1 + ɛ)(f 1 L ɛ )] = L E[F1 2 ] L + L E[F 1 ɛ ] +E[ɛF 1 ] L + E[ɛɛ ] }{{}}{{}}{{}}{{} φ 11 =1 0 0 Ψ = LL + Ψ Rotation 14 / 52
15 Factor Analytic The One Common Factor Analytic Unique Variables Unique Variables & Common Variables Implications for Data Implications for Data (continued) The Data is Covariance Matrix Victim Statistics One Factor Solution The Data is Covariance Matrix One common factor model implies : Σ = l 2 i1 = h2 i (l ψ 1) l 11 l l 11 l p1 l 11 l 21 (l ψ 2)... l 21 l p l 11 l I1 l 21 l I1... (l 2 I1 + ψ p) is the Communality of item i. Foreshadowing estimation methods: If Ψ were known, or Σ = (e 1 Σ Ψ = LL = (e 1 λ1 )( λ 1 e 1) λ 1 ) } {{ } L ( λ 1 e 1 ) } {{ } L Rotation 15 / 52 and Ψ = diag(σ L L )
16 Victim Statistics Factor Analytic The One Common Factor Analytic Unique Variables Unique Variables & Common Variables Implications for Data Implications for Data (continued) The Data is Covariance Matrix Victim Statistics One Factor Solution Descriptives Correlations Item Mean Std Dev q39 q40 q41 q42 q q q q Cronbach Coefficient Alpha with Deleted Variable Raw Variables Std Variables Correlation Correlation Deleted Variable with Total Alpha with Total Al q39: other students picked on me q40: students made fun of me q41: students called me names q42: got hit and pushed Rotation 16 / 52
17 One Factor Solution Factor Analytic The One Common Factor Analytic Unique Variables Unique Variables & Common Variables Implications for Data Implications for Data (continued) The Data is Covariance Matrix Victim Statistics One Factor Solution Factor Pattern Item Factor1 Communalities Specific l i1 h 2 i var: ψ i q39 other students picked on me q40 students made fun of me q41 students called me names q42 got hit and pushed (maximum likelihood estimation using the correlation matrix) Residual Correlation with Uniqueness ( ˆψ i ) on the diagonal: q39 q40 q41 q42 q q q q Root Mean Square Off-Diagonal Residuals: Overall = Rotation 17 / 52
18 A Closer Look at Implied Covariance Matrix Summary of Terminology and Components & Victim s Plot of Factor 18 / 52
19 Picture of this for p = 4 and m = 2. ɛ 1 X 1 ɛ 2 X 2 F 1 A Closer Look at Implied Covariance Matrix Summary of Terminology and Components & Victim s Plot of Factor ɛ 3 ɛ 4 X 3 X 4 F 2 19 / 52
20 A Closer Look at Implied Covariance Matrix Summary of Terminology and Components & Victim s Plot of Factor X 1 = µ 1 + l 11 F 1 + l 12 F l 1m F m + ɛ 1 X 2 = µ 2 + l 21 F 1 + l 22 F l 2m F m + ɛ 2.. X p = µ p + l p1 F 1 + l p2 F l pm F m + ɛ p X = µ + LF + ɛ E(F) = 0 and cov(f) = Φ = I... for now E(ɛ) = 0 and cov(ɛ) = Ψ = diag(ψ i ) cov(f,ɛ) = 0 L is matrix of Factor Loadings m q=1 l 2 iq = h2 i =Communality of item i. ψ i is Specific variance of item i. 20 / 52
21 A Closer Look at Implied Covariance Matrix Summary of Terminology and Components & Victim s Plot of Factor Using the model for i = 1,..., I observed variables and m common factors, the covariance matrix is X µ = LF + ɛ Mean of X µ= E(LF + ɛ) = LE(F) + E(ɛ) = 0 Covariance matrix for X or equivalently X µ: Σ = E[(X µ)(x µ) ] = E[(LF + ɛ)(lf ɛ) ] = E[(LF + ɛ)(f L ɛ )] = L E[FF ] }{{} I = LL + Ψ L + L E[F1ɛ ] +E[ɛF ] }{{}}{{} 0 I L + E[ɛɛ ] }{{} Ψ 21 / 52
22 The Common factor model implies : Σ = ( q l2 1q + ψ 1) q l 1ql 2q... q l 1ql pq q l 1ql 2q ( q l2 2q + ψ 2)... q l 2ql pq q l 1q l pq q l 2q l pq... ( q l 2 pq + ψ I ) A Closer Look at Implied Covariance Matrix Summary of Terminology and Components & Victim s Plot of Factor Variance for X i : σ ii = m q=1 l 2 im }{{} h 2 i +ψ i Covariance between X i and X k : m σ ik = l iq l kq q=1 22 / 52
23 A Closer Look at Implied Covariance Matrix A Closer Look at Implied Covariance Matrix Summary of Terminology and Components & Victim s Plot of Factor Covariance between X and F: cov(x, F) = cov((x µ), F) = E[(X µ)(f 0) ] = E[(LF + ɛ)(f) ] = LE(FF ) + E(ɛF ) }{{}}{{} I 0 = L = {l iq } p q The correlation between observed variables and the common factors equal cov(x i, F q ) l = iq h 2 i + ψ i 1 h 2 i + ψ i The correlations ρ(x i, F q ) are called Structure Coefficients. 23 / 52
24 Summary of Terminology and Components A Closer Look at Implied Covariance Matrix Summary of Terminology and Components & Victim s Plot of Factor F q are Common factors (latent variables). l iq are Factor Loadings. ɛ i are latent variables specific or Unique to item i. ψ i are the unique variance or Specifcities. h 2 i = m q=1 l iq are the common variance or Communalities. Correlation between X i and F q are Structure Coefficients: ρ(x i, F q ) = l iq mq l 2 iq + ψ i = l iq h 2 i + ψ i 24 / 52
25 A Closer Look at Implied Covariance Matrix Summary of Terminology and Components & Victim s Plot of Factor Basic Descriptive Statistics with a little item analysis: Corr Variable Mean Std Dev w/ Total Alph q36 upset students for fun q37 in group teased students q43 helped harass students q44 teased other students q49 mean to someone when angry q50 spread rumors q51 started arguments or conflicts q52 encouraged people to fight q53 excluded students from clique / 52
26 A Closer Look at Implied Covariance Matrix Summary of Terminology and Components & Victim s Plot of Factor Correlations: q36 q37 q43 q44 q49 q50 q51 q52 q53 q q q q q q q q q / 52
27 & Victim s A Closer Look at Implied Covariance Matrix Summary of Terminology and Components & Victim s Plot of Factor Correlations: q36 q37 q43 q44 q49 q50 q51 q52 q53 q39 q40 q What do you notice and what does this imply? 27 / 52
28 Plot of Factor Loadings A Closer Look at Implied Covariance Matrix Summary of Terminology and Components & Victim s Plot of Factor 28 / 52
29 A Closer Look at Implied Covariance Matrix Summary of Terminology and Components & Victim s Plot of Factor Factor Loadings Item Factor 1 Factor 2 Bully items q36 upset students for fun q37 in group teased students q43 helped harass students q44 teased other students q49 mean to someone when angry q50 spread rumors q51 started arguments or conflicts q52 encouraged people to fight q53 excluded students from clique Victim items q39 other students picked on me q40 students made fun of me q41 students called me names q42 got hit and pushed / 52
30 Residual Correlation Matrix w/ Uniqueness on Diagonal A Closer Look at Implied Covariance Matrix Summary of Terminology and Components & Victim s Plot of Factor q36 q37 q43 q44 q49 q50 q51 q52 q53 q q41 q42 q q Root Mean Square Off-Diagonal Residuals: Overall = / 52
31 Example Two: Fight A Closer Look at Implied Covariance Matrix Summary of Terminology and Components & Victim s Plot of Factor Deleted Correlation Variable with Total Alpha Label q fought students could beat q got in physical fight q threatened to hurt or hit student q physical fight because angry q hit back when hit first q38 q45 q46 q47 q38 fought students could beat q45 got in physical fight q46 threatened to hurt or hit student q47 physical fight because angry q48 hit back when hit first / 52
32 Correlations between Fight and Bully A Closer Look at Implied Covariance Matrix Summary of Terminology and Components & Victim s Plot of Factor Fight Bully q38 q45 q46 q47 q48 q q q q q q q q q Test of H o : Σ bully,fight = 0 versus H o : Σ bully,fight 0 F(45, 1367) = 8.72, p <.01, canonical correlation = / 52
33 Bully/Fight Factor Structure A Closer Look at Implied Covariance Matrix Summary of Terminology and Components & Victim s Plot of Factor 33 / 52
34 A Closer Look at Implied Covariance Matrix Summary of Terminology and Components & Victim s Plot of Factor Bully/Fight Factor Pattern Item Factor 1 Factor 2 Bully q Bully q Bully q Bully q Bully q Bully q Bully q Bully q Bully q Fight q Fight q Fight q Fight q Fight q Root Mean Square Off-Diagonal Residuals: Overall = / 52
35 Rotation Rotation Factor Pattern not Unique Desirable Structure Orthogonal Rotations Varimax for Victim & Bully Items Varimax Bully/Fight Factor Pattern Desirable (Oblique) Structure with Correlated Factors Oblique Rotation: Bully/Fight Factor Pattern Assessing Fit 35 / 52
36 Factor Pattern not Unique Two kinds: Rotation Factor Pattern not Unique Desirable Structure Orthogonal Rotations Varimax for Victim & Bully Items Varimax Bully/Fight Factor Pattern Desirable (Oblique) Structure with Correlated Factors Oblique Rotation: Bully/Fight Factor Pattern Assessing Fit Orthogonal Oblique Orthogonal: Let T be an orthogonal matrix such that Therefore, TT = T T = I Σ = LL + Ψ = LTT }{{} L + Ψ I = L L + Ψ Still keeping cov(f) = I and the model fit to data remains the same. 36 / 52
37 Desirable Structure for easier interpretation: ɛ 1 X 1 ɛ 2 X 2 F 1 Rotation Factor Pattern not Unique Desirable Structure Orthogonal Rotations Varimax for Victim & Bully Items Varimax Bully/Fight Factor Pattern Desirable (Oblique) Structure with Correlated Factors Oblique Rotation: Bully/Fight Factor Pattern ɛ 3 ɛ 4 X 3 X 4 F 2 Assessing Fit 37 / 52
38 Orthogonal Rotations Rotation Factor Pattern not Unique Desirable Structure Orthogonal Rotations Varimax for Victim & Bully Items Varimax Bully/Fight Factor Pattern Desirable (Oblique) Structure with Correlated Factors Oblique Rotation: Bully/Fight Factor Pattern Assessing Fit Change in coordinates corresponding to a rigid rotation of the axses. For interpretation, it s easier if there is a Simple Structure or Simple Solution such that each factor has either large or small (near zero) loadings. There are 10 methods for orthogonal rotation in SAS PROC FACTOR. The most commonly used one is varimax that is due to Henry Kaiser (1958). Maximize the sum of the variances of the vectors of loadings. Formally V = p (l 2 iq l 2 iq) 2 i=1 If you have a target matrix, then use procrustes. 38 / 52
39 Varimax for Victim & Bully Items Rotation Factor Pattern not Unique Desirable Structure Orthogonal Rotations Varimax for Victim & Bully Items Varimax Bully/Fight Factor Pattern Desirable (Oblique) Structure with Correlated Factors Oblique Rotation: Bully/Fight Factor Pattern Assessing Fit Initial Solution varimax Item Fac 1 Fac 2 Fac 1 Fac 2 Bully items q36 upset students for fun q37 in group teased students q43 helped harass students q44 teased other students q49 mean to someone when angry q50 spread rumors q51 started arguments or conflicts q52 encouraged people to fight q53 excluded students from clique Victim items q39 other students picked on me q40 students made fun of me q41 students called me names q42 got hit and pushed About the same. 39 / 52
40 Varimax Bully/Fight Factor Pattern Rotation Factor Pattern not Unique Desirable Structure Orthogonal Rotations Varimax for Victim & Bully Items Varimax Bully/Fight Factor Pattern Desirable (Oblique) Structure with Correlated Factors Oblique Rotation: Bully/Fight Factor Pattern Initial varimax Item Fac 1 Fac 2 Fac 1 Fac 2 Bully q Bully q Bully q Bully q Bully q Bully q Bully q Bully q Bully q Fight q Fight q Fight q Fight q Fight q Assessing Fit 40 / 52
41 Desirable (Oblique) Structure with Correlated Factors for easier interpretation: ɛ 1 X 1 ɛ 2 X 2 F 1 Rotation Factor Pattern not Unique Desirable Structure Orthogonal Rotations Varimax for Victim & Bully Items Varimax Bully/Fight Factor Pattern Desirable (Oblique) Structure with Correlated Factors Oblique Rotation: Bully/Fight Factor Pattern ɛ 3 ɛ 4 X 3 X 4 F 2 Assessing Fit 41 / 52
42 Oblique Rotation: Bully/Fight Factor Pattern Rotation Factor Pattern not Unique Desirable Structure Orthogonal Rotations Varimax for Victim & Bully Items Varimax Bully/Fight Factor Pattern Desirable (Oblique) Structure with Correlated Factors Oblique Rotation: Bully/Fight Factor Pattern Assessing Fit Initial varimax oblimin Item Fac 1 Fac 2 Fac 1 Fac 2 Fac 1 Fac 2 Bully q Bully q Bully q Bully q Bully q Bully q Bully q Bully q Bully q Fight q Fight q Fight q Fight q Fight q From oblimin, ˆρ(F 1, F 2 ) = / 52
43 Rotation Methods Principal Components Solution of Factor Principal Factor or Axis Solution A Modern Solution: Maximum Likelihood Maximum Likelihood MLE continued Rotation & Indeterminacy Assessing Fit 43 / 52
44 Methods Rotation Methods Principal Components Solution of Factor Principal Factor or Axis Solution A Modern Solution: Maximum Likelihood Maximum Likelihood MLE continued Rotation & Indeterminacy Assessing Fit method (Factor extraction) Eigen-decomposition based Eigen-decomposition of S or the principal components solution of the factor model. Iterative eigen-decompositions of S Ψ or the Principal factor or Principal Axis solution. Maximum likelihood estimation We now must assume that F and ɛ are multivariate normal. Tends to fit data better & yields scale invariance (ie., use either S or R). 44 / 52
45 Principal Components Solution of Factor Rotation Methods Principal Components Solution of Factor Principal Factor or Axis Solution A Modern Solution: Maximum Likelihood Maximum Likelihood MLE continued Rotation & Indeterminacy Assessing Fit Let ˆλ 1 ˆλ 2... ˆλ p be the eigenvalues of the sample covariance matrix S and ê 1, ê 2,...,ê p be corresponding eigenvectors. Let m < p (ie, the number of factors be less than the number of X variables), then the matrix of factor loadings equals ( ) L = ˆλ 1 ê 1, ˆλ 2 ê 2,..., ˆλ m ê m The estimate specific variances are provided by the diagonal elements of the matrix S L L, So Ψ = ψ ψ ψm where ψi = s ii m q=1 And estimates of communalities are h 2 i = m q=1 l2 iq Alternatively, use R instead of S (will get a different result). l 2 iq 45 / 52
46 Principal Factor or Axis Solution Rotation Methods Principal Components Solution of Factor Principal Factor or Axis Solution A Modern Solution: Maximum Likelihood Maximum Likelihood MLE continued Rotation & Indeterminacy Assessing Fit 1. Get an initial estimate of Ψ. Most common choice is the square of the multiple correlation (SMC) coefficient between X i and the other p 1 variables. The SMC is the diagonal element of R 1 and is our initial estimate of the ψ i s; that is, ψ i = the i th diagonal element of R 1 = SMC 2. Find the eigenvalues λ q and eigenvectors e i of R diag(ψ i ) and set L = ( λ 1e 1, λ 2e 2,..., λ me m). 3. New estimates of ψ ψ i = 1 m q=1 l 2 iq = 1 h 2 i. 4. Repeat steps 2 through 3 until convergence. 46 / 52
47 A Modern Solution: Maximum Likelihood Rotation Methods Principal Components Solution of Factor Principal Factor or Axis Solution A Modern Solution: Maximum Likelihood Maximum Likelihood MLE continued Rotation & Indeterminacy Assessing Fit With principal component solution, extraction of an additional factor does not effect the values already extracted; this is not the case with the MLE or the principal factor method. Both MLE and principal factor methods can run into problems (Heywood cases, improper solutions) where the ψs want to be negative. You get different results if you use S or R (ie, not scale invariant); however, this is not the case for MLE: If l iq is a loading from a (population) correlation matrix, then l iq σ ii is the corresponding loading from the (population) covariance matrix. You can get a better fit via MLE. Statistical tests become possible. With MLE it becomes possible to set values for l iq (and/or ψ i )...a confirmatory analysis / 52
48 Rotation Methods Principal Components Solution of Factor Principal Factor or Axis Solution A Modern Solution: Maximum Likelihood Maximum Likelihood MLE continued Rotation & Indeterminacy Assessing Fit Maximum Likelihood For MLE, we need to add the assumption that (or X µ). Note that X N(0,Σ) i.i.d where Σ = LL + Ψ. Σ = LΦL + Ψ Ψ} 1/2 ΣΨ {{ 1/2 } = Ψ} 1/2 {{ LΦ 1/2 } Φ} 1/2 L {{ Ψ 1/2 } +I Σ L L Σ Given Ψ, we can get the L s: = L L + I Σ I = Ψ 1/2 ΣΨ 1/2 I = PΛP = PΛ }{{ 1/2 } Λ} 1/2 {{ P } L L 48 / 52
49 MLE continued Starting with a initial guess for Ψ this gives us an initial estimate of L. Using an optimization algorithm, iteratively up-date estimates by maximizing L(L,Ψ) = n 2 (ln( Σ ) + Tr(SΣ 1 )) + constant Rotation Methods Principal Components Solution of Factor Principal Factor or Axis Solution A Modern Solution: Maximum Likelihood Maximum Likelihood MLE continued Rotation & Indeterminacy Assessing Fit or equivalently minimize F(L,Ψ) = ln( σ ) + Tr(SΣ 1 )) ln( S ) p 49 / 52
50 Rotation & Indeterminacy Rotation Methods Principal Components Solution of Factor Principal Factor or Axis Solution A Modern Solution: Maximum Likelihood Maximum Likelihood MLE continued Rotation & Indeterminacy Assessing Fit To improve interpretation, we can rotate factor loadings using either an orthogonal or oblique method. With oblique, Σ = LΦL The m m matrix Φ is the Factor Pattern Matrix that contains correlations between the factors. The p m matrix LΦ is the Factor Structure Matrix that contains the covariances (or correlations) between the X p 1 observed variables and the F m 1 latent variables. Due to the indeterminacy, computing factor scores is not a reasonable thing to do event though many computer programs will give them to you if asked. With MLE, you can fix parameters (e.g., set some loadings equal to 0) and this leads to confirmatory factor analysis. 50 / 52
51 Assessing Fit to Data Rotation Assessing Fit to Data Or How Many Factors Needed? 51 / 52
52 Or How Many Factors Needed? Rotation Assessing Fit to Data Or How Many Factors Needed? Data points: There are p variances and p(p 1)/2 covariances. Parameters to be estimated: There are pm Factor loadings and p unique variances. Residual correlation matrix Root mean squared errors Statistical tests for number of factors Proportion of total sample variance due to the q th factor Proportion of total sample variance accounted for by the m factors. Others. 52 / 52
9.1 Orthogonal factor model.
36 Chapter 9 Factor Analysis Factor analysis may be viewed as a refinement of the principal component analysis The objective is, like the PC analysis, to describe the relevant variables in study in terms
More informationChapter 4: Factor Analysis
Chapter 4: Factor Analysis In many studies, we may not be able to measure directly the variables of interest. We can merely collect data on other variables which may be related to the variables of interest.
More informationIntroduction to Factor Analysis
to Factor Analysis Lecture 10 August 2, 2011 Advanced Multivariate Statistical Methods ICPSR Summer Session #2 Lecture #10-8/3/2011 Slide 1 of 55 Today s Lecture Factor Analysis Today s Lecture Exploratory
More information2/26/2017. This is similar to canonical correlation in some ways. PSY 512: Advanced Statistics for Psychological and Behavioral Research 2
PSY 512: Advanced Statistics for Psychological and Behavioral Research 2 What is factor analysis? What are factors? Representing factors Graphs and equations Extracting factors Methods and criteria Interpreting
More informationIntroduction to Factor Analysis
to Factor Analysis Lecture 11 November 2, 2005 Multivariate Analysis Lecture #11-11/2/2005 Slide 1 of 58 Today s Lecture Factor Analysis. Today s Lecture Exploratory factor analysis (EFA). Confirmatory
More informationFactor Analysis. Robert L. Wolpert Department of Statistical Science Duke University, Durham, NC, USA
Factor Analysis Robert L. Wolpert Department of Statistical Science Duke University, Durham, NC, USA 1 Factor Models The multivariate regression model Y = XB +U expresses each row Y i R p as a linear combination
More informationOr, in terms of basic measurement theory, we could model it as:
1 Neuendorf Factor Analysis Assumptions: 1. Metric (interval/ratio) data 2. Linearity (in relationships among the variables--factors are linear constructions of the set of variables; the critical source
More information3.1. The probabilistic view of the principal component analysis.
301 Chapter 3 Principal Components and Statistical Factor Models This chapter of introduces the principal component analysis (PCA), briefly reviews statistical factor models PCA is among the most popular
More informationFactor Analysis. Qian-Li Xue
Factor Analysis Qian-Li Xue Biostatistics Program Harvard Catalyst The Harvard Clinical & Translational Science Center Short course, October 7, 06 Well-used latent variable models Latent variable scale
More informationPrincipal Component Analysis & Factor Analysis. Psych 818 DeShon
Principal Component Analysis & Factor Analysis Psych 818 DeShon Purpose Both are used to reduce the dimensionality of correlated measurements Can be used in a purely exploratory fashion to investigate
More informationApplied Multivariate Analysis
Department of Mathematics and Statistics, University of Vaasa, Finland Spring 2017 Dimension reduction Exploratory (EFA) Background While the motivation in PCA is to replace the original (correlated) variables
More information1 A factor can be considered to be an underlying latent variable: (a) on which people differ. (b) that is explained by unknown variables
1 A factor can be considered to be an underlying latent variable: (a) on which people differ (b) that is explained by unknown variables (c) that cannot be defined (d) that is influenced by observed variables
More informationMore Linear Algebra. Edps/Soc 584, Psych 594. Carolyn J. Anderson
More Linear Algebra Edps/Soc 584, Psych 594 Carolyn J. Anderson Department of Educational Psychology I L L I N O I S university of illinois at urbana-champaign c Board of Trustees, University of Illinois
More informationTAMS39 Lecture 10 Principal Component Analysis Factor Analysis
TAMS39 Lecture 10 Principal Component Analysis Factor Analysis Martin Singull Department of Mathematics Mathematical Statistics Linköping University, Sweden Content - Lecture Principal component analysis
More informationSTAT 730 Chapter 9: Factor analysis
STAT 730 Chapter 9: Factor analysis Timothy Hanson Department of Statistics, University of South Carolina Stat 730: Multivariate Data Analysis 1 / 15 Basic idea Factor analysis attempts to explain the
More informationFACTOR ANALYSIS AND MULTIDIMENSIONAL SCALING
FACTOR ANALYSIS AND MULTIDIMENSIONAL SCALING Vishwanath Mantha Department for Electrical and Computer Engineering Mississippi State University, Mississippi State, MS 39762 mantha@isip.msstate.edu ABSTRACT
More informationLinear Dimensionality Reduction
Outline Hong Chang Institute of Computing Technology, Chinese Academy of Sciences Machine Learning Methods (Fall 2012) Outline Outline I 1 Introduction 2 Principal Component Analysis 3 Factor Analysis
More informationVAR2 VAR3 VAR4 VAR5. Or, in terms of basic measurement theory, we could model it as:
1 Neuendorf Factor Analysis Assumptions: 1. Metric (interval/ratio) data 2. Linearity (in the relationships among the variables) -Factors are linear constructions of the set of variables (see #8 under
More informationInter Item Correlation Matrix (R )
7 1. I have the ability to influence my child s well-being. 2. Whether my child avoids injury is just a matter of luck. 3. Luck plays a big part in determining how healthy my child is. 4. I can do a lot
More informationExploratory Factor Analysis and Principal Component Analysis
Exploratory Factor Analysis and Principal Component Analysis Today s Topics: What are EFA and PCA for? Planning a factor analytic study Analysis steps: Extraction methods How many factors Rotation and
More informationExploratory Factor Analysis and Principal Component Analysis
Exploratory Factor Analysis and Principal Component Analysis Today s Topics: What are EFA and PCA for? Planning a factor analytic study Analysis steps: Extraction methods How many factors Rotation and
More informationFactor Analysis (10/2/13)
STA561: Probabilistic machine learning Factor Analysis (10/2/13) Lecturer: Barbara Engelhardt Scribes: Li Zhu, Fan Li, Ni Guan Factor Analysis Factor analysis is related to the mixture models we have studied.
More informationFE670 Algorithmic Trading Strategies. Stevens Institute of Technology
FE670 Algorithmic Trading Strategies Lecture 3. Factor Models and Their Estimation Steve Yang Stevens Institute of Technology 09/12/2012 Outline 1 The Notion of Factors 2 Factor Analysis via Maximum Likelihood
More informationI L L I N O I S UNIVERSITY OF ILLINOIS AT URBANA-CHAMPAIGN
Canonical Edps/Soc 584 and Psych 594 Applied Multivariate Statistics Carolyn J. Anderson Department of Educational Psychology I L L I N O I S UNIVERSITY OF ILLINOIS AT URBANA-CHAMPAIGN Canonical Slide
More informationThe Reliability of a Homogeneous Test. Measurement Methods Lecture 12
The Reliability of a Homogeneous Test Measurement Methods Lecture 12 Today s Class Estimating the reliability of a homogeneous test. McDonald s Omega. Guttman-Chronbach Alpha. Spearman-Brown. What to do
More informationFactor Analysis Continued. Psy 524 Ainsworth
Factor Analysis Continued Psy 524 Ainsworth Equations Extraction Principal Axis Factoring Variables Skiers Cost Lift Depth Powder S1 32 64 65 67 S2 61 37 62 65 S3 59 40 45 43 S4 36 62 34 35 S5 62 46 43
More informationMultivariate Statistics
Multivariate Statistics Chapter 4: Factor analysis Pedro Galeano Departamento de Estadística Universidad Carlos III de Madrid pedro.galeano@uc3m.es Course 2017/2018 Master in Mathematical Engineering Pedro
More informationFactor Analysis. -Applied Multivariate Analysis- Lecturer: Darren Homrighausen, PhD
Factor Analysis -Applied Multivariate Analysis- Lecturer: Darren Homrighausen, PhD 1 From PCA to factor analysis Remember: PCA tries to estimate a transformation of the data such that: 1. The maximum amount
More informationFactor analysis. George Balabanis
Factor analysis George Balabanis Key Concepts and Terms Deviation. A deviation is a value minus its mean: x - mean x Variance is a measure of how spread out a distribution is. It is computed as the average
More informationInferences about a Mean Vector
Inferences about a Mean Vector Edps/Soc 584, Psych 594 Carolyn J. Anderson Department of Educational Psychology I L L I N O I S university of illinois at urbana-champaign c Board of Trustees, University
More informationE X P L O R E R. R e l e a s e A Program for Common Factor Analysis and Related Models for Data Analysis
E X P L O R E R R e l e a s e 3. 2 A Program for Common Factor Analysis and Related Models for Data Analysis Copyright (c) 1990-2011, J. S. Fleming, PhD Date and Time : 23-Sep-2011, 13:59:56 Number of
More informationIntermediate Social Statistics
Intermediate Social Statistics Lecture 5. Factor Analysis Tom A.B. Snijders University of Oxford January, 2008 c Tom A.B. Snijders (University of Oxford) Intermediate Social Statistics January, 2008 1
More informationI L L I N O I S UNIVERSITY OF ILLINOIS AT URBANA-CHAMPAIGN
Principal Analysis Edps/Soc 584 and Psych 594 Applied Multivariate Statistics Carolyn J. Anderson Department of Educational Psychology I L L I N O I S UNIVERSITY OF ILLINOIS AT URBANA-CHAMPAIGN c Board
More informationPCA and admixture models
PCA and admixture models CM226: Machine Learning for Bioinformatics. Fall 2016 Sriram Sankararaman Acknowledgments: Fei Sha, Ameet Talwalkar, Alkes Price PCA and admixture models 1 / 57 Announcements HW1
More informationPrinciple Components Analysis (PCA) Relationship Between a Linear Combination of Variables and Axes Rotation for PCA
Principle Components Analysis (PCA) Relationship Between a Linear Combination of Variables and Axes Rotation for PCA Principle Components Analysis: Uses one group of variables (we will call this X) In
More informationMultilevel Models in Matrix Form. Lecture 7 July 27, 2011 Advanced Multivariate Statistical Methods ICPSR Summer Session #2
Multilevel Models in Matrix Form Lecture 7 July 27, 2011 Advanced Multivariate Statistical Methods ICPSR Summer Session #2 Today s Lecture Linear models from a matrix perspective An example of how to do
More informationI L L I N O I S UNIVERSITY OF ILLINOIS AT URBANA-CHAMPAIGN
Introduction Edps/Psych/Stat/ 584 Applied Multivariate Statistics Carolyn J Anderson Department of Educational Psychology I L L I N O I S UNIVERSITY OF ILLINOIS AT URBANA-CHAMPAIGN c Board of Trustees,
More information18.S096 Problem Set 7 Fall 2013 Factor Models Due Date: 11/14/2013. [ ] variance: E[X] =, and Cov[X] = Σ = =
18.S096 Problem Set 7 Fall 2013 Factor Models Due Date: 11/14/2013 1. Consider a bivariate random variable: [ ] X X = 1 X 2 with mean and co [ ] variance: [ ] [ α1 Σ 1,1 Σ 1,2 σ 2 ρσ 1 σ E[X] =, and Cov[X]
More informationLECTURE 4 PRINCIPAL COMPONENTS ANALYSIS / EXPLORATORY FACTOR ANALYSIS
LECTURE 4 PRINCIPAL COMPONENTS ANALYSIS / EXPLORATORY FACTOR ANALYSIS NOTES FROM PRE- LECTURE RECORDING ON PCA PCA and EFA have similar goals. They are substantially different in important ways. The goal
More informationMultivariate and Multivariable Regression. Stella Babalola Johns Hopkins University
Multivariate and Multivariable Regression Stella Babalola Johns Hopkins University Session Objectives At the end of the session, participants will be able to: Explain the difference between multivariable
More informationSample Geometry. Edps/Soc 584, Psych 594. Carolyn J. Anderson
Sample Geometry Edps/Soc 584, Psych 594 Carolyn J. Anderson Department of Educational Psychology I L L I N O I S university of illinois at urbana-champaign c Board of Trustees, University of Illinois Spring
More informationVectors and Matrices Statistics with Vectors and Matrices
Vectors and Matrices Statistics with Vectors and Matrices Lecture 3 September 7, 005 Analysis Lecture #3-9/7/005 Slide 1 of 55 Today s Lecture Vectors and Matrices (Supplement A - augmented with SAS proc
More information6348 Final, Fall 14. Closed book, closed notes, no electronic devices. Points (out of 200) in parentheses.
6348 Final, Fall 14. Closed book, closed notes, no electronic devices. Points (out of 200) in parentheses. 0 11 1 1.(5) Give the result of the following matrix multiplication: 1 10 1 Solution: 0 1 1 2
More informationFactor Analysis (1) Factor Analysis
Factor Analysis (1) Outlines: 1. Introduction of factor analysis 2. Principle component analysis 4. Factor rotation 5. Case Shan-Yu Chou 1 Factor Analysis Combines questions or variables to create new
More informationTHE UNIVERSITY OF CHICAGO Graduate School of Business Business 41912, Spring Quarter 2008, Mr. Ruey S. Tsay. Solutions to Final Exam
THE UNIVERSITY OF CHICAGO Graduate School of Business Business 41912, Spring Quarter 2008, Mr. Ruey S. Tsay Solutions to Final Exam 1. (13 pts) Consider the monthly log returns, in percentages, of five
More informationConflict Tactics Scales Kindergarten/Year 1 Fast Track Project Technical Report Cynthia Rains November 15, 2004
Conflict Tactics Scales Kindergarten/Year 1 Fast Track Project Technical Report Cynthia Rains November 15, 2004 Table of Contents I. Scale Description II. Report Sample III. Scaling IV. Differences Between
More informationUCLA STAT 233 Statistical Methods in Biomedical Imaging
UCLA STAT 233 Statistical Methods in Biomedical Imaging Instructor: Ivo Dinov, Asst. Prof. In Statistics and Neurology University of California, Los Angeles, Spring 2004 http://www.stat.ucla.edu/~dinov/
More informationCS281 Section 4: Factor Analysis and PCA
CS81 Section 4: Factor Analysis and PCA Scott Linderman At this point we have seen a variety of machine learning models, with a particular emphasis on models for supervised learning. In particular, we
More informationShort Answer Questions: Answer on your separate blank paper. Points are given in parentheses.
ISQS 6348 Final exam solutions. Name: Open book and notes, but no electronic devices. Answer short answer questions on separate blank paper. Answer multiple choice on this exam sheet. Put your name on
More informationData Mining. Dimensionality reduction. Hamid Beigy. Sharif University of Technology. Fall 1395
Data Mining Dimensionality reduction Hamid Beigy Sharif University of Technology Fall 1395 Hamid Beigy (Sharif University of Technology) Data Mining Fall 1395 1 / 42 Outline 1 Introduction 2 Feature selection
More informationFactor Analysis: An Introduction. What is Factor Analysis? 100+ years of Factor Analysis FACTOR ANALYSIS AN INTRODUCTION NILAM RAM
NILAM RAM 2018 PSYCHOLOGY R BOOTCAMP PENNSYLVANIA STATE UNIVERSITY AUGUST 16, 2018 FACTOR ANALYSIS https://psu-psychology.github.io/r-bootcamp-2018/index.html WITH ADDITIONAL MATERIALS AT https://quantdev.ssri.psu.edu/tutorials
More informationEstimation: Problems & Solutions
Estimation: Problems & Solutions Edps/Psych/Stat 587 Carolyn J. Anderson Department of Educational Psychology c Board of Trustees, University of Illinois Fall 2017 Outline 1. Introduction: Estimation of
More informationCapítulo 12 FACTOR ANALYSIS 12.1 INTRODUCTION
Capítulo 12 FACTOR ANALYSIS Charles Spearman (1863-1945) British psychologist. Spearman was an officer in the British Army in India and upon his return at the age of 40, and influenced by the work of Galton,
More informationCanonical Correlations
Canonical Correlations Like Principal Components Analysis, Canonical Correlation Analysis looks for interesting linear combinations of multivariate observations. In Canonical Correlation Analysis, a multivariate
More informationPrincipal Component Analysis (PCA) Principal Component Analysis (PCA)
Recall: Eigenvectors of the Covariance Matrix Covariance matrices are symmetric. Eigenvectors are orthogonal Eigenvectors are ordered by the magnitude of eigenvalues: λ 1 λ 2 λ p {v 1, v 2,..., v n } Recall:
More informationVAR Model. (k-variate) VAR(p) model (in the Reduced Form): Y t-2. Y t-1 = A + B 1. Y t + B 2. Y t-p. + ε t. + + B p. where:
VAR Model (k-variate VAR(p model (in the Reduced Form: where: Y t = A + B 1 Y t-1 + B 2 Y t-2 + + B p Y t-p + ε t Y t = (y 1t, y 2t,, y kt : a (k x 1 vector of time series variables A: a (k x 1 vector
More informationRandom Vectors 1. STA442/2101 Fall See last slide for copyright information. 1 / 30
Random Vectors 1 STA442/2101 Fall 2017 1 See last slide for copyright information. 1 / 30 Background Reading: Renscher and Schaalje s Linear models in statistics Chapter 3 on Random Vectors and Matrices
More informationVector Space Models. wine_spectral.r
Vector Space Models 137 wine_spectral.r Latent Semantic Analysis Problem with words Even a small vocabulary as in wine example is challenging LSA Reduce number of columns of DTM by principal components
More informationsimple if it completely specifies the density of x
3. Hypothesis Testing Pure significance tests Data x = (x 1,..., x n ) from f(x, θ) Hypothesis H 0 : restricts f(x, θ) Are the data consistent with H 0? H 0 is called the null hypothesis simple if it completely
More informationStat 206, Week 6: Factor Analysis
Stat 206, Week 6: Factor Analysis James Johndrow 2016-09-24 Introduction Factor analysis aims to explain correlation between a large set (p) of variables in terms of a smaller number (m) of underlying
More informationFactor Analysis of Data Matrices
Factor Analysis of Data Matrices PAUL HORST University of Washington HOLT, RINEHART AND WINSTON, INC. New York Chicago San Francisco Toronto London Contents Preface PART I. Introductory Background 1. The
More informationStatistics for Applications. Chapter 9: Principal Component Analysis (PCA) 1/16
Statistics for Applications Chapter 9: Principal Component Analysis (PCA) 1/16 Multivariate statistics and review of linear algebra (1) Let X be a d-dimensional random vector and X 1,..., X n be n independent
More informationUNIVERSITY OF TORONTO Faculty of Arts and Science
UNIVERSITY OF TORONTO Faculty of Arts and Science December 2013 Final Examination STA442H1F/2101HF Methods of Applied Statistics Jerry Brunner Duration - 3 hours Aids: Calculator Model(s): Any calculator
More informationLecture 7 Spectral methods
CSE 291: Unsupervised learning Spring 2008 Lecture 7 Spectral methods 7.1 Linear algebra review 7.1.1 Eigenvalues and eigenvectors Definition 1. A d d matrix M has eigenvalue λ if there is a d-dimensional
More informationPrincipal Component Analysis. Applied Multivariate Statistics Spring 2012
Principal Component Analysis Applied Multivariate Statistics Spring 2012 Overview Intuition Four definitions Practical examples Mathematical example Case study 2 PCA: Goals Goal 1: Dimension reduction
More informationRegularized Common Factor Analysis
New Trends in Psychometrics 1 Regularized Common Factor Analysis Sunho Jung 1 and Yoshio Takane 1 (1) Department of Psychology, McGill University, 1205 Dr. Penfield Avenue, Montreal, QC, H3A 1B1, Canada
More informationPsychology 405: Psychometric Theory Constructs, Components, and Factor Models
Psychology 405: Psychometric Theory Constructs, Components, and Factor Models William Revelle Department of Psychology Northwestern University Evanston, Illinois USA April, 2012 1 / 58 Outline 1 Preliminaries
More informationPrincipal Components Theory Notes
Principal Components Theory Notes Charles J. Geyer August 29, 2007 1 Introduction These are class notes for Stat 5601 (nonparametrics) taught at the University of Minnesota, Spring 2006. This not a theory
More informationA Study of Statistical Power and Type I Errors in Testing a Factor Analytic. Model for Group Differences in Regression Intercepts
A Study of Statistical Power and Type I Errors in Testing a Factor Analytic Model for Group Differences in Regression Intercepts by Margarita Olivera Aguilar A Thesis Presented in Partial Fulfillment of
More informationExploratory Factor Analysis and Canonical Correlation
Exploratory Factor Analysis and Canonical Correlation 3 Dec 2010 CPSY 501 Dr. Sean Ho Trinity Western University Please download: SAQ.sav Outline for today Factor analysis Latent variables Correlation
More information[y i α βx i ] 2 (2) Q = i=1
Least squares fits This section has no probability in it. There are no random variables. We are given n points (x i, y i ) and want to find the equation of the line that best fits them. We take the equation
More informationDimensionality Reduction Techniques (DRT)
Dimensionality Reduction Techniques (DRT) Introduction: Sometimes we have lot of variables in the data for analysis which create multidimensional matrix. To simplify calculation and to get appropriate,
More informationMULTIVARIATE TIME SERIES ANALYSIS AN ADAPTATION OF BOX-JENKINS METHODOLOGY Joseph N Ladalla University of Illinois at Springfield, Springfield, IL
MULTIVARIATE TIME SERIES ANALYSIS AN ADAPTATION OF BOX-JENKINS METHODOLOGY Joseph N Ladalla University of Illinois at Springfield, Springfield, IL KEYWORDS: Multivariate time series, Box-Jenkins ARIMA
More informationComparing IRT with Other Models
Comparing IRT with Other Models Lecture #14 ICPSR Item Response Theory Workshop Lecture #14: 1of 45 Lecture Overview The final set of slides will describe a parallel between IRT and another commonly used
More informationCS534 Machine Learning - Spring Final Exam
CS534 Machine Learning - Spring 2013 Final Exam Name: You have 110 minutes. There are 6 questions (8 pages including cover page). If you get stuck on one question, move on to others and come back to the
More information2. Matrix Algebra and Random Vectors
2. Matrix Algebra and Random Vectors 2.1 Introduction Multivariate data can be conveniently display as array of numbers. In general, a rectangular array of numbers with, for instance, n rows and p columns
More informationStructural Equation Modeling and Confirmatory Factor Analysis. Types of Variables
/4/04 Structural Equation Modeling and Confirmatory Factor Analysis Advanced Statistics for Researchers Session 3 Dr. Chris Rakes Website: http://csrakes.yolasite.com Email: Rakes@umbc.edu Twitter: @RakesChris
More informationDimensionality Reduction
Dimensionality Reduction Le Song Machine Learning I CSE 674, Fall 23 Unsupervised learning Learning from raw (unlabeled, unannotated, etc) data, as opposed to supervised data where a classification of
More informationLinear Methods for Prediction
Chapter 5 Linear Methods for Prediction 5.1 Introduction We now revisit the classification problem and focus on linear methods. Since our prediction Ĝ(x) will always take values in the discrete set G we
More informationSTATISTICAL LEARNING SYSTEMS
STATISTICAL LEARNING SYSTEMS LECTURE 8: UNSUPERVISED LEARNING: FINDING STRUCTURE IN DATA Institute of Computer Science, Polish Academy of Sciences Ph. D. Program 2013/2014 Principal Component Analysis
More informationLOG-MULTIPLICATIVE ASSOCIATION MODELS AS LATENT VARIABLE MODELS FOR NOMINAL AND0OR ORDINAL DATA. Carolyn J. Anderson* Jeroen K.
3 LOG-MULTIPLICATIVE ASSOCIATION MODELS AS LATENT VARIABLE MODELS FOR NOMINAL AND0OR ORDINAL DATA Carolyn J. Anderson* Jeroen K. Vermunt Associations between multiple discrete measures are often due to
More informationICS 6N Computational Linear Algebra Symmetric Matrices and Orthogonal Diagonalization
ICS 6N Computational Linear Algebra Symmetric Matrices and Orthogonal Diagonalization Xiaohui Xie University of California, Irvine xhx@uci.edu Xiaohui Xie (UCI) ICS 6N 1 / 21 Symmetric matrices An n n
More informationModels for Clustered Data
Models for Clustered Data Edps/Psych/Soc 589 Carolyn J Anderson Department of Educational Psychology c Board of Trustees, University of Illinois Spring 2019 Outline Notation NELS88 data Fixed Effects ANOVA
More informationLecture 13. Principal Component Analysis. Brett Bernstein. April 25, CDS at NYU. Brett Bernstein (CDS at NYU) Lecture 13 April 25, / 26
Principal Component Analysis Brett Bernstein CDS at NYU April 25, 2017 Brett Bernstein (CDS at NYU) Lecture 13 April 25, 2017 1 / 26 Initial Question Intro Question Question Let S R n n be symmetric. 1
More informationIntroduction to Machine Learning
10-701 Introduction to Machine Learning PCA Slides based on 18-661 Fall 2018 PCA Raw data can be Complex, High-dimensional To understand a phenomenon we measure various related quantities If we knew what
More informationModels for Clustered Data
Models for Clustered Data Edps/Psych/Stat 587 Carolyn J Anderson Department of Educational Psychology c Board of Trustees, University of Illinois Fall 2017 Outline Notation NELS88 data Fixed Effects ANOVA
More informationA Factor Analysis of Key Decision Factors for Implementing SOA-based Enterprise level Business Intelligence Systems
icbst 2014 International Conference on Business, Science and Technology which will be held at Hatyai, Thailand on the 25th and 26th of April 2014. AENSI Journals Australian Journal of Basic and Applied
More informationMultivariate Statistical Analysis
Multivariate Statistical Analysis Fall 2011 C. L. Williams, Ph.D. Lecture 4 for Applied Multivariate Analysis Outline 1 Eigen values and eigen vectors Characteristic equation Some properties of eigendecompositions
More informationData Analysis and Manifold Learning Lecture 6: Probabilistic PCA and Factor Analysis
Data Analysis and Manifold Learning Lecture 6: Probabilistic PCA and Factor Analysis Radu Horaud INRIA Grenoble Rhone-Alpes, France Radu.Horaud@inrialpes.fr http://perception.inrialpes.fr/ Outline of Lecture
More informationMultivariate Fundamentals: Rotation. Exploratory Factor Analysis
Multivariate Fundamentals: Rotation Exploratory Factor Analysis PCA Analysis A Review Precipitation Temperature Ecosystems PCA Analysis with Spatial Data Proportion of variance explained Comp.1 + Comp.2
More informationFactor Models for Asset Returns. Prof. Daniel P. Palomar
Factor Models for Asset Returns Prof. Daniel P. Palomar The Hong Kong University of Science and Technology (HKUST) MAFS6010R- Portfolio Optimization with R MSc in Financial Mathematics Fall 2018-19, HKUST,
More informationA Primer on Statistical Inference using Maximum Likelihood
A Primer on Statistical Inference using Maximum Likelihood November 3, 2017 1 Inference via Maximum Likelihood Statistical inference is the process of using observed data to estimate features of the population.
More informationIntelligent Data Analysis. Principal Component Analysis. School of Computer Science University of Birmingham
Intelligent Data Analysis Principal Component Analysis Peter Tiňo School of Computer Science University of Birmingham Discovering low-dimensional spatial layout in higher dimensional spaces - 1-D/3-D example
More informationPrincipal Component Analysis
Principal Component Analysis Giorgos Korfiatis Alfa-Informatica University of Groningen Seminar in Statistics and Methodology, 2007 What Is PCA? Dimensionality reduction technique Aim: Extract relevant
More informationStat 5101 Lecture Notes
Stat 5101 Lecture Notes Charles J. Geyer Copyright 1998, 1999, 2000, 2001 by Charles J. Geyer May 7, 2001 ii Stat 5101 (Geyer) Course Notes Contents 1 Random Variables and Change of Variables 1 1.1 Random
More information1 Data Arrays and Decompositions
1 Data Arrays and Decompositions 1.1 Variance Matrices and Eigenstructure Consider a p p positive definite and symmetric matrix V - a model parameter or a sample variance matrix. The eigenstructure is
More informationFactor Analysis and Kalman Filtering (11/2/04)
CS281A/Stat241A: Statistical Learning Theory Factor Analysis and Kalman Filtering (11/2/04) Lecturer: Michael I. Jordan Scribes: Byung-Gon Chun and Sunghoon Kim 1 Factor Analysis Factor analysis is used
More informationProbability reminders
CS246 Winter 204 Mining Massive Data Sets Probability reminders Sammy El Ghazzal selghazz@stanfordedu Disclaimer These notes may contain typos, mistakes or confusing points Please contact the author so
More informationImproper Solutions in Exploratory Factor Analysis: Causes and Treatments
Improper Solutions in Exploratory Factor Analysis: Causes and Treatments Yutaka Kano Faculty of Human Sciences, Osaka University Suita, Osaka 565, Japan. email: kano@hus.osaka-u.ac.jp Abstract: There are
More informationFor more information about how to cite these materials visit
Author(s): Kerby Shedden, Ph.D., 2010 License: Unless otherwise noted, this material is made available under the terms of the Creative Commons Attribution Share Alike 3.0 License: http://creativecommons.org/licenses/by-sa/3.0/
More information