B. Weaver (18-Oct-2001) Factor analysis Chapter 7: Factor Analysis

Size: px
Start display at page:

Download "B. Weaver (18-Oct-2001) Factor analysis Chapter 7: Factor Analysis"

Transcription

1 B Weaver (18-Oct-2001) Factor analysis 1 Chapter 7: Factor Analysis 71 Introduction Factor analysis (FA) was developed by C Spearman It is a technique for examining the interrelationships in a set of variables As Ferguson and Takane (1989) put it, concern is with the description and interpretation of interdependencies within a single set of variables The usefulness of this technique becomes apparent when the number of variables under consideration is large For example, Table 71 shows the means and standard deviations for 24 psychological tests that were administered to 145 children; and Table 72 shows a correlation matrix for the 24 tests Clearly, it is very difficult to glean any useful information from inspection of these tables The information needs to be summarized somehow to make it more readily understandable Table 71 Means and standard deviations for 24 psychological tests Standard deviation Test Mean 1 Visual perception Cubes Paper form board Flags General information Paragraph comprehension Sentence completion Word classification Word meaning Addition Code Counting dots Straight-curved capitals Word recognition Number recognition Figure recognition Object-number Number-figure Figure-word Deduction Numerical puzzles Problem reasoning Series completion Arithmetic problems

2 B Weaver (18-Oct-2001) Factor analysis 2 Table 72 Intercorrelations of 24 psychological tests for 145 children Test

3 B Weaver (18-Oct-2001) Factor analysis 3 Tables 71 and 72 are taken from Ferguson and Takane (1989) The results of a factor analysis they carried out on these data are also reproduced in Table 73 They identified 4 underlying factors, which they identified as verbal, speed, deduction, and memory The numbers in the table are factor loadings The factor loadings are correlation coefficients Factor loadings equal to or greater than 4 are shown in bold in Table 73 Thus we can see, for example, that tests 5-9 correlate very highly with the verbal factor, but not with any of the other factors Similarly tests correlate highly with the speed factor, but apart from test 13, their correlations with all other factors are less than 4 And so on The point is that it is much easier to see structure in the data in Table 73 than it is in Tables 71 and 72 Table 73 Normal varimax solution for 24 psychological tests (following initial extraction using principal factor analysis) Test Verbal F 1 Speed F 2 Deduction F 3 Memory F 4 Calculated communality V p Percent Basic concepts of factor analysis Afifi and Clark (1990) do a masterful job of introducing the basic concepts of factor analysis If it were feasible, I would simply ask all of you to read Chapters 14 and 15 of their

4 B Weaver (18-Oct-2001) Factor analysis 4 book However, the library has only one copy, and the book is no doubt quite expensive Therefore, I will quote extensively from certain sections of it The shaded area that follows is Afifi and Clark s Section 154 (pp ): In factor analysis we begin with a set of variables X 1, X 2,, X P These variables are usually standardized by the computer program so that their variances are each equal to one and their covariances are correlation coefficients In the remainder of this chapter we therefore assume that each x i is a standardized variable, ie, x i = (X i Xi ) S i In the jargon of factor analysis the x i s are called the original or response variables The object of factor analysis is to represent each of these variables as a linear combination of a smaller set of common factors plus a factor unique to each of the response variables We express this representation as x 1 = l 11 F 1 + l 12 F l 1m F m + e 1 x 2 = l 21 F 1 + l 22 F l 2m F m + e 2 x P = l P1 F 1 + l P2 F l Pm F m + e P where the following assumptions are made: 1 m is the number of common factors (typically this number is much smaller than P) 2 F 1, F 2,, F m are the common factors These factors are assumed to have zero means and unit variances 3 l ij is the coefficient of F j in the linear combination describing x i This term is called the loading of the ith variable on the jth common factor 4 e 1, e 2,, e P are unique factors, each relating to one of the original variables The above equations and assumptions constitute the so-called factor model Thus each of the response variables is composed of a part due to the common factors and a part due to its own unique factor [emphasis added] The part due to the common factors is assumed to be a linear combination of these factors As an example, suppose that x 1, x 2, x 3, x 4, x 5 are the standardized scores on an individual on five tests If m = 2 [ie, there are 2 common factors], we assume the following model: x 1 = l 11 F 1 + l 12 F 2 + e 1 x 2 = l 21 F 1 + l 22 F 2 + e 2 x 5 = l 51 F 1 + l 52 F 2 + e P

5 B Weaver (18-Oct-2001) Factor analysis 5 Each of the five scores consists of two parts: a part due the common factors F 1 and F 2 and a part due to the unique factor for that test The common factors F 1 and F 2 might be considered the individual s verbal and quantitative abilities respectively [for example] The unique factors express the individual variation on each test score The unique factor includes all other effects that keep the common factors from completely defining a particular x i In a sample of N individuals we can express the equations by adding a subscript to each x i, F j, and e i to represent the individual For the sake of simplicity this subscript was omitted in the model presented here The factor model is, in a sense, the mirror image of the principal components model, where each principal component is expressed as a linear combination of the variables Also, the number of principal components is equal to the number of original variables (although we may not use all of the principal components) [I will discuss principal components analysis in more detail later] On the other hand, in factor analysis we choose the number of factors to be smaller than the number of response variables Ideally, the number of factors should be known in advance, although this is often not the case However, as we will discuss later, it is possible to allow the data themselves to determine this number The factor model, by breaking each response variable x i into two parts, also breaks the variance of x i into two parts [emphasis added] Since x i is standardized, its variance is one This variance of one is composed of the following two parts: 1 The communality, ie, the part of the variance that is due to the common factors 2 The specificity, ie, the part of the variance that is due to the unique factor e i Denoting the communality of x i by h i 2 and the specificity by u i 2, we can write the variance of x i as Var x i = 1 = h i 2 + u i 2 In words, the variance of x i equals the communality plus the specificity 73 Two steps in factor analysis There are two steps in factor analysis: 1) initial factor extraction, and 2) rotation of the factors There are a number of techniques for carrying out each of these steps, and as Tabachnick and Fidell (1989) point out, these may be used in various combinations: None of the extraction techniques routinely provides an interpretable solution without rotation All types of extraction may be rotated by any of the procedures described in Section 1252 [of Tabachnick & Fidell, 1989], except Kaiser's Second Little Jiffy Extraction, which has its own rotational procedure We do not have enough time, nor do I have enough knowledge (but let s not go into that!) to discuss all of the available methods of initial factor extraction and factor rotation Therefore, let us focus on the most common techniques, and attempt to understand the basic principles The two most common factor extraction techniques are principal components analysis (PCA) and principal factor analysis (PFA) (PFA was the method used by Ferguson & Takane to generate the solution shown in Table 73) PCA is a bit easier to understand, so I will start with it I will first discuss PCA as a statistical technique in its own

6 B Weaver (18-Oct-2001) Factor analysis 6 right, and follow that with discussion of how it is used for initial factor extraction in factor analysis 74 Principal Components Analysis Principal components analysis (PCA) is a method for transforming a set of intercorrelated variables into a new set of uncorrelated variables These new, uncorrelated variables are called the principal components Each of the principal components is a linear combination of the original variables According to Afifi and Clark (1990, p 372), One measure of the amount of information conveyed by each principal component is its variance For this reason the principal components are arranged in order of decreasing variance Thus the most informative principal component is the first, and the least informative is the last (a variable with zero variance does not distinguish between members of the population) The number of principal components is equal to the number of original variables However, it is often possible to reduce the number of variables that need to be considered by analyzing only the first few principal components In essence, this allows one to analyze a relatively small number of uncorrelated principal components rather than a large number of original variables with complex interrelationships, and in so doing to explain most (but not all) of the variance in the original set of variables (Afifi & Clark, 1990, p 372) 741 A simple example of Principal Components Analysis Afifi and Clark (1990) introduce the basic concepts of PCA in the following simple (and hypothetical) situation Two variables, X 1 and X 2, are both normally distributed in the population, and we have a sample of N = 100 paired scores The population mean and variance for X 1 are 100 and 100; the population mean and variance for X 2 are 50 and 50 The sample statistics are as shown in Table 74 Finally the population correlation between X 1 and X 2 is 0707, and the sample correlation is r = 757 Table 74 Sample statistics for hypothetical data set Statistic X 1 X 2 N Mean Standard deviation Variance It will make interpretation of the results somewhat easier if we work with deviation scores rather than with the original data Recall that a deviation score is obtained by subtracting the mean from each raw score Thus, we can define variables x 1 and x 2 as follows: x 1 = X 1 X1 (71)

7 B Weaver (18-Oct-2001) Factor analysis 7 x 2 = X 2 X2 (72) Note that the means of x 1 and x 2 will both be equal to zero (because the mean of the deviations about the mean is equal to zero for any distribution); but the sample variances (S 2 1 and S 2 2) and r 12 will not change What we need to do next is create two new variables called C 1 and C 2 These are the principal components They are linear combinations of x 1 and x 2 : C 1 = a 11 x 1 + a 12 x 2 (73) C 2 = a 21 x 1 + a 22 x 2 (74) The coefficients (a 11, a 12, a 21, and a 22 ) are chosen to meet the following three requirements: 1 The variance of C 1 must be as large as possible 2 The N pairs of C 1 and C 2 values must be uncorrelated 3 a a 2 12 = a a 22 = 1 Afifi and Clark (1990) go on to say this: The mathematical solution for the coefficients was derived by Hotelling (1933) The solution is illustrated graphically in Figure 142 [see below] Principal components analysis amounts to rotating the original x 1 and x 2 axes to new C 1 and C 2 axes The angle of rotation is determined uniquely by the requirements just stated For a given point x 1, x 2 (see Figure 142) the values of C 1 and C 2 are found by drawing perpendicular lines to the new C 1 and C 2 axes The N values of C 1 thus obtained will have the largest variance according to requirement 1 The N values of C 1 and C 2 will have a zero correlation C 2 x 2 x1,x2 C 1 x 1 Afifi & Clark Figure 142 Illustration of Two Principal Components C 1 and C 2 For the simple example we have been examining, the two principal components are:

8 B Weaver (18-Oct-2001) Factor analysis 8 C 1 = 0841x x 2 (75) C 2 = 0539x x 2 (76) Note that these coefficients do meet the third requirement listed above That is, apart from some rounding error, = 1, and (-0539) = 1 Note as well that a 11 = a 22, and a 12 = -a 21 Afifi and Clark point out that this is so for the two-variable case only The original x 1 and x 2 axes, and the rotated C 1 and C 2 principal component axes are shown in Figure 143 from Afifi and Clark (1990) The figure also shows an ellipse of concentration of the original bivariate normal distribution (Afifi & Clark, 1990) (The ellipse defines an area within which most of the original x 1, x 2 pairs fall) NORM 2 x 2 C 1 C 2 X2 x 1 X1 NORM 1 Afifi & Clark Figure 143 Plot of Principal Components for Hypothetical Bivariate Data To calculate values of C 1 and C 2 for a given subject, you would have to plug that subject s x 1 and x 2 scores into equations 75 and 76 If you did this for each subject, you would end up with N pairs of C 1 and C 2 scores The means and variances C 1 and C 2 of are as follows: C1 = C 2 = 0 (77) Var C 1 = a 2 11 S a 2 12 S a 11 a 22 r 12 S 1 S 2 (78) Var C 2 = a 2 21 S a 2 22 S a 21 a 22 r 12 S 1 S 2 (79) where S i 2 = the variance of X i For our data, the variance of C 1 is 14744, and the variance of C 2 is 1759 These two variances have special names: They are known as the first and second eigenvalues (You could also see the terms characteristic root, latent root, or proper value used in place of eigenvalue) The sum of variances of C 1 and C 2, or the sum of the eigenvalues, is equal to Note that the sum of the variances of the original variables (x 1 and x 2 ) is also equal to 16503: = According to Afifi and Clark (1990), this result will always be the

9 B Weaver (18-Oct-2001) Factor analysis 9 case; ie, the total variance is preserved under rotation of the principal components Afifi and Clark also point out that: The lengths of the axes of the ellipse of concentration (Figure 143) are proportional to the sample standard deviations of C 1 and C 2, respectively These standard deviations are the square roots of the eigenvalues It is therefore easily seen from Figure 143 that C 1 has a larger variance than C 2 In fact, C 1 has a larger variance than either of the original variables x 1 and x 2 In other words, the total variance remains the same, but it is partitioned in a different way In the original variables, the variance of x 1 accounted for 6643% of the total variance, and x 2 accounted for the remaining 3357% But when you look at the principal components, C 1 accounts for 8934% of the total variance, and C 2 for only 1066% 742 PCA with more than 2 variables The basic principles we have looked at in a simple two-variable context can be extended to the case of P variables As in the simple case, each principal component (C i ) is a linear combination of variables x 1, x 2, x P And according to Afifi and Clark (1990, p 378), the following must be true: 1 Var C 1 Var C 2 Var C P 2 The values of any two principal components are uncorrelated 3 For any principal component the sum of the squares of the coefficients is one As in the two-variable case, the variances of the principal components (Var C i ) are the eigenvalues The sum of the eigenvalues is equal to the total variance in the original variables (ie, the sum of the variances of the original variables) Finally, you may also encounter the term eigenvector The ith eigenvector is just the set of coefficients of the linear combination for the ith principal component It is also known as the characteristic vector, or latent vector 743 Principal Components Analysis using Standardized Variables It is often preferable to standardize the x variables before carrying out principal components analysis (You may recall that to standardize a variable, you must subtract the mean from each score, and divide the difference by the standard deviation The x variables are already deviation scores, so all that we need to do to standardize them is divide each one by its standard deviation) According to Afifi and Clark (1990, p 380), This analysis [using standardized variables] is then equivalent to analyzing the correlation matrix instead of the covariance matrix When we derive the principal components from the correlation matrix, the interpretation becomes easier in two ways: 1 The total variance is simply the number of variables P, and the proportion explained by each principal component is the corresponding eigenvalue divided by P 2 The correlation between the i th principal component C i and the j th variable x j is

10 B Weaver (18-Oct-2001) Factor analysis 10 r ij = a ij Var _C i Therefore for a given C i we can compare the a ij to quantify the relative degree of dependence of C i on each of the standardized variables Finally, it is important to note that the covariance matrix and the correlation matrix yield different solutions Furthermore, according to Afifi and Clark (1990), there is no easy way to convert the results based on the covariance matrix into those based on the correlation matrix They also suggest that use of the correlation matrix is preferred because all variables are measured in standard deviation units But if it is used, they caution, then all interpretations must be made in terms of the standardized variables 744 Reduction of dimensionality Principal components analysis can be useful in its own right, not just as a first step (ie, initial factor extraction) in factor analysis A major reason for using PCA is that it may allow reduction of dimensionality That is, if the original variables are not all mutually independent, one can probably reduce the problem from dealing with a large number of interrelated variables to dealing with a relatively small number of uncorrelated principal components (Remember that the principal components are just linear combinations of the original variables) The question is how many of the principal components should be retained for further analysis? There are no hard and fast rules to apply here, only guidelines One rule of thumb that can be used for either the correlation matrix or the covariance matrix is to omit those principal components that do not explain at least 100/P percent of the total variance (Afifi & Clark, 1990, p 385) For example, if you had P = 20 variables, you would omit those principal components that do not explain at least 5% of the total variance If you are using the correlation matrix, recall that the total variance is equal to the number of variables, P; and the variance of the i th principal component (Var C i ) is the eigenvalue for that principal component Therefore, the percentage of the total variance explained by C i is equal to the eigenvalue for C i times 100 divided by P: % total variance explained by C i = (i th eigenvalue)(100)/p (710) where P = the number of standardized x variables If you use this method for deciding how many principal components to retain, bear in mind these words of caution from Afifi and Clark (1990, p 385): It should be explained that eigenvalues are estimated variances of the principal components and are therefore subject to large sample variations Arbitrary cutoff points should thus not be taken too seriously Ideally, the investigator should make the selection of the number of principal components on the basis of some underlying theory 75 Initial factor extraction using PCA Section 155 from Afifi and Clark (1990) gives a very clear explanation of how initial factor extraction may be carried out using principal components analysis They use a hypothetical set of data with 5 variables (X 1, X 2, X 3, X 4, X 5 ), and N = 100 observations The correlation matrix for these data is shown in Table 75

11 B Weaver (18-Oct-2001) Factor analysis 11 Table 75 Correlation matrix for 100 hypothetical data points X 1 X 2 X 3 X 4 X 5 X X X X X Afifi and Clark (1990) note that there exists a high correlation between X 4 and X 5, a moderately high correlation between X 1 and X 2, and a moderate correlation between X 3 and each of X 4 and X 5 Bear this in mind when we get to the final solution later The following shaded block is taken directly (with some minor omissions) from Section 155 of Afifi and Clark (1990) The basic idea is to choose the first m principal components and modify them to fit the factor model defined [earlier] The reason for choosing the first m principal components, rather than any others, is that they explain the greatest proportion of variance and are therefore the most important Note that the principal components are also uncorrelated and thus present an attractive choice as factors To satisfy the assumption of unit variances of the factors, we divide each principal component by its standard deviation That is, we define the jth common factor F j as F j = C j /(Var C j ) 1/2, where C j is the jth principal component [Note that (Var C j ) 1/2 = the square root of (Var C j ), or the standard deviation of C j ] To express each variable x i in terms of the F j s, we first recall the relationship between the variables x i and the principal components C j Specifically, C 1 = a 11 x 1 + a 12 x a 1P x P C 2 = a 21 x 1 + a 22 x a 2P x P C P = a P1 x 1 + a P2 x a PP x P It may be shown mathematically that this set of equations can be inverted to express the x i s as functions of the C j s The result is (see Harmon 1976 or Afifi and Azen 1979): x 1 = a 11 C 1 + a 21 C a P1 C P x 2 = a 12 C 1 + a 22 C a P2 x P x P = a 1P C 1 + a 2P C a PP x P

12 B Weaver (18-Oct-2001) Factor analysis 12 Note that the rows of the first set of equations become the columns of the second set of equations Now since F j = C j /(Var C j ) 1/2, it follows that C j = F j (Var C j ) 1/2, and we can then express the ith equation as x i = a 1i F 1 (Var C 1 ) 1/2 + a 2i F 2 (Var C 2 ) 1/2 + + a Pi F P (Var C P ) 1/2 This last equation is now modified in two ways: 1 We use the notation l ij = a ji (Var C j ) 1/2 for the first m components 2 We combine the last P-m terms and denote the result by e i With these manipulations we have now expressed each variable x i as x i = l i1 F 1 + l i2 F l im F m + e i for i = 1,2,, P In other words, we have transformed the principal components model to produce the factor model For later use, note also that when the original variables are standardized, the factor loading l ij turns out to be the correlation between x i and F j (see Section 145) The matrix of factor loadings is sometimes called the pattern matrix When the factor loadings are correlations between the x i s and F j s as they are here, it is also called the factor structure matrix (see Dillon and Goldstein 1984) Furthermore, it can be shown mathematically that the communality of x i is h i 2 = l i1 2 + l i l im 2 For example, in our hypothetical data set the eigenvalues of the correlation matrix (or Var C i ) are Var C 1 = 2578 Var C 2 = 1567 Var C 3 = 0571 Var C 4 = 0241 Var C 5 = total = 5000 Note that the sum 50 is equal to P, the total number of variables Based on the rule of thumb of selecting only those principal components corresponding to eigenvalues of one or more, we select m = 2 Using BMDP4M [a stats package], we obtain the principal components analysis factor loadings, the l ij s shown in Table 153 For example, the loading of x 1 on F 1 is l 11 = 0511 and on F 2 is l 12 = 0782 Thus the first equation of the factor model is x 1 = 0511F F 2 + e 1 Table 153 also shows the variance explained by each factor For example, the variance explained by F 1 is 2578, or 516% of the total variance of 5

13 B Weaver (18-Oct-2001) Factor analysis 13 TABLE 153 Initial Factor Analysis Summary for Hypothetical Data Set from Principal Components Extraction Method Variable Factor F 1 Loadings F 2 Communality 2 h i Specificity 2 u i x x x x x Variance Explained h 2 i = 4145 u 2 i = 0855 Percentage 516% 313% 829% 171% The communality column, h i 2 in Table 153 shows the part of the variance of each variable explained by the common factors For example, h 1 2 = = 0873 Finally, the specificity u i 2 is the part of the variance [of x i ] not explained by the common factors In this example we use standardized variables x i whose variances are equal to one Therefore for each x i, u i 2 = 1 - h i 2 A valuable graphical aid to interpreting the factors is the factor diagram shown in Figure 151 [see below] Unlike the usual scatter diagrams where each point represents an individual case and each axis a variable, in this graph each point represents a response variable [x i ] and each axis a common factor For example, the point labeled 1 represents the response variable x 1 The coordinates of that point are the loadings of x 1 on F 1 and F 2, respectively The other four points represent the remaining four variables Since the factor loadings are correlations between the standardized variables and the factors, the range of values shown on the axes of the factor diagram is -1 to +1 It can be seen from Figure 151 that x 1 and x 2 load more on F 2 than they do on F 1 Conversely, x 3, x 4, and x 5 load more on F 1 than on F 2 However, these distinctions are not very clear-cut, and the technique of factor rotation will produce clearer results (see Section 157) [emphasis added]

14 B Weaver (18-Oct-2001) Factor analysis Factor Factor 1 10 Afifi & Clark Figure 151 Factor Diagram for Principal Compenents Extraction Method 76 Rotation of the extracted factors The factors obtained via initial factor extraction are almost always difficult to interpret, no matter which method of initial factor extraction one uses Take the data in Figure 151, for example It is clear that variables 1 and 2 are more highly correlated with Factor 2 than with Factor 1 (Another way of saying this is that they load more heavily on Factor 2 than they do on Factor 1) Note however that the difference between these correlations (or factor loadings) is not that great For variable 1, the two factor loadings are 0511 and 0782; and for variable 2, 0553 and 0754 Similarly, it is clear that variables 3, 4, and 5 load more heavily on Factor 1 than on Factor 2 Again though, the difference in factor loadings for the two factors is not very big for any of these three variables (see Table 153 from Afifi and Clark) Interpretation would be much easier if each of the variables had a high loading on only one of the two factors, and a very low loading on the other factor Rotation of the factors is carried out in an effort to bring about exactly this situation, which is often referred to as simple structure There are many factor rotation techniques to choose from As I suggested earlier, I have neither the time nor the expertise to discuss all of these I must, however, draw your attention to the important distinction between orthogonal and oblique rotations Figure 153 (from Afifi & Clark, 1990) shows one orthogonal rotation solution (the Varimax Orthogonal Rotation solution) for the data shown in Figure 151; and Figure 155 shows one oblique rotation solution (the Direct Quartimin Oblique Rotation solution) for the same data Before going on to the differences between these solutions, note the similarities In both cases, variables 1 and 2 have very high loadings on Factor 2, and very low loadings

15 B Weaver (18-Oct-2001) Factor analysis 15 on Factor 1; and variables 3, 4, and 5 have very high loadings on Factor 1, and very low loadings on Factor 2 Therefore, both solutions have achieved the goal of factor rotation, and made interpretation of the factors easier For our purposes, the important difference between orthogonal and oblique rotation techniques is that the angle between the rotated factors must be 90 in the former case, but not in the latter (In geometry, two lines that intersect are said to be oblique if the angle of intersection is not a right angle) You may recall that orthogonal means independent, or uncorrelated What you may not realize is that when two Factors are plotted on X-Y axes, the correlation between them is equal to the cosine of the angle between their axes Note that cos 90 = 0 Thus, when you use an orthogonal rotation technique, the rotated factors, like the initially extracted factors, are all mutually independent When you use an oblique rotation technique, on the other hand, you relax this restriction, and allow the factors to be correlated with each other For the oblique solution shown in Figure 155, for example, the correlation between Factors 1 and 2 is equal to cos 805 = 0165 Both types of solutions have certain advantages You may recall from earlier parts of the course that orthogonal variance components add up neatly to a total, and that this is not so when the components are not orthogonal On the other hand, oblique solutions offer more flexibility, which can be desirable in certain situations Factor Rotated Factor Factor 1 Rotated Factor 1 10 Afifi & Clark Figure 153 Varimax Orthogonal Rotation for Factor Diagram of Figure 151: Rotated axes go through clusters of variables and are perpendicular to each other

16 B Weaver (18-Oct-2001) Factor analysis Factor Rotated Factor Factor 1 Rotated Factor 1 10 Afifi & Clark Figure 155 Direct Quartimin Oblique Rotation for Factor Diagram of Figure 151: Rotated axes go through clusters of variables but are not required to be perpendicular to each other 77 Principal Factor Analysis Earlier I mentioned that there are two main methods of initial factor extraction, principal components analysis (PCA) and principal factor analysis (PFA) PCA has already been explained PFA is similar to PCA, but the solution is iterative In fact, Afifi and Clark (1990) call PFA the iterated principal components method They go on to explain how it works as follows: To understand this method [PFA], you should recall that the communality is the part of the variance of each variable associated with the common factors The principle underlying the iterated solution states that we should perform the factor analysis by using the communalities in place of the original variance This principle entails substituting communality estimates for the 1 s representing the variances of the standardized variables along the diagonal of the correlation matrix With 1 s in the diagonal we are factoring the total variance of the variables; with communalities in the diagonal we are factoring the variance associated with the common factors [emphasis added] [Conversely, we are excluding that part of the variance that cannot be explained by the common factors] Thus with communalities along the diagonal we select those common factors that maximize the total communality Many factor analysts consider maximizing the total communality a more attractive objective than maximizing the total proportion of explained variance, as is done in the principal components method The problem is that communalities are not known before the factor analysis is performed [emphasis added] Some initial estimates of the

17 B Weaver (18-Oct-2001) Factor analysis 17 communalities must be obtained prior to the analysis Various procedures exist, and we recommend, in the absence of a priori estimates, that the investigator use the default option in the particular program since the resulting factor solution is usually little affected by the initial communality estimates The steps performed by a packaged program in carrying out the iterated factor extraction are summarized as follows: 1 Find the initial communality estimates 2 Substitute the communalities for the diagonal elements (1 s) in the correlation matrix 3 Extract m principal components from the modified matrix 4 Multiply the principal components coefficients by the standard deviation of the respective principal components to obtain factor loadings 5 Compute new communalities from the computed factor loadings 6 Replace the communalities in step 2 with these new communalities and repeat steps 3, 4, and 5 This step constitutes an iteration 7 Continue iterating, stopping when the communalities stay essentially the same in the last two iterations The initial factor extraction using PFA is summarized in Table 154 (from Afifi and Clark) If you compare these results to those obtained using PCA (see Table 153), you will note that the total communality is lower for this solution (746% versus 829%) According to Afifi and Clark (p 406), This result is generally the case since in the iterative method we are factoring the total communality, which is by necessity smaller than the total variance TABLE 154 Initial Factor Analysis Summary for Hypothetical Data Set from Iterated Principal Factor Extraction Method Variable Factor F 1 Loadings F 2 Communality 2 h i Specificity 2 u i x x x x x Variance Explained h 2 i = 3730 u 2 i = 1270 Percentage 483% 263% 746% 254% Finally, note that the final rotated solutions using the (orthogonal) varimax procedure are also very similar no matter whether initial factor extraction is carried out with principal components analysis (Figure 153) or principal factor analysis (Figure 154)

18 B Weaver (18-Oct-2001) Factor analysis Factor Rotated Factor Factor 1 Rotated Factor 1 10 Afifi & Clark Figure 153 Varimax Orthogonal Rotation for Factor Diagram of Figure How many factors should be rotated? One of the questions that faces any factor analysts is how many factors ought to be rotated Under ideal circumstances, the number of factors is dictated by theory When this is not so, one must use the data themselves to determine the number of factors Unfortunately, there are no hard and fast rules to apply in attempting to determine this number, just as there are no hard and fast rules for deciding how many principal components to retain for further analysis in PCA (see Section 744) As in that case, there are only guidelines and rules of thumb One commonly used guideline is to use only those factors whose eigenvalues are equal to or greater than 1 According to Afifi and Clark (1990), this method yields roughly one factor for every three to five variables They go on to say: It appears to correctly estimate the number of factors when the communalities are high and the number of variables is not too large If the communalities are low, then the number of factors obtained with a cutoff equal to the average communality tends to be similar to that obtained by the principals components method with a cutoff of 1 According to Cattell (1978), however, this method severely overestimates the number of factors in large matrices An alternative, and fairly common method for determining the number of factors to rotate is Cattell s Scree method Kline (1994) certainly endorses the Scree method According to him, There is now agreement among most factor analysts of any repute that Cattell s Scree test is just about the best solution to selecting the correct number of factors

19 B Weaver (18-Oct-2001) Factor analysis 19 (Does this means that anyone who dares to disagree with Kline is a disreputable factor analyst?!) To carry out the Scree test, one must make a Scree plot You can do this by plotting the principal component numbers (on the X axis) against their eigenvalues (on the Y axis) Figure 54 from Kline (1994) shows a typical Scree plot (see next page) Kline says that the cutoff point for factor rotation is where the line changes slope ; and he suggests that in this case, 5 factors would be rotated Note that if you were to apply the first rule of thumb, and take those principal components with eigenvalues equal to or greater than 1, you would rotate 7, or maybe even 8 factors In this example therefore, Cattell s warning about overestimation of the number of factors appears to be borne out 7 Figure 54 (from Kline, 1994) A Scree test 6 Eigenvalues Slope Change Factors 79 Criticisms of Factor Analysis Factor analysis has been criticized for several reasons One criticism is that it lacks rigor, because many of the decisions that are made along the way (eg, how many factors to rotate, whether to use orthogonal or oblique rotation, etc) are completely arbitrary Perhaps an even more serious criticism that there are an infinite number of ways to rotate the initially extracted factors--and that all of these solutions are equivalent from a purely mathematical point of view For this reason, the results of factor analytic studies have been very difficult to replicate In response to these kinds of criticisms, Kline (1994, p 182) says this: As was argued [earlier], rotation to simple structure usually yields replicable factors and the technical rules for reaching simple structure have been explicated [emphasis added] [A matrix has simple structure when each variable has moderate to high loadings on only one or two factors, and very low loadings on the other factors] In exploratory factor analysis simple structure and factor replicability is the answer to the problem of the indeterminacy of factor analysis An infinity of solutions there may be, but the simple structure solution is best [emphasis added] Certainly, as Cattell

20 B Weaver (18-Oct-2001) Factor analysis 20 (1978) has demonstrated, failure to reach simple structure is the main cause of discrepant results in the field of personality testing Thus it is essential to reach simple structure, and this can easily be done by following the technical rules for adequate factor analyses 710 Miscellaneous points that didn t fit anywhere else! As we have seen, factor analysis and principal components analysis are closely related techniques Note however that both of these techniques differ from regression analysis in that we do not have a dependent variable to be explained by a set of independent variables (Afifi & Clark, 1990, p 396) Rather, principal components analysis is concerned with explaining as much of the total variance as possible; and factor analysis is concerned primarily with explaining the interrelationships amongst the original variables Finally, I should point out the possibility of computing factor scores for individuals For example, if the two factors are labeled Verbal and Quantitative, you may want to have two scores for each individual that reflect their scores on these factors For more information on how to go about doing this, see Section 158 in Afifi and Clark (1990)

21 B Weaver (18-Oct-2001) Factor analysis 21 Review Questions 1 What are the two main goals of factor analysis? 2 What are common factors? 3 What are unique factors? 4 Define communality and specificity 5 What is a principal component? 6 What is an eigenvalue? 7 There are two steps in factor analysis What are they? 8 What is a Scree plot? How is one used to choose the number of factors to rotate? 9 What is the difference between orthogonal and oblique rotation? 10 One of the major criticisms of factor analysis is that there are an infinite number of solutions to any problem, all of which are equally valid from a mathematical point of view A defense against this criticism is that some solutions are better (or simpler) than others Explain what is meant by a good solution [HINT: Think about simple structure]

Principal Component Analysis & Factor Analysis. Psych 818 DeShon

Principal Component Analysis & Factor Analysis. Psych 818 DeShon Principal Component Analysis & Factor Analysis Psych 818 DeShon Purpose Both are used to reduce the dimensionality of correlated measurements Can be used in a purely exploratory fashion to investigate

More information

Principle Components Analysis (PCA) Relationship Between a Linear Combination of Variables and Axes Rotation for PCA

Principle Components Analysis (PCA) Relationship Between a Linear Combination of Variables and Axes Rotation for PCA Principle Components Analysis (PCA) Relationship Between a Linear Combination of Variables and Axes Rotation for PCA Principle Components Analysis: Uses one group of variables (we will call this X) In

More information

UCLA STAT 233 Statistical Methods in Biomedical Imaging

UCLA STAT 233 Statistical Methods in Biomedical Imaging UCLA STAT 233 Statistical Methods in Biomedical Imaging Instructor: Ivo Dinov, Asst. Prof. In Statistics and Neurology University of California, Los Angeles, Spring 2004 http://www.stat.ucla.edu/~dinov/

More information

Exploratory Factor Analysis and Principal Component Analysis

Exploratory Factor Analysis and Principal Component Analysis Exploratory Factor Analysis and Principal Component Analysis Today s Topics: What are EFA and PCA for? Planning a factor analytic study Analysis steps: Extraction methods How many factors Rotation and

More information

Exploratory Factor Analysis and Principal Component Analysis

Exploratory Factor Analysis and Principal Component Analysis Exploratory Factor Analysis and Principal Component Analysis Today s Topics: What are EFA and PCA for? Planning a factor analytic study Analysis steps: Extraction methods How many factors Rotation and

More information

LECTURE 4 PRINCIPAL COMPONENTS ANALYSIS / EXPLORATORY FACTOR ANALYSIS

LECTURE 4 PRINCIPAL COMPONENTS ANALYSIS / EXPLORATORY FACTOR ANALYSIS LECTURE 4 PRINCIPAL COMPONENTS ANALYSIS / EXPLORATORY FACTOR ANALYSIS NOTES FROM PRE- LECTURE RECORDING ON PCA PCA and EFA have similar goals. They are substantially different in important ways. The goal

More information

2/26/2017. This is similar to canonical correlation in some ways. PSY 512: Advanced Statistics for Psychological and Behavioral Research 2

2/26/2017. This is similar to canonical correlation in some ways. PSY 512: Advanced Statistics for Psychological and Behavioral Research 2 PSY 512: Advanced Statistics for Psychological and Behavioral Research 2 What is factor analysis? What are factors? Representing factors Graphs and equations Extracting factors Methods and criteria Interpreting

More information

1 A factor can be considered to be an underlying latent variable: (a) on which people differ. (b) that is explained by unknown variables

1 A factor can be considered to be an underlying latent variable: (a) on which people differ. (b) that is explained by unknown variables 1 A factor can be considered to be an underlying latent variable: (a) on which people differ (b) that is explained by unknown variables (c) that cannot be defined (d) that is influenced by observed variables

More information

VAR2 VAR3 VAR4 VAR5. Or, in terms of basic measurement theory, we could model it as:

VAR2 VAR3 VAR4 VAR5. Or, in terms of basic measurement theory, we could model it as: 1 Neuendorf Factor Analysis Assumptions: 1. Metric (interval/ratio) data 2. Linearity (in the relationships among the variables) -Factors are linear constructions of the set of variables (see #8 under

More information

STAT 730 Chapter 9: Factor analysis

STAT 730 Chapter 9: Factor analysis STAT 730 Chapter 9: Factor analysis Timothy Hanson Department of Statistics, University of South Carolina Stat 730: Multivariate Data Analysis 1 / 15 Basic idea Factor analysis attempts to explain the

More information

Multivariate Fundamentals: Rotation. Exploratory Factor Analysis

Multivariate Fundamentals: Rotation. Exploratory Factor Analysis Multivariate Fundamentals: Rotation Exploratory Factor Analysis PCA Analysis A Review Precipitation Temperature Ecosystems PCA Analysis with Spatial Data Proportion of variance explained Comp.1 + Comp.2

More information

Introduction to Factor Analysis

Introduction to Factor Analysis to Factor Analysis Lecture 10 August 2, 2011 Advanced Multivariate Statistical Methods ICPSR Summer Session #2 Lecture #10-8/3/2011 Slide 1 of 55 Today s Lecture Factor Analysis Today s Lecture Exploratory

More information

FACTOR ANALYSIS AND MULTIDIMENSIONAL SCALING

FACTOR ANALYSIS AND MULTIDIMENSIONAL SCALING FACTOR ANALYSIS AND MULTIDIMENSIONAL SCALING Vishwanath Mantha Department for Electrical and Computer Engineering Mississippi State University, Mississippi State, MS 39762 mantha@isip.msstate.edu ABSTRACT

More information

Introduction to Machine Learning

Introduction to Machine Learning 10-701 Introduction to Machine Learning PCA Slides based on 18-661 Fall 2018 PCA Raw data can be Complex, High-dimensional To understand a phenomenon we measure various related quantities If we knew what

More information

Or, in terms of basic measurement theory, we could model it as:

Or, in terms of basic measurement theory, we could model it as: 1 Neuendorf Factor Analysis Assumptions: 1. Metric (interval/ratio) data 2. Linearity (in relationships among the variables--factors are linear constructions of the set of variables; the critical source

More information

Applied Multivariate Analysis

Applied Multivariate Analysis Department of Mathematics and Statistics, University of Vaasa, Finland Spring 2017 Dimension reduction Exploratory (EFA) Background While the motivation in PCA is to replace the original (correlated) variables

More information

Canonical Correlation & Principle Components Analysis

Canonical Correlation & Principle Components Analysis Canonical Correlation & Principle Components Analysis Aaron French Canonical Correlation Canonical Correlation is used to analyze correlation between two sets of variables when there is one set of IVs

More information

Quantitative Understanding in Biology Principal Components Analysis

Quantitative Understanding in Biology Principal Components Analysis Quantitative Understanding in Biology Principal Components Analysis Introduction Throughout this course we have seen examples of complex mathematical phenomena being represented as linear combinations

More information

7. Variable extraction and dimensionality reduction

7. Variable extraction and dimensionality reduction 7. Variable extraction and dimensionality reduction The goal of the variable selection in the preceding chapter was to find least useful variables so that it would be possible to reduce the dimensionality

More information

[Disclaimer: This is not a complete list of everything you need to know, just some of the topics that gave people difficulty.]

[Disclaimer: This is not a complete list of everything you need to know, just some of the topics that gave people difficulty.] Math 43 Review Notes [Disclaimer: This is not a complete list of everything you need to know, just some of the topics that gave people difficulty Dot Product If v (v, v, v 3 and w (w, w, w 3, then the

More information

PRINCIPAL COMPONENTS ANALYSIS (PCA)

PRINCIPAL COMPONENTS ANALYSIS (PCA) PRINCIPAL COMPONENTS ANALYSIS (PCA) Introduction PCA is considered an exploratory technique that can be used to gain a better understanding of the interrelationships between variables. PCA is performed

More information

Introduction to Factor Analysis

Introduction to Factor Analysis to Factor Analysis Lecture 11 November 2, 2005 Multivariate Analysis Lecture #11-11/2/2005 Slide 1 of 58 Today s Lecture Factor Analysis. Today s Lecture Exploratory factor analysis (EFA). Confirmatory

More information

ROBERTO BATTITI, MAURO BRUNATO. The LION Way: Machine Learning plus Intelligent Optimization. LIONlab, University of Trento, Italy, Apr 2015

ROBERTO BATTITI, MAURO BRUNATO. The LION Way: Machine Learning plus Intelligent Optimization. LIONlab, University of Trento, Italy, Apr 2015 ROBERTO BATTITI, MAURO BRUNATO. The LION Way: Machine Learning plus Intelligent Optimization. LIONlab, University of Trento, Italy, Apr 2015 http://intelligentoptimization.org/lionbook Roberto Battiti

More information

This appendix provides a very basic introduction to linear algebra concepts.

This appendix provides a very basic introduction to linear algebra concepts. APPENDIX Basic Linear Algebra Concepts This appendix provides a very basic introduction to linear algebra concepts. Some of these concepts are intentionally presented here in a somewhat simplified (not

More information

Principal Component Analysis (PCA) Theory, Practice, and Examples

Principal Component Analysis (PCA) Theory, Practice, and Examples Principal Component Analysis (PCA) Theory, Practice, and Examples Data Reduction summarization of data with many (p) variables by a smaller set of (k) derived (synthetic, composite) variables. p k n A

More information

Principal component analysis

Principal component analysis Principal component analysis Angela Montanari 1 Introduction Principal component analysis (PCA) is one of the most popular multivariate statistical methods. It was first introduced by Pearson (1901) and

More information

Structure in Data. A major objective in data analysis is to identify interesting features or structure in the data.

Structure in Data. A major objective in data analysis is to identify interesting features or structure in the data. Structure in Data A major objective in data analysis is to identify interesting features or structure in the data. The graphical methods are very useful in discovering structure. There are basically two

More information

The 3 Indeterminacies of Common Factor Analysis

The 3 Indeterminacies of Common Factor Analysis The 3 Indeterminacies of Common Factor Analysis James H. Steiger Department of Psychology and Human Development Vanderbilt University James H. Steiger (Vanderbilt University) The 3 Indeterminacies of Common

More information

Chapter 4: Factor Analysis

Chapter 4: Factor Analysis Chapter 4: Factor Analysis In many studies, we may not be able to measure directly the variables of interest. We can merely collect data on other variables which may be related to the variables of interest.

More information

I L L I N O I S UNIVERSITY OF ILLINOIS AT URBANA-CHAMPAIGN

I L L I N O I S UNIVERSITY OF ILLINOIS AT URBANA-CHAMPAIGN Introduction Edps/Psych/Stat/ 584 Applied Multivariate Statistics Carolyn J Anderson Department of Educational Psychology I L L I N O I S UNIVERSITY OF ILLINOIS AT URBANA-CHAMPAIGN c Board of Trustees,

More information

Basics of Multivariate Modelling and Data Analysis

Basics of Multivariate Modelling and Data Analysis Basics of Multivariate Modelling and Data Analysis Kurt-Erik Häggblom 6. Principal component analysis (PCA) 6.1 Overview 6.2 Essentials of PCA 6.3 Numerical calculation of PCs 6.4 Effects of data preprocessing

More information

Table of Contents. Multivariate methods. Introduction II. Introduction I

Table of Contents. Multivariate methods. Introduction II. Introduction I Table of Contents Introduction Antti Penttilä Department of Physics University of Helsinki Exactum summer school, 04 Construction of multinormal distribution Test of multinormality with 3 Interpretation

More information

Intermediate Social Statistics

Intermediate Social Statistics Intermediate Social Statistics Lecture 5. Factor Analysis Tom A.B. Snijders University of Oxford January, 2008 c Tom A.B. Snijders (University of Oxford) Intermediate Social Statistics January, 2008 1

More information

Factor Analysis (10/2/13)

Factor Analysis (10/2/13) STA561: Probabilistic machine learning Factor Analysis (10/2/13) Lecturer: Barbara Engelhardt Scribes: Li Zhu, Fan Li, Ni Guan Factor Analysis Factor analysis is related to the mixture models we have studied.

More information

Machine Learning 2nd Edition

Machine Learning 2nd Edition INTRODUCTION TO Lecture Slides for Machine Learning 2nd Edition ETHEM ALPAYDIN, modified by Leonardo Bobadilla and some parts from http://www.cs.tau.ac.il/~apartzin/machinelearning/ The MIT Press, 2010

More information

Factor analysis. George Balabanis

Factor analysis. George Balabanis Factor analysis George Balabanis Key Concepts and Terms Deviation. A deviation is a value minus its mean: x - mean x Variance is a measure of how spread out a distribution is. It is computed as the average

More information

PRINCIPAL COMPONENT ANALYSIS

PRINCIPAL COMPONENT ANALYSIS PRINCIPAL COMPONENT ANALYSIS 1 INTRODUCTION One of the main problems inherent in statistics with more than two variables is the issue of visualising or interpreting data. Fortunately, quite often the problem

More information

7 Principal Components and Factor Analysis

7 Principal Components and Factor Analysis 7 Principal Components and actor nalysis 7.1 Principal Components a oal. Relationships between two variables can be graphically well captured in a meaningful way. or three variables this is also possible,

More information

Unconstrained Ordination

Unconstrained Ordination Unconstrained Ordination Sites Species A Species B Species C Species D Species E 1 0 (1) 5 (1) 1 (1) 10 (4) 10 (4) 2 2 (3) 8 (3) 4 (3) 12 (6) 20 (6) 3 8 (6) 20 (6) 10 (6) 1 (2) 3 (2) 4 4 (5) 11 (5) 8 (5)

More information

Multivariate Statistics (I) 2. Principal Component Analysis (PCA)

Multivariate Statistics (I) 2. Principal Component Analysis (PCA) Multivariate Statistics (I) 2. Principal Component Analysis (PCA) 2.1 Comprehension of PCA 2.2 Concepts of PCs 2.3 Algebraic derivation of PCs 2.4 Selection and goodness-of-fit of PCs 2.5 Algebraic derivation

More information

June If you want, you may scan your assignment and convert it to a.pdf file and it to me.

June If you want, you may scan your assignment and convert it to a.pdf file and  it to me. Summer Assignment Pre-Calculus Honors June 2016 Dear Student: This assignment is a mandatory part of the Pre-Calculus Honors course. Students who do not complete the assignment will be placed in the regular

More information

Factor Analysis. Robert L. Wolpert Department of Statistical Science Duke University, Durham, NC, USA

Factor Analysis. Robert L. Wolpert Department of Statistical Science Duke University, Durham, NC, USA Factor Analysis Robert L. Wolpert Department of Statistical Science Duke University, Durham, NC, USA 1 Factor Models The multivariate regression model Y = XB +U expresses each row Y i R p as a linear combination

More information

Quantitative Understanding in Biology Short Course Session 9 Principal Components Analysis

Quantitative Understanding in Biology Short Course Session 9 Principal Components Analysis Quantitative Understanding in Biology Short Course Session 9 Principal Components Analysis Jinhyun Ju Jason Banfelder Luce Skrabanek June 21st, 218 1 Preface For the last session in this course, we ll

More information

Chapter 11. Correlation and Regression

Chapter 11. Correlation and Regression Chapter 11. Correlation and Regression The word correlation is used in everyday life to denote some form of association. We might say that we have noticed a correlation between foggy days and attacks of

More information

1 Measurement Uncertainties

1 Measurement Uncertainties 1 Measurement Uncertainties (Adapted stolen, really from work by Amin Jaziri) 1.1 Introduction No measurement can be perfectly certain. No measuring device is infinitely sensitive or infinitely precise.

More information

Neuroscience Introduction

Neuroscience Introduction Neuroscience Introduction The brain As humans, we can identify galaxies light years away, we can study particles smaller than an atom. But we still haven t unlocked the mystery of the three pounds of matter

More information

CHAPTER 3 THE COMMON FACTOR MODEL IN THE POPULATION. From Exploratory Factor Analysis Ledyard R Tucker and Robert C. MacCallum

CHAPTER 3 THE COMMON FACTOR MODEL IN THE POPULATION. From Exploratory Factor Analysis Ledyard R Tucker and Robert C. MacCallum CHAPTER 3 THE COMMON FACTOR MODEL IN THE POPULATION From Exploratory Factor Analysis Ledyard R Tucker and Robert C. MacCallum 1997 19 CHAPTER 3 THE COMMON FACTOR MODEL IN THE POPULATION 3.0. Introduction

More information

GEOG 4110/5100 Advanced Remote Sensing Lecture 15

GEOG 4110/5100 Advanced Remote Sensing Lecture 15 GEOG 4110/5100 Advanced Remote Sensing Lecture 15 Principal Component Analysis Relevant reading: Richards. Chapters 6.3* http://www.ce.yildiz.edu.tr/personal/songul/file/1097/principal_components.pdf *For

More information

College Algebra Through Problem Solving (2018 Edition)

College Algebra Through Problem Solving (2018 Edition) City University of New York (CUNY) CUNY Academic Works Open Educational Resources Queensborough Community College Winter 1-25-2018 College Algebra Through Problem Solving (2018 Edition) Danielle Cifone

More information

ALGEBRA 1 CURRICULUM COMMON CORE BASED

ALGEBRA 1 CURRICULUM COMMON CORE BASED ALGEBRA 1 CURRICULUM COMMON CORE BASED (Supplemented with 8th grade PSSA anchors ) UPPER MERION AREA SCHOOL DISTRICT 435 CROSSFIELD ROAD KING OF PRUSSIA, PA 19406 8/20/2012 PA COMMON CORE ALIGNED MATHEMATICS

More information

PCA, Kernel PCA, ICA

PCA, Kernel PCA, ICA PCA, Kernel PCA, ICA Learning Representations. Dimensionality Reduction. Maria-Florina Balcan 04/08/2015 Big & High-Dimensional Data High-Dimensions = Lot of Features Document classification Features per

More information

Data Preprocessing Tasks

Data Preprocessing Tasks Data Tasks 1 2 3 Data Reduction 4 We re here. 1 Dimensionality Reduction Dimensionality reduction is a commonly used approach for generating fewer features. Typically used because too many features can

More information

Machine Learning (Spring 2012) Principal Component Analysis

Machine Learning (Spring 2012) Principal Component Analysis 1-71 Machine Learning (Spring 1) Principal Component Analysis Yang Xu This note is partly based on Chapter 1.1 in Chris Bishop s book on PRML and the lecture slides on PCA written by Carlos Guestrin in

More information

PRINCIPAL COMPONENTS ANALYSIS

PRINCIPAL COMPONENTS ANALYSIS 121 CHAPTER 11 PRINCIPAL COMPONENTS ANALYSIS We now have the tools necessary to discuss one of the most important concepts in mathematical statistics: Principal Components Analysis (PCA). PCA involves

More information

Math Precalculus I University of Hawai i at Mānoa Spring

Math Precalculus I University of Hawai i at Mānoa Spring Math 135 - Precalculus I University of Hawai i at Mānoa Spring - 2014 Created for Math 135, Spring 2008 by Lukasz Grabarek and Michael Joyce Send comments and corrections to lukasz@math.hawaii.edu Contents

More information

CS168: The Modern Algorithmic Toolbox Lecture #7: Understanding Principal Component Analysis (PCA)

CS168: The Modern Algorithmic Toolbox Lecture #7: Understanding Principal Component Analysis (PCA) CS68: The Modern Algorithmic Toolbox Lecture #7: Understanding Principal Component Analysis (PCA) Tim Roughgarden & Gregory Valiant April 0, 05 Introduction. Lecture Goal Principal components analysis

More information

The Principal Component Analysis

The Principal Component Analysis The Principal Component Analysis Philippe B. Laval KSU Fall 2017 Philippe B. Laval (KSU) PCA Fall 2017 1 / 27 Introduction Every 80 minutes, the two Landsat satellites go around the world, recording images

More information

PY 351 Modern Physics Short assignment 4, Nov. 9, 2018, to be returned in class on Nov. 15.

PY 351 Modern Physics Short assignment 4, Nov. 9, 2018, to be returned in class on Nov. 15. PY 351 Modern Physics Short assignment 4, Nov. 9, 2018, to be returned in class on Nov. 15. You may write your answers on this sheet or on a separate paper. Remember to write your name on top. Please note:

More information

Writing proofs for MATH 61CM, 61DM Week 1: basic logic, proof by contradiction, proof by induction

Writing proofs for MATH 61CM, 61DM Week 1: basic logic, proof by contradiction, proof by induction Writing proofs for MATH 61CM, 61DM Week 1: basic logic, proof by contradiction, proof by induction written by Sarah Peluse, revised by Evangelie Zachos and Lisa Sauermann September 27, 2016 1 Introduction

More information

Factor Analysis. -Applied Multivariate Analysis- Lecturer: Darren Homrighausen, PhD

Factor Analysis. -Applied Multivariate Analysis- Lecturer: Darren Homrighausen, PhD Factor Analysis -Applied Multivariate Analysis- Lecturer: Darren Homrighausen, PhD 1 From PCA to factor analysis Remember: PCA tries to estimate a transformation of the data such that: 1. The maximum amount

More information

A Tutorial on Data Reduction. Principal Component Analysis Theoretical Discussion. By Shireen Elhabian and Aly Farag

A Tutorial on Data Reduction. Principal Component Analysis Theoretical Discussion. By Shireen Elhabian and Aly Farag A Tutorial on Data Reduction Principal Component Analysis Theoretical Discussion By Shireen Elhabian and Aly Farag University of Louisville, CVIP Lab November 2008 PCA PCA is A backbone of modern data

More information

B. Weaver (24-Mar-2005) Multiple Regression Chapter 5: Multiple Regression Y ) (5.1) Deviation score = (Y i

B. Weaver (24-Mar-2005) Multiple Regression Chapter 5: Multiple Regression Y ) (5.1) Deviation score = (Y i B. Weaver (24-Mar-2005) Multiple Regression... 1 Chapter 5: Multiple Regression 5.1 Partial and semi-partial correlation Before starting on multiple regression per se, we need to consider the concepts

More information

FACULTY OF APPLIED ECONOMICS

FACULTY OF APPLIED ECONOMICS FACULTY OF APPLIED ECONOMICS DEPARTMENT OF MATHEMATICS, STATISTICS AND ACTUARIAL SCIENCES French Regional Wheat Prices : 1756-1870 An Illustrated Statistical Analysis E. Borghers RESEARCH PAPER 06-003

More information

Relationships between variables. Visualizing Bivariate Distributions: Scatter Plots

Relationships between variables. Visualizing Bivariate Distributions: Scatter Plots SFBS Course Notes Part 7: Correlation Bivariate relationships (p. 1) Linear transformations (p. 3) Pearson r : Measuring a relationship (p. 5) Interpretation of correlations (p. 10) Relationships between

More information

Implementing Horn s parallel analysis for principal component analysis and factor analysis

Implementing Horn s parallel analysis for principal component analysis and factor analysis The Stata Journal (2009) 9, Number 2, pp. 291 298 Implementing Horn s parallel analysis for principal component analysis and factor analysis Alexis Dinno Department of Biological Sciences California State

More information

Principal Components. Summary. Sample StatFolio: pca.sgp

Principal Components. Summary. Sample StatFolio: pca.sgp Principal Components Summary... 1 Statistical Model... 4 Analysis Summary... 5 Analysis Options... 7 Scree Plot... 8 Component Weights... 9 D and 3D Component Plots... 10 Data Table... 11 D and 3D Component

More information

Finite Mathematics Chapter 2. where a, b, c, d, h, and k are real numbers and neither a and b nor c and d are both zero.

Finite Mathematics Chapter 2. where a, b, c, d, h, and k are real numbers and neither a and b nor c and d are both zero. Finite Mathematics Chapter 2 Section 2.1 Systems of Linear Equations: An Introduction Systems of Equations Recall that a system of two linear equations in two variables may be written in the general form

More information

Regression Analysis. Ordinary Least Squares. The Linear Model

Regression Analysis. Ordinary Least Squares. The Linear Model Regression Analysis Linear regression is one of the most widely used tools in statistics. Suppose we were jobless college students interested in finding out how big (or small) our salaries would be 20

More information

Interaction effects for continuous predictors in regression modeling

Interaction effects for continuous predictors in regression modeling Interaction effects for continuous predictors in regression modeling Testing for interactions The linear regression model is undoubtedly the most commonly-used statistical model, and has the advantage

More information

CS281 Section 4: Factor Analysis and PCA

CS281 Section 4: Factor Analysis and PCA CS81 Section 4: Factor Analysis and PCA Scott Linderman At this point we have seen a variety of machine learning models, with a particular emphasis on models for supervised learning. In particular, we

More information

Introduction to Machine Learning. PCA and Spectral Clustering. Introduction to Machine Learning, Slides: Eran Halperin

Introduction to Machine Learning. PCA and Spectral Clustering. Introduction to Machine Learning, Slides: Eran Halperin 1 Introduction to Machine Learning PCA and Spectral Clustering Introduction to Machine Learning, 2013-14 Slides: Eran Halperin Singular Value Decomposition (SVD) The singular value decomposition (SVD)

More information

Linear Dimensionality Reduction

Linear Dimensionality Reduction Outline Hong Chang Institute of Computing Technology, Chinese Academy of Sciences Machine Learning Methods (Fall 2012) Outline Outline I 1 Introduction 2 Principal Component Analysis 3 Factor Analysis

More information

April 2016 Draft. DRAFT New Louisiana Standards for Correlation to Eureka Math Page 1. eureka math.org 2016 Great Minds

April 2016 Draft. DRAFT New Louisiana Standards for Correlation to Eureka Math Page 1. eureka math.org 2016 Great Minds DRAFT New Louisiana Standards for 2016 2017 Correlation to Eureka Math Grade 8 April 2016 Draft DRAFT New Louisiana Standards for 2016 2017 Correlation to Eureka Math Page 1 Grade 8 Mathematics The majority

More information

Getting Started with Communications Engineering. Rows first, columns second. Remember that. R then C. 1

Getting Started with Communications Engineering. Rows first, columns second. Remember that. R then C. 1 1 Rows first, columns second. Remember that. R then C. 1 A matrix is a set of real or complex numbers arranged in a rectangular array. They can be any size and shape (provided they are rectangular). A

More information

Pre-sessional Mathematics for Big Data MSc Class 2: Linear Algebra

Pre-sessional Mathematics for Big Data MSc Class 2: Linear Algebra Pre-sessional Mathematics for Big Data MSc Class 2: Linear Algebra Yuri Kalnishkan September 22, 2018 Linear algebra is vitally important for applied mathematics We will approach linear algebra from a

More information

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra. DS-GA 1002 Lecture notes 0 Fall 2016 Linear Algebra These notes provide a review of basic concepts in linear algebra. 1 Vector spaces You are no doubt familiar with vectors in R 2 or R 3, i.e. [ ] 1.1

More information

VAR Model. (k-variate) VAR(p) model (in the Reduced Form): Y t-2. Y t-1 = A + B 1. Y t + B 2. Y t-p. + ε t. + + B p. where:

VAR Model. (k-variate) VAR(p) model (in the Reduced Form): Y t-2. Y t-1 = A + B 1. Y t + B 2. Y t-p. + ε t. + + B p. where: VAR Model (k-variate VAR(p model (in the Reduced Form: where: Y t = A + B 1 Y t-1 + B 2 Y t-2 + + B p Y t-p + ε t Y t = (y 1t, y 2t,, y kt : a (k x 1 vector of time series variables A: a (k x 1 vector

More information

Do not copy, post, or distribute

Do not copy, post, or distribute 14 CORRELATION ANALYSIS AND LINEAR REGRESSION Assessing the Covariability of Two Quantitative Properties 14.0 LEARNING OBJECTIVES In this chapter, we discuss two related techniques for assessing a possible

More information

Linear Programming Redux

Linear Programming Redux Linear Programming Redux Jim Bremer May 12, 2008 The purpose of these notes is to review the basics of linear programming and the simplex method in a clear, concise, and comprehensive way. The book contains

More information

Factor Analysis Continued. Psy 524 Ainsworth

Factor Analysis Continued. Psy 524 Ainsworth Factor Analysis Continued Psy 524 Ainsworth Equations Extraction Principal Axis Factoring Variables Skiers Cost Lift Depth Powder S1 32 64 65 67 S2 61 37 62 65 S3 59 40 45 43 S4 36 62 34 35 S5 62 46 43

More information

REVIEW 8/2/2017 陈芳华东师大英语系

REVIEW 8/2/2017 陈芳华东师大英语系 REVIEW Hypothesis testing starts with a null hypothesis and a null distribution. We compare what we have to the null distribution, if the result is too extreme to belong to the null distribution (p

More information

Unit 8 - Polynomial and Rational Functions Classwork

Unit 8 - Polynomial and Rational Functions Classwork Unit 8 - Polynomial and Rational Functions Classwork This unit begins with a study of polynomial functions. Polynomials are in the form: f ( x) = a n x n + a n 1 x n 1 + a n 2 x n 2 +... + a 2 x 2 + a

More information

An Introduction to Path Analysis

An Introduction to Path Analysis An Introduction to Path Analysis PRE 905: Multivariate Analysis Lecture 10: April 15, 2014 PRE 905: Lecture 10 Path Analysis Today s Lecture Path analysis starting with multivariate regression then arriving

More information

Appendix A: Matrices

Appendix A: Matrices Appendix A: Matrices A matrix is a rectangular array of numbers Such arrays have rows and columns The numbers of rows and columns are referred to as the dimensions of a matrix A matrix with, say, 5 rows

More information

Principal Components Analysis using R Francis Huang / November 2, 2016

Principal Components Analysis using R Francis Huang / November 2, 2016 Principal Components Analysis using R Francis Huang / huangf@missouri.edu November 2, 2016 Principal components analysis (PCA) is a convenient way to reduce high dimensional data into a smaller number

More information

Eigenvalues, Eigenvectors, and an Intro to PCA

Eigenvalues, Eigenvectors, and an Intro to PCA Eigenvalues, Eigenvectors, and an Intro to PCA Eigenvalues, Eigenvectors, and an Intro to PCA Changing Basis We ve talked so far about re-writing our data using a new set of variables, or a new basis.

More information

Eigenvalues, Eigenvectors, and an Intro to PCA

Eigenvalues, Eigenvectors, and an Intro to PCA Eigenvalues, Eigenvectors, and an Intro to PCA Eigenvalues, Eigenvectors, and an Intro to PCA Changing Basis We ve talked so far about re-writing our data using a new set of variables, or a new basis.

More information

e 2 e 1 (a) (b) (d) (c)

e 2 e 1 (a) (b) (d) (c) 2.13 Rotated principal component analysis [Book, Sect. 2.2] Fig.: PCA applied to a dataset composed of (a) 1 cluster, (b) 2 clusters, (c) and (d) 4 clusters. In (c), an orthonormal rotation and (d) an

More information

CS 246 Review of Linear Algebra 01/17/19

CS 246 Review of Linear Algebra 01/17/19 1 Linear algebra In this section we will discuss vectors and matrices. We denote the (i, j)th entry of a matrix A as A ij, and the ith entry of a vector as v i. 1.1 Vectors and vector operations A vector

More information

Fairfield Public Schools

Fairfield Public Schools Mathematics Fairfield Public Schools Pre-Algebra 8 Pre-Algebra 8 BOE Approved 05/21/2013 1 PRE-ALGEBRA 8 Critical Areas of Focus In the Pre-Algebra 8 course, instructional time should focus on three critical

More information

ACCESS TO SCIENCE, ENGINEERING AND AGRICULTURE: MATHEMATICS 1 MATH00030 SEMESTER /2018

ACCESS TO SCIENCE, ENGINEERING AND AGRICULTURE: MATHEMATICS 1 MATH00030 SEMESTER /2018 ACCESS TO SCIENCE, ENGINEERING AND AGRICULTURE: MATHEMATICS 1 MATH00030 SEMESTER 1 2017/2018 DR. ANTHONY BROWN 1. Arithmetic and Algebra 1.1. Arithmetic of Numbers. While we have calculators and computers

More information

Karhunen-Loève Transform KLT. JanKees van der Poel D.Sc. Student, Mechanical Engineering

Karhunen-Loève Transform KLT. JanKees van der Poel D.Sc. Student, Mechanical Engineering Karhunen-Loève Transform KLT JanKees van der Poel D.Sc. Student, Mechanical Engineering Karhunen-Loève Transform Has many names cited in literature: Karhunen-Loève Transform (KLT); Karhunen-Loève Decomposition

More information

Vectors and their uses

Vectors and their uses Vectors and their uses Sharon Goldwater Institute for Language, Cognition and Computation School of Informatics, University of Edinburgh DRAFT Version 0.95: 3 Sep 2015. Do not redistribute without permission.

More information

Math 381 Midterm Practice Problem Solutions

Math 381 Midterm Practice Problem Solutions Math 381 Midterm Practice Problem Solutions Notes: -Many of the exercises below are adapted from Operations Research: Applications and Algorithms by Winston. -I have included a list of topics covered on

More information

An Introduction to Mplus and Path Analysis

An Introduction to Mplus and Path Analysis An Introduction to Mplus and Path Analysis PSYC 943: Fundamentals of Multivariate Modeling Lecture 10: October 30, 2013 PSYC 943: Lecture 10 Today s Lecture Path analysis starting with multivariate regression

More information

Principal Component Analysis

Principal Component Analysis Principal Component Analysis Laurenz Wiskott Institute for Theoretical Biology Humboldt-University Berlin Invalidenstraße 43 D-10115 Berlin, Germany 11 March 2004 1 Intuition Problem Statement Experimental

More information

Grade 7 Honors Yearlong Mathematics Map

Grade 7 Honors Yearlong Mathematics Map rade 7 Honors Yearlong Mathematics Map Resources: Approved from Board of Education Assessments: PARCC Assessments, Performance Series, District Benchmark Assessments NS NS Common Core State Standards Standards

More information

The Matrix Algebra of Sample Statistics

The Matrix Algebra of Sample Statistics The Matrix Algebra of Sample Statistics James H. Steiger Department of Psychology and Human Development Vanderbilt University James H. Steiger (Vanderbilt University) The Matrix Algebra of Sample Statistics

More information

Linear Algebra Basics

Linear Algebra Basics Linear Algebra Basics For the next chapter, understanding matrices and how to do computations with them will be crucial. So, a good first place to start is perhaps What is a matrix? A matrix A is an array

More information

9 Correlation and Regression

9 Correlation and Regression 9 Correlation and Regression SW, Chapter 12. Suppose we select n = 10 persons from the population of college seniors who plan to take the MCAT exam. Each takes the test, is coached, and then retakes the

More information