Principal Components Analysis

Size: px
Start display at page:

Download "Principal Components Analysis"

Transcription

1 Principal Components Analysis Lecture 9 August 2, 2011 Advanced Multivariate Statistical Methods ICPSR Summer Session #2 Lecture #9-8/2/2011 Slide 1 of 54

2 Today s Lecture Principal Components Analysis (PCA) Today s Lecture What it is How it works How to do such an analysis Examples of uses of PCA Lecture #9-8/2/2011 Slide 2 of 54

3 Principal components analysis is concerned with explaining the variance-covariance structure of a set of variables Variability PCA Features This explanation comes from a few linear combinations of the original variables Generally speaking, PCA has two objectives: Data reduction - moving from many original variables down to a few composite variables (linear combinations of the original variables) - which variables play a larger role in the explanation of total variance Think of the new variables, called the principal components, as composite variables consisting of a mixture of the original variables Com pos ite: A structure or an entity made up of distinct components (dictionary.com) Lecture #9-8/2/2011 Slide 3 of 54

4 Variability When talking about the joint variability of a set of variables, the concept of the total sample variance is important Variability PCA Features Essentially, the total sample variance provides a way to describe the sample covariance matrix, S, with a single number Recall from previous lectures the concept of total sample variance: Another way to characterize the sample variance is with the total variance Total variance equals tr(s) = s 11 + s s pp Like generalized variance, the total sample variance reflects the overall spread of the data Many multivariate techniques refer use total sample variance in computation of variance accounted for Lecture #9-8/2/2011 Slide 4 of 54

5 PCA and Variability In PCA, the goal is to find a set of k principal components (composite variables) that: Variability PCA Features 1. Is much smaller than the original set of p variables and 2. Accounts for nearly all of the total sample variance If these two goals can be accomplished, then the set of k principal components contains almost as much information as the original p variables This means that the k principal components can then replace the p original variables The original data set is thereby reduced, from n measurements on p variables to n measurements on k variables Lecture #9-8/2/2011 Slide 5 of 54

6 PCA Features PCA often reveals relationships between variables that were not previously suspected Variability PCA Features Because of such relationships, new interpretations of the data and variables often stem from PCA PCA usually serves more of a means to an end rather than an end in themselves in that the composite variables (the principal components) are often used in larger investigations or in other statistical techniques such as: Multiple regression Cluster analysis Scaled principal components are one factoring of the covariance matrix for the exploratory factor analytic approaches that will be discussed next class Lecture #9-8/2/2011 Slide 6 of 54

7 Linear Combinations Prinicpal Components Example #1 Principal components, the composite variables created by combinations of the original variables, are formed by a set of linear combinations For p variables, a total of p components can be formed using PCA (although much fewer are usually used) Letting Y stand for the new composite variables, or principal components, the linear combinations look like: Y 1 = a 1X = a 11 X 1 + a 12 X a 1p X p Y 2 = a 2X = a 21 X 1 + a 22 X a 2p X p. Y p = a px = a p1 X 1 + a p2 X a pp X p Lecture #9-8/2/2011 Slide 7 of 54

8 What are the Linear Combinations? Linear Combinations Prinicpal Components Example #1 In PCA, the linear combinations, each formed by weighting each original variable by a pp are formed so that the following conditions are met: The variance of each successive component is smaller than the previous component: Var(Y 1 ) > Var(Y 2 ) >... > Var(Y p ) The covariance (or correlation) between any two different principal components (i and j) is zero: Cov(Y i, Y j ) = 0 The sum of the variances of the principal components is equal to the total sample variance: p Var(Y p ) = Tr(S) = s 11 + s s pp i=1 Lecture #9-8/2/2011 Slide 8 of 54

9 What are the Linear Combinations? Principal components depend solely on the covariance matrix S or the correlation matrix R Linear Combinations Prinicpal Components Example #1 Linear combination weights in PCA aren t typically either ones or zeros (although there is one special case where that is the case) Rather, all the linear combination weights for each principal component come directly from the eigenvectors of S or R Recall that for p variables, the p p covariance/correlation matrix has a set of: p eigenvalues - {λ 1, λ 2,...,λ p } p eigenvectors - {e 1, e 2,...,e p } Lecture #9-8/2/2011 Slide 9 of 54

10 Forming the Principal Components Each principal component is formed by taking the values of the elements of the eigenvalues as the weights of the linear combination Linear Combinations Prinicpal Components Example #1 If each eigenvector has elements e ik : e 1 = e 11 e 21.., e 2 = e 12 e 22..,...,e p = e 1p e 2p.. e p1 e p2 e pp Then the principal components are formed by: Y 1 = e 11 X 1 + e 21 X e p1 X p Y 2 = e 12 X 1 + e 22 X e p2 X p. Y p = e 1p X 1 + e 2p X e pp X p Lecture #9-8/2/2011 Slide 10 of 54

11 SAS Example #1 Imagine that somehow (via the wonders of simulation) I collected data SAT test scores for both the Math (SATM) and Verbal (SATV) sections of 1,000 students Linear Combinations Prinicpal Components Example #1 The descriptive statistics of this data set are given below: Statistic SATV SATM Mean SD Correlation SATV SATM Note that this example is given for purposes of demonstration with a small number of variables In practice, this type of application of PCA with two variables is unrealistic Lecture #9-8/2/2011 Slide 11 of 54

12 SAS Example #1 The scatterplot of the data is shown below: Linear Combinations Prinicpal Components Example #1 SATM SATV Lecture #9-8/2/2011 Slide 12 of 54

13 SAS Code proc princomp cov data=sat out=pcsat; var satv satm; run; For more SAS code, see the proc princomp part of the SAS Manual at: Lecture #9-8/2/2011 Slide 13 of 54

14 13-1

15 SAS New Data Set The new SAS data set contains the two principal components: Lecture #9-8/2/2011 Slide 14 of 54

16 Principal Component Computation The way the principal components are often constructed (and are constructed in SAS) is to use mean-centered variables Linear Combinations Prinicpal Components Example #1 In the context of our example, the two principal components are formed by: Y 1 = 0.475(X 1 X 1 ) (X 2 X 2 ) Y 2 = 0.880(X 1 X 1 ) (X 2 X 2 ) For example, let s take the first individual s principle components, this person had a SATM of 580 and a SATV of 520: Y 1 = 0.475( ) ( ) = Y 2 = 0.880( ) ( ) = Lecture #9-8/2/2011 Slide 15 of 54

17 Principal Component of principal components can be with respect to two possible types of information: PC Variance Correlation 1. The importance of each component Measured by the proportion of total sample variance accounted for by the component 2. The importance of each original variable within each component The weight of the component for each variable (for interpretation of the relative importance of the original variables) The correlation between each of the original variables and the new principal components Lecture #9-8/2/2011 Slide 16 of 54

18 Principal Component Variance The variance of each principal component is equal to the corresponding eigenvalue of that component Var(Y p ) = s 2 Y p = λ p PC Variance Correlation The sum of the eigenvalues is equal to the total sample variance: λ 1 + λ λ p = p λ i = Tr(S) i=1 The corresponding percentage of total sample variance accounted for by each principal component, p, is then: λ p λ 1 + λ λ p Lecture #9-8/2/2011 Slide 17 of 54

19 Principal Component Variance From our SAT Example, the variance of each principal component (and the subsequent relationship to the total sample variance) is given by: PC Variance Correlation Lecture #9-8/2/2011 Slide 18 of 54

20 Component Weight Magnitude To interpret the relative importance of the original variables, the magnitude of the weights are examined PC Variance Correlation The magnitude of e ik measures the importance of the k th variable to the i th principal component, irrespective of the other variables For example, take the first weights of the first principal component from the SAT example: [ ] [ ] SATV e 1 = = SATM This shows that the SAT Math score plays a larger role in the development of the first principal component This also shows that the values of the weights are sensitive to the scale of the measures (SAT Math had a larger variance) Lecture #9-8/2/2011 Slide 19 of 54

21 Correlation of X and Y Occasionally, you will find people judging the relative importance of the original variables by use of their correlation with each of the principal components PC Variance Correlation The correlation between variable X k and principal component Y i can be computed by: ρ Yi,X k = e ik λi s 2 X k Many statisticians do not like this method for evaluation of the importance of a variable because it measures only the univariate contribution of a variable X to a component Y Some recommend only using the coefficients of the principal component to evaluate the relative importance of the original variables Lecture #9-8/2/2011 Slide 20 of 54

22 PCA provides a geometric shift and rescaling of the axes that the original variables were measured The two-variable example of SAT shows how this rotation occurs Original Variables Principal Components Lecture #9-8/2/2011 Slide 21 of 54

23 Original Variables Original Variables Principal Components Lecture #9-8/2/2011 Slide 22 of 54

24 Principal Components Original Variables Principal Components Lecture #9-8/2/2011 Slide 23 of 54

25 Principal Components Original Variables Principal Components Lecture #9-8/2/2011 Slide 24 of 54

26 Example #2 Rather than using the raw data, and hence, the covariance matrix to attain the eigenvalues and eigenvectors used in PCA, another alternative is to use the correlation matrix Using the correlation matrix is equivalent to using the covariance matrix of standardized observations The resulting principal component values (the Y i ), are calculated based on the standardized observations The results of such an analysis can produce different interpretations than a PCA with the original variables It is advised, however, that PCA be used with correlations when variables with widely different scales of measurement are included in the analysis To demonstrate, here is SAS Example #2... Lecture #9-8/2/2011 Slide 25 of 54

27 Example #2 Note the lack of cov in the proc line: proc princomp data=sat out=pcsat_cor; var satv satm; run; Example #2 Lecture #9-8/2/2011 Slide 26 of 54

28 26-1

29 Differing Component Values Also note that the values of the principal components are now based on the standardized variables: Y 1 = X 1 X 1 s X X 2 X 2 s X2 Y 2 = X 1 X 1 s X X 2 X 2 s X2 Lecture #9-8/2/2011 Slide 27 of 54

30 Differing Eigenvalues/Eigenvectors Note the comparison of the eigenvalues and eigenvectors (along with the comparison of the variances): Example #2 Covariance Matrix Variable e 1 e 2 Variance SATV SATM Correlation Variable e 1 e 2 Variance SATV SATM Look at the ratio of eigenvector weights between SATV and SATM and how it corresponds to the ratio of the variance between the two variables. Lecture #9-8/2/2011 Slide 28 of 54

31 PCA Assumptions PCA does not specifically presume any type of data for the analysis PCA Assumptions How Many PC? Uncorrelated Variables Many people prefer to think of using PCA for only continuous variables (although there are numerous examples of this not being the case) If the variables happen to be MVN, then the principal components will also be MVN, with a zero mean vector, and a covariance matrix that has zero off-diagonal elements and diagonal elements equal to the eigenvalues of the principal components Lecture #9-8/2/2011 Slide 29 of 54

32 How Many Principal Components? Although our example had only two variables, more often PCA is used to reduce the number of dimensions PCA Assumptions How Many PC? Uncorrelated Variables Used in that context, the main question of PCA regards how many principal components are to be used to describe the original data set Although no rules exist, most often people try to have a set of components that account for at least 70% to 80% of the total sample variance As a guide to where to decide upon the number of components, a scree plot is sometime used (I will show one in the next example) From dictionary.com: Scree (n): a sloping mass of loose rocks at the base of a cliff Lecture #9-8/2/2011 Slide 30 of 54

33 Uncorrelated Variables in PCA To finalize the theoretical topics of this lecture, consider what would happen to the PCA we had an uncorrelated variable in the analysis PCA Assumptions How Many PC? Uncorrelated Variables For instance, think about how the analysis would be if we included, for each person, the time it took to tie their shoes? The uncorrelated variable essentially cannot be summarized by the new principal components, and therefore becomes its own principal component Code for SAS Example #3: proc princomp data=sat3var out=pcsat3v_cor; var satv satm time; run; Lecture #9-8/2/2011 Slide 31 of 54

34 31-1

35 Gambling Explosion GRI Sample SAS Code SAS Output Accessability of gambling has increased exponentially in the past several years: State lotteries Native American Tribal Casinos Riverboat gambling Internet gambling As accessability has increased, incidences of pathological gambling has increased There is a need to be able to measure pathological gambling and to be able to understand its etiology Lecture #9-8/2/2011 Slide 32 of 54

36 Pathological Gambling The DSM-IV-TR defines pathological gambling as an impulse-control disorder (not elsewhere classified) Persistent and recurrent maladaptive gambling behavior as indicated by five (or more) of the following: A1 Is preoccupied with gambling A2 Needs to gamble with increasing amounts of money in order to achieve the desired excitement A3 Has repeated unsuccessful efforts to control, cut back, or stop gambling A4 Is restless or irritable when attempting to cut down or stop gambling A5 Gambles as a way of escaping from problems or of relieving a dysphoric mood A6 After losing money gambling, often returns another day to get even A7 Lies to family members, therapist, or others to conceal the extend of involvement with gambling A8 Has committed illegal acts such as forgery, fraud, theft, or embezzlement to finance gambling A9 Has jeopardized or lost a significant relationship, job, or educational or career opportunity because of gambling A10 Relies on others to provide money to relieve a desperate financial situation caused by gambling GRI Sample SAS Code SAS Output Behind the 10 criteria are a set of three personality traits (higher order factors): Dependence, Disruption, Loss of Control (Henson et al., 2000) Lecture #9-8/2/2011 Slide 33 of 54

37 Gambling Research Instrument GRI Sample SAS Code SAS Output Henson, Feasel, & Jones (2000) Forty-one Likert scale items Six response levels (Strongly Agree - Strongly Disagree) Twenty-five items used to create sum-scores for each of the 10 DSM criteria Increased variability in responses enabled structure of pathological gambling to be studied GRI Example Items 1. I would like to cut back on my gambling 2. There are few things I would rather do than gamble 3. If I lost a lot of money gambling one day, I would be more likely to want to play again the following day 4. I enjoy talking with my family and friends about my past gambling experiences (R) 5. I find it necessary to gamble with larger amounts of money (than when I first gambled) for gambling to be exciting 6. I have gone to great lengths to obtain money for gambling 7. I feel "high" when I gamble 8. I worry that I am spending too much money gambling 9. I feel restless when I try to cut down or stop gambling 10. It bothers me when I have no money to gamble. 11. I gamble to take my mind off my worries 12. When I lose money gambling, it is a long time before I gamble again (R) 13. I find it difficult to stop gambling 14. I am drawn more by the thrill of gambling than by the money I could win Lecture #9-8/2/2011 Slide 34 of 54

38 Sample Characteristics In a study of the gambling tendencies of college students, an online version of the GRI was created Study included 287 college students recruited from a large midwestern university Average age 19 GRI Sample SAS Code SAS Output Subjects provided responses to the GRI and also responded to an item asking about their overall gambling experience Lecture #9-8/2/2011 Slide 35 of 54

39 SAS Code *SAS ; PROC IMPORT OUT= WORK.gri DATAFILE= "C:\gri.xls" DBMS=EXCEL2000 REPLACE; RANGE="gri$"; GETNAMES=YES; RUN; GRI Sample SAS Code SAS Output proc princomp data=gri out=pcgri_cor outstat=pc_gri; var gri1-gri41; run; Lecture #9-8/2/2011 Slide 36 of 54

40 SAS Output GRI Sample SAS Code SAS Output Lecture #9-8/2/2011 Slide 37 of 54

41 SAS Output To account for 70% of the total sample variance, some 10 principal components must be used: Eigenvalues of the Correlation Matrix Eigenvalue Difference Proportion Cumulative Lecture #9-8/2/2011 Slide 38 of 54

42 SAS Output Highest magnitude component weights: GRI Sample SAS Code SAS Output Component 1 Variable Weight GRI GRI GRI Component 2 Variable Weight GRI GRI GRI GRI Lecture #9-8/2/2011 Slide 39 of 54

43 To demonstrate the use of the principal components in another analysis, consider trying to predict a person s gambling experience rating by the answers to the survey questions In the first part of this example, we will use the original 41 variables in a multiple regression In the second part of this example, we will use the first 10 principle components in a multiple regression Finally, to demonstrate the properties of the principal components (recovery of the R 2 ), all 41 principal components will be used in the regression Lecture #9-8/2/2011 Slide 40 of 54

44 Part 1 Code/Output proc glm data=pcgri_cor; model gambling=gri1-gri41/ss3; run; Lecture #9-8/2/2011 Slide 41 of 54

45 Part 2 Code/Output proc glm data=pcgri_cor; model gambling=prin1-prin10/ss3; run; Lecture #9-8/2/2011 Slide 42 of 54

46 Part 3 Code/Output proc glm data=pcgri_cor; model gambling=prin1-prin41/ss3; run; Lecture #9-8/2/2011 Slide 43 of 54

47 As a final example, consider the following data set (from the SAS user s guide): Police officers were rated by their supervisors in 14 categories as part of standard police departmental administrative procedure. Each variable contains the job ratings, using a scale measurement from 1 to 10 (1=fail to comply, 10=exceptional). Lecture #9-8/2/2011 Slide 44 of 54

48 The variables collected are: Communication Skills Problem Solving Learning Ability Judgment Under Pressure Observational Skills Willingness to Confront Problems Interest in People Interpersonal Sensitivity Desire for Self-Improvement Appearance Dependability Physical Ability Integrity Lecture #9-8/2/2011 Slide 45 of 54

49 Lecture #9-8/2/2011 Slide 46 of 54

50 Lecture #9-8/2/2011 Slide 47 of 54

51 Lecture #9-8/2/2011 Slide 48 of 54

52 Lecture #9-8/2/2011 Slide 49 of 54

53 Lecture #9-8/2/2011 Slide 50 of 54

54 Lecture #9-8/2/2011 Slide 51 of 54

55 Lecture #9-8/2/2011 Slide 52 of 54

56 Lecture #9-8/2/2011 Slide 53 of 54

57 Final Thought Final Thought Principal components analysis is a multivariate statistical technique for reducing the number of dimensions of a data set Often, PCA is an intermediate step between, providing variables that will be used in subsequent statistical analyses The method of gaining meaning of the components through use of their weights is a method that will carry over throughout the rest of this week Many multivariate techniques create new variables that can be interpreted by the nature of the values of the original variables Up next: Exploratory Factor Analysis Lecture #9-8/2/2011 Slide 54 of 54

Principal Components Analysis and Exploratory Factor Analysis

Principal Components Analysis and Exploratory Factor Analysis Principal Components Analysis and Exploratory Factor Analysis PRE 905: Multivariate Analysis Lecture 12: May 6, 2014 PRE 905: PCA and EFA (with CFA) Today s Class Advanced matrix operations Principal Components

More information

Confirmatory Factor Analysis

Confirmatory Factor Analysis Confirmatory Factor Analysis Latent Trait Measurement and Structural Equation Models Lecture #6 February 13, 2013 PSYC 948: Lecture #6 Today s Class An introduction to confirmatory factor analysis The

More information

A Introduction to Matrix Algebra and the Multivariate Normal Distribution

A Introduction to Matrix Algebra and the Multivariate Normal Distribution A Introduction to Matrix Algebra and the Multivariate Normal Distribution PRE 905: Multivariate Analysis Spring 2014 Lecture 6 PRE 905: Lecture 7 Matrix Algebra and the MVN Distribution Today s Class An

More information

An Introduction to Matrix Algebra

An Introduction to Matrix Algebra An Introduction to Matrix Algebra EPSY 905: Fundamentals of Multivariate Modeling Online Lecture #8 EPSY 905: Matrix Algebra In This Lecture An introduction to matrix algebra Ø Scalars, vectors, and matrices

More information

UCLA STAT 233 Statistical Methods in Biomedical Imaging

UCLA STAT 233 Statistical Methods in Biomedical Imaging UCLA STAT 233 Statistical Methods in Biomedical Imaging Instructor: Ivo Dinov, Asst. Prof. In Statistics and Neurology University of California, Los Angeles, Spring 2004 http://www.stat.ucla.edu/~dinov/

More information

Introduction to Matrix Algebra and the Multivariate Normal Distribution

Introduction to Matrix Algebra and the Multivariate Normal Distribution Introduction to Matrix Algebra and the Multivariate Normal Distribution Introduction to Structural Equation Modeling Lecture #2 January 18, 2012 ERSH 8750: Lecture 2 Motivation for Learning the Multivariate

More information

Introduction to Factor Analysis

Introduction to Factor Analysis to Factor Analysis Lecture 10 August 2, 2011 Advanced Multivariate Statistical Methods ICPSR Summer Session #2 Lecture #10-8/3/2011 Slide 1 of 55 Today s Lecture Factor Analysis Today s Lecture Exploratory

More information

Multilevel Models in Matrix Form. Lecture 7 July 27, 2011 Advanced Multivariate Statistical Methods ICPSR Summer Session #2

Multilevel Models in Matrix Form. Lecture 7 July 27, 2011 Advanced Multivariate Statistical Methods ICPSR Summer Session #2 Multilevel Models in Matrix Form Lecture 7 July 27, 2011 Advanced Multivariate Statistical Methods ICPSR Summer Session #2 Today s Lecture Linear models from a matrix perspective An example of how to do

More information

Vectors and Matrices Statistics with Vectors and Matrices

Vectors and Matrices Statistics with Vectors and Matrices Vectors and Matrices Statistics with Vectors and Matrices Lecture 3 September 7, 005 Analysis Lecture #3-9/7/005 Slide 1 of 55 Today s Lecture Vectors and Matrices (Supplement A - augmented with SAS proc

More information

An Introduction to Path Analysis

An Introduction to Path Analysis An Introduction to Path Analysis PRE 905: Multivariate Analysis Lecture 10: April 15, 2014 PRE 905: Lecture 10 Path Analysis Today s Lecture Path analysis starting with multivariate regression then arriving

More information

An Introduction to Mplus and Path Analysis

An Introduction to Mplus and Path Analysis An Introduction to Mplus and Path Analysis PSYC 943: Fundamentals of Multivariate Modeling Lecture 10: October 30, 2013 PSYC 943: Lecture 10 Today s Lecture Path analysis starting with multivariate regression

More information

Eigenvalues, Eigenvectors, and an Intro to PCA

Eigenvalues, Eigenvectors, and an Intro to PCA Eigenvalues, Eigenvectors, and an Intro to PCA Eigenvalues, Eigenvectors, and an Intro to PCA Changing Basis We ve talked so far about re-writing our data using a new set of variables, or a new basis.

More information

Eigenvalues, Eigenvectors, and an Intro to PCA

Eigenvalues, Eigenvectors, and an Intro to PCA Eigenvalues, Eigenvectors, and an Intro to PCA Eigenvalues, Eigenvectors, and an Intro to PCA Changing Basis We ve talked so far about re-writing our data using a new set of variables, or a new basis.

More information

PRINCIPAL COMPONENTS ANALYSIS

PRINCIPAL COMPONENTS ANALYSIS 121 CHAPTER 11 PRINCIPAL COMPONENTS ANALYSIS We now have the tools necessary to discuss one of the most important concepts in mathematical statistics: Principal Components Analysis (PCA). PCA involves

More information

Exploratory Factor Analysis and Principal Component Analysis

Exploratory Factor Analysis and Principal Component Analysis Exploratory Factor Analysis and Principal Component Analysis Today s Topics: What are EFA and PCA for? Planning a factor analytic study Analysis steps: Extraction methods How many factors Rotation and

More information

Statistical Distribution Assumptions of General Linear Models

Statistical Distribution Assumptions of General Linear Models Statistical Distribution Assumptions of General Linear Models Applied Multilevel Models for Cross Sectional Data Lecture 4 ICPSR Summer Workshop University of Colorado Boulder Lecture 4: Statistical Distributions

More information

Introduction to Factor Analysis

Introduction to Factor Analysis to Factor Analysis Lecture 11 November 2, 2005 Multivariate Analysis Lecture #11-11/2/2005 Slide 1 of 58 Today s Lecture Factor Analysis. Today s Lecture Exploratory factor analysis (EFA). Confirmatory

More information

LECTURE 4 PRINCIPAL COMPONENTS ANALYSIS / EXPLORATORY FACTOR ANALYSIS

LECTURE 4 PRINCIPAL COMPONENTS ANALYSIS / EXPLORATORY FACTOR ANALYSIS LECTURE 4 PRINCIPAL COMPONENTS ANALYSIS / EXPLORATORY FACTOR ANALYSIS NOTES FROM PRE- LECTURE RECORDING ON PCA PCA and EFA have similar goals. They are substantially different in important ways. The goal

More information

Introduction to Confirmatory Factor Analysis

Introduction to Confirmatory Factor Analysis Introduction to Confirmatory Factor Analysis Multivariate Methods in Education ERSH 8350 Lecture #12 November 16, 2011 ERSH 8350: Lecture 12 Today s Class An Introduction to: Confirmatory Factor Analysis

More information

Principal Component Analysis. Applied Multivariate Statistics Spring 2012

Principal Component Analysis. Applied Multivariate Statistics Spring 2012 Principal Component Analysis Applied Multivariate Statistics Spring 2012 Overview Intuition Four definitions Practical examples Mathematical example Case study 2 PCA: Goals Goal 1: Dimension reduction

More information

Exploratory Factor Analysis and Principal Component Analysis

Exploratory Factor Analysis and Principal Component Analysis Exploratory Factor Analysis and Principal Component Analysis Today s Topics: What are EFA and PCA for? Planning a factor analytic study Analysis steps: Extraction methods How many factors Rotation and

More information

Univariate Normal Distribution; GLM with the Univariate Normal; Least Squares Estimation

Univariate Normal Distribution; GLM with the Univariate Normal; Least Squares Estimation Univariate Normal Distribution; GLM with the Univariate Normal; Least Squares Estimation PRE 905: Multivariate Analysis Spring 2014 Lecture 4 Today s Class The building blocks: The basics of mathematical

More information

Chapter 4: Factor Analysis

Chapter 4: Factor Analysis Chapter 4: Factor Analysis In many studies, we may not be able to measure directly the variables of interest. We can merely collect data on other variables which may be related to the variables of interest.

More information

Principal Component Analysis (PCA) Principal Component Analysis (PCA)

Principal Component Analysis (PCA) Principal Component Analysis (PCA) Recall: Eigenvectors of the Covariance Matrix Covariance matrices are symmetric. Eigenvectors are orthogonal Eigenvectors are ordered by the magnitude of eigenvalues: λ 1 λ 2 λ p {v 1, v 2,..., v n } Recall:

More information

Applied Multivariate Analysis

Applied Multivariate Analysis Department of Mathematics and Statistics, University of Vaasa, Finland Spring 2017 Dimension reduction Principal Component Analysis (PCA) The problem in exploratory multivariate data analysis usually is

More information

SPIRITUAL GIFTS. ( ) ( ) 1. Would you describe yourself as an effective public speaker?

SPIRITUAL GIFTS. ( ) ( ) 1. Would you describe yourself as an effective public speaker? SPIRITUAL GIFTS QUESTIONNAIRE: SPIRITUAL GIFTS ( ) ( ) 1. Would you describe yourself as an effective public speaker? ( ) ( ) 2. Do you find it easy and enjoyable to spend time in intense study and research

More information

Principal Component Analysis, A Powerful Scoring Technique

Principal Component Analysis, A Powerful Scoring Technique Principal Component Analysis, A Powerful Scoring Technique George C. J. Fernandez, University of Nevada - Reno, Reno NV 89557 ABSTRACT Data mining is a collection of analytical techniques to uncover new

More information

Multivariate Statistics (I) 2. Principal Component Analysis (PCA)

Multivariate Statistics (I) 2. Principal Component Analysis (PCA) Multivariate Statistics (I) 2. Principal Component Analysis (PCA) 2.1 Comprehension of PCA 2.2 Concepts of PCs 2.3 Algebraic derivation of PCs 2.4 Selection and goodness-of-fit of PCs 2.5 Algebraic derivation

More information

MATH 10 INTRODUCTORY STATISTICS

MATH 10 INTRODUCTORY STATISTICS MATH 10 INTRODUCTORY STATISTICS Tommy Khoo Your friendly neighbourhood graduate student. Week 1 Chapter 1 Introduction What is Statistics? Why do you need to know Statistics? Technical lingo and concepts:

More information

Principal Component Analysis & Factor Analysis. Psych 818 DeShon

Principal Component Analysis & Factor Analysis. Psych 818 DeShon Principal Component Analysis & Factor Analysis Psych 818 DeShon Purpose Both are used to reduce the dimensionality of correlated measurements Can be used in a purely exploratory fashion to investigate

More information

PRINCIPAL COMPONENTS ANALYSIS (PCA)

PRINCIPAL COMPONENTS ANALYSIS (PCA) PRINCIPAL COMPONENTS ANALYSIS (PCA) Introduction PCA is considered an exploratory technique that can be used to gain a better understanding of the interrelationships between variables. PCA is performed

More information

Multiple Linear Regression

Multiple Linear Regression Multiple Linear Regression Simple linear regression tries to fit a simple line between two variables Y and X. If X is linearly related to Y this explains some of the variability in Y. In most cases, there

More information

Generalized Linear Models for Non-Normal Data

Generalized Linear Models for Non-Normal Data Generalized Linear Models for Non-Normal Data Today s Class: 3 parts of a generalized model Models for binary outcomes Complications for generalized multivariate or multilevel models SPLH 861: Lecture

More information

Do not copy, post, or distribute

Do not copy, post, or distribute 14 CORRELATION ANALYSIS AND LINEAR REGRESSION Assessing the Covariability of Two Quantitative Properties 14.0 LEARNING OBJECTIVES In this chapter, we discuss two related techniques for assessing a possible

More information

Mean Vector Inferences

Mean Vector Inferences Mean Vector Inferences Lecture 5 September 21, 2005 Multivariate Analysis Lecture #5-9/21/2005 Slide 1 of 34 Today s Lecture Inferences about a Mean Vector (Chapter 5). Univariate versions of mean vector

More information

DIMENSION REDUCTION AND CLUSTER ANALYSIS

DIMENSION REDUCTION AND CLUSTER ANALYSIS DIMENSION REDUCTION AND CLUSTER ANALYSIS EECS 833, 6 March 2006 Geoff Bohling Assistant Scientist Kansas Geological Survey geoff@kgs.ku.edu 864-2093 Overheads and resources available at http://people.ku.edu/~gbohling/eecs833

More information

Applied Multivariate Analysis

Applied Multivariate Analysis Department of Mathematics and Statistics, University of Vaasa, Finland Spring 2017 Dimension reduction Exploratory (EFA) Background While the motivation in PCA is to replace the original (correlated) variables

More information

Introduction to Bayesian Statistics and Markov Chain Monte Carlo Estimation. EPSY 905: Multivariate Analysis Spring 2016 Lecture #10: April 6, 2016

Introduction to Bayesian Statistics and Markov Chain Monte Carlo Estimation. EPSY 905: Multivariate Analysis Spring 2016 Lecture #10: April 6, 2016 Introduction to Bayesian Statistics and Markov Chain Monte Carlo Estimation EPSY 905: Multivariate Analysis Spring 2016 Lecture #10: April 6, 2016 EPSY 905: Intro to Bayesian and MCMC Today s Class An

More information

Principal component analysis

Principal component analysis Principal component analysis Angela Montanari 1 Introduction Principal component analysis (PCA) is one of the most popular multivariate statistical methods. It was first introduced by Pearson (1901) and

More information

Linear Regression. Linear Regression. Linear Regression. Did You Mean Association Or Correlation?

Linear Regression. Linear Regression. Linear Regression. Did You Mean Association Or Correlation? Did You Mean Association Or Correlation? AP Statistics Chapter 8 Be careful not to use the word correlation when you really mean association. Often times people will incorrectly use the word correlation

More information

Unconstrained Ordination

Unconstrained Ordination Unconstrained Ordination Sites Species A Species B Species C Species D Species E 1 0 (1) 5 (1) 1 (1) 10 (4) 10 (4) 2 2 (3) 8 (3) 4 (3) 12 (6) 20 (6) 3 8 (6) 20 (6) 10 (6) 1 (2) 3 (2) 4 4 (5) 11 (5) 8 (5)

More information

Introduction To Confirmatory Factor Analysis and Item Response Theory

Introduction To Confirmatory Factor Analysis and Item Response Theory Introduction To Confirmatory Factor Analysis and Item Response Theory Lecture 23 May 3, 2005 Applied Regression Analysis Lecture #23-5/3/2005 Slide 1 of 21 Today s Lecture Confirmatory Factor Analysis.

More information

Course Introduction and Overview Descriptive Statistics Conceptualizations of Variance Review of the General Linear Model

Course Introduction and Overview Descriptive Statistics Conceptualizations of Variance Review of the General Linear Model Course Introduction and Overview Descriptive Statistics Conceptualizations of Variance Review of the General Linear Model PSYC 943 (930): Fundamentals of Multivariate Modeling Lecture 1: August 22, 2012

More information

Path Analysis. PRE 906: Structural Equation Modeling Lecture #5 February 18, PRE 906, SEM: Lecture 5 - Path Analysis

Path Analysis. PRE 906: Structural Equation Modeling Lecture #5 February 18, PRE 906, SEM: Lecture 5 - Path Analysis Path Analysis PRE 906: Structural Equation Modeling Lecture #5 February 18, 2015 PRE 906, SEM: Lecture 5 - Path Analysis Key Questions for Today s Lecture What distinguishes path models from multivariate

More information

Exst7037 Multivariate Analysis Cancorr interpretation Page 1

Exst7037 Multivariate Analysis Cancorr interpretation Page 1 Exst7037 Multivariate Analysis Cancorr interpretation Page 1 1 *** C03S3D1 ***; 2 ****************************************************************************; 3 *** The data set insulin shows data from

More information

How to write maths (well)

How to write maths (well) How to write maths (well) Dr Euan Spence 29 September 2017 These are the slides from a talk I gave to the new first-year students at Bath, annotated with some of the things I said (which appear in boxes

More information

I L L I N O I S UNIVERSITY OF ILLINOIS AT URBANA-CHAMPAIGN

I L L I N O I S UNIVERSITY OF ILLINOIS AT URBANA-CHAMPAIGN Canonical Edps/Soc 584 and Psych 594 Applied Multivariate Statistics Carolyn J. Anderson Department of Educational Psychology I L L I N O I S UNIVERSITY OF ILLINOIS AT URBANA-CHAMPAIGN Canonical Slide

More information

Lecture 13. Principal Component Analysis. Brett Bernstein. April 25, CDS at NYU. Brett Bernstein (CDS at NYU) Lecture 13 April 25, / 26

Lecture 13. Principal Component Analysis. Brett Bernstein. April 25, CDS at NYU. Brett Bernstein (CDS at NYU) Lecture 13 April 25, / 26 Principal Component Analysis Brett Bernstein CDS at NYU April 25, 2017 Brett Bernstein (CDS at NYU) Lecture 13 April 25, 2017 1 / 26 Initial Question Intro Question Question Let S R n n be symmetric. 1

More information

Lecture 11 Multiple Linear Regression

Lecture 11 Multiple Linear Regression Lecture 11 Multiple Linear Regression STAT 512 Spring 2011 Background Reading KNNL: 6.1-6.5 11-1 Topic Overview Review: Multiple Linear Regression (MLR) Computer Science Case Study 11-2 Multiple Regression

More information

Subject CS1 Actuarial Statistics 1 Core Principles

Subject CS1 Actuarial Statistics 1 Core Principles Institute of Actuaries of India Subject CS1 Actuarial Statistics 1 Core Principles For 2019 Examinations Aim The aim of the Actuarial Statistics 1 subject is to provide a grounding in mathematical and

More information

CS 361: Probability & Statistics

CS 361: Probability & Statistics February 19, 2018 CS 361: Probability & Statistics Random variables Markov s inequality This theorem says that for any random variable X and any value a, we have A random variable is unlikely to have an

More information

Business Statistics. Lecture 9: Simple Regression

Business Statistics. Lecture 9: Simple Regression Business Statistics Lecture 9: Simple Regression 1 On to Model Building! Up to now, class was about descriptive and inferential statistics Numerical and graphical summaries of data Confidence intervals

More information

MACHINE LEARNING. Methods for feature extraction and reduction of dimensionality: Probabilistic PCA and kernel PCA

MACHINE LEARNING. Methods for feature extraction and reduction of dimensionality: Probabilistic PCA and kernel PCA 1 MACHINE LEARNING Methods for feature extraction and reduction of dimensionality: Probabilistic PCA and kernel PCA 2 Practicals Next Week Next Week, Practical Session on Computer Takes Place in Room GR

More information

MATH 10 INTRODUCTORY STATISTICS

MATH 10 INTRODUCTORY STATISTICS MATH 10 INTRODUCTORY STATISTICS Tommy Khoo Your friendly neighbourhood graduate student. It is Time for Homework! ( ω `) First homework + data will be posted on the website, under the homework tab. And

More information

Q-Matrix Development. NCME 2009 Workshop

Q-Matrix Development. NCME 2009 Workshop Q-Matrix Development NCME 2009 Workshop Introduction We will define the Q-matrix Then we will discuss method of developing your own Q-matrix Talk about possible problems of the Q-matrix to avoid The Q-matrix

More information

EPSY 905: Fundamentals of Multivariate Modeling Online Lecture #7

EPSY 905: Fundamentals of Multivariate Modeling Online Lecture #7 Introduction to Generalized Univariate Models: Models for Binary Outcomes EPSY 905: Fundamentals of Multivariate Modeling Online Lecture #7 EPSY 905: Intro to Generalized In This Lecture A short review

More information

Quantitative Understanding in Biology Principal Components Analysis

Quantitative Understanding in Biology Principal Components Analysis Quantitative Understanding in Biology Principal Components Analysis Introduction Throughout this course we have seen examples of complex mathematical phenomena being represented as linear combinations

More information

Alternatives to Difference Scores: Polynomial Regression and Response Surface Methodology. Jeffrey R. Edwards University of North Carolina

Alternatives to Difference Scores: Polynomial Regression and Response Surface Methodology. Jeffrey R. Edwards University of North Carolina Alternatives to Difference Scores: Polynomial Regression and Response Surface Methodology Jeffrey R. Edwards University of North Carolina 1 Outline I. Types of Difference Scores II. Questions Difference

More information

Special Theory of Relativity Prof. Shiva Prasad Department of Physics Indian Institute of Technology, Bombay. Lecture - 15 Momentum Energy Four Vector

Special Theory of Relativity Prof. Shiva Prasad Department of Physics Indian Institute of Technology, Bombay. Lecture - 15 Momentum Energy Four Vector Special Theory of Relativity Prof. Shiva Prasad Department of Physics Indian Institute of Technology, Bombay Lecture - 15 Momentum Energy Four Vector We had started discussing the concept of four vectors.

More information

Eigenvalues, Eigenvectors, and an Intro to PCA

Eigenvalues, Eigenvectors, and an Intro to PCA Eigenvalues, Eigenvectors, and an Intro to PCA Eigenvalues, Eigenvectors, and an Intro to PCA Changing Basis We ve talked so far about re-writing our data using a new set of variables, or a new basis.

More information

Math 223 Lecture Notes 3/15/04 From The Basic Practice of Statistics, bymoore

Math 223 Lecture Notes 3/15/04 From The Basic Practice of Statistics, bymoore Math 223 Lecture Notes 3/15/04 From The Basic Practice of Statistics, bymoore Chapter 3 continued Describing distributions with numbers Measuring spread of data: Quartiles Definition 1: The interquartile

More information

Quantitative Understanding in Biology Short Course Session 9 Principal Components Analysis

Quantitative Understanding in Biology Short Course Session 9 Principal Components Analysis Quantitative Understanding in Biology Short Course Session 9 Principal Components Analysis Jinhyun Ju Jason Banfelder Luce Skrabanek June 21st, 218 1 Preface For the last session in this course, we ll

More information

Table of Contents. Multivariate methods. Introduction II. Introduction I

Table of Contents. Multivariate methods. Introduction II. Introduction I Table of Contents Introduction Antti Penttilä Department of Physics University of Helsinki Exactum summer school, 04 Construction of multinormal distribution Test of multinormality with 3 Interpretation

More information

Interactions among Continuous Predictors

Interactions among Continuous Predictors Interactions among Continuous Predictors Today s Class: Simple main effects within two-way interactions Conquering TEST/ESTIMATE/LINCOM statements Regions of significance Three-way interactions (and beyond

More information

Academic Affairs Assessment of Student Learning Report for Academic Year

Academic Affairs Assessment of Student Learning Report for Academic Year Academic Affairs Assessment of Student Learning Report for Academic Year 2017-2018. Department/Program Chemistry Assessment Coordinator s Name: Micheal Fultz Assessment Coordinator s Email Address: mfultz@wvstateu.edu

More information

Principle Components Analysis (PCA) Relationship Between a Linear Combination of Variables and Axes Rotation for PCA

Principle Components Analysis (PCA) Relationship Between a Linear Combination of Variables and Axes Rotation for PCA Principle Components Analysis (PCA) Relationship Between a Linear Combination of Variables and Axes Rotation for PCA Principle Components Analysis: Uses one group of variables (we will call this X) In

More information

Lecture 2: Linear and Mixed Models

Lecture 2: Linear and Mixed Models Lecture 2: Linear and Mixed Models Bruce Walsh lecture notes Introduction to Mixed Models SISG, Seattle 18 20 July 2018 1 Quick Review of the Major Points The general linear model can be written as y =

More information

6 THE NORMAL DISTRIBUTION

6 THE NORMAL DISTRIBUTION CHAPTER 6 THE NORMAL DISTRIBUTION 341 6 THE NORMAL DISTRIBUTION Figure 6.1 If you ask enough people about their shoe size, you will find that your graphed data is shaped like a bell curve and can be described

More information

Machine Learning 2nd Edition

Machine Learning 2nd Edition INTRODUCTION TO Lecture Slides for Machine Learning 2nd Edition ETHEM ALPAYDIN, modified by Leonardo Bobadilla and some parts from http://www.cs.tau.ac.il/~apartzin/machinelearning/ The MIT Press, 2010

More information

Short Answer Questions: Answer on your separate blank paper. Points are given in parentheses.

Short Answer Questions: Answer on your separate blank paper. Points are given in parentheses. ISQS 6348 Final exam solutions. Name: Open book and notes, but no electronic devices. Answer short answer questions on separate blank paper. Answer multiple choice on this exam sheet. Put your name on

More information

Review of Multiple Regression

Review of Multiple Regression Ronald H. Heck 1 Let s begin with a little review of multiple regression this week. Linear models [e.g., correlation, t-tests, analysis of variance (ANOVA), multiple regression, path analysis, multivariate

More information

B. Weaver (24-Mar-2005) Multiple Regression Chapter 5: Multiple Regression Y ) (5.1) Deviation score = (Y i

B. Weaver (24-Mar-2005) Multiple Regression Chapter 5: Multiple Regression Y ) (5.1) Deviation score = (Y i B. Weaver (24-Mar-2005) Multiple Regression... 1 Chapter 5: Multiple Regression 5.1 Partial and semi-partial correlation Before starting on multiple regression per se, we need to consider the concepts

More information

Principal Component Analysis

Principal Component Analysis Principal Component Analysis Yingyu Liang yliang@cs.wisc.edu Computer Sciences Department University of Wisconsin, Madison [based on slides from Nina Balcan] slide 1 Goals for the lecture you should understand

More information

ANCOVA. Lecture 9 Andrew Ainsworth

ANCOVA. Lecture 9 Andrew Ainsworth ANCOVA Lecture 9 Andrew Ainsworth What is ANCOVA? Analysis of covariance an extension of ANOVA in which main effects and interactions are assessed on DV scores after the DV has been adjusted for by the

More information

Simple Regression Model. January 24, 2011

Simple Regression Model. January 24, 2011 Simple Regression Model January 24, 2011 Outline Descriptive Analysis Causal Estimation Forecasting Regression Model We are actually going to derive the linear regression model in 3 very different ways

More information

Comparing IRT with Other Models

Comparing IRT with Other Models Comparing IRT with Other Models Lecture #14 ICPSR Item Response Theory Workshop Lecture #14: 1of 45 Lecture Overview The final set of slides will describe a parallel between IRT and another commonly used

More information

I L L I N O I S UNIVERSITY OF ILLINOIS AT URBANA-CHAMPAIGN

I L L I N O I S UNIVERSITY OF ILLINOIS AT URBANA-CHAMPAIGN Principal Analysis Edps/Soc 584 and Psych 594 Applied Multivariate Statistics Carolyn J. Anderson Department of Educational Psychology I L L I N O I S UNIVERSITY OF ILLINOIS AT URBANA-CHAMPAIGN c Board

More information

TESTING AND MEASUREMENT IN PSYCHOLOGY. Merve Denizci Nazlıgül, M.S.

TESTING AND MEASUREMENT IN PSYCHOLOGY. Merve Denizci Nazlıgül, M.S. TESTING AND MEASUREMENT IN PSYCHOLOGY Merve Denizci Nazlıgül, M.S. PREPARING THE DATA FOR ANALYSIS 1. ACCURACY OF DATA FILE 1st step SPSS FREQUENCIES For continuous variables, are all the values within

More information

Recitation 9: Probability Matrices and Real Symmetric Matrices. 3 Probability Matrices: Definitions and Examples

Recitation 9: Probability Matrices and Real Symmetric Matrices. 3 Probability Matrices: Definitions and Examples Math b TA: Padraic Bartlett Recitation 9: Probability Matrices and Real Symmetric Matrices Week 9 Caltech 20 Random Question Show that + + + + +... = ϕ, the golden ratio, which is = + 5. 2 2 Homework comments

More information

Student Questionnaire (s) Main Survey

Student Questionnaire (s) Main Survey School: Class: Student: Identification Label IEA Third International Mathematics and Science Study - Repeat Student Questionnaire (s) Main Survey TIMSS Study Center Boston College Chestnut Hill, MA 02467

More information

Methods of Mathematics

Methods of Mathematics Methods of Mathematics Kenneth A. Ribet UC Berkeley Math 10B April 19, 2016 There is a new version of the online textbook file Matrix_Algebra.pdf. The next breakfast will be two days from today, April

More information

Course Introduction and Overview Descriptive Statistics Conceptualizations of Variance Review of the General Linear Model

Course Introduction and Overview Descriptive Statistics Conceptualizations of Variance Review of the General Linear Model Course Introduction and Overview Descriptive Statistics Conceptualizations of Variance Review of the General Linear Model EPSY 905: Multivariate Analysis Lecture 1 20 January 2016 EPSY 905: Lecture 1 -

More information

Principal Component Analysis!! Lecture 11!

Principal Component Analysis!! Lecture 11! Principal Component Analysis Lecture 11 1 Eigenvectors and Eigenvalues g Consider this problem of spreading butter on a bread slice 2 Eigenvectors and Eigenvalues g Consider this problem of stretching

More information

Multiple Regression. Inference for Multiple Regression and A Case Study. IPS Chapters 11.1 and W.H. Freeman and Company

Multiple Regression. Inference for Multiple Regression and A Case Study. IPS Chapters 11.1 and W.H. Freeman and Company Multiple Regression Inference for Multiple Regression and A Case Study IPS Chapters 11.1 and 11.2 2009 W.H. Freeman and Company Objectives (IPS Chapters 11.1 and 11.2) Multiple regression Data for multiple

More information

36-309/749 Experimental Design for Behavioral and Social Sciences. Dec 1, 2015 Lecture 11: Mixed Models (HLMs)

36-309/749 Experimental Design for Behavioral and Social Sciences. Dec 1, 2015 Lecture 11: Mixed Models (HLMs) 36-309/749 Experimental Design for Behavioral and Social Sciences Dec 1, 2015 Lecture 11: Mixed Models (HLMs) Independent Errors Assumption An error is the deviation of an individual observed outcome (DV)

More information

START: Read 1 Guide for Repeated Interactive Read-Alouds

START: Read 1 Guide for Repeated Interactive Read-Alouds Push-In and Connect Key Events START: Read 1 Guide for Repeated Interactive Read-Alouds The Snowy Day By: Ezra Jack Keats Push-In Story Problem Target Vocabulary Read 1: STATE STATE: Show cover illustration

More information

Generative Clustering, Topic Modeling, & Bayesian Inference

Generative Clustering, Topic Modeling, & Bayesian Inference Generative Clustering, Topic Modeling, & Bayesian Inference INFO-4604, Applied Machine Learning University of Colorado Boulder December 12-14, 2017 Prof. Michael Paul Unsupervised Naïve Bayes Last week

More information

STATISTICS 407 METHODS OF MULTIVARIATE ANALYSIS TOPICS

STATISTICS 407 METHODS OF MULTIVARIATE ANALYSIS TOPICS STATISTICS 407 METHODS OF MULTIVARIATE ANALYSIS TOPICS Principal Component Analysis (PCA): Reduce the, summarize the sources of variation in the data, transform the data into a new data set where the variables

More information

Principal Components Analysis using R Francis Huang / November 2, 2016

Principal Components Analysis using R Francis Huang / November 2, 2016 Principal Components Analysis using R Francis Huang / huangf@missouri.edu November 2, 2016 Principal components analysis (PCA) is a convenient way to reduce high dimensional data into a smaller number

More information

Teaching a Prestatistics Course: Propelling Non-STEM Students Forward

Teaching a Prestatistics Course: Propelling Non-STEM Students Forward Teaching a Prestatistics Course: Propelling Non-STEM Students Forward Jay Lehmann College of San Mateo MathNerdJay@aol.com www.pearsonhighered.com/lehmannseries Learning Is in the Details Detailing concepts

More information

, (1) e i = ˆσ 1 h ii. c 2016, Jeffrey S. Simonoff 1

, (1) e i = ˆσ 1 h ii. c 2016, Jeffrey S. Simonoff 1 Regression diagnostics As is true of all statistical methodologies, linear regression analysis can be a very effective way to model data, as along as the assumptions being made are true. For the regression

More information

MATH-0955: BEGINNING ALGEBRA

MATH-0955: BEGINNING ALGEBRA MATH-0955: Beginning Algebra 1 MATH-0955: BEGINNING ALGEBRA Cuyahoga Community College Viewing:MATH-0955 : Beginning Algebra Board of Trustees: 2016-03-31 Academic Term: 2016-08-23 Subject Code MATH -

More information

MATH 10 INTRODUCTORY STATISTICS

MATH 10 INTRODUCTORY STATISTICS MATH 10 INTRODUCTORY STATISTICS Ramesh Yapalparvi It is Time for Homework! ( ω `) First homework + data will be posted on the website, under the homework tab. And also sent out via email. 30% weekly homework.

More information

2/26/2017. This is similar to canonical correlation in some ways. PSY 512: Advanced Statistics for Psychological and Behavioral Research 2

2/26/2017. This is similar to canonical correlation in some ways. PSY 512: Advanced Statistics for Psychological and Behavioral Research 2 PSY 512: Advanced Statistics for Psychological and Behavioral Research 2 What is factor analysis? What are factors? Representing factors Graphs and equations Extracting factors Methods and criteria Interpreting

More information

Balancing Chemical Equations

Balancing Chemical Equations Lesson Created by: Lauryn Atwood Length of lesson: 1 week Description of the class: Heterogeneous Name of course: Chemistry Grade level: 10-12 Honors or regular: Regular Balancing Chemical Equations Source

More information

Dimension Reduction and Classification Using PCA and Factor. Overview

Dimension Reduction and Classification Using PCA and Factor. Overview Dimension Reduction and Classification Using PCA and - A Short Overview Laboratory for Interdisciplinary Statistical Analysis Department of Statistics Virginia Tech http://www.stat.vt.edu/consult/ March

More information

MATH-0965: INTERMEDIATE ALGEBRA

MATH-0965: INTERMEDIATE ALGEBRA MATH-0965: Intermediate Algebra 1 MATH-0965: INTERMEDIATE ALGEBRA Cuyahoga Community College Viewing:MATH-0965 : Intermediate Algebra Board of Trustees: 2016-03-31 Academic Term: 2016-08-23 Subject Code

More information

SESSION 5 Descriptive Statistics

SESSION 5 Descriptive Statistics SESSION 5 Descriptive Statistics Descriptive statistics are used to describe the basic features of the data in a study. They provide simple summaries about the sample and the measures. Together with simple

More information

Grace College Department of Graduate Counseling Vital Statistics and Program Evaluation Outcomes Vital Statistics

Grace College Department of Graduate Counseling Vital Statistics and Program Evaluation Outcomes Vital Statistics Grace College Department of Graduate Counseling Vital Statistics and Program Evaluation Outcomes 2017-2018 Vital Statistics 2017-2018 Enrollment as of 4/1/2018 (full-time) 113 Graduates (Summer 2017-Spring

More information

Lecture Wigner-Ville Distributions

Lecture Wigner-Ville Distributions Introduction to Time-Frequency Analysis and Wavelet Transforms Prof. Arun K. Tangirala Department of Chemical Engineering Indian Institute of Technology, Madras Lecture - 6.1 Wigner-Ville Distributions

More information