Correspondence Analysis (CA)
|
|
- Reynard Lee
- 6 years ago
- Views:
Transcription
1 Correspondence Analysis (CA) François Husson & Magalie Houée-Bigot Department o Applied Mathematics - Rennes Agrocampus husson@agrocampus-ouest.r / 43
2 Correspondence Analysis (CA) Data 2 Independence model 3 Point clouds and what they mean 4 Percentage o inertia, and inertia in CA 5 Simultaneous representation o rows and columns 6 Interpretation aids 2 / 43
3 Correspondence Analysis (CA) Data 2 Independence model 3 Point clouds and what they mean 4 Percentage o inertia, and inertia in CA 5 Simultaneous representation o rows and columns 6 Interpretation aids 3 / 43
4 Contingency tables Variable V 2 Variable V i j J x ij : number o individuals with category i o the variable I category j o the variable J I Characters in Words Number o times character i uses Phèdre (Racine) the word j Perume Descriptors Number o times perume i was described with the word j Biodiversity Species Abundance o species j in place i = Examples where a χ 2 test or independence can be applied 4 / 43
5 Nobel prize data Chemistry Economic Literature Medicine Peace Physics Sum sciences Canada France Germany Italy Japan Russia UK USA Sum Is there a link between countries and prize categories? Do some countries specialize in certain prizes? 5 / 43
6 Data n individuals and 2 qualitative variables V V 2 Categories o V 2 j J Individuals l i j Categories o V i n I n Distribution o the n individual in the I J boxes 6 / 43
7 From contingency tables to probability tables Categories o V 2 Marginal column (marginal probability) j J Σ ij = x ij n Categories o V Marginal row (marginal probability) i Σ I ij. j =. j ij i= I i. J i. = ij Link between V and V2: Deviation o the observed data rom the independence model j= 7 / 43
8 Correspondence Analysis (CA) Data 2 Independence model 3 Point clouds and what they mean 4 Percentage o inertia, and inertia in CA 5 Simultaneous representation o rows and columns 6 Interpretation aids 8 / 43
9 Links and independence between qualitative variables Independence model: Independent events: P(A and B) = P(A) P(B) Independent qualitative variables: i, j, ij = i..j Joint probability = product o marginal probabilities ij ij Another way to write it: =.j = i. i..j Conditional probability = marginal probability 9 / 43
10 Links and independence between qualitative variables Deviation o observed data ( ij ) rom independence model ( i..j ) Signiicance o the link/deviation: χ 2 test χ 2 obs = I J (obs. num. theor. num.) 2 theor. num i= j= = I J (n ij n i..j ) 2 n i..j i= j= χ 2 obs = I J n i= j= (observed probability theoretical probability)2 theoretical probability = n Φ 2 2 Strength o the link = Φ 2 = deviation o observed probabilities rom theoretical ones 3 Type o link = connections between certain categories CA works with the table o probabilities says nothing about signiicance visualizes the sorts o links there are between the variables 0 / 43
11 How does CA detect a deviation rom independence? Analysis by row: ij i. =.j Categories o V 2 j J Σ Row proile i = conditional distribution o V2 knowing that we are the category i or V Categories o V i ij i. I G I. j Mean row proile = marginal distribution o V2 Proile o ALL the individuals Deviation rom independence using a multidimensional approach / 43
12 Comparing the row proile to the mean proile Chemistry Eco Lit. Medicine Peace Physics Sum Canada France Germany Italy Japan Russia UK USA Mean proile Canada France Germany Italy Japan Russia UK USA Mean proile Do Italians win particular categories o Nobel prize? 2 / 43
13 How does CA detect a deviation rom independence? Analysis by column: Categories o V 2 j J G J ij.j = i. Categories o V i ij. j i. I Σ Column proile j = conditional distribution o V knowing that we are j or V2 Mean column proile = marginal distribution o V Proile o ALL the individuals CA compares the column proile to the mean proile Deviation rom independence using a multidimensional approach 3 / 43
14 Comparing the column proile to the mean proile Chem Eco Lit Med Peace Phys Mean proile Canada France Chemistry Economic sc. Literature Medicine Peace Physics Mean proile Germany Italy Japan Russia UK USA Sum Is the country distribution or literature prizes the same as the country distribution or total prizes? 4 / 43
15 Correspondence Analysis (CA) Data 2 Independence model 3 Point clouds and what they mean 4 Percentage o inertia, and inertia in CA 5 Simultaneous representation o rows and columns 6 Interpretation aids 5 / 43
16 The cloud o row proiles J RI d χ 2 ( i, i' ) Categories o V 2 j J Σ i i (weight i. ) N I G I Categories o V i ij i. I G I. j. j i' j i'. ij i. Category j o V 2 Distance between two proiles: d 2 χ 2 (i, i ) = Distance to the mean proile G I : d 2 χ 2 (i, G I ) = J j=.j J ( ij i. i j j=.j i. ) 2 ( ) 2 ij.j i. 6 / 43
17 The cloud o column proiles Categories o V 2 j J G J J RI I d χ 2 ( j, j' ) j (weight.j ) j G J Categories o V i ij. j i. I Σ i. ij'. j' ij. j Category i o V Distance between two proiles: d 2 χ 2 (j, j ) = Distance to the mean proile G J : d 2 χ 2 (j, G J ) = I i= i. I ( ij i= i. ) 2 ij.j.j ( ij i..j ) 2 7 / 43
18 What happens when there is independence? For all i, ij i. =.j the proiles are the same as the mean proile N I becomes just G I (the cloud has zero inertia) RI J i N I G I. j ij i. Category j o V 2 Same or the columns: or all j, ij.j = i / 43
19 Deviation rom independence, and inertia The urther the data is rom independence, the more the proiles spread rom the origin I I Inertia(N I /G I ) = Inertia(i/G I ) = i. dχ 2 2 (i, G I ) = = i= I J i. i= j=.j I J i= j= φ 2 measures the strength o the link ( ij i..j ( ij i..j ) 2 i..j i= ) 2 = χ2 n = φ2 Studying the inertia o N I turns out to be the same as studying deviation rom independence Same or N J : Inertia(N J /G J ) = Inertia(N I /G I ) (duality) 9 / 43
20 Visualizing the row (or column) cloud Decompose the inertia o N I using actor analysis Project N I onto a sequence o orthogonal axes with maximal inertia RI J M i u N I O = G I H i projection o M i on P u 2 P I Find P such that i. (OH i ) 2 is maximal i= u axis o maximal inertia u 2 axis o maximal inertia such that u 2 u Inertia associated with the s-th axis: I i. (OHi s )2 = λ s i= 20 / 43
21 How to interpret? Our example: Dim 2 (24.60%) Economic science Medicine USA Canada UK Physics Chemistry Peace Germany Japan Russia France Italy Literature st axis: contrasts science - other categories 2nd axis: contrasts physics/chemistry - economic sciences Dim (54.75%) Economic Chemistry Literature Medicine Peace Physics Sum science Italy UK Row margin / 43
22 Correspondence Analysis (CA) Data 2 Independence model 3 Point clouds and what they mean 4 Percentage o inertia, and inertia in CA 5 Simultaneous representation o rows and columns 6 Interpretation aids 22 / 43
23 Percentage o inertia Quality o representation o N I on the sth axis I ( ) Projected inertia o N I on u s i= = i. OH s 2 i Total inertia o N = λ s I I i= i. (OM i ) 2 30 descriptors K i= λ k Inertia Inertia (%) F F F F F Sum Deviation rom independence wellsummarized by the irst two axes (79 %) 2 Projected inertia can be summed across axes (because orthogonal) K k= λ k = Inertia (N I ) = Φ 2 Here nφ 2 = = χ 2 = How the inertia decreases can help choose number o axes to keep % o inertia p-value = % o inertia CA on a contingency table acing o 0 white wines with Rank o the axis CA on a contingency table acing o 0 white wines with 30 descriptors Rank o the axis 23 / 43
24 Inertia (= eigenvalues) FC CA 0 λ ss En ACP λ elle structure Indes CA: données 0 λ s correspond Inle PCA maximum (normalized):? λ What structure does an eigenvalue o correspond to? J J 22 II II J I J 2 I 2 s-axis Partition into two classes o rows and columns Exclusive associations between classes Partition en deux classes des lignes 24 / 43
25 Inertia (= eigenvalues) Data: recognizing three lavors (sweet, sour, bitter) For each lavor, we asked 0 people to try to recognize taste o a sample Perc. Perc. Perc. Perc. sweet sweet sour sour bitter bitter Sweet Sweet Sour Sour Bitter Bitter Perc. Perc. Perc. Perc. sweet sweet sour sour bitter bitter Sweet Sweet Sour Sour Bitter Bitter CA CAEigenvalue Eigenvalue % % Axis Axis Axis 2 Axis Sum Sum CA CA Eigenvalue Eigenvalue % % Axis Axis Axis 2 Axis Sum Sum Dim 2 (27.27%) Perc. bitter Perc. bitter Bitter Bitter Sweet Sweet Perc. sweet Perc. sweet Dim 2 (27.27%) Perc. sour Perc. sour Sour Sour Dim 2 (4.00%) Perc. bitter Perc. bitter Bitter Bitter Sweet Sweet Perc. sweet Perc. sweet Perc. sour Perc. sour Sour Sour Dim 2 (4.00%) Dim (72.73%) Dim (72.73%) Dim (96.00%) Dim (96.00%) 25 / 43
26 Inertia (= eigenvalues) Chemistery Economic Literature science Mathematics Medicine Peace Physics Canada France Germany Italy Japan Russia UK USA Inertia Inertia (%) F F F F F Sum λ = i.e., we are ar rom an exclusive association between certain rows and columns Φ 2 = we are ar rom a perect link. i.e., ar rom exclusive association between categories o the two variables 26 / 43
27 Correspondence Analysis (CA) Data 2 Independence model 3 Point clouds and what they mean 4 Percentage o inertia, and inertia in CA 5 Simultaneous representation o rows and columns 6 Interpretation aids 27 / 43
28 Simultaneous representation o rows and columns Transition ormulas = barycentric properties F s (i) = λs J ij i. G s (j) j= }{{} F s(i): coord. o row i on the s-th axis ij i. : j-th element o proile i G s(j): coord. o column j on the s-th axis λ s: inertia associated with s-th axis (in CA, λ s ) Along the s-th axis, we calculate the barycenter o each column, with column j given a weight ij / i. The smaller the λ s, the urther the barycenter rom the origin: / λ s G s (j) = λs I i= ij.j F s (i) 28 / 43
29 Simultaneous representation and inertia G s (j) = λs I i= ij.j F s (i) Sweet Sour Bitter Perc. Perc. Perc. Perc. sweet sour sweet bitter sour bitter Sweet Sweet Sour Sour Bitter Bitter Perc. Perc. Perc. Perc. sweet sour sweet bitter sour bitter Sweet Sour Bitter CA Eigenvalue CA Eigenvalue % % Axis Axis Axis 2 Axis Sum Sum CA Eigenvalue CA Eigenvalue % % Axis Axis Axis 2 Axis Sum Sum Dim 2 (27.27%) Perc. bitter Perc. bitter Bitter Bitter Sweet Sweet Perc. sweet Perc. sweet Dim 2 (27.27%) Perc. sour Perc. sour Sour Sour Dim 2 (4.00%) Perc. bitter Perc. bitter Bitter Bitter Sweet Sweet Perc. sour Perc. sweet Perc. sweet Perc. sour Sour Sour Dim 2 (4.00%) Dim (72.73%) Dim (72.73%) Dim (96.00%) Dim (96.00%) 29 / 43
30 Simultaneous representation and inertia G s (j) = λs I i= ij.j F s (i) Perc. Perc. sour bitter bitter (5/8 = 33 /.3 ) sour (3/8 = 23 /.3 ) = Perc. bitter Bitter (7/8 = 33 /.3 ) Perc. Perc. Perc. sweet sour bitter Sweet Sour 0 9 Bitter Perc. sour Sour (/8 = 23 /.3 ).6 λ =.375 = Perc. Perc. Perc. sweet sour bitter Sweet Sour Bitter λ =.042 = 2 Perc. bitter Bitter (5/8 = 33 /.3 ) Perc. sour Sour (3/8 = 23 /.3 ) Perc. Bitter Sweet Sour Bitter Perc. Sour ( λ = / 43
31 The barycentric property Dim 2 (24.60%) Economic science 0. Medicine 0.25 Physics 0.26 Chemistery 0.2 Peace 0.09 Italy Literature 0.09 Japan Dim (54.75%) Chemistery Economic Literature Medicine Peace Physics Italy Japan Mean proile / 43
32 The barycentric property Dim 2 (24.60%) Economic science 0.00 Medicine 0.3 Chemistery Peace Physics 0.04 Italy Literature 0.09 Japan Dim (54.75%) Chemistery Economic Literature Medicine Peace Physics Italy Japan Mean proile / 43
33 The barycentric property Dim 2 (24.60%) Economic science 0.05 Medicine 0.26 Physics Chemistery Peace 0.05 Italy Literature 0.32 Japan Dim (54.75%) Chemistery Economic Literature Medicine Peace Physics Italy Japan Mean proile / 43
34 The barycentric property Dim 2 (24.60%) Economic science Medicine USA Canada UK Physics Chemistry Peace Germany Japan Russia France Italy Literature Dim (54.75%) 32 / 43
35 Correspondence Analysis (CA) Data 2 Independence model 3 Point clouds and what they mean 4 Percentage o inertia, and inertia in CA 5 Simultaneous representation o rows and columns 6 Interpretation aids 33 / 43
36 Interpretation aid: quality o the representation An indicator o the quality o representation o a point (or cloud): projected inertia o M i on u s total inertia o M i = i.(oh s i )2 i. (OM i ) 2 = cos2 ( OM i, u s ) N I RI O = G I J M i s H i u s Unit vector o the axis s projection o M i on P This indicator show how much the deviation o a proile rom the mean proile is shown in an axis or plane 34 / 43
37 The quality o representation: example Sweet Sour Bitter Perc. Perc. Perc. sweet sour bitter CA Eigenvalue % Axis 96 Axis Sum Dim 2 (4.00%) Perc. bitter Bitter Sweet Perc. sweet Perc. sour Sour Quality o representation (cos²) Axis Axis2 Sweet Sour Bitter Perc. sweet Perc. sour Perc. bitter Dim (96.00%) Interpretation o the graph based on extreme points with good quality representations 35 / 43
38 Interpretation aid: contributions Absolute indicator: projected inertia o M i on u s = i. (OH s i )2 Relative indicator: proj. inertia o M i on u s inertia o axis s = i.(oh s i )2 λ s We can sum the contributions o several elements This shows how much we can consider that an axis is due to one or several elements Practical compromise between distance to the origin, and weights Useul in big tables or selecting a subset o elements when starting interpretation (jointly with the quality o representation) 36 / 43
39 Contributions : exemple Contributions: example X X2 X3 X4 a 0 0 b c d 0 0 Dim 2 (.54%) a X b X2 X3 c d X4 Inertia % % cum. Axis Axis Axis Axis Axis 2 a b c d Σ Dim (83.50%) The extreme points are not necessarily the ones that contribute most to axis construction / 43
40 Supplementary inormation G s (j) = λs I i= ij.j F s (i) Mathematics is on the French and Russian side, also the side o literature and peace, but opposite the sciences Dim 2 (24.60%) Economic science Canada USA Medicine UK Chemistry Physics Peace Germany Japan Mathematics Italy France Russia Literature Dim (54.75%) 38 / 43
41 Distributional equivalence Distributional equivalence: i rows with the same values are grouped, the CA results are totally equivalent (same or columns) Application to analysis o texts: Thanks to distributional equivalence, i two (or more) words are used in the same circumstances, their coordinates will be close together, so doing the analysis with both, or just one, is entirely equivalent very useul (to group singular and plural versions o words, verb conjugations, etc.) 39 / 43
42 Maximum number o axes, and Cramer s V Point cloud or rows: I points in a J-dimensional space J dim. but constraint (proiles) S J I points in at most I dim. S I = Φ 2 = min(i,j ) k= λ k min(i, J ) } S min(i, J ) leading to the idea o a bounded indicator o the link between 2 variables: Cramer s V = Φ 2 [0; ] min(i, J ) Nobel Prize Three tastes Three tastes Cramer s V 0.522/5 = /2 = /2 = / 43
43 Conclusions or our example Chemistery Economic science Literature Mathematics Medicine Peace Physics Canada France Germany Italy Japan Russia UK USA Dim 2 (24.60%) Economic science Peace Medicine USA Canada UK Physics Chemistry Germany Japan Russia France Italy Literature Dim (54.75%) CA gives a good summary visual o the deviation rom independence, which helps to understand the data table (and especially, big data tables) As or this data: Most o the deviation rom independence appears in the separation o Science vs The rest. And to a lesser extent: physics/chemistry vs economics The position o each country shows its strengths in getting various prizes 4 / 43
44 Conclusion To study the link between qualitative variables, we build a contingency table The link is ound in the dierence between this table, and what it would look like i there was independence Correspondence analysis: build a point cloud o rows (and o columns) whose total inertia measures the strength o the deviation rom independence break down this total inertia into a sequence o axes o decreasing importance, each representing some eature o the link between the variables visually represent the rows and columns in such a way that their position on the graph relects their participation in the deviation rom independence 42 / 43
45 Data Independence model Point clouds Inertia Simultaneous representation Interpretation aids Bibliography For more inormation in the same vein, have a look at this video: nalysis Series Husson Lê Pagès Exploratory Multivariate Analysis by Example Using R Second Edition 36 Chapman & Hall/CRC Computer Science & Data Analysis Series Exploratory Multivariate Analysis by Example Using R SECOND EDITION Husson F., Lê S. & Pagès J. (207) Exploratory Multivariate Analysis by Example Using R 2nd edition, 230 p., CRC/Press. François Husson Sébastien Lê Jérôme Pagès / 43
Correspondence Analysis
Correspondence Analysis Julie Josse, François Husson, Sébastien Lê Applied Mathematics Department, Agrocampus Ouest user-2008 Dortmund, August 11th 2008 1 / 22 History Theoretical principles: Fisher (1940)
More informationBiplots in Practice MICHAEL GREENACRE. Professor of Statistics at the Pompeu Fabra University. Chapter 6 Offprint
Biplots in Practice MICHAEL GREENACRE Proessor o Statistics at the Pompeu Fabra University Chapter 6 Oprint Principal Component Analysis Biplots First published: September 010 ISBN: 978-84-93846-8-6 Supporting
More informationAudio transcription of the Principal Component Analysis course
Audio transcription of the Principal Component Analysis course Part I. Data - Practicalities Slides 1 to 8 Pages 2 to 5 Part II. Studying individuals and variables Slides 9 to 26 Pages 6 to 13 Part III.
More informationHandling missing values in Multiple Factor Analysis
Handling missing values in Multiple Factor Analysis François Husson, Julie Josse Applied mathematics department, Agrocampus Ouest, Rennes, France Rabat, 26 March 2014 1 / 10 Multi-blocks data set Groups
More informationCORRESPONDENCE ANALYSIS
CORRESPONDENCE ANALYSIS INTUITIVE THEORETICAL PRESENTATION BASIC RATIONALE DATA PREPARATION INITIAL TRANSFORAMATION OF THE INPUT MATRIX INTO PROFILES DEFINITION OF GEOMETRIC CONCEPTS (MASS, DISTANCE AND
More informationReview (Probability & Linear Algebra)
Review (Probability & Linear Algebra) CE-725 : Statistical Pattern Recognition Sharif University of Technology Spring 2013 M. Soleymani Outline Axioms of probability theory Conditional probability, Joint
More informationDiscriminant Correspondence Analysis
Discriminant Correspondence Analysis Hervé Abdi Overview As the name indicates, discriminant correspondence analysis (DCA) is an extension of discriminant analysis (DA) and correspondence analysis (CA).
More informationSpace-time data. Simple space-time analyses. PM10 in space. PM10 in time
Space-time data Observations taken over space and over time Z(s, t): indexed by space, s, and time, t Here, consider geostatistical/time data Z(s, t) exists for all locations and all times May consider
More informationLecture 2: Data Analytics of Narrative
Lecture 2: Data Analytics of Narrative Data Analytics of Narrative: Pattern Recognition in Text, and Text Synthesis, Supported by the Correspondence Analysis Platform. This Lecture is presented in three
More informationAssignment 1 Math 5341 Linear Algebra Review. Give complete answers to each of the following questions. Show all of your work.
Assignment 1 Math 5341 Linear Algebra Review Give complete answers to each of the following questions Show all of your work Note: You might struggle with some of these questions, either because it has
More informationMultivariate Gaussian Analysis
BS2 Statistical Inference, Lecture 7, Hilary Term 2009 February 13, 2009 Marginal and conditional distributions For a positive definite covariance matrix Σ, the multivariate Gaussian distribution has density
More informationPHY 335 Data Analysis for Physicists
PHY 335 Data Analysis or Physicists Instructor: Kok Wai Ng Oice CP 171 Telephone 7 1782 e mail: kng@uky.edu Oice hour: Thursday 11:00 12:00 a.m. or by appointment Time: Tuesday and Thursday 9:30 10:45
More informationRATIONAL FUNCTIONS. Finding Asymptotes..347 The Domain Finding Intercepts Graphing Rational Functions
RATIONAL FUNCTIONS Finding Asymptotes..347 The Domain....350 Finding Intercepts.....35 Graphing Rational Functions... 35 345 Objectives The ollowing is a list o objectives or this section o the workbook.
More informationRegularized PCA to denoise and visualise data
Regularized PCA to denoise and visualise data Marie Verbanck Julie Josse François Husson Laboratoire de statistique, Agrocampus Ouest, Rennes, France CNAM, Paris, 16 janvier 2013 1 / 30 Outline 1 PCA 2
More informationCSC 576: Linear System
CSC 576: Linear System Ji Liu Department of Computer Science, University of Rochester September 3, 206 Linear Equations Consider solving linear equations where A R m n and b R n m and n could be extremely
More informationSupplementary Information Reconstructing propagation networks with temporal similarity
Supplementary Inormation Reconstructing propagation networks with temporal similarity Hao Liao and An Zeng I. SI NOTES A. Special range. The special range is actually due to two reasons: the similarity
More information(Analyse des Correspondances (AFC) (Simple) Correspondence Analysis (CA) Hervé Abdi CNAM STA201
u u (Analyse des Correspondances (AFC) (Simple) Correspondence Analysis (CA) How to Analye the... Hervé Abdi CNAM-04. STA0 www.utdallas.edu/ herve herve@utdallas.edu Page of 85 . u u Refresher: Weight
More informationData Preprocessing Tasks
Data Tasks 1 2 3 Data Reduction 4 We re here. 1 Dimensionality Reduction Dimensionality reduction is a commonly used approach for generating fewer features. Typically used because too many features can
More informationMultiple Correspondence Analysis
Multiple Correspondence Analysis 18 Up to now we have analysed the association between two categorical variables or between two sets of categorical variables where the row variables are different from
More informationmissmda: a package to handle missing values in Multivariate exploratory Data Analysis methods
missmda: a package to handle missing values in Multivariate exploratory Data Analysis methods Julie Josse & Francois Husson Applied Mathematics Department Agrocampus Rennes - France user! 2011, Warwick,
More informationMSE405 Microstructure Characterization XRD-1 Lab X-ray diffraction in crystallography
X-ray diraction in crystallography I. Goals Crystallography is the science that studies the structure (and structure-derived properties) o crystals. Among its many tools, X-ray diraction (XRD) has had
More informationHypothesis Testing hypothesis testing approach
Hypothesis Testing In this case, we d be trying to form an inference about that neighborhood: Do people there shop more often those people who are members of the larger population To ascertain this, we
More informationSingular Value Decomposition. 1 Singular Value Decomposition and the Four Fundamental Subspaces
Singular Value Decomposition This handout is a review of some basic concepts in linear algebra For a detailed introduction, consult a linear algebra text Linear lgebra and its pplications by Gilbert Strang
More informationAlgebra II Notes Inverse Functions Unit 1.2. Inverse of a Linear Function. Math Background
Algebra II Notes Inverse Functions Unit 1. Inverse o a Linear Function Math Background Previously, you Perormed operations with linear unctions Identiied the domain and range o linear unctions In this
More informationOrthogonality. 6.1 Orthogonal Vectors and Subspaces. Chapter 6
Chapter 6 Orthogonality 6.1 Orthogonal Vectors and Subspaces Recall that if nonzero vectors x, y R n are linearly independent then the subspace of all vectors αx + βy, α, β R (the space spanned by x and
More informationRoberto s Notes on Differential Calculus Chapter 8: Graphical analysis Section 1. Extreme points
Roberto s Notes on Dierential Calculus Chapter 8: Graphical analysis Section 1 Extreme points What you need to know already: How to solve basic algebraic and trigonometric equations. All basic techniques
More informationReview (probability, linear algebra) CE-717 : Machine Learning Sharif University of Technology
Review (probability, linear algebra) CE-717 : Machine Learning Sharif University of Technology M. Soleymani Fall 2012 Some slides have been adopted from Prof. H.R. Rabiee s and also Prof. R. Gutierrez-Osuna
More informationAlgebra of Principal Component Analysis
Algebra of Principal Component Analysis 3 Data: Y = 5 Centre each column on its mean: Y c = 7 6 9 y y = 3..6....6.8 3. 3.8.6 Covariance matrix ( variables): S = -----------Y n c ' Y 8..6 c =.6 5.8 Equation
More informationCOMP 408/508. Computer Vision Fall 2017 PCA for Recognition
COMP 408/508 Computer Vision Fall 07 PCA or Recognition Recall: Color Gradient by PCA v λ ( G G, ) x x x R R v, v : eigenvectors o D D with v ^v (, ) x x λ, λ : eigenvalues o D D with λ >λ v λ ( B B, )
More informationMa/CS 6b Class 23: Eigenvalues in Regular Graphs
Ma/CS 6b Class 3: Eigenvalues in Regular Graphs By Adam Sheffer Recall: The Spectrum of a Graph Consider a graph G = V, E and let A be the adjacency matrix of G. The eigenvalues of G are the eigenvalues
More informationX-ray Diffraction. Interaction of Waves Reciprocal Lattice and Diffraction X-ray Scattering by Atoms The Integrated Intensity
X-ray Diraction Interaction o Waves Reciprocal Lattice and Diraction X-ray Scattering by Atoms The Integrated Intensity Basic Principles o Interaction o Waves Periodic waves characteristic: Frequency :
More informationCorrespondence Analysis
STATGRAPHICS Rev. 7/6/009 Correspondence Analysis The Correspondence Analysis procedure creates a map of the rows and columns in a two-way contingency table for the purpose of providing insights into the
More informationGROUP THEORY PRIMER. New terms: tensor, rank k tensor, Young tableau, Young diagram, hook, hook length, factors over hooks rule
GROUP THEORY PRIMER New terms: tensor, rank k tensor, Young tableau, Young diagram, hook, hook length, factors over hooks rule 1. Tensor methods for su(n) To study some aspects of representations of a
More information-Principal components analysis is by far the oldest multivariate technique, dating back to the early 1900's; ecologists have used PCA since the
1 2 3 -Principal components analysis is by far the oldest multivariate technique, dating back to the early 1900's; ecologists have used PCA since the 1950's. -PCA is based on covariance or correlation
More informationStatistical Machine Learning
Statistical Machine Learning Christoph Lampert Spring Semester 2015/2016 // Lecture 12 1 / 36 Unsupervised Learning Dimensionality Reduction 2 / 36 Dimensionality Reduction Given: data X = {x 1,..., x
More informationThe Singular Value Decomposition (SVD) and Principal Component Analysis (PCA)
Chapter 5 The Singular Value Decomposition (SVD) and Principal Component Analysis (PCA) 5.1 Basics of SVD 5.1.1 Review of Key Concepts We review some key definitions and results about matrices that will
More informationNext is material on matrix rank. Please see the handout
B90.330 / C.005 NOTES for Wednesday 0.APR.7 Suppose that the model is β + ε, but ε does not have the desired variance matrix. Say that ε is normal, but Var(ε) σ W. The form of W is W w 0 0 0 0 0 0 w 0
More information7. Symmetric Matrices and Quadratic Forms
Linear Algebra 7. Symmetric Matrices and Quadratic Forms CSIE NCU 1 7. Symmetric Matrices and Quadratic Forms 7.1 Diagonalization of symmetric matrices 2 7.2 Quadratic forms.. 9 7.4 The singular value
More informationMATHEMATICS: PAPER I TRIAL EXAMINATION 28 AUGUST 2015
MATHEMATICS: PAPER I TRIAL EXAMINATION 8 AUGUST 015 TIME: 3 HOURS TOTAL: 150 MARKS EXAMINATION NUMBER: PLEASE READ THE FOLLOWING INSTRUCTIONS CAREFULLY 1. Write your examination number on the paper.. This
More information1. Select the unique answer (choice) for each problem. Write only the answer.
MATH 5 Practice Problem Set Spring 7. Select the unique answer (choice) for each problem. Write only the answer. () Determine all the values of a for which the system has infinitely many solutions: x +
More informationProbabilistic Model of Error in Fixed-Point Arithmetic Gaussian Pyramid
Probabilistic Model o Error in Fixed-Point Arithmetic Gaussian Pyramid Antoine Méler John A. Ruiz-Hernandez James L. Crowley INRIA Grenoble - Rhône-Alpes 655 avenue de l Europe 38 334 Saint Ismier Cedex
More informationMachine learning comes from Bayesian decision theory in statistics. There we want to minimize the expected value of the loss function.
Bayesian learning: Machine learning comes from Bayesian decision theory in statistics. There we want to minimize the expected value of the loss function. Let y be the true label and y be the predicted
More information14 Singular Value Decomposition
14 Singular Value Decomposition For any high-dimensional data analysis, one s first thought should often be: can I use an SVD? The singular value decomposition is an invaluable analysis tool for dealing
More informationExample Linear Algebra Competency Test
Example Linear Algebra Competency Test The 4 questions below are a combination of True or False, multiple choice, fill in the blank, and computations involving matrices and vectors. In the latter case,
More informationGAUSSIAN PROCESS TRANSFORMS
GAUSSIAN PROCESS TRANSFORMS Philip A. Chou Ricardo L. de Queiroz Microsoft Research, Redmond, WA, USA pachou@microsoft.com) Computer Science Department, Universidade de Brasilia, Brasilia, Brazil queiroz@ieee.org)
More informationFunctions. Essential Question What are some of the characteristics of the graph of a logarithmic function?
5. Logarithms and Logarithmic Functions Essential Question What are some o the characteristics o the graph o a logarithmic unction? Ever eponential unction o the orm () = b, where b is a positive real
More informationMATH 583A REVIEW SESSION #1
MATH 583A REVIEW SESSION #1 BOJAN DURICKOVIC 1. Vector Spaces Very quick review of the basic linear algebra concepts (see any linear algebra textbook): (finite dimensional) vector space (or linear space),
More informationFinal Review Sheet. B = (1, 1 + 3x, 1 + x 2 ) then 2 + 3x + 6x 2
Final Review Sheet The final will cover Sections Chapters 1,2,3 and 4, as well as sections 5.1-5.4, 6.1-6.2 and 7.1-7.3 from chapters 5,6 and 7. This is essentially all material covered this term. Watch
More informationSample Geometry. Edps/Soc 584, Psych 594. Carolyn J. Anderson
Sample Geometry Edps/Soc 584, Psych 594 Carolyn J. Anderson Department of Educational Psychology I L L I N O I S university of illinois at urbana-champaign c Board of Trustees, University of Illinois Spring
More informationToday. Introduction to optimization Definition and motivation 1-dimensional methods. Multi-dimensional methods. General strategies, value-only methods
Optimization Last time Root inding: deinition, motivation Algorithms: Bisection, alse position, secant, Newton-Raphson Convergence & tradeos Eample applications o Newton s method Root inding in > 1 dimension
More informationL2: Review of probability and statistics
Probability L2: Review of probability and statistics Definition of probability Axioms and properties Conditional probability Bayes theorem Random variables Definition of a random variable Cumulative distribution
More informationChapter 7: Symmetric Matrices and Quadratic Forms
Chapter 7: Symmetric Matrices and Quadratic Forms (Last Updated: December, 06) These notes are derived primarily from Linear Algebra and its applications by David Lay (4ed). A few theorems have been moved
More informationPhysics 2B Chapter 17 Notes - First Law of Thermo Spring 2018
Internal Energy o a Gas Work Done by a Gas Special Processes The First Law o Thermodynamics p Diagrams The First Law o Thermodynamics is all about the energy o a gas: how much energy does the gas possess,
More informationTable of Contents. Multivariate methods. Introduction II. Introduction I
Table of Contents Introduction Antti Penttilä Department of Physics University of Helsinki Exactum summer school, 04 Construction of multinormal distribution Test of multinormality with 3 Interpretation
More informationChapter 4: Factor Analysis
Chapter 4: Factor Analysis In many studies, we may not be able to measure directly the variables of interest. We can merely collect data on other variables which may be related to the variables of interest.
More informationExam 2. Jeremy Morris. March 23, 2006
Exam Jeremy Morris March 3, 006 4. Consider a bivariate normal population with µ 0, µ, σ, σ and ρ.5. a Write out the bivariate normal density. The multivariate normal density is defined by the following
More informationCurve Sketching. The process of curve sketching can be performed in the following steps:
Curve Sketching So ar you have learned how to ind st and nd derivatives o unctions and use these derivatives to determine where a unction is:. Increasing/decreasing. Relative extrema 3. Concavity 4. Points
More informationConsideration on Sensitivity for Multiple Correspondence Analysis
Consideration on Sensitivity for Multiple Correspondence Analysis Masaaki Ida Abstract Correspondence analysis is a popular research technique applicable to data and text mining, because the technique
More informationPrincipal Component Analysis
Principal Component Analysis Anders Øland David Christiansen 1 Introduction Principal Component Analysis, or PCA, is a commonly used multi-purpose technique in data analysis. It can be used for feature
More informationThe Deutsch-Jozsa Problem: De-quantization and entanglement
The Deutsch-Jozsa Problem: De-quantization and entanglement Alastair A. Abbott Department o Computer Science University o Auckland, New Zealand May 31, 009 Abstract The Deustch-Jozsa problem is one o the
More informationThe Singular Value Decomposition
The Singular Value Decomposition Philippe B. Laval KSU Fall 2015 Philippe B. Laval (KSU) SVD Fall 2015 1 / 13 Review of Key Concepts We review some key definitions and results about matrices that will
More informationFocus was on solving matrix inversion problems Now we look at other properties of matrices Useful when A represents a transformations.
Previously Focus was on solving matrix inversion problems Now we look at other properties of matrices Useful when A represents a transformations y = Ax Or A simply represents data Notion of eigenvectors,
More informationReduction of Random Variables in Structural Reliability Analysis
Reduction of Random Variables in Structural Reliability Analysis S. Adhikari and R. S. Langley Department of Engineering University of Cambridge Trumpington Street Cambridge CB2 1PZ (U.K.) February 21,
More informationComments on Problems. 3.1 This problem offers some practice in deriving utility functions from indifference curve specifications.
CHAPTER 3 PREFERENCES AND UTILITY These problems provide some practice in eamining utilit unctions b looking at indierence curve maps and at a ew unctional orms. The primar ocus is on illustrating the
More informationGRAPH SIGNAL PROCESSING: A STATISTICAL VIEWPOINT
GRAPH SIGNAL PROCESSING: A STATISTICAL VIEWPOINT Cha Zhang Joint work with Dinei Florêncio and Philip A. Chou Microsoft Research Outline Gaussian Markov Random Field Graph construction Graph transform
More informationMath 520 Exam 2 Topic Outline Sections 1 3 (Xiao/Dumas/Liaw) Spring 2008
Math 520 Exam 2 Topic Outline Sections 1 3 (Xiao/Dumas/Liaw) Spring 2008 Exam 2 will be held on Tuesday, April 8, 7-8pm in 117 MacMillan What will be covered The exam will cover material from the lectures
More informationLinear Dimensionality Reduction
Outline Hong Chang Institute of Computing Technology, Chinese Academy of Sciences Machine Learning Methods (Fall 2012) Outline Outline I 1 Introduction 2 Principal Component Analysis 3 Factor Analysis
More informationWhat is Principal Component Analysis?
What is Principal Component Analysis? Principal component analysis (PCA) Reduce the dimensionality of a data set by finding a new set of variables, smaller than the original set of variables Retains most
More informationSTATISTICAL LEARNING SYSTEMS
STATISTICAL LEARNING SYSTEMS LECTURE 8: UNSUPERVISED LEARNING: FINDING STRUCTURE IN DATA Institute of Computer Science, Polish Academy of Sciences Ph. D. Program 2013/2014 Principal Component Analysis
More informationExponential and Logarithmic. Functions CHAPTER The Algebra of Functions; Composite
CHAPTER 9 Exponential and Logarithmic Functions 9. The Algebra o Functions; Composite Functions 9.2 Inverse Functions 9.3 Exponential Functions 9.4 Exponential Growth and Decay Functions 9.5 Logarithmic
More informationNOTES ON MATRICES OF FULL COLUMN (ROW) RANK. Shayle R. Searle ABSTRACT
NOTES ON MATRICES OF FULL COLUMN (ROW) RANK Shayle R. Searle Biometrics Unit, Cornell University, Ithaca, N.Y. 14853 BU-1361-M August 1996 ABSTRACT A useful left (right) inverse of a full column (row)
More informationChapter 6 - Orthogonality
Chapter 6 - Orthogonality Maggie Myers Robert A. van de Geijn The University of Texas at Austin Orthogonality Fall 2009 http://z.cs.utexas.edu/wiki/pla.wiki/ 1 Orthogonal Vectors and Subspaces http://z.cs.utexas.edu/wiki/pla.wiki/
More informationTypes of Information. Topic 2 - Descriptive Statistics. Examples. Sample and Sample Size. Background Reading. Variables classified as STAT 511
Topic 2 - Descriptive Statistics STAT 511 Professor Bruce Craig Types of Information Variables classified as Categorical (qualitative) - variable classifies individual into one of several groups or categories
More informationInterpolation via Barycentric Coordinates
Interpolation via Barycentric Coordinates Pierre Alliez Inria Some material from D. Anisimov Outline Barycenter Convexity Barycentric coordinates For Simplices For point sets Inverse distance (Shepard)
More informationLecture 9: Vector Algebra
Lecture 9: Vector Algebra Linear combination of vectors Geometric interpretation Interpreting as Matrix-Vector Multiplication Span of a set of vectors Vector Spaces and Subspaces Linearly Independent/Dependent
More informationSHORT COMMUNICATION. Diversity partitions in 3-way sorting: functions, Venn diagram mappings, typical additive series, and examples
COMMUNITY ECOLOGY 7(): 53-57, 006 585-8553/$0.00 Akadémiai Kiadó, Budapest DOI: 0.556/ComEc.7.006.. SHORT COMMUNICATION Diversity partitions in 3-way sorting: unctions, Venn diagram mappings, typical additive
More informationMACHINE LEARNING. Methods for feature extraction and reduction of dimensionality: Probabilistic PCA and kernel PCA
1 MACHINE LEARNING Methods for feature extraction and reduction of dimensionality: Probabilistic PCA and kernel PCA 2 Practicals Next Week Next Week, Practical Session on Computer Takes Place in Room GR
More informationSolution: a) Let us consider the matrix form of the system, focusing on the augmented matrix: 0 k k + 1. k 2 1 = 0. k = 1
Exercise. Given the system of equations 8 < : x + y + z x + y + z k x + k y + z k where k R. a) Study the system in terms of the values of k. b) Find the solutions when k, using the reduced row echelon
More informationFitting functions to data
1 Fitting functions to data 1.1 Exact fitting 1.1.1 Introduction Suppose we have a set of real-number data pairs x i, y i, i = 1, 2,, N. These can be considered to be a set of points in the xy-plane. They
More informationMTH 464: Computational Linear Algebra
MTH 464: Computational Linear Algebra Lecture Outlines Exam 2 Material Prof. M. Beauregard Department of Mathematics & Statistics Stephen F. Austin State University March 2, 2018 Linear Algebra (MTH 464)
More informationAn Introduction to Ordination Connie Clark
An Introduction to Ordination Connie Clark Ordination is a collective term for multivariate techniques that adapt a multidimensional swarm of data points in such a way that when it is projected onto a
More informationMATH 829: Introduction to Data Mining and Analysis Principal component analysis
1/11 MATH 829: Introduction to Data Mining and Analysis Principal component analysis Dominique Guillot Departments of Mathematical Sciences University of Delaware April 4, 2016 Motivation 2/11 High-dimensional
More information2. ETA EVALUATIONS USING WEBER FUNCTIONS. Introduction
. ETA EVALUATIONS USING WEBER FUNCTIONS Introduction So ar we have seen some o the methods or providing eta evaluations that appear in the literature and we have seen some o the interesting properties
More informationLeast-Squares Spectral Analysis Theory Summary
Least-Squares Spectral Analysis Theory Summary Reerence: Mtamakaya, J. D. (2012). Assessment o Atmospheric Pressure Loading on the International GNSS REPRO1 Solutions Periodic Signatures. Ph.D. dissertation,
More informationGaussian random variables inr n
Gaussian vectors Lecture 5 Gaussian random variables inr n One-dimensional case One-dimensional Gaussian density with mean and standard deviation (called N, ): fx x exp. Proposition If X N,, then ax b
More informationMS-E2112 Multivariate Statistical Analysis (5cr) Lecture 5: Bivariate Correspondence Analysis
MS-E2112 Multivariate Statistical (5cr) Lecture 5: Bivariate Contents analysis is a PCA-type method appropriate for analyzing categorical variables. The aim in bivariate correspondence analysis is to
More informationUnit of Study: Physical Geography & Settlement Patterns; Culture & Civilizations; and The Spread of Ideas
6 th Grade Social Studies 1 st Nine Weeks TEKS Unit of Study: Physical Geography & Settlement Patterns; Culture & Civilizations; and The Spread of Ideas 6.1) History. The student understands that historical
More informationSyllabus Objective: 2.9 The student will sketch the graph of a polynomial, radical, or rational function.
Precalculus Notes: Unit Polynomial Functions Syllabus Objective:.9 The student will sketch the graph o a polynomial, radical, or rational unction. Polynomial Function: a unction that can be written in
More informationMultivariate Statistical Analysis
Multivariate Statistical Analysis Fall 2011 C. L. Williams, Ph.D. Lecture 4 for Applied Multivariate Analysis Outline 1 Eigen values and eigen vectors Characteristic equation Some properties of eigendecompositions
More informationSIO 211B, Rudnick, adapted from Davis 1
SIO 211B, Rudnick, adapted from Davis 1 XVII.Empirical orthogonal functions Often in oceanography we collect large data sets that are time series at a group of locations. Moored current meter arrays do
More informationAPPENDIX 1 ERROR ESTIMATION
1 APPENDIX 1 ERROR ESTIMATION Measurements are always subject to some uncertainties no matter how modern and expensive equipment is used or how careully the measurements are perormed These uncertainties
More informationPrinciple Components Analysis (PCA) Relationship Between a Linear Combination of Variables and Axes Rotation for PCA
Principle Components Analysis (PCA) Relationship Between a Linear Combination of Variables and Axes Rotation for PCA Principle Components Analysis: Uses one group of variables (we will call this X) In
More information14 : Theory of Variational Inference: Inner and Outer Approximation
10-708: Probabilistic Graphical Models 10-708, Spring 2014 14 : Theory of Variational Inference: Inner and Outer Approximation Lecturer: Eric P. Xing Scribes: Yu-Hsin Kuo, Amos Ng 1 Introduction Last lecture
More informationConceptual Questions for Review
Conceptual Questions for Review Chapter 1 1.1 Which vectors are linear combinations of v = (3, 1) and w = (4, 3)? 1.2 Compare the dot product of v = (3, 1) and w = (4, 3) to the product of their lengths.
More informationMultivariate Analysis of Ecological Data
Multivariate Analysis of Ecological Data MICHAEL GREENACRE Professor of Statistics at the Pompeu Fabra University in Barcelona, Spain RAUL PRIMICERIO Associate Professor of Ecology, Evolutionary Biology
More information(a) If A is a 3 by 4 matrix, what does this tell us about its nullspace? Solution: dim N(A) 1, since rank(a) 3. Ax =
. (5 points) (a) If A is a 3 by 4 matrix, what does this tell us about its nullspace? dim N(A), since rank(a) 3. (b) If we also know that Ax = has no solution, what do we know about the rank of A? C(A)
More informationStructure in Data. A major objective in data analysis is to identify interesting features or structure in the data.
Structure in Data A major objective in data analysis is to identify interesting features or structure in the data. The graphical methods are very useful in discovering structure. There are basically two
More informationUnit 7 Probability M2 13.1,2,4, 5,6
+ Unit 7 Probability M2 13.1,2,4, 5,6 7.1 Probability n Obj.: I will be able to determine the experimental and theoretical probabilities of an event, or its complement, occurring. n Vocabulary o Outcome
More informationA linear algebraic view of partition regular matrices
A linear algebraic view of partition regular matrices Leslie Hogben Jillian McLeod June 7, 3 4 5 6 7 8 9 Abstract Rado showed that a rational matrix is partition regular over N if and only if it satisfies
More information( x) f = where P and Q are polynomials.
9.8 Graphing Rational Functions Lets begin with a deinition. Deinition: Rational Function A rational unction is a unction o the orm ( ) ( ) ( ) P where P and Q are polynomials. Q An eample o a simple rational
More information