Karhunen-Loève Transform KLT. JanKees van der Poel D.Sc. Student, Mechanical Engineering

Size: px
Start display at page:

Download "Karhunen-Loève Transform KLT. JanKees van der Poel D.Sc. Student, Mechanical Engineering"

Transcription

1 Karhunen-Loève Transform KLT JanKees van der Poel D.Sc. Student, Mechanical Engineering

2 Karhunen-Loève Transform Has many names cited in literature: Karhunen-Loève Transform (KLT); Karhunen-Loève Decomposition (or Expansion); Principal (or Principle) Component Analysis (PCA); Principal (or Principle) Factor Analysis (PFA); Singular Value decomposition (SVD); Proper Orthogonal Decomposition (POD); 2 de janeiro de 2012 JanKees van der Poel, D.Sc. Student, Mechanical Enginering 2

3 Karhunen-Loève Transform Has many names cited in literature: Galerkin Method (this variation is used to find solutions to certain types of Partial Differential Equations, PDEs, specially in the field of Mechanical Engineering and electromechanically coupled systems); Hotelling Transform; and Collective Coordinates. 2 de janeiro de 2012 JanKees van der Poel, D.Sc. Student, Mechanical Enginering 3

4 Karhunen-Loève Transform Karhunen-Loève Transform (KLT) takes a given collection of data (an input collection) and creates an orthogonal basis (the KLT basis) for the data. An orthogonal basis for a space V is a set of mutually orthogonal vectors (in other words, they are linearly independent) {b i } that span the space V. Here is provided an overview of KLT for some specific type of input collections. 2 de janeiro de 2012 JanKees van der Poel, D.Sc. Student, Mechanical Enginering 4

5 Karhunen-Loève Transform Pearson (1901), Hotelling (1933), Kosambi (1943), Loève (1945), Karhunen (1946), Pougachev (1953) and Obukhov (1954) have been independently credited to the discovery of KLT under one of its many titles. KLT has applications in almost any scientific field. 2 de janeiro de 2012 JanKees van der Poel, D.Sc. Student, Mechanical Enginering 5

6 Karhunen-Loève Transform KLT has been widely used in: Studies of turbulence; Thermal/chemical reactions; Feed-forward and feedback control design applications (KLT is used to obtain a reduced order model for simulations or control design); Data analysis or compression (characterization of human faces, map generation by robots and freight traffic prediction); 2 de janeiro de 2012 JanKees van der Poel, D.Sc. Student, Mechanical Enginering 6

7 Karhunen-Loève Transform One of the most important mathematical matrix factorizations is what is called the Singular Value Decomposition (SVD). The Singular Value Decomposition has many useful properties desirable in many applications. The Principle Components Analysis (PCA) is an application of the SVD. It identifies patterns in data, expressing this data in a way as to highlight their similarities and differences. 2 de janeiro de 2012 JanKees van der Poel, D.Sc. Student, Mechanical Enginering 7

8 Karhunen-Loève Transform To make things easy, the name Principal Component Analysis (PCA) will be used from now on, instead of KLT or SVD. In our field of signal/image processing, this is the known name for the Karhunen-Loève Transform What is Principal Component Analysis? Patterns in data can be hard to find in high dimensional data (where the luxury of graphical representation is not available). 2 de janeiro de 2012 JanKees van der Poel, D.Sc. Student, Mechanical Enginering 8

9 Principal Component Analysis So, use PCA for analyzing the data. Once the data patterns where found, reduce the number of data dimensions (without much loss of information), by compressing the data (this makes more easy to visualize the hidden data pattern). The PCA basically analyzes the data in order to reduce its dimensions, eliminate superpositions and it better using linear combinations obtained from the original variables. 2 de janeiro de 2012 JanKees van der Poel, D.Sc. Student, Mechanical Enginering 9

10 Value Data Presentation Example: 53 blood and urine measurements from 65 people (33 alcoholics, 32 non-alcoholics). H-WBC H-RBC H-Hgb H-Hct H-MCV H-MCH H-MCHC A A A A A A A A A Matrix Format Measurement measurement Spectral Format 2 de janeiro de 2012 JanKees van der Poel, D.Sc. Student, Mechanical Enginering 10

11 M-EPI H-Bands C-LDH Data Presentation Univariate Person Trivariate Bivariate C-Triglycerides C-LDH C-Triglycerides 2 de janeiro de 2012 JanKees van der Poel, D.Sc. Student, Mechanical Enginering 11

12 Data Presentation Is there a better presentation than the common Cartesian axes? That is, do we really need a space with 53 dimensions to view the data? This rises the question of how to find the best low dimension space that conveys maximum useful information. The answer is Find the Principal Components! 2 de janeiro de 2012 JanKees van der Poel, D.Sc. Student, Mechanical Enginering 12

13 Principal Components All of the Principal Components (PCs) start at the origin of the ordinate axes. The first PC is the direction of maximum variance from origin. All subsequent PCs are orthogonal to the first PC, describing maximum residual variance. 2 de janeiro de 2012 JanKees van der Poel, D.Sc. Student, Mechanical Enginering 13

14 Algebraic Interpretation nd Case Let's say that m points in a space with n (n large) dimensions are given. Now, how does one project these m points on to a low dimensional space while preserving broad trends in the data, while also allowing it to be visualized? 2 de janeiro de 2012 JanKees van der Poel, D.Sc. Student, Mechanical Enginering 14

15 Algebraic Interpretation 1D Case Given m points in a n (n large) dimensional space, how does one project these m points on to a one dimensional space? Simply choose a line that fits the data so the points are spread out well along the line. 2 de janeiro de 2012 JanKees van der Poel, D.Sc. Student, Mechanical Enginering 15

16 Algebraic Interpretation 1D Case Formally, minimize the sum of squares of distances to the line. Why sum of squares? Because it allows fast minimization, assuming the line passes through zero! 2 de janeiro de 2012 JanKees van der Poel, D.Sc. Student, Mechanical Enginering 16

17 Algebraic Interpretation 1D Case Minimizing the sum of squares of distances to the line is the same as maximizing the sum of squares of the projections on that line. Many thanks to Pythagoras! 2 de janeiro de 2012 JanKees van der Poel, D.Sc. Student, Mechanical Enginering 17

18 Basic Mathematical Concepts Before getting to a description of PCA, this tutorial first introduces mathematical concepts that will be used in PCA: Standard deviation, covariance, and eigenvectors and eigenvalues This background knowledge is meant to make the PCA section very easy, but can be skipped if the concepts are already familiar. 2 de janeiro de 2012 JanKees van der Poel, D.Sc. Student, Mechanical Enginering 18

19 Standard Deviation The Standard Deviation (SD) of a data set is a measure of how spread out the data is. The average distance from the mean of the data set to a point. The datasets [0, 8, 12, 20] and [8, 9, 11, 12] have the same mean (that is 10) but are quite different. 2 de janeiro de 2012 JanKees van der Poel, D.Sc. Student, Mechanical Enginering 19

20 Standard Deviation By means of the Standard Deviation it is possible, in some way, to differentiate these two sets As expected, the first set ([0, 8, 12, 20]) has a much larger standard deviation than the second set ([8, 9, 11, 12]) due to the fact that the data is much more spread out from the mean. 2 de janeiro de 2012 JanKees van der Poel, D.Sc. Student, Mechanical Enginering 20

21 Variance Variance is another measure of the spread of data in a data set. In fact it is almost identical to the standard deviation. The only difference is that the variance is simply the standard deviation squared. Variance, in addition to Standard Deviation, was introduced to provide a solid platform from which the next section, covariance, can be launched. 2 de janeiro de 2012 JanKees van der Poel, D.Sc. Student, Mechanical Enginering 21

22 Covariance Both standard deviation and variance are purely one dimensional measures. However many data sets have more than one dimension. The aim of the statistical analysis of these kind of data sets is usually to see if there is any relationship between its dimensions. 2 de janeiro de 2012 JanKees van der Poel, D.Sc. Student, Mechanical Enginering 22

23 Covariance Standard deviation and variance only operate on one dimensional data, so it is only possible to calculate the standard deviation for each dimension of the data set independently of the other dimensions. However, it is useful to have a similar measure to find out how much the dimensions vary from the mean with respect to each other. 2 de janeiro de 2012 JanKees van der Poel, D.Sc. Student, Mechanical Enginering 23

24 Covariance Covariance is always calculated between two dimensions. With a 3D data (X, Y, Z), covariance is calculated between (X, Y), (X, Z) and (Y, Z). With a nd data set, [n!/2*(n-2)!] different covariance values can be calculated. The covariance calculated between a dimension and itself gives the variance. The covariance between (X, X), (Y, Y) and (Z, Z) gives the variance of the X, Y and Z dimensions. 2 de janeiro de 2012 JanKees van der Poel, D.Sc. Student, Mechanical Enginering 24

25 Covariance Matrix As an example, let s make up the covariance matrix for an imaginary 3 dimensional data set, with the usual dimensions x, y and z. In this case, the covariance matrix has three rows and three columns with these values: C cov( x, x) cov( y, x) cov( z, x) cov( x, cov( y, cov( z, cov( x, z) cov( y, z) cov( z, z) 2 de janeiro de 2012 JanKees van der Poel, D.Sc. Student, Mechanical Enginering 25 y) y) y)

26 Covariance Matrix Down the main diagonal, one can see that the covariance value is computed between one of the dimensions and itself (which are the variances for that dimension). Since cov(a,b) = cov(b,a), the covariance matrix is symmetrical about the main diagonal. 2 de janeiro de 2012 JanKees van der Poel, D.Sc. Student, Mechanical Enginering 26

27 Eigenvectors and Eigenvalues A vector v is an eigenvector of a square matrix (m by m) M if M*v (multiplication of the matrix M by the vector v) gives a multiple of v, i.e., a *v (multiplication of the scalar by the vector v). In this case, is called the eigenvalue of M that is associated to the eigenvector v. 2 de janeiro de 2012 JanKees van der Poel, D.Sc. Student, Mechanical Enginering 27

28 Eigenvector Properties Eigenvectors can only be found for square matrices. Not every square matrix has eigenvectors. An m by m matrix has m eigenvectors, given that they exist. For example, given a 3 by 3 matrix that has eigenvectors, there are three of them. 2 de janeiro de 2012 JanKees van der Poel, D.Sc. Student, Mechanical Enginering 28

29 Eigenvector Properties Even if the eigenvector is scaled by some amount before being multiplied, one still gets the same multiple of it as a result. This is because if a vector is scaled by some amount, all it is done is to make it longer, not changing its direction 2 de janeiro de 2012 JanKees van der Poel, D.Sc. Student, Mechanical Enginering 29

30 Eigenvector Properties All the eigenvectors of a matrix are perpendicular (orthogonal), i.e., at right angles to each other, no matter how many dimensions the matrix have. This is important because it means that the data can be expressed in terms of these perpendicular eigenvectors, instead of expressing them in terms of their axes. 2 de janeiro de 2012 JanKees van der Poel, D.Sc. Student, Mechanical Enginering 30

31 The PCA Method Step 1: Get some data to use in a simple example. I am going to use my own two dimensional data set. I have chosen a two dimensional data set because I can provide plots of the data to show what the PCA analysis is doing at each step. The data I have used is found in the next slide. 2 de janeiro de 2012 JanKees van der Poel, D.Sc. Student, Mechanical Enginering 31

32 The PCA Method The data used in this example is shown here. Data = alturas pesos de janeiro de 2012 JanKees van der Poel, D.Sc. Student, Mechanical Enginering 32

33 The PCA Method Step 2: Subtract the mean. For PCA to work properly, you have to subtract the mean from each of the data dimensions. The mean subtracted is the average across each dimension. All the x values have their mean value subtracted from them, as well as all the y values have their mean value subtracted from them. This produces a data set whose mean is zero. 2 de janeiro de 2012 JanKees van der Poel, D.Sc. Student, Mechanical Enginering 33

34 The PCA Method The data with its mean subtracted (adjusted data) is shown here. Both the data and the adjusted data are plotted in the next slide. Data = alturas pesos de janeiro de 2012 JanKees van der Poel, D.Sc. Student, Mechanical Enginering 34

35 The PCA Method 2 de janeiro de 2012 JanKees van der Poel, D.Sc. Student, Mechanical Enginering 35

36 The PCA Method Step 3: Calculate the covariance matrix. Since the data is two dimensional, the covariance matrix will have two rows and two columns: C One should notice that heights and weights do normally increase together. As the non-diagonal e l e m e n t s in t h i s covariance matrix are positive, we should expect that both x and y v a r i a b l e s increase together. 2 de janeiro de 2012 JanKees van der Poel, D.Sc. Student, Mechanical Enginering 36

37 The PCA Method Step 4: Calculate the eigenvectors and eigenvalues of the data matrix. In Matlab, this step is performed using eig (only for square matrices) or svd (matrices with any shape) commands. As the data matrix is not square, we only can use the svd command. The eigenvectors and eigenvalues are rather important, giving useful information about the data. 2 de janeiro de 2012 JanKees van der Poel, D.Sc. Student, Mechanical Enginering 37

38 The PCA Method Step 4: Calculate the eigenvectors and eigenvalues of the data matrix. Here are the eigenvectors, which are found along the diagonal of the matrix S, diag(s) in Matlab, and the eigenvalues: eigenvalues eigenvectors de janeiro de 2012 JanKees van der Poel, D.Sc. Student, Mechanical Enginering 38

39 The PCA Method Looking at the plot of the adjusted data shown here, one can see how it has quite a strong pattern. As expected from the covariance matrix (and from the common sense), both of the variables increase together. 2 de janeiro de 2012 JanKees van der Poel, D.Sc. Student, Mechanical Enginering 39

40 The PCA Method On top of the adjusted data I have plotted both eigenvectors as well (appearing as a red and a green line). As stated earlier, they are perpendicular to each other. More important than this is that they provide information about the data patterns. One of the eigenvectors goes right through the middle of the points, drawing a line of best fit. 2 de janeiro de 2012 JanKees van der Poel, D.Sc. Student, Mechanical Enginering 40

41 The PCA Method The first eigenvector (the one plotted in green) shows us that these two data sets are very related to each other along that line. The second eigenvector (the one plotted in red) gives the other, and less important, pattern in the data. It shows that all the points follow the main line, but are off to its side by some amount. 2 de janeiro de 2012 JanKees van der Poel, D.Sc. Student, Mechanical Enginering 41

42 The PCA Method By the process of taking the eigenvectors of the covariance matrix, we have been able to extract lines that characterize the data. The rest of the steps involve transforming the data so that this data is expressed in terms of these lines. 2 de janeiro de 2012 JanKees van der Poel, D.Sc. Student, Mechanical Enginering 42

43 The PCA Method Recalling the important aspects from the previous figure: Two lines are perpendicular to each other, being interchangeably orthogonal ; The eigenvectors provides us a way to see hidden patterns of the data; One of the eigenvectors draws a line which best fits to the data. 2 de janeiro de 2012 JanKees van der Poel, D.Sc. Student, Mechanical Enginering 43

44 The PCA Method Step 5: Choosing components and forming a feature vector. Here comes the notion of data compression and reduced dimensionality. Eigenvalues have different values: the highest one corresponds to the eigenvector that is the principal component of the data set (the most significant relationship between the data dimensions). 2 de janeiro de 2012 JanKees van der Poel, D.Sc. Student, Mechanical Enginering 44

45 The PCA Method Once the eigenvectors are found from the data matrix, they are ordered by their eigenvalues, from the highest to the lowest. This gives the components in order of significance. The components which are less significant can be ignored. Some information is lost but, if the eigenvalues are small, the amount lost is not too much. 2 de janeiro de 2012 JanKees van der Poel, D.Sc. Student, Mechanical Enginering 45

46 The PCA Method If some components are left out, the final data set will have less dimensions than the original. If the original data set has n dimensions and n eigenvectors are calculated (together with their eigenvalues) and only the first p eigenvectors are chosen, then the final data set will have only p dimensions. 2 de janeiro de 2012 JanKees van der Poel, D.Sc. Student, Mechanical Enginering 46

47 The PCA Method Now, what needs to be done is to form a feature vector (a fancy name for a matrix of vectors). This feature vector is constructed by taking the eigenvectors that are to be kept from the list of eigenvectors and form a matrix with them in the columns. eigenvector 1, Feature _ Vector eigenvector 2,, eigenvector n T 2 de janeiro de 2012 JanKees van der Poel, D.Sc. Student, Mechanical Enginering 47

48 The PCA Method Using the data set seen before, and the fact that there are two eigenvectors, there are two choices. One is to form a feature vector with both of the eigenvectors: eigenvectors de janeiro de 2012 JanKees van der Poel, D.Sc. Student, Mechanical Enginering 48

49 The PCA Method The other is to form a feature vector leaving out the smaller, less significant, component and only have a single column: eigenvalues eigenvectors Most significant eigenvector Most significant eigenvalue Less significant eigenvalue Less significant eigenvector 2 de janeiro de 2012 JanKees van der Poel, D.Sc. Student, Mechanical Enginering 49

50 The PCA Method In other words, the result is a feature vector with p vectors, selected from n eigenvectors (where p < n). This is the most common option. eigenvalue Most significant eigenvalue eigenvector Most significant eigenvector 2 de janeiro de 2012 JanKees van der Poel, D.Sc. Student, Mechanical Enginering 50

51 The PCA Method Step 6: Deriving the new data set. This the final step in PCA (and the easiest one). Chose the components (eigenvectors) to be kept in the data set and form a feature vector. Just remember that the eigenvector with the highest eigenvalue is the principal component of the data set. Take the transpose of the vector and multiply it on the left of the transposed original data set. 2 de janeiro de 2012 JanKees van der Poel, D.Sc. Student, Mechanical Enginering 51

52 The PCA Method Final _ Data RowFeatureVector RowDataAdjusted The matrix called RowFeaureVector has the transposed eigenvectors in its columns. The eigenvectors are now in the rows, with the most significant one at the top. The matrix called RowDataAdusted has the transposed mean adjusted data in its columns. The data items are in each column, each row holding a separate dimension. 2 de janeiro de 2012 JanKees van der Poel, D.Sc. Student, Mechanical Enginering 52

53 The PCA Method This sudden transpose of all data is confusing, but equations from now on are easier if the transpose of the feature vector and the data is taken first. Better that having to always carry a little T symbol above their names! Final_Data is the final data set, with data items in columns, and dimensions along rows. 2 de janeiro de 2012 JanKees van der Poel, D.Sc. Student, Mechanical Enginering 53

54 The PCA Method The original data is now only given in terms of the chosen vectors. The original data set was written in terms of the x and y axes. The data can be expressed in terms of any axes, but the expression is most efficient if these axes are perpendicular. This is why it was important that eigenvectors are always perpendicular to each other. 2 de janeiro de 2012 JanKees van der Poel, D.Sc. Student, Mechanical Enginering 54

55 The PCA Method So, the original data (expressed in terms of the x and y axes ) is now expressed in terms of the eigenvectors found. If a reduced dimension is needed (throwing some of the eigenvectors out), the new data will be expressed in terms of the vectors that were kept. 2 de janeiro de 2012 JanKees van der Poel, D.Sc. Student, Mechanical Enginering 55

56 The PCA Method 2 de janeiro de 2012 JanKees van der Poel, D.Sc. Student, Mechanical Enginering 56

57 The PCA Method Among all possible orthogonal transforms, PCA is optimal in the following sense: KLT completely decorrelates the signal; and KLT maximally compacts the energy (in other words, the information) contained in the signal. But the PCA is computationally expensive and is not supposed to be used carelessly. Instead, one can use the Discrete Cosine Transform, DCT, which approaches the KLT in this sense. 2 de janeiro de 2012 JanKees van der Poel, D.Sc. Student, Mechanical Enginering 57

58 The PCA Method Examples Here, we switch to Matlab in order to run some examples that (I sincerely hope) may clarify the things to you: Project the data into the principal component axis, show the rank one approximation, and compress an image by reducing the number of its coefficients (PCA.m), pretty much as by using the DCT. Show the difference between the least squares and the PCA and do the alignment of 3D models using the PCA properties (SVD.m). 2 de janeiro de 2012 JanKees van der Poel, D.Sc. Student, Mechanical Enginering 58

59 The PCA Method Examples Some things should be noticed about the power of PCA to compress an image (as seen in the PCA.m example). The amount of memory required to store an uncompressed image of size m n is M image = m*n. So, notice that the amount of memory we need to store an image increases exponentially as its dimensions get larger. 2 de janeiro de 2012 JanKees van der Poel, D.Sc. Student, Mechanical Enginering 59

60 The PCA Method Examples But, the amount of memory required to store an SVD image (also of size m n) approximation using rank k is M approx = k(m + n + 1). So, notice that the amount of memory required increases linearly as the dimensions get larger, as opposed to exponentially. Thus, as the image gets larger, more memory is saved by using SVD. 2 de janeiro de 2012 JanKees van der Poel, D.Sc. Student, Mechanical Enginering 60

61 The PCA Method Examples Perform face recognition using the Principal Component Analysis approach! This is accomplished using a technique known in the literature by the Eigenface Technique. We will see an example of how to do it using a well known Face Database called The AT & T Faces Database. Two Matlab functions: facerecognitionexample.m and loadfacedatabase.m. 2 de janeiro de 2012 JanKees van der Poel, D.Sc. Student, Mechanical Enginering 61

62 What is the Eigenface Technique? The idea is that face images can be economically represented by their projection onto a small number of basis images derived by finding the most significant eigenvectors of the pixel wise covariance matrix for a set of training images. A lot of people like to play with this technique, but in my tutorial I will simply show how to get some eigenfaces and play with them in Matlab. 2 de janeiro de 2012 JanKees van der Poel, D.Sc. Student, Mechanical Enginering 62

63 AT&T Database of Faces AT&T Database of Faces contains a set of face images. Database used in the context of a face recognition project. Ten different images of 40 distinct subjects taken at different times (varying lighting, facial details and expressions) and against a dark homogeneous background with subjects in an upright, frontal position (some side movement was tolerated). 2 de janeiro de 2012 JanKees van der Poel, D.Sc. Student, Mechanical Enginering 63

64 AT&T Database of Faces The images have a size of 92x112 pixels (in other words, pixels) and 256 grey levels per pixel, organized in 40 directories (one for each subject) and each directory contains ten different images of a subject. Matlab can read PNG files and other formats without help. So, it is relatively easy to load all face database into Matlab s workspace and process it. 2 de janeiro de 2012 JanKees van der Poel, D.Sc. Student, Mechanical Enginering 64

65 Getting The Faces Into One Big Matrix First of all, we need to put all the faces of the database in one huge matrix with a size of 112*92 = lines and 400 columns. This step is done by the function called loadfacedatabase.m. It reads a bunch of images, makes column vectors out of each of one of them, put all together and return the result. 2 de janeiro de 2012 JanKees van der Poel, D.Sc. Student, Mechanical Enginering 65

66 Getting the Recognition to Work Here we change to Matlab directly, because the steps we do to perform the face recognition task are better explained seeing the function called facerecognitionexample.m. All the steps necessary to perform this task are done in this function and it is ready to be executed and commented. 2 de janeiro de 2012 JanKees van der Poel, D.Sc. Student, Mechanical Enginering 66

67 Cases When PCA Fail (1) PCA projects data onto a set of orthogonal vectors (principle components). This restricts the new input components to be a linear combination of old ones. However, there are cases where the intrinsic freedom of data can not be expressed as a linear combination of input components In such cases PCA will overestimate the input dimensionality. 2 de janeiro de 2012 JanKees van der Poel, D.Sc. Student, Mechanical Enginering 67

68 Cases When PCA Fail (1) So, PCA does is not capable to find the nonlinear intrinsic dimension of data (like the angle between the two vectors in the example above). Instead, it will find out two components with equal importance. 2 de janeiro de 2012 JanKees van der Poel, D.Sc. Student, Mechanical Enginering 68

69 Cases When PCA Fail (2) In cases when components with small variability really matter, PCA will make mistakes due to its unsupervised nature. In such cases, if we only consider the projections of two classes of data as input, they will become indistinguishable. 2 de janeiro de 2012 JanKees van der Poel, D.Sc. Student, Mechanical Enginering 69

70 Any (Reasonable) Doubts? 2 de janeiro de 2012 JanKees van der Poel, D.Sc. Student, Mechanical Enginering 70

Robot Image Credit: Viktoriya Sukhanova 123RF.com. Dimensionality Reduction

Robot Image Credit: Viktoriya Sukhanova 123RF.com. Dimensionality Reduction Robot Image Credit: Viktoriya Sukhanova 13RF.com Dimensionality Reduction Feature Selection vs. Dimensionality Reduction Feature Selection (last time) Select a subset of features. When classifying novel

More information

Lecture 24: Principal Component Analysis. Aykut Erdem May 2016 Hacettepe University

Lecture 24: Principal Component Analysis. Aykut Erdem May 2016 Hacettepe University Lecture 4: Principal Component Analysis Aykut Erdem May 016 Hacettepe University This week Motivation PCA algorithms Applications PCA shortcomings Autoencoders Kernel PCA PCA Applications Data Visualization

More information

Advanced Introduction to Machine Learning CMU-10715

Advanced Introduction to Machine Learning CMU-10715 Advanced Introduction to Machine Learning CMU-10715 Principal Component Analysis Barnabás Póczos Contents Motivation PCA algorithms Applications Some of these slides are taken from Karl Booksh Research

More information

A Tutorial on Data Reduction. Principal Component Analysis Theoretical Discussion. By Shireen Elhabian and Aly Farag

A Tutorial on Data Reduction. Principal Component Analysis Theoretical Discussion. By Shireen Elhabian and Aly Farag A Tutorial on Data Reduction Principal Component Analysis Theoretical Discussion By Shireen Elhabian and Aly Farag University of Louisville, CVIP Lab November 2008 PCA PCA is A backbone of modern data

More information

A tutorial on Principal Components Analysis

A tutorial on Principal Components Analysis A tutorial on Principal Components Analysis Lindsay I Smith February 26, 2002 Chapter 1 Introduction This tutorial is designed to give the reader an understanding of Principal Components Analysis (PCA).

More information

Unsupervised Learning: K- Means & PCA

Unsupervised Learning: K- Means & PCA Unsupervised Learning: K- Means & PCA Unsupervised Learning Supervised learning used labeled data pairs (x, y) to learn a func>on f : X Y But, what if we don t have labels? No labels = unsupervised learning

More information

Dimensionality Reduction: PCA. Nicholas Ruozzi University of Texas at Dallas

Dimensionality Reduction: PCA. Nicholas Ruozzi University of Texas at Dallas Dimensionality Reduction: PCA Nicholas Ruozzi University of Texas at Dallas Eigenvalues λ is an eigenvalue of a matrix A R n n if the linear system Ax = λx has at least one non-zero solution If Ax = λx

More information

Principal Component Analysis

Principal Component Analysis Principal Component Analysis Anders Øland David Christiansen 1 Introduction Principal Component Analysis, or PCA, is a commonly used multi-purpose technique in data analysis. It can be used for feature

More information

Principal Component Analysis

Principal Component Analysis B: Chapter 1 HTF: Chapter 1.5 Principal Component Analysis Barnabás Póczos University of Alberta Nov, 009 Contents Motivation PCA algorithms Applications Face recognition Facial expression recognition

More information

Principal Component Analysis -- PCA (also called Karhunen-Loeve transformation)

Principal Component Analysis -- PCA (also called Karhunen-Loeve transformation) Principal Component Analysis -- PCA (also called Karhunen-Loeve transformation) PCA transforms the original input space into a lower dimensional space, by constructing dimensions that are linear combinations

More information

CS4495/6495 Introduction to Computer Vision. 8B-L2 Principle Component Analysis (and its use in Computer Vision)

CS4495/6495 Introduction to Computer Vision. 8B-L2 Principle Component Analysis (and its use in Computer Vision) CS4495/6495 Introduction to Computer Vision 8B-L2 Principle Component Analysis (and its use in Computer Vision) Wavelength 2 Wavelength 2 Principal Components Principal components are all about the directions

More information

1 Singular Value Decomposition and Principal Component

1 Singular Value Decomposition and Principal Component Singular Value Decomposition and Principal Component Analysis In these lectures we discuss the SVD and the PCA, two of the most widely used tools in machine learning. Principal Component Analysis (PCA)

More information

ISSN: (Online) Volume 3, Issue 5, May 2015 International Journal of Advance Research in Computer Science and Management Studies

ISSN: (Online) Volume 3, Issue 5, May 2015 International Journal of Advance Research in Computer Science and Management Studies ISSN: 2321-7782 (Online) Volume 3, Issue 5, May 2015 International Journal of Advance Research in Computer Science and Management Studies Research Article / Survey Paper / Case Study Available online at:

More information

Keywords Eigenface, face recognition, kernel principal component analysis, machine learning. II. LITERATURE REVIEW & OVERVIEW OF PROPOSED METHODOLOGY

Keywords Eigenface, face recognition, kernel principal component analysis, machine learning. II. LITERATURE REVIEW & OVERVIEW OF PROPOSED METHODOLOGY Volume 6, Issue 3, March 2016 ISSN: 2277 128X International Journal of Advanced Research in Computer Science and Software Engineering Research Paper Available online at: www.ijarcsse.com Eigenface and

More information

Data Preprocessing Tasks

Data Preprocessing Tasks Data Tasks 1 2 3 Data Reduction 4 We re here. 1 Dimensionality Reduction Dimensionality reduction is a commonly used approach for generating fewer features. Typically used because too many features can

More information

Computational paradigms for the measurement signals processing. Metodologies for the development of classification algorithms.

Computational paradigms for the measurement signals processing. Metodologies for the development of classification algorithms. Computational paradigms for the measurement signals processing. Metodologies for the development of classification algorithms. January 5, 25 Outline Methodologies for the development of classification

More information

Singular Value Decomposition. 1 Singular Value Decomposition and the Four Fundamental Subspaces

Singular Value Decomposition. 1 Singular Value Decomposition and the Four Fundamental Subspaces Singular Value Decomposition This handout is a review of some basic concepts in linear algebra For a detailed introduction, consult a linear algebra text Linear lgebra and its pplications by Gilbert Strang

More information

Deriving Principal Component Analysis (PCA)

Deriving Principal Component Analysis (PCA) -0 Mathematical Foundations for Machine Learning Machine Learning Department School of Computer Science Carnegie Mellon University Deriving Principal Component Analysis (PCA) Matt Gormley Lecture 11 Oct.

More information

Principal Components Analysis (PCA)

Principal Components Analysis (PCA) Principal Components Analysis (PCA) Principal Components Analysis (PCA) a technique for finding patterns in data of high dimension Outline:. Eigenvectors and eigenvalues. PCA: a) Getting the data b) Centering

More information

Principal Component Analysis CS498

Principal Component Analysis CS498 Principal Component Analysis CS498 Today s lecture Adaptive Feature Extraction Principal Component Analysis How, why, when, which A dual goal Find a good representation The features part Reduce redundancy

More information

Linear Algebra & Geometry why is linear algebra useful in computer vision?

Linear Algebra & Geometry why is linear algebra useful in computer vision? Linear Algebra & Geometry why is linear algebra useful in computer vision? References: -Any book on linear algebra! -[HZ] chapters 2, 4 Some of the slides in this lecture are courtesy to Prof. Octavia

More information

Lecture: Face Recognition and Feature Reduction

Lecture: Face Recognition and Feature Reduction Lecture: Face Recognition and Feature Reduction Juan Carlos Niebles and Ranjay Krishna Stanford Vision and Learning Lab 1 Recap - Curse of dimensionality Assume 5000 points uniformly distributed in the

More information

7. Variable extraction and dimensionality reduction

7. Variable extraction and dimensionality reduction 7. Variable extraction and dimensionality reduction The goal of the variable selection in the preceding chapter was to find least useful variables so that it would be possible to reduce the dimensionality

More information

Principal Component Analysis (PCA)

Principal Component Analysis (PCA) Principal Component Analysis (PCA) Salvador Dalí, Galatea of the Spheres CSC411/2515: Machine Learning and Data Mining, Winter 2018 Michael Guerzhoy and Lisa Zhang Some slides from Derek Hoiem and Alysha

More information

Linear Algebra Review. Fei-Fei Li

Linear Algebra Review. Fei-Fei Li Linear Algebra Review Fei-Fei Li 1 / 37 Vectors Vectors and matrices are just collections of ordered numbers that represent something: movements in space, scaling factors, pixel brightnesses, etc. A vector

More information

CS168: The Modern Algorithmic Toolbox Lecture #8: How PCA Works

CS168: The Modern Algorithmic Toolbox Lecture #8: How PCA Works CS68: The Modern Algorithmic Toolbox Lecture #8: How PCA Works Tim Roughgarden & Gregory Valiant April 20, 206 Introduction Last lecture introduced the idea of principal components analysis (PCA). The

More information

Focus was on solving matrix inversion problems Now we look at other properties of matrices Useful when A represents a transformations.

Focus was on solving matrix inversion problems Now we look at other properties of matrices Useful when A represents a transformations. Previously Focus was on solving matrix inversion problems Now we look at other properties of matrices Useful when A represents a transformations y = Ax Or A simply represents data Notion of eigenvectors,

More information

Machine Learning (Spring 2012) Principal Component Analysis

Machine Learning (Spring 2012) Principal Component Analysis 1-71 Machine Learning (Spring 1) Principal Component Analysis Yang Xu This note is partly based on Chapter 1.1 in Chris Bishop s book on PRML and the lecture slides on PCA written by Carlos Guestrin in

More information

Eigenimaging for Facial Recognition

Eigenimaging for Facial Recognition Eigenimaging for Facial Recognition Aaron Kosmatin, Clayton Broman December 2, 21 Abstract The interest of this paper is Principal Component Analysis, specifically its area of application to facial recognition

More information

PCA, Kernel PCA, ICA

PCA, Kernel PCA, ICA PCA, Kernel PCA, ICA Learning Representations. Dimensionality Reduction. Maria-Florina Balcan 04/08/2015 Big & High-Dimensional Data High-Dimensions = Lot of Features Document classification Features per

More information

Principal Component Analysis

Principal Component Analysis Principal Component Analysis Yingyu Liang yliang@cs.wisc.edu Computer Sciences Department University of Wisconsin, Madison [based on slides from Nina Balcan] slide 1 Goals for the lecture you should understand

More information

Lecture: Face Recognition and Feature Reduction

Lecture: Face Recognition and Feature Reduction Lecture: Face Recognition and Feature Reduction Juan Carlos Niebles and Ranjay Krishna Stanford Vision and Learning Lab Lecture 11-1 Recap - Curse of dimensionality Assume 5000 points uniformly distributed

More information

14 Singular Value Decomposition

14 Singular Value Decomposition 14 Singular Value Decomposition For any high-dimensional data analysis, one s first thought should often be: can I use an SVD? The singular value decomposition is an invaluable analysis tool for dealing

More information

CS 4495 Computer Vision Principle Component Analysis

CS 4495 Computer Vision Principle Component Analysis CS 4495 Computer Vision Principle Component Analysis (and it s use in Computer Vision) Aaron Bobick School of Interactive Computing Administrivia PS6 is out. Due *** Sunday, Nov 24th at 11:55pm *** PS7

More information

Linear Algebra Review. Fei-Fei Li

Linear Algebra Review. Fei-Fei Li Linear Algebra Review Fei-Fei Li 1 / 51 Vectors Vectors and matrices are just collections of ordered numbers that represent something: movements in space, scaling factors, pixel brightnesses, etc. A vector

More information

Image Registration Lecture 2: Vectors and Matrices

Image Registration Lecture 2: Vectors and Matrices Image Registration Lecture 2: Vectors and Matrices Prof. Charlene Tsai Lecture Overview Vectors Matrices Basics Orthogonal matrices Singular Value Decomposition (SVD) 2 1 Preliminary Comments Some of this

More information

Introduction to Machine Learning

Introduction to Machine Learning 10-701 Introduction to Machine Learning PCA Slides based on 18-661 Fall 2018 PCA Raw data can be Complex, High-dimensional To understand a phenomenon we measure various related quantities If we knew what

More information

Lecture 13. Principal Component Analysis. Brett Bernstein. April 25, CDS at NYU. Brett Bernstein (CDS at NYU) Lecture 13 April 25, / 26

Lecture 13. Principal Component Analysis. Brett Bernstein. April 25, CDS at NYU. Brett Bernstein (CDS at NYU) Lecture 13 April 25, / 26 Principal Component Analysis Brett Bernstein CDS at NYU April 25, 2017 Brett Bernstein (CDS at NYU) Lecture 13 April 25, 2017 1 / 26 Initial Question Intro Question Question Let S R n n be symmetric. 1

More information

Announcements (repeat) Principal Components Analysis

Announcements (repeat) Principal Components Analysis 4/7/7 Announcements repeat Principal Components Analysis CS 5 Lecture #9 April 4 th, 7 PA4 is due Monday, April 7 th Test # will be Wednesday, April 9 th Test #3 is Monday, May 8 th at 8AM Just hour long

More information

Quantitative Understanding in Biology Principal Components Analysis

Quantitative Understanding in Biology Principal Components Analysis Quantitative Understanding in Biology Principal Components Analysis Introduction Throughout this course we have seen examples of complex mathematical phenomena being represented as linear combinations

More information

Signal Analysis. Principal Component Analysis

Signal Analysis. Principal Component Analysis Multi dimensional Signal Analysis Lecture 2E Principal Component Analysis Subspace representation Note! Given avector space V of dimension N a scalar product defined by G 0 a subspace U of dimension M

More information

Introduction to Machine Learning. PCA and Spectral Clustering. Introduction to Machine Learning, Slides: Eran Halperin

Introduction to Machine Learning. PCA and Spectral Clustering. Introduction to Machine Learning, Slides: Eran Halperin 1 Introduction to Machine Learning PCA and Spectral Clustering Introduction to Machine Learning, 2013-14 Slides: Eran Halperin Singular Value Decomposition (SVD) The singular value decomposition (SVD)

More information

Principal Component Analysis (PCA)

Principal Component Analysis (PCA) Principal Component Analysis (PCA) Additional reading can be found from non-assessed exercises (week 8) in this course unit teaching page. Textbooks: Sect. 6.3 in [1] and Ch. 12 in [2] Outline Introduction

More information

Linear Algebra & Geometry why is linear algebra useful in computer vision?

Linear Algebra & Geometry why is linear algebra useful in computer vision? Linear Algebra & Geometry why is linear algebra useful in computer vision? References: -Any book on linear algebra! -[HZ] chapters 2, 4 Some of the slides in this lecture are courtesy to Prof. Octavia

More information

Singular Value Decomposition and Digital Image Compression

Singular Value Decomposition and Digital Image Compression Singular Value Decomposition and Digital Image Compression Chris Bingham December 1, 016 Page 1 of Abstract The purpose of this document is to be a very basic introduction to the singular value decomposition

More information

PCA FACE RECOGNITION

PCA FACE RECOGNITION PCA FACE RECOGNITION The slides are from several sources through James Hays (Brown); Srinivasa Narasimhan (CMU); Silvio Savarese (U. of Michigan); Shree Nayar (Columbia) including their own slides. Goal

More information

The Singular Value Decomposition (SVD) and Principal Component Analysis (PCA)

The Singular Value Decomposition (SVD) and Principal Component Analysis (PCA) Chapter 5 The Singular Value Decomposition (SVD) and Principal Component Analysis (PCA) 5.1 Basics of SVD 5.1.1 Review of Key Concepts We review some key definitions and results about matrices that will

More information

15 Singular Value Decomposition

15 Singular Value Decomposition 15 Singular Value Decomposition For any high-dimensional data analysis, one s first thought should often be: can I use an SVD? The singular value decomposition is an invaluable analysis tool for dealing

More information

PCA & ICA. CE-717: Machine Learning Sharif University of Technology Spring Soleymani

PCA & ICA. CE-717: Machine Learning Sharif University of Technology Spring Soleymani PCA & ICA CE-717: Machine Learning Sharif University of Technology Spring 2015 Soleymani Dimensionality Reduction: Feature Selection vs. Feature Extraction Feature selection Select a subset of a given

More information

CSC 411 Lecture 12: Principal Component Analysis

CSC 411 Lecture 12: Principal Component Analysis CSC 411 Lecture 12: Principal Component Analysis Roger Grosse, Amir-massoud Farahmand, and Juan Carrasquilla University of Toronto UofT CSC 411: 12-PCA 1 / 23 Overview Today we ll cover the first unsupervised

More information

Example: Face Detection

Example: Face Detection Announcements HW1 returned New attendance policy Face Recognition: Dimensionality Reduction On time: 1 point Five minutes or more late: 0.5 points Absent: 0 points Biometrics CSE 190 Lecture 14 CSE190,

More information

Covariance and Principal Components

Covariance and Principal Components COMP3204/COMP6223: Computer Vision Covariance and Principal Components Jonathon Hare jsh2@ecs.soton.ac.uk Variance and Covariance Random Variables and Expected Values Mathematicians talk variance (and

More information

PRINCIPAL COMPONENT ANALYSIS

PRINCIPAL COMPONENT ANALYSIS PRINCIPAL COMPONENT ANALYSIS Dimensionality Reduction Tzompanaki Katerina Dimensionality Reduction Unsupervised learning Goal: Find hidden patterns in the data. Used for Visualization Data compression

More information

22.3. Repeated Eigenvalues and Symmetric Matrices. Introduction. Prerequisites. Learning Outcomes

22.3. Repeated Eigenvalues and Symmetric Matrices. Introduction. Prerequisites. Learning Outcomes Repeated Eigenvalues and Symmetric Matrices. Introduction In this Section we further develop the theory of eigenvalues and eigenvectors in two distinct directions. Firstly we look at matrices where one

More information

CPSC 340: Machine Learning and Data Mining. More PCA Fall 2017

CPSC 340: Machine Learning and Data Mining. More PCA Fall 2017 CPSC 340: Machine Learning and Data Mining More PCA Fall 2017 Admin Assignment 4: Due Friday of next week. No class Monday due to holiday. There will be tutorials next week on MAP/PCA (except Monday).

More information

EE731 Lecture Notes: Matrix Computations for Signal Processing

EE731 Lecture Notes: Matrix Computations for Signal Processing EE731 Lecture Notes: Matrix Computations for Signal Processing James P. Reilly c Department of Electrical and Computer Engineering McMaster University September 22, 2005 0 Preface This collection of ten

More information

Dimensionality Reduction

Dimensionality Reduction Lecture 5 1 Outline 1. Overview a) What is? b) Why? 2. Principal Component Analysis (PCA) a) Objectives b) Explaining variability c) SVD 3. Related approaches a) ICA b) Autoencoders 2 Example 1: Sportsball

More information

Conceptual Questions for Review

Conceptual Questions for Review Conceptual Questions for Review Chapter 1 1.1 Which vectors are linear combinations of v = (3, 1) and w = (4, 3)? 1.2 Compare the dot product of v = (3, 1) and w = (4, 3) to the product of their lengths.

More information

Vectors To begin, let us describe an element of the state space as a point with numerical coordinates, that is x 1. x 2. x =

Vectors To begin, let us describe an element of the state space as a point with numerical coordinates, that is x 1. x 2. x = Linear Algebra Review Vectors To begin, let us describe an element of the state space as a point with numerical coordinates, that is x 1 x x = 2. x n Vectors of up to three dimensions are easy to diagram.

More information

Dimensionality Reduction with Principal Component Analysis

Dimensionality Reduction with Principal Component Analysis 10 Dimensionality Reduction with Principal Component Analysis Working directly with high-dimensional data, such as images, comes with some difficulties: it is hard to analyze, interpretation is difficult,

More information

System 1 (last lecture) : limited to rigidly structured shapes. System 2 : recognition of a class of varying shapes. Need to:

System 1 (last lecture) : limited to rigidly structured shapes. System 2 : recognition of a class of varying shapes. Need to: System 2 : Modelling & Recognising Modelling and Recognising Classes of Classes of Shapes Shape : PDM & PCA All the same shape? System 1 (last lecture) : limited to rigidly structured shapes System 2 :

More information

Linear Subspace Models

Linear Subspace Models Linear Subspace Models Goal: Explore linear models of a data set. Motivation: A central question in vision concerns how we represent a collection of data vectors. The data vectors may be rasterized images,

More information

Background Mathematics (2/2) 1. David Barber

Background Mathematics (2/2) 1. David Barber Background Mathematics (2/2) 1 David Barber University College London Modified by Samson Cheung (sccheung@ieee.org) 1 These slides accompany the book Bayesian Reasoning and Machine Learning. The book and

More information

Dot Products. K. Behrend. April 3, Abstract A short review of some basic facts on the dot product. Projections. The spectral theorem.

Dot Products. K. Behrend. April 3, Abstract A short review of some basic facts on the dot product. Projections. The spectral theorem. Dot Products K. Behrend April 3, 008 Abstract A short review of some basic facts on the dot product. Projections. The spectral theorem. Contents The dot product 3. Length of a vector........................

More information

Vectors and Matrices Statistics with Vectors and Matrices

Vectors and Matrices Statistics with Vectors and Matrices Vectors and Matrices Statistics with Vectors and Matrices Lecture 3 September 7, 005 Analysis Lecture #3-9/7/005 Slide 1 of 55 Today s Lecture Vectors and Matrices (Supplement A - augmented with SAS proc

More information

Principal Component Analysis (PCA) CSC411/2515 Tutorial

Principal Component Analysis (PCA) CSC411/2515 Tutorial Principal Component Analysis (PCA) CSC411/2515 Tutorial Harris Chan Based on previous tutorial slides by Wenjie Luo, Ladislav Rampasek University of Toronto hchan@cs.toronto.edu October 19th, 2017 (UofT)

More information

Math 1553, Introduction to Linear Algebra

Math 1553, Introduction to Linear Algebra Learning goals articulate what students are expected to be able to do in a course that can be measured. This course has course-level learning goals that pertain to the entire course, and section-level

More information

Dimensionality reduction

Dimensionality reduction Dimensionality Reduction PCA continued Machine Learning CSE446 Carlos Guestrin University of Washington May 22, 2013 Carlos Guestrin 2005-2013 1 Dimensionality reduction n Input data may have thousands

More information

Machine Learning. Principal Components Analysis. Le Song. CSE6740/CS7641/ISYE6740, Fall 2012

Machine Learning. Principal Components Analysis. Le Song. CSE6740/CS7641/ISYE6740, Fall 2012 Machine Learning CSE6740/CS7641/ISYE6740, Fall 2012 Principal Components Analysis Le Song Lecture 22, Nov 13, 2012 Based on slides from Eric Xing, CMU Reading: Chap 12.1, CB book 1 2 Factor or Component

More information

What is Principal Component Analysis?

What is Principal Component Analysis? What is Principal Component Analysis? Principal component analysis (PCA) Reduce the dimensionality of a data set by finding a new set of variables, smaller than the original set of variables Retains most

More information

ROBERTO BATTITI, MAURO BRUNATO. The LION Way: Machine Learning plus Intelligent Optimization. LIONlab, University of Trento, Italy, Apr 2015

ROBERTO BATTITI, MAURO BRUNATO. The LION Way: Machine Learning plus Intelligent Optimization. LIONlab, University of Trento, Italy, Apr 2015 ROBERTO BATTITI, MAURO BRUNATO. The LION Way: Machine Learning plus Intelligent Optimization. LIONlab, University of Trento, Italy, Apr 2015 http://intelligentoptimization.org/lionbook Roberto Battiti

More information

Linear Algebra and Matrices

Linear Algebra and Matrices Linear Algebra and Matrices 4 Overview In this chapter we studying true matrix operations, not element operations as was done in earlier chapters. Working with MAT- LAB functions should now be fairly routine.

More information

Structure in Data. A major objective in data analysis is to identify interesting features or structure in the data.

Structure in Data. A major objective in data analysis is to identify interesting features or structure in the data. Structure in Data A major objective in data analysis is to identify interesting features or structure in the data. The graphical methods are very useful in discovering structure. There are basically two

More information

Numerical Methods I Singular Value Decomposition

Numerical Methods I Singular Value Decomposition Numerical Methods I Singular Value Decomposition Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 MATH-GA 2011.003 / CSCI-GA 2945.003, Fall 2014 October 9th, 2014 A. Donev (Courant Institute)

More information

PCA Review. CS 510 February 25 th, 2013

PCA Review. CS 510 February 25 th, 2013 PCA Review CS 510 February 25 th, 2013 Recall the goal: image matching Probe image, registered to gallery Registered Gallery of Images 3/7/13 CS 510, Image Computa5on, Ross Beveridge & Bruce Draper 2 Getting

More information

Principal Components Analysis (PCA) and Singular Value Decomposition (SVD) with applications to Microarrays

Principal Components Analysis (PCA) and Singular Value Decomposition (SVD) with applications to Microarrays Principal Components Analysis (PCA) and Singular Value Decomposition (SVD) with applications to Microarrays Prof. Tesler Math 283 Fall 2015 Prof. Tesler Principal Components Analysis Math 283 / Fall 2015

More information

Lecture 3: Review of Linear Algebra

Lecture 3: Review of Linear Algebra ECE 83 Fall 2 Statistical Signal Processing instructor: R Nowak Lecture 3: Review of Linear Algebra Very often in this course we will represent signals as vectors and operators (eg, filters, transforms,

More information

Eigenvalues, Eigenvectors, and an Intro to PCA

Eigenvalues, Eigenvectors, and an Intro to PCA Eigenvalues, Eigenvectors, and an Intro to PCA Eigenvalues, Eigenvectors, and an Intro to PCA Changing Basis We ve talked so far about re-writing our data using a new set of variables, or a new basis.

More information

Principal Component Analysis. Applied Multivariate Statistics Spring 2012

Principal Component Analysis. Applied Multivariate Statistics Spring 2012 Principal Component Analysis Applied Multivariate Statistics Spring 2012 Overview Intuition Four definitions Practical examples Mathematical example Case study 2 PCA: Goals Goal 1: Dimension reduction

More information

Getting Started with Communications Engineering

Getting Started with Communications Engineering 1 Linear algebra is the algebra of linear equations: the term linear being used in the same sense as in linear functions, such as: which is the equation of a straight line. y ax c (0.1) Of course, if we

More information

PRINCIPAL COMPONENTS ANALYSIS

PRINCIPAL COMPONENTS ANALYSIS 121 CHAPTER 11 PRINCIPAL COMPONENTS ANALYSIS We now have the tools necessary to discuss one of the most important concepts in mathematical statistics: Principal Components Analysis (PCA). PCA involves

More information

3.3 Eigenvalues and Eigenvectors

3.3 Eigenvalues and Eigenvectors .. EIGENVALUES AND EIGENVECTORS 27. Eigenvalues and Eigenvectors In this section, we assume A is an n n matrix and x is an n vector... Definitions In general, the product Ax results is another n vector

More information

Topic 14 Notes Jeremy Orloff

Topic 14 Notes Jeremy Orloff Topic 4 Notes Jeremy Orloff 4 Row reduction and subspaces 4. Goals. Be able to put a matrix into row reduced echelon form (RREF) using elementary row operations.. Know the definitions of null and column

More information

Eigenvalues, Eigenvectors, and an Intro to PCA

Eigenvalues, Eigenvectors, and an Intro to PCA Eigenvalues, Eigenvectors, and an Intro to PCA Eigenvalues, Eigenvectors, and an Intro to PCA Changing Basis We ve talked so far about re-writing our data using a new set of variables, or a new basis.

More information

Principal Component Analysis (PCA) Theory, Practice, and Examples

Principal Component Analysis (PCA) Theory, Practice, and Examples Principal Component Analysis (PCA) Theory, Practice, and Examples Data Reduction summarization of data with many (p) variables by a smaller set of (k) derived (synthetic, composite) variables. p k n A

More information

STATISTICAL SHAPE MODELS (SSM)

STATISTICAL SHAPE MODELS (SSM) STATISTICAL SHAPE MODELS (SSM) Medical Image Analysis Serena Bonaretti serena.bonaretti@istb.unibe.ch ISTB - Institute for Surgical Technology and Biomechanics University of Bern Overview > Introduction

More information

Eigenvalues, Eigenvectors, and an Intro to PCA

Eigenvalues, Eigenvectors, and an Intro to PCA Eigenvalues, Eigenvectors, and an Intro to PCA Eigenvalues, Eigenvectors, and an Intro to PCA Changing Basis We ve talked so far about re-writing our data using a new set of variables, or a new basis.

More information

Applied Linear Algebra in Geoscience Using MATLAB

Applied Linear Algebra in Geoscience Using MATLAB Applied Linear Algebra in Geoscience Using MATLAB Contents Getting Started Creating Arrays Mathematical Operations with Arrays Using Script Files and Managing Data Two-Dimensional Plots Programming in

More information

Principle Components Analysis (PCA) Relationship Between a Linear Combination of Variables and Axes Rotation for PCA

Principle Components Analysis (PCA) Relationship Between a Linear Combination of Variables and Axes Rotation for PCA Principle Components Analysis (PCA) Relationship Between a Linear Combination of Variables and Axes Rotation for PCA Principle Components Analysis: Uses one group of variables (we will call this X) In

More information

Designing Information Devices and Systems II Fall 2018 Elad Alon and Miki Lustig Homework 9

Designing Information Devices and Systems II Fall 2018 Elad Alon and Miki Lustig Homework 9 EECS 16B Designing Information Devices and Systems II Fall 18 Elad Alon and Miki Lustig Homework 9 This homework is due Wednesday, October 31, 18, at 11:59pm. Self grades are due Monday, November 5, 18,

More information

Review problems for MA 54, Fall 2004.

Review problems for MA 54, Fall 2004. Review problems for MA 54, Fall 2004. Below are the review problems for the final. They are mostly homework problems, or very similar. If you are comfortable doing these problems, you should be fine on

More information

4 ORTHOGONALITY ORTHOGONALITY OF THE FOUR SUBSPACES 4.1

4 ORTHOGONALITY ORTHOGONALITY OF THE FOUR SUBSPACES 4.1 4 ORTHOGONALITY ORTHOGONALITY OF THE FOUR SUBSPACES 4.1 Two vectors are orthogonal when their dot product is zero: v w = orv T w =. This chapter moves up a level, from orthogonal vectors to orthogonal

More information

LINEAR ALGEBRA KNOWLEDGE SURVEY

LINEAR ALGEBRA KNOWLEDGE SURVEY LINEAR ALGEBRA KNOWLEDGE SURVEY Instructions: This is a Knowledge Survey. For this assignment, I am only interested in your level of confidence about your ability to do the tasks on the following pages.

More information

Data Mining Lecture 4: Covariance, EVD, PCA & SVD

Data Mining Lecture 4: Covariance, EVD, PCA & SVD Data Mining Lecture 4: Covariance, EVD, PCA & SVD Jo Houghton ECS Southampton February 25, 2019 1 / 28 Variance and Covariance - Expectation A random variable takes on different values due to chance The

More information

Main matrix factorizations

Main matrix factorizations Main matrix factorizations A P L U P permutation matrix, L lower triangular, U upper triangular Key use: Solve square linear system Ax b. A Q R Q unitary, R upper triangular Key use: Solve square or overdetrmined

More information

Maximum variance formulation

Maximum variance formulation 12.1. Principal Component Analysis 561 Figure 12.2 Principal component analysis seeks a space of lower dimensionality, known as the principal subspace and denoted by the magenta line, such that the orthogonal

More information

Face Recognition. Face Recognition. Subspace-Based Face Recognition Algorithms. Application of Face Recognition

Face Recognition. Face Recognition. Subspace-Based Face Recognition Algorithms. Application of Face Recognition ace Recognition Identify person based on the appearance of face CSED441:Introduction to Computer Vision (2017) Lecture10: Subspace Methods and ace Recognition Bohyung Han CSE, POSTECH bhhan@postech.ac.kr

More information

Face Detection and Recognition

Face Detection and Recognition Face Detection and Recognition Face Recognition Problem Reading: Chapter 18.10 and, optionally, Face Recognition using Eigenfaces by M. Turk and A. Pentland Queryimage face query database Face Verification

More information

Eigenvalues and Eigenvectors

Eigenvalues and Eigenvectors Eigenvalues and Eigenvectors Philippe B. Laval KSU Fall 2015 Philippe B. Laval (KSU) Eigenvalues and Eigenvectors Fall 2015 1 / 14 Introduction We define eigenvalues and eigenvectors. We discuss how to

More information

Lecture: Face Recognition

Lecture: Face Recognition Lecture: Face Recognition Juan Carlos Niebles and Ranjay Krishna Stanford Vision and Learning Lab Lecture 12-1 What we will learn today Introduction to face recognition The Eigenfaces Algorithm Linear

More information