The Principal Component Analysis

Size: px
Start display at page:

Download "The Principal Component Analysis"

Transcription

1 The Principal Component Analysis Philippe B. Laval KSU Fall 2017 Philippe B. Laval (KSU) PCA Fall / 27

2 Introduction Every 80 minutes, the two Landsat satellites go around the world, recording images of our planet along a 185 Km wide path. Every 16 days, each satellite covers the entire surface of the planet. Every 8 days, every single location on the planet can be monitored. These images can be used by urban planners to study the rate and direction of population growth. They can be used to analyze soil moisture, vegetation growth, lakes and rivers. Governments can use them to detect and assess damages from natural disasters... Sensors on board the satellites acquire seven simultaneous images of every region by recording energy from separate wavelength bands. Each image is digitized and stored as a matrix., each number indicating the signal intensity of the corresponding pixel. Each of the seven image is one channel of a multichannel or multispectral image. The seven images of one fixed region typically contains redundant information. Yet, certain features will only appear in one or two images because their color or temperature was only captured by two of the seven sensors. Philippe B. Laval (KSU) PCA Fall / 27

3 Introduction Principal Component Analysis (PCA) is an effective way to suppress redundant information and provide in only one or two composite images most of the information from the initial data. Roughly speaking, the goal is to find a special linear combination of the images that combines each pixel of the seven images into one single new pixel. PCA can be applied to any data that consists of lists of measurements made on a collections of objects or individuals. For example, consider a chemical process that produces a plastic material. To monitor the process, 300 samples of the material produced are taken and each sample is subject to a battery of eight tests. The lab report for each sample is a vector in R 8 (8 tests). Since there are 300 samples, the set of these vectors is a matrix. Such a matrix is called the matrix of observations. Loosely speaking, we say that the process control data is eighth-dimensional. Philippe B. Laval (KSU) PCA Fall / 27

4 Introduction Example Suppose we have N college students for which we record their weights and heights. This is an example of two-dimensional data. Let [ X j be ] the observation vector in R 2 for the j th wj student that is X j = where w j is the weight (in Kg) and h j is the height (in m) of the jth student. Then, the [ matrix of observations ] has the form w1 w 2 w N = [X h 1 h 2 h 1 X 2 X N ]. Here is a possible matrix of N observations (randomly generated with height in meters and weight in kilograms) for N = 25 visualized on a scatter graph shown on the next slide. h j Philippe B. Laval (KSU) PCA Fall / 27

5 Introduction Figure: Scatter Plot of Observations Philippe B. Laval (KSU) PCA Fall / 27

6 Introduction We will use [X 1 X 2 X N ] to denote the matrix of observations. It will be a p N matrix meaning that we are observing (measuring) p different variables and we collect N samples of the measurements of these variables. x 1 x 2 We will let X = be the generic vector of all the variables we. x p are observing. Each X i will denote one measure of these variables, the ith measure. x i 1 x i 2 We will denote its components X i =.. When the number of variables being observed is small, we can use different names for the variables as we did in the example above. Philippe B. Laval (KSU) PCA Fall / 27 x i p

7 Mean and Covariance Definition (Sample Mean) Let [X 1 X 2 X N ] be a p N matrix of observations (meaning that for each object, we have p observations, each X i is a p 1 vector). The sample mean M of the observation vectors X 1, X 2,..., X N is given by M = 1 N N i=1 X i The reader will note that M is also a p 1 vector. Geometrically, the sample mean is the point at the "center" of the scatter plot. For i = 1, 2,..., N, we define X ] i = X i M and B = [ X1 X2 XN. The columns of B have a zero sample mean. Philippe B. Laval (KSU) PCA Fall / 27

8 Mean and Covariance Definition If the columns of a matrix have a zero sample mean, then B is said to be in mean-deviation form. Example The scatter plot of the observation matrix of the previous example put in mean-deviation form is shown in the picture on the next slide. You will note that the "center" of the data is now at the origin. Philippe B. Laval (KSU) PCA Fall / 27

9 Mean and Covariance Figure: Observation Data in Mean-Deviation Form Philippe B. Laval (KSU) PCA Fall / 27

10 Mean and Covariance Definition The sample covariance matrix or the covariance matrix is the p p matrix defined by S = 1 N 1 BBT Since any matrix of the form BB T is symmetric and positive semidefinite, so is S. Recall that a matrix A is positive definite (positive semidefinite) if for every nonzero vector x, x T Ax > 0 (x T Ax 0). Philippe B. Laval (KSU) PCA Fall / 27

11 Mean and Covariance Example Suppose that measurements are made on 4 individuals and the observation vectors are X 1 = 2, X 2 = 2, X 3 = 8, X 4 = Compute the sample mean and the covariance matrix. Philippe B. Laval (KSU) PCA Fall / 27

12 Mean and Covariance We should have found that the sample mean is M = The corresponding matrixin mean-deviation form is B = The covariance matrix is S = Philippe B. Laval (KSU) PCA Fall / 27

13 Mean and Covariance We now discuss the meaning of the entries in S, the covariance matrix, assuming that S = [s ij ]. We continue using the notation developed in this section. Philippe B. Laval (KSU) PCA Fall / 27

14 Mean and Covariance Definition (variance) For i = 1, 2,..., p, the diagonal entry s ii is called the variance of x i The variance of x j measures the spread of the values of x j. In the example above, from our computations, the variance of x 1 is 10, the variance of x 2 is 8 and the variance of x 3 is 32. What is important here is the relative size of these numbers. The fact that 32 > 10 indicates that the set of third entries contains a wider spread than the set of first entries. Definition (total variance) The total variance of the data is the sum of the diagonal entries in S. For a square matrix S (recall S is p p), the sum of the diagonal entries is called the trace of S and is denoted tr (S). Thus total variance = tr (S) Philippe B. Laval (KSU) PCA Fall / 27

15 Mean and Covariance Definition (covariance) The entry s ij for i i, in S is called the covariance of x i and x j. In the example above, looking at S, we see that the covariance between entries 1 and 3 is 0. This means that x 1 and x 3 are uncorrelated. Philippe B. Laval (KSU) PCA Fall / 27

16 PCA Using Eigenvalues As above, let us assume that our original variables are represented by the p 1 vector X and we perform N measurements we denote X 1, X 2,..., X N. For simplicity, we will assume that the p N matrix of observations [X 1 X 2 X N ] is already in mean-deviation form. If this were not the case, the first step would be to put it in that form. The goal of PCA is to find an orthogonal p p matrix, P = [u 1 u 2... u p ] that determines a change of variables X = PY that is x 1 x 2. x p = [u 1 u 2... u p ] with the property that the new variables y 1, y 2,..., y p are uncorrelated and are arranged in order of decreasing variance. y 1 y 2. y p Philippe B. Laval (KSU) PCA Fall / 27

17 PCA Using Eigenvalues A Question: From the statements on the previous slide, what can we say about the covariance matrix corresponding to the new variables? Another Question? How can we achieve the answer to the previous question? Philippe B. Laval (KSU) PCA Fall / 27

18 PCA Using Eigenvalues The orthogonal change of variable X = PY means that each observation vector X i is transformed into a new vector we call Y i such that X i = PY i. Therefore, Y i = P 1 X i = P T X i since P is supposed to be orthogonal. This is true for i = 1, 2,..., N. It is not hard to see (see homework) that for each orthogonal matrix P, the covariance matrix for Y 1, Y 2,..., Y N is P T SP where S is the covariance matrix for X 1, X 2,..., X N. So, the desired orthogonal matrix P is the one that makes P T SP diagonal (why?). Let D be a diagonal matrix with the eigenvalues λ 1, λ 2,..., λ p of S on the diagonal arranged so that λ 1 λ 2... λ p 0 and let P be an orthogonal matrix whose columns are the unit eigenvectors u 1, u 2,..., u p corresponding to the eigenvalues λ 1 λ 2... λ p. Then, S = PDP T and D = P T SP. Philippe B. Laval (KSU) PCA Fall / 27

19 PCA Using Eigenvalues Definition The unit eigenvectors u 1, u 2,..., u p of the covariance matrix S are called the principal components of the data in the matrix of observations. The first principal component is the eigenvector corresponding to the largest eigenvalue of S. The second principal component is the eigenvector corresponding to the second largest eigenvalue of S and so on. The first principal component u 1 determines the new variable y 1 in the following way. Let c 1, c 2,..., c p be the entries in u 1. Since u T 1 is the first row of P T, the equation Y = P T X shows that p y 1 = u T 1 X = c i x i Thus y 1 is a linear combination of the original variable using the entries in u 1 as weights. In a similar way, u 2 determines the entries in y 2, and so on. Philippe B. Laval (KSU) PCA Fall / 27 i=1

20 PCA Using Eigenvalues: Summary Given a matrix of observation B = [X 1 X 2 X N ] which is assumed to be in mean-deviation form, to perform a principal component analysis, we do the following: 1 Find the covariance matrix S = 1 N 1 BBT 2 Diagonalize S using eigenvalues. We let λ 1, λ 2,..., λ p be the eigenvalues of S arranged so that λ 1 λ 2... λ p 0 and let P be an orthogonal matrix whose columns are the unit eigenvectors u 1, u 2,..., u p corresponding to the eigenvalues λ 1 λ 2... λ p. 3 The unit eigenvectors u 1, u 2,..., u p of the covariance matrix S are called the principal components of the data in the matrix of observations. 4 The new uncorrelated variables are defined by Y = P T X where Y,like X is p 1. The covariance matrix of Y is P T SP. 5 The components y 1, y 2,..., y p of y can be expressed in terms of the components x 1, x 2,..., x p of X by y i = u T i X. Philippe B. Laval (KSU) PCA Fall / 27

21 PCA Using Eigenvalues Example Suppose the covariance matrix of some data is S = Find the principal components of the data, list the new variable determined by the first principal component and give the diagonal form of S. Philippe B. Laval (KSU) PCA Fall / 27

22 Dimension Reduction PCA is valuable when most of the variation in the data is due to variations in only a few of the new variables y 1, y 2,..., y p. It can be shown that an orthogonal change of variables X = PY does not change the total variance of the data. This means that if S = PDP T then { } { } total variance total variance = of x 1, x 2,..., x p of y 1, y 2,..., y p = tr (D) = λ 1 + λ λ p The variance of y i is λ i, and the quotient measures the tr (S) fraction of the total variance captured by y i. λ i We can eliminate the variables for which is small. tr (S) Philippe B. Laval (KSU) PCA Fall / 27 λ i

23 Dimension Reduction Example In the example above, compute the various percentages of variance captured by each variable. Example Consider the following data: Boy #1 #2 #3 #4 #5 Weight (lb) Height (in) Find the covariance matrix for the data. 2 Make a principal component analysis of the data to find a single size index that explains most of the variation in the data. Philippe B. Laval (KSU) PCA Fall / 27

24 PCA Using SVD The goal is the same as above, what changes is how we achieve it. As above, we will assume that the p N matrix of observations [X 1 X 2 X N ] is already in mean-deviation form. If this were not the case, the first step would be to put it in that form. Assume the SVD of the matrix of observations [X 1 X 2 X N ] is UΣV T. Here, U is p p, Σ is p N and V T is N N hence V is N N. We define the new variable Y by X = UY or Y = U T X. It can be shown that the covariance matrix 1 corresponding to Y is Y is N 1 Σ2 which is already in diagonal form; this is what we wanted. In this case, the principal components will be the columns of U. If we let U = [u 1 u 2... u p ] then the equation Y = P T X p shows that y 1 = u T 1 X = c i x i where c 1, c 2,..., c p are the entries in u 1. i=1 Thus y 1 is a linear combination of the original variable using the entries in u 1 as weights. In a similar way, u 2 determines the entries in y 2, and so on. Philippe B. Laval (KSU) PCA Fall / 27

25 PCA Using SVD: Summary Given a matrix of observation B = [X 1 X 2 X N ] which is assumed to be in mean-deviation form, to perform a principal component analysis, we do the following: 1 Find a SVD for B, say B = UΣV T. 2 If we let U = [u 1 u 2... u p ] then we define Y = U T X. The covariance 1 matrix of Y is N 1 Σ2. Y is the new uncorrelated variable. 1 3 We let λ 1, λ 2,..., λ p be the diagonal elements of N 1 Σ2 (automatically arranged so that λ 1 λ 2... λ p 0). Then λ i is the varaince of y i which is also the variance of x i. 4 The unit vectors u 1, u 2,..., u p are called the principal components of the data in the matrix of observations. We can get all the components we were able to get with the eigenvalue diagonalization. We simply got them differently. MATLAB uses this technique as it is a little bit more robust than the eigenvalue method. Philippe B. Laval (KSU) PCA Fall / 27

26 PCA and MATLAB Newer versions of MATLAB have a function to perform this, it is called pca. You can get help for it in MATLAB. Two important things to remember about MATLAB s pca: 1 It uses the SVD approach to find the principal components and related information. 2 It works differently than what is explained in these notes in the following sense. MATLAB s matrix of observation is the transpose of the matrix of observations used in these notes. In other words, the columns of our matrix of observations contain the various variables used for our data. For MATLAB, these variables are stored as the rows of the matrix of observations. Suppose that A is our matrix of observations. To use MATLAB s PCA, use pca(a ), where A is the transpose of A. Philippe B. Laval (KSU) PCA Fall / 27

27 Exercises See the problems at the end of the notes on Principal Component Analysis. Philippe B. Laval (KSU) PCA Fall / 27

Independent Component Analysis

Independent Component Analysis Independent Component Analysis Philippe B. Laval KSU Fall 2017 Philippe B. Laval (KSU) ICA Fall 2017 1 / 18 Introduction Independent Component Analysis (ICA) falls under the broader topic of Blind Source

More information

The Singular Value Decomposition

The Singular Value Decomposition The Singular Value Decomposition Philippe B. Laval KSU Fall 2015 Philippe B. Laval (KSU) SVD Fall 2015 1 / 13 Review of Key Concepts We review some key definitions and results about matrices that will

More information

GEOG 4110/5100 Advanced Remote Sensing Lecture 15

GEOG 4110/5100 Advanced Remote Sensing Lecture 15 GEOG 4110/5100 Advanced Remote Sensing Lecture 15 Principal Component Analysis Relevant reading: Richards. Chapters 6.3* http://www.ce.yildiz.edu.tr/personal/songul/file/1097/principal_components.pdf *For

More information

Introduction to Machine Learning

Introduction to Machine Learning 10-701 Introduction to Machine Learning PCA Slides based on 18-661 Fall 2018 PCA Raw data can be Complex, High-dimensional To understand a phenomenon we measure various related quantities If we knew what

More information

Lecture: Face Recognition and Feature Reduction

Lecture: Face Recognition and Feature Reduction Lecture: Face Recognition and Feature Reduction Juan Carlos Niebles and Ranjay Krishna Stanford Vision and Learning Lab 1 Recap - Curse of dimensionality Assume 5000 points uniformly distributed in the

More information

Lecture: Face Recognition and Feature Reduction

Lecture: Face Recognition and Feature Reduction Lecture: Face Recognition and Feature Reduction Juan Carlos Niebles and Ranjay Krishna Stanford Vision and Learning Lab Lecture 11-1 Recap - Curse of dimensionality Assume 5000 points uniformly distributed

More information

Eigenvalues and Eigenvectors

Eigenvalues and Eigenvectors Eigenvalues and Eigenvectors Philippe B. Laval KSU Fall 2015 Philippe B. Laval (KSU) Eigenvalues and Eigenvectors Fall 2015 1 / 14 Introduction We define eigenvalues and eigenvectors. We discuss how to

More information

12.2 Dimensionality Reduction

12.2 Dimensionality Reduction 510 Chapter 12 of this dimensionality problem, regularization techniques such as SVD are almost always needed to perform the covariance matrix inversion. Because it appears to be a fundamental property

More information

Singular Value Decomposition and Digital Image Compression

Singular Value Decomposition and Digital Image Compression Singular Value Decomposition and Digital Image Compression Chris Bingham December 1, 016 Page 1 of Abstract The purpose of this document is to be a very basic introduction to the singular value decomposition

More information

A A x i x j i j (i, j) (j, i) Let. Compute the value of for and

A A x i x j i j (i, j) (j, i) Let. Compute the value of for and 7.2 - Quadratic Forms quadratic form on is a function defined on whose value at a vector in can be computed by an expression of the form, where is an symmetric matrix. The matrix R n Q R n x R n Q(x) =

More information

Dimensionality Reduction: PCA. Nicholas Ruozzi University of Texas at Dallas

Dimensionality Reduction: PCA. Nicholas Ruozzi University of Texas at Dallas Dimensionality Reduction: PCA Nicholas Ruozzi University of Texas at Dallas Eigenvalues λ is an eigenvalue of a matrix A R n n if the linear system Ax = λx has at least one non-zero solution If Ax = λx

More information

The Singular Value Decomposition (SVD) and Principal Component Analysis (PCA)

The Singular Value Decomposition (SVD) and Principal Component Analysis (PCA) Chapter 5 The Singular Value Decomposition (SVD) and Principal Component Analysis (PCA) 5.1 Basics of SVD 5.1.1 Review of Key Concepts We review some key definitions and results about matrices that will

More information

Review of Linear Algebra

Review of Linear Algebra Review of Linear Algebra Definitions An m n (read "m by n") matrix, is a rectangular array of entries, where m is the number of rows and n the number of columns. 2 Definitions (Con t) A is square if m=

More information

Notes on singular value decomposition for Math 54. Recall that if A is a symmetric n n matrix, then A has real eigenvalues A = P DP 1 A = P DP T.

Notes on singular value decomposition for Math 54. Recall that if A is a symmetric n n matrix, then A has real eigenvalues A = P DP 1 A = P DP T. Notes on singular value decomposition for Math 54 Recall that if A is a symmetric n n matrix, then A has real eigenvalues λ 1,, λ n (possibly repeated), and R n has an orthonormal basis v 1,, v n, where

More information

Singular Value Decompsition

Singular Value Decompsition Singular Value Decompsition Massoud Malek One of the most useful results from linear algebra, is a matrix decomposition known as the singular value decomposition It has many useful applications in almost

More information

Solving Linear Systems

Solving Linear Systems Solving Linear Systems Iterative Solutions Methods Philippe B. Laval KSU Fall 207 Philippe B. Laval (KSU) Linear Systems Fall 207 / 2 Introduction We continue looking how to solve linear systems of the

More information

Review of Functions. Functions. Philippe B. Laval. Current Semester KSU. Philippe B. Laval (KSU) Functions Current Semester 1 / 12

Review of Functions. Functions. Philippe B. Laval. Current Semester KSU. Philippe B. Laval (KSU) Functions Current Semester 1 / 12 Review of Functions Functions Philippe B. Laval KSU Current Semester Philippe B. Laval (KSU) Functions Current Semester 1 / 12 Introduction Students are expected to know the following concepts about functions:

More information

Gopalkrishna Veni. Project 4 (Active Shape Models)

Gopalkrishna Veni. Project 4 (Active Shape Models) Gopalkrishna Veni Project 4 (Active Shape Models) Introduction Active shape Model (ASM) is a technique of building a model by learning the variability patterns from training datasets. ASMs try to deform

More information

Linear Algebra Review. Fei-Fei Li

Linear Algebra Review. Fei-Fei Li Linear Algebra Review Fei-Fei Li 1 / 51 Vectors Vectors and matrices are just collections of ordered numbers that represent something: movements in space, scaling factors, pixel brightnesses, etc. A vector

More information

1 Principal component analysis and dimensional reduction

1 Principal component analysis and dimensional reduction Linear Algebra Working Group :: Day 3 Note: All vector spaces will be finite-dimensional vector spaces over the field R. 1 Principal component analysis and dimensional reduction Definition 1.1. Given an

More information

Singular Value Decomposition

Singular Value Decomposition Singular Value Decomposition Motivatation The diagonalization theorem play a part in many interesting applications. Unfortunately not all matrices can be factored as A = PDP However a factorization A =

More information

MLCC 2015 Dimensionality Reduction and PCA

MLCC 2015 Dimensionality Reduction and PCA MLCC 2015 Dimensionality Reduction and PCA Lorenzo Rosasco UNIGE-MIT-IIT June 25, 2015 Outline PCA & Reconstruction PCA and Maximum Variance PCA and Associated Eigenproblem Beyond the First Principal Component

More information

Appendix A: Matrices

Appendix A: Matrices Appendix A: Matrices A matrix is a rectangular array of numbers Such arrays have rows and columns The numbers of rows and columns are referred to as the dimensions of a matrix A matrix with, say, 5 rows

More information

Linear Algebra Review. Fei-Fei Li

Linear Algebra Review. Fei-Fei Li Linear Algebra Review Fei-Fei Li 1 / 37 Vectors Vectors and matrices are just collections of ordered numbers that represent something: movements in space, scaling factors, pixel brightnesses, etc. A vector

More information

PRINCIPAL COMPONENTS ANALYSIS

PRINCIPAL COMPONENTS ANALYSIS 121 CHAPTER 11 PRINCIPAL COMPONENTS ANALYSIS We now have the tools necessary to discuss one of the most important concepts in mathematical statistics: Principal Components Analysis (PCA). PCA involves

More information

Numerical Linear Algebra

Numerical Linear Algebra Numerical Linear Algebra Direct Methods Philippe B. Laval KSU Fall 2017 Philippe B. Laval (KSU) Linear Systems: Direct Solution Methods Fall 2017 1 / 14 Introduction The solution of linear systems is one

More information

Conceptual Questions for Review

Conceptual Questions for Review Conceptual Questions for Review Chapter 1 1.1 Which vectors are linear combinations of v = (3, 1) and w = (4, 3)? 1.2 Compare the dot product of v = (3, 1) and w = (4, 3) to the product of their lengths.

More information

Linear Algebra & Geometry why is linear algebra useful in computer vision?

Linear Algebra & Geometry why is linear algebra useful in computer vision? Linear Algebra & Geometry why is linear algebra useful in computer vision? References: -Any book on linear algebra! -[HZ] chapters 2, 4 Some of the slides in this lecture are courtesy to Prof. Octavia

More information

Linear Algebra & Geometry why is linear algebra useful in computer vision?

Linear Algebra & Geometry why is linear algebra useful in computer vision? Linear Algebra & Geometry why is linear algebra useful in computer vision? References: -Any book on linear algebra! -[HZ] chapters 2, 4 Some of the slides in this lecture are courtesy to Prof. Octavia

More information

15 Singular Value Decomposition

15 Singular Value Decomposition 15 Singular Value Decomposition For any high-dimensional data analysis, one s first thought should often be: can I use an SVD? The singular value decomposition is an invaluable analysis tool for dealing

More information

Principle Components Analysis (PCA) Relationship Between a Linear Combination of Variables and Axes Rotation for PCA

Principle Components Analysis (PCA) Relationship Between a Linear Combination of Variables and Axes Rotation for PCA Principle Components Analysis (PCA) Relationship Between a Linear Combination of Variables and Axes Rotation for PCA Principle Components Analysis: Uses one group of variables (we will call this X) In

More information

CHAPTER 4 PRINCIPAL COMPONENT ANALYSIS-BASED FUSION

CHAPTER 4 PRINCIPAL COMPONENT ANALYSIS-BASED FUSION 59 CHAPTER 4 PRINCIPAL COMPONENT ANALYSIS-BASED FUSION 4. INTRODUCTION Weighted average-based fusion algorithms are one of the widely used fusion methods for multi-sensor data integration. These methods

More information

Background Mathematics (2/2) 1. David Barber

Background Mathematics (2/2) 1. David Barber Background Mathematics (2/2) 1 David Barber University College London Modified by Samson Cheung (sccheung@ieee.org) 1 These slides accompany the book Bayesian Reasoning and Machine Learning. The book and

More information

Principal Component Analysis (PCA) Our starting point consists of T observations from N variables, which will be arranged in an T N matrix R,

Principal Component Analysis (PCA) Our starting point consists of T observations from N variables, which will be arranged in an T N matrix R, Principal Component Analysis (PCA) PCA is a widely used statistical tool for dimension reduction. The objective of PCA is to find common factors, the so called principal components, in form of linear combinations

More information

Lecture 13. Principal Component Analysis. Brett Bernstein. April 25, CDS at NYU. Brett Bernstein (CDS at NYU) Lecture 13 April 25, / 26

Lecture 13. Principal Component Analysis. Brett Bernstein. April 25, CDS at NYU. Brett Bernstein (CDS at NYU) Lecture 13 April 25, / 26 Principal Component Analysis Brett Bernstein CDS at NYU April 25, 2017 Brett Bernstein (CDS at NYU) Lecture 13 April 25, 2017 1 / 26 Initial Question Intro Question Question Let S R n n be symmetric. 1

More information

CS168: The Modern Algorithmic Toolbox Lecture #8: How PCA Works

CS168: The Modern Algorithmic Toolbox Lecture #8: How PCA Works CS68: The Modern Algorithmic Toolbox Lecture #8: How PCA Works Tim Roughgarden & Gregory Valiant April 20, 206 Introduction Last lecture introduced the idea of principal components analysis (PCA). The

More information

Introduction to Machine Learning. PCA and Spectral Clustering. Introduction to Machine Learning, Slides: Eran Halperin

Introduction to Machine Learning. PCA and Spectral Clustering. Introduction to Machine Learning, Slides: Eran Halperin 1 Introduction to Machine Learning PCA and Spectral Clustering Introduction to Machine Learning, 2013-14 Slides: Eran Halperin Singular Value Decomposition (SVD) The singular value decomposition (SVD)

More information

Functions of Several Variables

Functions of Several Variables Functions of Several Variables Extreme Values Philippe B Laval KSU April 9, 2012 Philippe B Laval (KSU) Functions of Several Variables April 9, 2012 1 / 13 Introduction In Calculus I (differential calculus

More information

7. Symmetric Matrices and Quadratic Forms

7. Symmetric Matrices and Quadratic Forms Linear Algebra 7. Symmetric Matrices and Quadratic Forms CSIE NCU 1 7. Symmetric Matrices and Quadratic Forms 7.1 Diagonalization of symmetric matrices 2 7.2 Quadratic forms.. 9 7.4 The singular value

More information

GNR401 Principles of Satellite Image Processing

GNR401 Principles of Satellite Image Processing Principles of Satellite Image Processing Instructor: Prof. CSRE, IIT Bombay bkmohan@csre.iitb.ac.in Slot 5 Guest Lecture PCT and Band Arithmetic November 07, 2012 9.30 AM 10.55 AM IIT Bombay Slide 1 November

More information

Vectors To begin, let us describe an element of the state space as a point with numerical coordinates, that is x 1. x 2. x =

Vectors To begin, let us describe an element of the state space as a point with numerical coordinates, that is x 1. x 2. x = Linear Algebra Review Vectors To begin, let us describe an element of the state space as a point with numerical coordinates, that is x 1 x x = 2. x n Vectors of up to three dimensions are easy to diagram.

More information

Principal Component Analysis

Principal Component Analysis Principal Component Analysis Anders Øland David Christiansen 1 Introduction Principal Component Analysis, or PCA, is a commonly used multi-purpose technique in data analysis. It can be used for feature

More information

Data Preprocessing Tasks

Data Preprocessing Tasks Data Tasks 1 2 3 Data Reduction 4 We re here. 1 Dimensionality Reduction Dimensionality reduction is a commonly used approach for generating fewer features. Typically used because too many features can

More information

EECS490: Digital Image Processing. Lecture #26

EECS490: Digital Image Processing. Lecture #26 Lecture #26 Moments; invariant moments Eigenvector, principal component analysis Boundary coding Image primitives Image representation: trees, graphs Object recognition and classes Minimum distance classifiers

More information

LEC 2: Principal Component Analysis (PCA) A First Dimensionality Reduction Approach

LEC 2: Principal Component Analysis (PCA) A First Dimensionality Reduction Approach LEC 2: Principal Component Analysis (PCA) A First Dimensionality Reduction Approach Dr. Guangliang Chen February 9, 2016 Outline Introduction Review of linear algebra Matrix SVD PCA Motivation The digits

More information

Covariance to PCA. CS 510 Lecture #8 February 17, 2014

Covariance to PCA. CS 510 Lecture #8 February 17, 2014 Covariance to PCA CS 510 Lecture 8 February 17, 2014 Status Update Programming Assignment 2 is due March 7 th Expect questions about your progress at the start of class I still owe you Assignment 1 back

More information

Designing Information Devices and Systems II Fall 2018 Elad Alon and Miki Lustig Homework 9

Designing Information Devices and Systems II Fall 2018 Elad Alon and Miki Lustig Homework 9 EECS 16B Designing Information Devices and Systems II Fall 18 Elad Alon and Miki Lustig Homework 9 This homework is due Wednesday, October 31, 18, at 11:59pm. Self grades are due Monday, November 5, 18,

More information

Linear Subspace Models

Linear Subspace Models Linear Subspace Models Goal: Explore linear models of a data set. Motivation: A central question in vision concerns how we represent a collection of data vectors. The data vectors may be rasterized images,

More information

Principal Component Analysis (PCA) Theory, Practice, and Examples

Principal Component Analysis (PCA) Theory, Practice, and Examples Principal Component Analysis (PCA) Theory, Practice, and Examples Data Reduction summarization of data with many (p) variables by a smaller set of (k) derived (synthetic, composite) variables. p k n A

More information

A Peak to the World of Multivariate Statistical Analysis

A Peak to the World of Multivariate Statistical Analysis A Peak to the World of Multivariate Statistical Analysis Real Contents Real Real Real Why is it important to know a bit about the theory behind the methods? Real 5 10 15 20 Real 10 15 20 Figure: Multivariate

More information

Principal Component Analysis (PCA)

Principal Component Analysis (PCA) Principal Component Analysis (PCA) Additional reading can be found from non-assessed exercises (week 8) in this course unit teaching page. Textbooks: Sect. 6.3 in [1] and Ch. 12 in [2] Outline Introduction

More information

A Tutorial on Data Reduction. Principal Component Analysis Theoretical Discussion. By Shireen Elhabian and Aly Farag

A Tutorial on Data Reduction. Principal Component Analysis Theoretical Discussion. By Shireen Elhabian and Aly Farag A Tutorial on Data Reduction Principal Component Analysis Theoretical Discussion By Shireen Elhabian and Aly Farag University of Louisville, CVIP Lab November 2008 PCA PCA is A backbone of modern data

More information

Solving Linear Systems

Solving Linear Systems Solving Linear Systems Iterative Solutions Methods Philippe B. Laval KSU Fall 2015 Philippe B. Laval (KSU) Linear Systems Fall 2015 1 / 12 Introduction We continue looking how to solve linear systems of

More information

7. Variable extraction and dimensionality reduction

7. Variable extraction and dimensionality reduction 7. Variable extraction and dimensionality reduction The goal of the variable selection in the preceding chapter was to find least useful variables so that it would be possible to reduce the dimensionality

More information

PCA, Kernel PCA, ICA

PCA, Kernel PCA, ICA PCA, Kernel PCA, ICA Learning Representations. Dimensionality Reduction. Maria-Florina Balcan 04/08/2015 Big & High-Dimensional Data High-Dimensions = Lot of Features Document classification Features per

More information

Principal Component Analysis -- PCA (also called Karhunen-Loeve transformation)

Principal Component Analysis -- PCA (also called Karhunen-Loeve transformation) Principal Component Analysis -- PCA (also called Karhunen-Loeve transformation) PCA transforms the original input space into a lower dimensional space, by constructing dimensions that are linear combinations

More information

Functions of Several Variables

Functions of Several Variables Functions of Several Variables Extreme Values Philippe B. Laval KSU Today Philippe B. Laval (KSU) Extreme Values Today 1 / 18 Introduction In Calculus I (differential calculus for functions of one variable),

More information

Linear Algebra Methods for Data Mining

Linear Algebra Methods for Data Mining Linear Algebra Methods for Data Mining Saara Hyvönen, Saara.Hyvonen@cs.helsinki.fi Spring 2007 Linear Discriminant Analysis Linear Algebra Methods for Data Mining, Spring 2007, University of Helsinki Principal

More information

Singular Value Decomposition

Singular Value Decomposition Chapter 6 Singular Value Decomposition In Chapter 5, we derived a number of algorithms for computing the eigenvalues and eigenvectors of matrices A R n n. Having developed this machinery, we complete our

More information

Principal Components Analysis (PCA)

Principal Components Analysis (PCA) Principal Components Analysis (PCA) Principal Components Analysis (PCA) a technique for finding patterns in data of high dimension Outline:. Eigenvectors and eigenvalues. PCA: a) Getting the data b) Centering

More information

B553 Lecture 5: Matrix Algebra Review

B553 Lecture 5: Matrix Algebra Review B553 Lecture 5: Matrix Algebra Review Kris Hauser January 19, 2012 We have seen in prior lectures how vectors represent points in R n and gradients of functions. Matrices represent linear transformations

More information

Machine Learning (Spring 2012) Principal Component Analysis

Machine Learning (Spring 2012) Principal Component Analysis 1-71 Machine Learning (Spring 1) Principal Component Analysis Yang Xu This note is partly based on Chapter 1.1 in Chris Bishop s book on PRML and the lecture slides on PCA written by Carlos Guestrin in

More information

Vectors and Matrices Statistics with Vectors and Matrices

Vectors and Matrices Statistics with Vectors and Matrices Vectors and Matrices Statistics with Vectors and Matrices Lecture 3 September 7, 005 Analysis Lecture #3-9/7/005 Slide 1 of 55 Today s Lecture Vectors and Matrices (Supplement A - augmented with SAS proc

More information

Representation of Functions as Power Series

Representation of Functions as Power Series Representation of Functions as Power Series Philippe B. Laval KSU Today Philippe B. Laval (KSU) Functions as Power Series Today / Introduction In this section and the next, we develop several techniques

More information

UNIT 6: The singular value decomposition.

UNIT 6: The singular value decomposition. UNIT 6: The singular value decomposition. María Barbero Liñán Universidad Carlos III de Madrid Bachelor in Statistics and Business Mathematical methods II 2011-2012 A square matrix is symmetric if A T

More information

CS168: The Modern Algorithmic Toolbox Lecture #7: Understanding Principal Component Analysis (PCA)

CS168: The Modern Algorithmic Toolbox Lecture #7: Understanding Principal Component Analysis (PCA) CS68: The Modern Algorithmic Toolbox Lecture #7: Understanding Principal Component Analysis (PCA) Tim Roughgarden & Gregory Valiant April 0, 05 Introduction. Lecture Goal Principal components analysis

More information

14 Singular Value Decomposition

14 Singular Value Decomposition 14 Singular Value Decomposition For any high-dimensional data analysis, one s first thought should often be: can I use an SVD? The singular value decomposition is an invaluable analysis tool for dealing

More information

Linear Algebra Methods for Data Mining

Linear Algebra Methods for Data Mining Linear Algebra Methods for Data Mining Saara Hyvönen, Saara.Hyvonen@cs.helsinki.fi Spring 2007 The Singular Value Decomposition (SVD) continued Linear Algebra Methods for Data Mining, Spring 2007, University

More information

ANOVA: Analysis of Variance - Part I

ANOVA: Analysis of Variance - Part I ANOVA: Analysis of Variance - Part I The purpose of these notes is to discuss the theory behind the analysis of variance. It is a summary of the definitions and results presented in class with a few exercises.

More information

Principal Component Analysis

Principal Component Analysis Principal Component Analysis Yingyu Liang yliang@cs.wisc.edu Computer Sciences Department University of Wisconsin, Madison [based on slides from Nina Balcan] slide 1 Goals for the lecture you should understand

More information

Announcements (repeat) Principal Components Analysis

Announcements (repeat) Principal Components Analysis 4/7/7 Announcements repeat Principal Components Analysis CS 5 Lecture #9 April 4 th, 7 PA4 is due Monday, April 7 th Test # will be Wednesday, April 9 th Test #3 is Monday, May 8 th at 8AM Just hour long

More information

Singular Value Decomposition and Principal Component Analysis (PCA) I

Singular Value Decomposition and Principal Component Analysis (PCA) I Singular Value Decomposition and Principal Component Analysis (PCA) I Prof Ned Wingreen MOL 40/50 Microarray review Data per array: 0000 genes, I (green) i,i (red) i 000 000+ data points! The expression

More information

Matrix Algebra: Summary

Matrix Algebra: Summary May, 27 Appendix E Matrix Algebra: Summary ontents E. Vectors and Matrtices.......................... 2 E.. Notation.................................. 2 E..2 Special Types of Vectors.........................

More information

PRINCIPAL COMPONENT ANALYSIS

PRINCIPAL COMPONENT ANALYSIS PRINCIPAL COMPONENT ANALYSIS Dimensionality Reduction Tzompanaki Katerina Dimensionality Reduction Unsupervised learning Goal: Find hidden patterns in the data. Used for Visualization Data compression

More information

Dimensionality Reduction

Dimensionality Reduction 394 Chapter 11 Dimensionality Reduction There are many sources of data that can be viewed as a large matrix. We saw in Chapter 5 how the Web can be represented as a transition matrix. In Chapter 9, the

More information

PCA Review. CS 510 February 25 th, 2013

PCA Review. CS 510 February 25 th, 2013 PCA Review CS 510 February 25 th, 2013 Recall the goal: image matching Probe image, registered to gallery Registered Gallery of Images 3/7/13 CS 510, Image Computa5on, Ross Beveridge & Bruce Draper 2 Getting

More information

Numerical Methods I Singular Value Decomposition

Numerical Methods I Singular Value Decomposition Numerical Methods I Singular Value Decomposition Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 MATH-GA 2011.003 / CSCI-GA 2945.003, Fall 2014 October 9th, 2014 A. Donev (Courant Institute)

More information

Principal Components Analysis (PCA) and Singular Value Decomposition (SVD) with applications to Microarrays

Principal Components Analysis (PCA) and Singular Value Decomposition (SVD) with applications to Microarrays Principal Components Analysis (PCA) and Singular Value Decomposition (SVD) with applications to Microarrays Prof. Tesler Math 283 Fall 2015 Prof. Tesler Principal Components Analysis Math 283 / Fall 2015

More information

Differentiation - Important Theorems

Differentiation - Important Theorems Differentiation - Important Theorems Philippe B Laval KSU Spring 2012 Philippe B Laval (KSU) Differentiation - Important Theorems Spring 2012 1 / 10 Introduction We study several important theorems related

More information

CS4495/6495 Introduction to Computer Vision. 8B-L2 Principle Component Analysis (and its use in Computer Vision)

CS4495/6495 Introduction to Computer Vision. 8B-L2 Principle Component Analysis (and its use in Computer Vision) CS4495/6495 Introduction to Computer Vision 8B-L2 Principle Component Analysis (and its use in Computer Vision) Wavelength 2 Wavelength 2 Principal Components Principal components are all about the directions

More information

Linear Algebra Review. Vectors

Linear Algebra Review. Vectors Linear Algebra Review 9/4/7 Linear Algebra Review By Tim K. Marks UCSD Borrows heavily from: Jana Kosecka http://cs.gmu.edu/~kosecka/cs682.html Virginia de Sa (UCSD) Cogsci 8F Linear Algebra review Vectors

More information

Introduction to SVD and Applications

Introduction to SVD and Applications Introduction to SVD and Applications Eric Kostelich and Dave Kuhl MSRI Climate Change Summer School July 18, 2008 Introduction The goal of this exercise is to familiarize you with the basics of the singular

More information

Numerical Linear Algebra Homework Assignment - Week 2

Numerical Linear Algebra Homework Assignment - Week 2 Numerical Linear Algebra Homework Assignment - Week 2 Đoàn Trần Nguyên Tùng Student ID: 1411352 8th October 2016 Exercise 2.1: Show that if a matrix A is both triangular and unitary, then it is diagonal.

More information

Principal Components Theory Notes

Principal Components Theory Notes Principal Components Theory Notes Charles J. Geyer August 29, 2007 1 Introduction These are class notes for Stat 5601 (nonparametrics) taught at the University of Minnesota, Spring 2006. This not a theory

More information

Homework 1. Yuan Yao. September 18, 2011

Homework 1. Yuan Yao. September 18, 2011 Homework 1 Yuan Yao September 18, 2011 1. Singular Value Decomposition: The goal of this exercise is to refresh your memory about the singular value decomposition and matrix norms. A good reference to

More information

Machine Learning. Principal Components Analysis. Le Song. CSE6740/CS7641/ISYE6740, Fall 2012

Machine Learning. Principal Components Analysis. Le Song. CSE6740/CS7641/ISYE6740, Fall 2012 Machine Learning CSE6740/CS7641/ISYE6740, Fall 2012 Principal Components Analysis Le Song Lecture 22, Nov 13, 2012 Based on slides from Eric Xing, CMU Reading: Chap 12.1, CB book 1 2 Factor or Component

More information

More Linear Algebra. Edps/Soc 584, Psych 594. Carolyn J. Anderson

More Linear Algebra. Edps/Soc 584, Psych 594. Carolyn J. Anderson More Linear Algebra Edps/Soc 584, Psych 594 Carolyn J. Anderson Department of Educational Psychology I L L I N O I S university of illinois at urbana-champaign c Board of Trustees, University of Illinois

More information

Principal Component Analysis

Principal Component Analysis Principal Component Analysis Yuanzhen Shao MA 26500 Yuanzhen Shao PCA 1 / 13 Data as points in R n Assume that we have a collection of data in R n. x 11 x 21 x 12 S = {X 1 =., X x 22 2 =.,, X x m2 m =.

More information

Probabilistic Latent Semantic Analysis

Probabilistic Latent Semantic Analysis Probabilistic Latent Semantic Analysis Seungjin Choi Department of Computer Science and Engineering Pohang University of Science and Technology 77 Cheongam-ro, Nam-gu, Pohang 37673, Korea seungjin@postech.ac.kr

More information

Multivariate Statistics (I) 2. Principal Component Analysis (PCA)

Multivariate Statistics (I) 2. Principal Component Analysis (PCA) Multivariate Statistics (I) 2. Principal Component Analysis (PCA) 2.1 Comprehension of PCA 2.2 Concepts of PCs 2.3 Algebraic derivation of PCs 2.4 Selection and goodness-of-fit of PCs 2.5 Algebraic derivation

More information

Neuroscience Introduction

Neuroscience Introduction Neuroscience Introduction The brain As humans, we can identify galaxies light years away, we can study particles smaller than an atom. But we still haven t unlocked the mystery of the three pounds of matter

More information

Eigenimaging for Facial Recognition

Eigenimaging for Facial Recognition Eigenimaging for Facial Recognition Aaron Kosmatin, Clayton Broman December 2, 21 Abstract The interest of this paper is Principal Component Analysis, specifically its area of application to facial recognition

More information

Math 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination

Math 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination Math 0, Winter 07 Final Exam Review Chapter. Matrices and Gaussian Elimination { x + x =,. Different forms of a system of linear equations. Example: The x + 4x = 4. [ ] [ ] [ ] vector form (or the column

More information

IR-MAD Iteratively Re-weighted Multivariate Alteration Detection

IR-MAD Iteratively Re-weighted Multivariate Alteration Detection IR-MAD Iteratively Re-weighted Multivariate Alteration Detection Nielsen, A. A., Conradsen, K., & Simpson, J. J. (1998). Multivariate Alteration Detection (MAD) and MAF Postprocessing in Multispectral,

More information

Covariance and Principal Components

Covariance and Principal Components COMP3204/COMP6223: Computer Vision Covariance and Principal Components Jonathon Hare jsh2@ecs.soton.ac.uk Variance and Covariance Random Variables and Expected Values Mathematicians talk variance (and

More information

Image Registration Lecture 2: Vectors and Matrices

Image Registration Lecture 2: Vectors and Matrices Image Registration Lecture 2: Vectors and Matrices Prof. Charlene Tsai Lecture Overview Vectors Matrices Basics Orthogonal matrices Singular Value Decomposition (SVD) 2 1 Preliminary Comments Some of this

More information

MIT Final Exam Solutions, Spring 2017

MIT Final Exam Solutions, Spring 2017 MIT 8.6 Final Exam Solutions, Spring 7 Problem : For some real matrix A, the following vectors form a basis for its column space and null space: C(A) = span,, N(A) = span,,. (a) What is the size m n of

More information

CS246: Mining Massive Data Sets Winter Only one late period is allowed for this homework (11:59pm 2/14). General Instructions

CS246: Mining Massive Data Sets Winter Only one late period is allowed for this homework (11:59pm 2/14). General Instructions CS246: Mining Massive Data Sets Winter 2017 Problem Set 2 Due 11:59pm February 9, 2017 Only one late period is allowed for this homework (11:59pm 2/14). General Instructions Submission instructions: These

More information

1 Inner Product and Orthogonality

1 Inner Product and Orthogonality CSCI 4/Fall 6/Vora/GWU/Orthogonality and Norms Inner Product and Orthogonality Definition : The inner product of two vectors x and y, x x x =.., y =. x n y y... y n is denoted x, y : Note that n x, y =

More information

Recall the convention that, for us, all vectors are column vectors.

Recall the convention that, for us, all vectors are column vectors. Some linear algebra Recall the convention that, for us, all vectors are column vectors. 1. Symmetric matrices Let A be a real matrix. Recall that a complex number λ is an eigenvalue of A if there exists

More information