The Principal Component Analysis
|
|
- Hester Chase
- 5 years ago
- Views:
Transcription
1 The Principal Component Analysis Philippe B. Laval KSU Fall 2017 Philippe B. Laval (KSU) PCA Fall / 27
2 Introduction Every 80 minutes, the two Landsat satellites go around the world, recording images of our planet along a 185 Km wide path. Every 16 days, each satellite covers the entire surface of the planet. Every 8 days, every single location on the planet can be monitored. These images can be used by urban planners to study the rate and direction of population growth. They can be used to analyze soil moisture, vegetation growth, lakes and rivers. Governments can use them to detect and assess damages from natural disasters... Sensors on board the satellites acquire seven simultaneous images of every region by recording energy from separate wavelength bands. Each image is digitized and stored as a matrix., each number indicating the signal intensity of the corresponding pixel. Each of the seven image is one channel of a multichannel or multispectral image. The seven images of one fixed region typically contains redundant information. Yet, certain features will only appear in one or two images because their color or temperature was only captured by two of the seven sensors. Philippe B. Laval (KSU) PCA Fall / 27
3 Introduction Principal Component Analysis (PCA) is an effective way to suppress redundant information and provide in only one or two composite images most of the information from the initial data. Roughly speaking, the goal is to find a special linear combination of the images that combines each pixel of the seven images into one single new pixel. PCA can be applied to any data that consists of lists of measurements made on a collections of objects or individuals. For example, consider a chemical process that produces a plastic material. To monitor the process, 300 samples of the material produced are taken and each sample is subject to a battery of eight tests. The lab report for each sample is a vector in R 8 (8 tests). Since there are 300 samples, the set of these vectors is a matrix. Such a matrix is called the matrix of observations. Loosely speaking, we say that the process control data is eighth-dimensional. Philippe B. Laval (KSU) PCA Fall / 27
4 Introduction Example Suppose we have N college students for which we record their weights and heights. This is an example of two-dimensional data. Let [ X j be ] the observation vector in R 2 for the j th wj student that is X j = where w j is the weight (in Kg) and h j is the height (in m) of the jth student. Then, the [ matrix of observations ] has the form w1 w 2 w N = [X h 1 h 2 h 1 X 2 X N ]. Here is a possible matrix of N observations (randomly generated with height in meters and weight in kilograms) for N = 25 visualized on a scatter graph shown on the next slide. h j Philippe B. Laval (KSU) PCA Fall / 27
5 Introduction Figure: Scatter Plot of Observations Philippe B. Laval (KSU) PCA Fall / 27
6 Introduction We will use [X 1 X 2 X N ] to denote the matrix of observations. It will be a p N matrix meaning that we are observing (measuring) p different variables and we collect N samples of the measurements of these variables. x 1 x 2 We will let X = be the generic vector of all the variables we. x p are observing. Each X i will denote one measure of these variables, the ith measure. x i 1 x i 2 We will denote its components X i =.. When the number of variables being observed is small, we can use different names for the variables as we did in the example above. Philippe B. Laval (KSU) PCA Fall / 27 x i p
7 Mean and Covariance Definition (Sample Mean) Let [X 1 X 2 X N ] be a p N matrix of observations (meaning that for each object, we have p observations, each X i is a p 1 vector). The sample mean M of the observation vectors X 1, X 2,..., X N is given by M = 1 N N i=1 X i The reader will note that M is also a p 1 vector. Geometrically, the sample mean is the point at the "center" of the scatter plot. For i = 1, 2,..., N, we define X ] i = X i M and B = [ X1 X2 XN. The columns of B have a zero sample mean. Philippe B. Laval (KSU) PCA Fall / 27
8 Mean and Covariance Definition If the columns of a matrix have a zero sample mean, then B is said to be in mean-deviation form. Example The scatter plot of the observation matrix of the previous example put in mean-deviation form is shown in the picture on the next slide. You will note that the "center" of the data is now at the origin. Philippe B. Laval (KSU) PCA Fall / 27
9 Mean and Covariance Figure: Observation Data in Mean-Deviation Form Philippe B. Laval (KSU) PCA Fall / 27
10 Mean and Covariance Definition The sample covariance matrix or the covariance matrix is the p p matrix defined by S = 1 N 1 BBT Since any matrix of the form BB T is symmetric and positive semidefinite, so is S. Recall that a matrix A is positive definite (positive semidefinite) if for every nonzero vector x, x T Ax > 0 (x T Ax 0). Philippe B. Laval (KSU) PCA Fall / 27
11 Mean and Covariance Example Suppose that measurements are made on 4 individuals and the observation vectors are X 1 = 2, X 2 = 2, X 3 = 8, X 4 = Compute the sample mean and the covariance matrix. Philippe B. Laval (KSU) PCA Fall / 27
12 Mean and Covariance We should have found that the sample mean is M = The corresponding matrixin mean-deviation form is B = The covariance matrix is S = Philippe B. Laval (KSU) PCA Fall / 27
13 Mean and Covariance We now discuss the meaning of the entries in S, the covariance matrix, assuming that S = [s ij ]. We continue using the notation developed in this section. Philippe B. Laval (KSU) PCA Fall / 27
14 Mean and Covariance Definition (variance) For i = 1, 2,..., p, the diagonal entry s ii is called the variance of x i The variance of x j measures the spread of the values of x j. In the example above, from our computations, the variance of x 1 is 10, the variance of x 2 is 8 and the variance of x 3 is 32. What is important here is the relative size of these numbers. The fact that 32 > 10 indicates that the set of third entries contains a wider spread than the set of first entries. Definition (total variance) The total variance of the data is the sum of the diagonal entries in S. For a square matrix S (recall S is p p), the sum of the diagonal entries is called the trace of S and is denoted tr (S). Thus total variance = tr (S) Philippe B. Laval (KSU) PCA Fall / 27
15 Mean and Covariance Definition (covariance) The entry s ij for i i, in S is called the covariance of x i and x j. In the example above, looking at S, we see that the covariance between entries 1 and 3 is 0. This means that x 1 and x 3 are uncorrelated. Philippe B. Laval (KSU) PCA Fall / 27
16 PCA Using Eigenvalues As above, let us assume that our original variables are represented by the p 1 vector X and we perform N measurements we denote X 1, X 2,..., X N. For simplicity, we will assume that the p N matrix of observations [X 1 X 2 X N ] is already in mean-deviation form. If this were not the case, the first step would be to put it in that form. The goal of PCA is to find an orthogonal p p matrix, P = [u 1 u 2... u p ] that determines a change of variables X = PY that is x 1 x 2. x p = [u 1 u 2... u p ] with the property that the new variables y 1, y 2,..., y p are uncorrelated and are arranged in order of decreasing variance. y 1 y 2. y p Philippe B. Laval (KSU) PCA Fall / 27
17 PCA Using Eigenvalues A Question: From the statements on the previous slide, what can we say about the covariance matrix corresponding to the new variables? Another Question? How can we achieve the answer to the previous question? Philippe B. Laval (KSU) PCA Fall / 27
18 PCA Using Eigenvalues The orthogonal change of variable X = PY means that each observation vector X i is transformed into a new vector we call Y i such that X i = PY i. Therefore, Y i = P 1 X i = P T X i since P is supposed to be orthogonal. This is true for i = 1, 2,..., N. It is not hard to see (see homework) that for each orthogonal matrix P, the covariance matrix for Y 1, Y 2,..., Y N is P T SP where S is the covariance matrix for X 1, X 2,..., X N. So, the desired orthogonal matrix P is the one that makes P T SP diagonal (why?). Let D be a diagonal matrix with the eigenvalues λ 1, λ 2,..., λ p of S on the diagonal arranged so that λ 1 λ 2... λ p 0 and let P be an orthogonal matrix whose columns are the unit eigenvectors u 1, u 2,..., u p corresponding to the eigenvalues λ 1 λ 2... λ p. Then, S = PDP T and D = P T SP. Philippe B. Laval (KSU) PCA Fall / 27
19 PCA Using Eigenvalues Definition The unit eigenvectors u 1, u 2,..., u p of the covariance matrix S are called the principal components of the data in the matrix of observations. The first principal component is the eigenvector corresponding to the largest eigenvalue of S. The second principal component is the eigenvector corresponding to the second largest eigenvalue of S and so on. The first principal component u 1 determines the new variable y 1 in the following way. Let c 1, c 2,..., c p be the entries in u 1. Since u T 1 is the first row of P T, the equation Y = P T X shows that p y 1 = u T 1 X = c i x i Thus y 1 is a linear combination of the original variable using the entries in u 1 as weights. In a similar way, u 2 determines the entries in y 2, and so on. Philippe B. Laval (KSU) PCA Fall / 27 i=1
20 PCA Using Eigenvalues: Summary Given a matrix of observation B = [X 1 X 2 X N ] which is assumed to be in mean-deviation form, to perform a principal component analysis, we do the following: 1 Find the covariance matrix S = 1 N 1 BBT 2 Diagonalize S using eigenvalues. We let λ 1, λ 2,..., λ p be the eigenvalues of S arranged so that λ 1 λ 2... λ p 0 and let P be an orthogonal matrix whose columns are the unit eigenvectors u 1, u 2,..., u p corresponding to the eigenvalues λ 1 λ 2... λ p. 3 The unit eigenvectors u 1, u 2,..., u p of the covariance matrix S are called the principal components of the data in the matrix of observations. 4 The new uncorrelated variables are defined by Y = P T X where Y,like X is p 1. The covariance matrix of Y is P T SP. 5 The components y 1, y 2,..., y p of y can be expressed in terms of the components x 1, x 2,..., x p of X by y i = u T i X. Philippe B. Laval (KSU) PCA Fall / 27
21 PCA Using Eigenvalues Example Suppose the covariance matrix of some data is S = Find the principal components of the data, list the new variable determined by the first principal component and give the diagonal form of S. Philippe B. Laval (KSU) PCA Fall / 27
22 Dimension Reduction PCA is valuable when most of the variation in the data is due to variations in only a few of the new variables y 1, y 2,..., y p. It can be shown that an orthogonal change of variables X = PY does not change the total variance of the data. This means that if S = PDP T then { } { } total variance total variance = of x 1, x 2,..., x p of y 1, y 2,..., y p = tr (D) = λ 1 + λ λ p The variance of y i is λ i, and the quotient measures the tr (S) fraction of the total variance captured by y i. λ i We can eliminate the variables for which is small. tr (S) Philippe B. Laval (KSU) PCA Fall / 27 λ i
23 Dimension Reduction Example In the example above, compute the various percentages of variance captured by each variable. Example Consider the following data: Boy #1 #2 #3 #4 #5 Weight (lb) Height (in) Find the covariance matrix for the data. 2 Make a principal component analysis of the data to find a single size index that explains most of the variation in the data. Philippe B. Laval (KSU) PCA Fall / 27
24 PCA Using SVD The goal is the same as above, what changes is how we achieve it. As above, we will assume that the p N matrix of observations [X 1 X 2 X N ] is already in mean-deviation form. If this were not the case, the first step would be to put it in that form. Assume the SVD of the matrix of observations [X 1 X 2 X N ] is UΣV T. Here, U is p p, Σ is p N and V T is N N hence V is N N. We define the new variable Y by X = UY or Y = U T X. It can be shown that the covariance matrix 1 corresponding to Y is Y is N 1 Σ2 which is already in diagonal form; this is what we wanted. In this case, the principal components will be the columns of U. If we let U = [u 1 u 2... u p ] then the equation Y = P T X p shows that y 1 = u T 1 X = c i x i where c 1, c 2,..., c p are the entries in u 1. i=1 Thus y 1 is a linear combination of the original variable using the entries in u 1 as weights. In a similar way, u 2 determines the entries in y 2, and so on. Philippe B. Laval (KSU) PCA Fall / 27
25 PCA Using SVD: Summary Given a matrix of observation B = [X 1 X 2 X N ] which is assumed to be in mean-deviation form, to perform a principal component analysis, we do the following: 1 Find a SVD for B, say B = UΣV T. 2 If we let U = [u 1 u 2... u p ] then we define Y = U T X. The covariance 1 matrix of Y is N 1 Σ2. Y is the new uncorrelated variable. 1 3 We let λ 1, λ 2,..., λ p be the diagonal elements of N 1 Σ2 (automatically arranged so that λ 1 λ 2... λ p 0). Then λ i is the varaince of y i which is also the variance of x i. 4 The unit vectors u 1, u 2,..., u p are called the principal components of the data in the matrix of observations. We can get all the components we were able to get with the eigenvalue diagonalization. We simply got them differently. MATLAB uses this technique as it is a little bit more robust than the eigenvalue method. Philippe B. Laval (KSU) PCA Fall / 27
26 PCA and MATLAB Newer versions of MATLAB have a function to perform this, it is called pca. You can get help for it in MATLAB. Two important things to remember about MATLAB s pca: 1 It uses the SVD approach to find the principal components and related information. 2 It works differently than what is explained in these notes in the following sense. MATLAB s matrix of observation is the transpose of the matrix of observations used in these notes. In other words, the columns of our matrix of observations contain the various variables used for our data. For MATLAB, these variables are stored as the rows of the matrix of observations. Suppose that A is our matrix of observations. To use MATLAB s PCA, use pca(a ), where A is the transpose of A. Philippe B. Laval (KSU) PCA Fall / 27
27 Exercises See the problems at the end of the notes on Principal Component Analysis. Philippe B. Laval (KSU) PCA Fall / 27
Independent Component Analysis
Independent Component Analysis Philippe B. Laval KSU Fall 2017 Philippe B. Laval (KSU) ICA Fall 2017 1 / 18 Introduction Independent Component Analysis (ICA) falls under the broader topic of Blind Source
More informationThe Singular Value Decomposition
The Singular Value Decomposition Philippe B. Laval KSU Fall 2015 Philippe B. Laval (KSU) SVD Fall 2015 1 / 13 Review of Key Concepts We review some key definitions and results about matrices that will
More informationGEOG 4110/5100 Advanced Remote Sensing Lecture 15
GEOG 4110/5100 Advanced Remote Sensing Lecture 15 Principal Component Analysis Relevant reading: Richards. Chapters 6.3* http://www.ce.yildiz.edu.tr/personal/songul/file/1097/principal_components.pdf *For
More informationIntroduction to Machine Learning
10-701 Introduction to Machine Learning PCA Slides based on 18-661 Fall 2018 PCA Raw data can be Complex, High-dimensional To understand a phenomenon we measure various related quantities If we knew what
More informationLecture: Face Recognition and Feature Reduction
Lecture: Face Recognition and Feature Reduction Juan Carlos Niebles and Ranjay Krishna Stanford Vision and Learning Lab 1 Recap - Curse of dimensionality Assume 5000 points uniformly distributed in the
More informationLecture: Face Recognition and Feature Reduction
Lecture: Face Recognition and Feature Reduction Juan Carlos Niebles and Ranjay Krishna Stanford Vision and Learning Lab Lecture 11-1 Recap - Curse of dimensionality Assume 5000 points uniformly distributed
More informationEigenvalues and Eigenvectors
Eigenvalues and Eigenvectors Philippe B. Laval KSU Fall 2015 Philippe B. Laval (KSU) Eigenvalues and Eigenvectors Fall 2015 1 / 14 Introduction We define eigenvalues and eigenvectors. We discuss how to
More information12.2 Dimensionality Reduction
510 Chapter 12 of this dimensionality problem, regularization techniques such as SVD are almost always needed to perform the covariance matrix inversion. Because it appears to be a fundamental property
More informationSingular Value Decomposition and Digital Image Compression
Singular Value Decomposition and Digital Image Compression Chris Bingham December 1, 016 Page 1 of Abstract The purpose of this document is to be a very basic introduction to the singular value decomposition
More informationA A x i x j i j (i, j) (j, i) Let. Compute the value of for and
7.2 - Quadratic Forms quadratic form on is a function defined on whose value at a vector in can be computed by an expression of the form, where is an symmetric matrix. The matrix R n Q R n x R n Q(x) =
More informationDimensionality Reduction: PCA. Nicholas Ruozzi University of Texas at Dallas
Dimensionality Reduction: PCA Nicholas Ruozzi University of Texas at Dallas Eigenvalues λ is an eigenvalue of a matrix A R n n if the linear system Ax = λx has at least one non-zero solution If Ax = λx
More informationThe Singular Value Decomposition (SVD) and Principal Component Analysis (PCA)
Chapter 5 The Singular Value Decomposition (SVD) and Principal Component Analysis (PCA) 5.1 Basics of SVD 5.1.1 Review of Key Concepts We review some key definitions and results about matrices that will
More informationReview of Linear Algebra
Review of Linear Algebra Definitions An m n (read "m by n") matrix, is a rectangular array of entries, where m is the number of rows and n the number of columns. 2 Definitions (Con t) A is square if m=
More informationNotes on singular value decomposition for Math 54. Recall that if A is a symmetric n n matrix, then A has real eigenvalues A = P DP 1 A = P DP T.
Notes on singular value decomposition for Math 54 Recall that if A is a symmetric n n matrix, then A has real eigenvalues λ 1,, λ n (possibly repeated), and R n has an orthonormal basis v 1,, v n, where
More informationSingular Value Decompsition
Singular Value Decompsition Massoud Malek One of the most useful results from linear algebra, is a matrix decomposition known as the singular value decomposition It has many useful applications in almost
More informationSolving Linear Systems
Solving Linear Systems Iterative Solutions Methods Philippe B. Laval KSU Fall 207 Philippe B. Laval (KSU) Linear Systems Fall 207 / 2 Introduction We continue looking how to solve linear systems of the
More informationReview of Functions. Functions. Philippe B. Laval. Current Semester KSU. Philippe B. Laval (KSU) Functions Current Semester 1 / 12
Review of Functions Functions Philippe B. Laval KSU Current Semester Philippe B. Laval (KSU) Functions Current Semester 1 / 12 Introduction Students are expected to know the following concepts about functions:
More informationGopalkrishna Veni. Project 4 (Active Shape Models)
Gopalkrishna Veni Project 4 (Active Shape Models) Introduction Active shape Model (ASM) is a technique of building a model by learning the variability patterns from training datasets. ASMs try to deform
More informationLinear Algebra Review. Fei-Fei Li
Linear Algebra Review Fei-Fei Li 1 / 51 Vectors Vectors and matrices are just collections of ordered numbers that represent something: movements in space, scaling factors, pixel brightnesses, etc. A vector
More information1 Principal component analysis and dimensional reduction
Linear Algebra Working Group :: Day 3 Note: All vector spaces will be finite-dimensional vector spaces over the field R. 1 Principal component analysis and dimensional reduction Definition 1.1. Given an
More informationSingular Value Decomposition
Singular Value Decomposition Motivatation The diagonalization theorem play a part in many interesting applications. Unfortunately not all matrices can be factored as A = PDP However a factorization A =
More informationMLCC 2015 Dimensionality Reduction and PCA
MLCC 2015 Dimensionality Reduction and PCA Lorenzo Rosasco UNIGE-MIT-IIT June 25, 2015 Outline PCA & Reconstruction PCA and Maximum Variance PCA and Associated Eigenproblem Beyond the First Principal Component
More informationAppendix A: Matrices
Appendix A: Matrices A matrix is a rectangular array of numbers Such arrays have rows and columns The numbers of rows and columns are referred to as the dimensions of a matrix A matrix with, say, 5 rows
More informationLinear Algebra Review. Fei-Fei Li
Linear Algebra Review Fei-Fei Li 1 / 37 Vectors Vectors and matrices are just collections of ordered numbers that represent something: movements in space, scaling factors, pixel brightnesses, etc. A vector
More informationPRINCIPAL COMPONENTS ANALYSIS
121 CHAPTER 11 PRINCIPAL COMPONENTS ANALYSIS We now have the tools necessary to discuss one of the most important concepts in mathematical statistics: Principal Components Analysis (PCA). PCA involves
More informationNumerical Linear Algebra
Numerical Linear Algebra Direct Methods Philippe B. Laval KSU Fall 2017 Philippe B. Laval (KSU) Linear Systems: Direct Solution Methods Fall 2017 1 / 14 Introduction The solution of linear systems is one
More informationConceptual Questions for Review
Conceptual Questions for Review Chapter 1 1.1 Which vectors are linear combinations of v = (3, 1) and w = (4, 3)? 1.2 Compare the dot product of v = (3, 1) and w = (4, 3) to the product of their lengths.
More informationLinear Algebra & Geometry why is linear algebra useful in computer vision?
Linear Algebra & Geometry why is linear algebra useful in computer vision? References: -Any book on linear algebra! -[HZ] chapters 2, 4 Some of the slides in this lecture are courtesy to Prof. Octavia
More informationLinear Algebra & Geometry why is linear algebra useful in computer vision?
Linear Algebra & Geometry why is linear algebra useful in computer vision? References: -Any book on linear algebra! -[HZ] chapters 2, 4 Some of the slides in this lecture are courtesy to Prof. Octavia
More information15 Singular Value Decomposition
15 Singular Value Decomposition For any high-dimensional data analysis, one s first thought should often be: can I use an SVD? The singular value decomposition is an invaluable analysis tool for dealing
More informationPrinciple Components Analysis (PCA) Relationship Between a Linear Combination of Variables and Axes Rotation for PCA
Principle Components Analysis (PCA) Relationship Between a Linear Combination of Variables and Axes Rotation for PCA Principle Components Analysis: Uses one group of variables (we will call this X) In
More informationCHAPTER 4 PRINCIPAL COMPONENT ANALYSIS-BASED FUSION
59 CHAPTER 4 PRINCIPAL COMPONENT ANALYSIS-BASED FUSION 4. INTRODUCTION Weighted average-based fusion algorithms are one of the widely used fusion methods for multi-sensor data integration. These methods
More informationBackground Mathematics (2/2) 1. David Barber
Background Mathematics (2/2) 1 David Barber University College London Modified by Samson Cheung (sccheung@ieee.org) 1 These slides accompany the book Bayesian Reasoning and Machine Learning. The book and
More informationPrincipal Component Analysis (PCA) Our starting point consists of T observations from N variables, which will be arranged in an T N matrix R,
Principal Component Analysis (PCA) PCA is a widely used statistical tool for dimension reduction. The objective of PCA is to find common factors, the so called principal components, in form of linear combinations
More informationLecture 13. Principal Component Analysis. Brett Bernstein. April 25, CDS at NYU. Brett Bernstein (CDS at NYU) Lecture 13 April 25, / 26
Principal Component Analysis Brett Bernstein CDS at NYU April 25, 2017 Brett Bernstein (CDS at NYU) Lecture 13 April 25, 2017 1 / 26 Initial Question Intro Question Question Let S R n n be symmetric. 1
More informationCS168: The Modern Algorithmic Toolbox Lecture #8: How PCA Works
CS68: The Modern Algorithmic Toolbox Lecture #8: How PCA Works Tim Roughgarden & Gregory Valiant April 20, 206 Introduction Last lecture introduced the idea of principal components analysis (PCA). The
More informationIntroduction to Machine Learning. PCA and Spectral Clustering. Introduction to Machine Learning, Slides: Eran Halperin
1 Introduction to Machine Learning PCA and Spectral Clustering Introduction to Machine Learning, 2013-14 Slides: Eran Halperin Singular Value Decomposition (SVD) The singular value decomposition (SVD)
More informationFunctions of Several Variables
Functions of Several Variables Extreme Values Philippe B Laval KSU April 9, 2012 Philippe B Laval (KSU) Functions of Several Variables April 9, 2012 1 / 13 Introduction In Calculus I (differential calculus
More information7. Symmetric Matrices and Quadratic Forms
Linear Algebra 7. Symmetric Matrices and Quadratic Forms CSIE NCU 1 7. Symmetric Matrices and Quadratic Forms 7.1 Diagonalization of symmetric matrices 2 7.2 Quadratic forms.. 9 7.4 The singular value
More informationGNR401 Principles of Satellite Image Processing
Principles of Satellite Image Processing Instructor: Prof. CSRE, IIT Bombay bkmohan@csre.iitb.ac.in Slot 5 Guest Lecture PCT and Band Arithmetic November 07, 2012 9.30 AM 10.55 AM IIT Bombay Slide 1 November
More informationVectors To begin, let us describe an element of the state space as a point with numerical coordinates, that is x 1. x 2. x =
Linear Algebra Review Vectors To begin, let us describe an element of the state space as a point with numerical coordinates, that is x 1 x x = 2. x n Vectors of up to three dimensions are easy to diagram.
More informationPrincipal Component Analysis
Principal Component Analysis Anders Øland David Christiansen 1 Introduction Principal Component Analysis, or PCA, is a commonly used multi-purpose technique in data analysis. It can be used for feature
More informationData Preprocessing Tasks
Data Tasks 1 2 3 Data Reduction 4 We re here. 1 Dimensionality Reduction Dimensionality reduction is a commonly used approach for generating fewer features. Typically used because too many features can
More informationEECS490: Digital Image Processing. Lecture #26
Lecture #26 Moments; invariant moments Eigenvector, principal component analysis Boundary coding Image primitives Image representation: trees, graphs Object recognition and classes Minimum distance classifiers
More informationLEC 2: Principal Component Analysis (PCA) A First Dimensionality Reduction Approach
LEC 2: Principal Component Analysis (PCA) A First Dimensionality Reduction Approach Dr. Guangliang Chen February 9, 2016 Outline Introduction Review of linear algebra Matrix SVD PCA Motivation The digits
More informationCovariance to PCA. CS 510 Lecture #8 February 17, 2014
Covariance to PCA CS 510 Lecture 8 February 17, 2014 Status Update Programming Assignment 2 is due March 7 th Expect questions about your progress at the start of class I still owe you Assignment 1 back
More informationDesigning Information Devices and Systems II Fall 2018 Elad Alon and Miki Lustig Homework 9
EECS 16B Designing Information Devices and Systems II Fall 18 Elad Alon and Miki Lustig Homework 9 This homework is due Wednesday, October 31, 18, at 11:59pm. Self grades are due Monday, November 5, 18,
More informationLinear Subspace Models
Linear Subspace Models Goal: Explore linear models of a data set. Motivation: A central question in vision concerns how we represent a collection of data vectors. The data vectors may be rasterized images,
More informationPrincipal Component Analysis (PCA) Theory, Practice, and Examples
Principal Component Analysis (PCA) Theory, Practice, and Examples Data Reduction summarization of data with many (p) variables by a smaller set of (k) derived (synthetic, composite) variables. p k n A
More informationA Peak to the World of Multivariate Statistical Analysis
A Peak to the World of Multivariate Statistical Analysis Real Contents Real Real Real Why is it important to know a bit about the theory behind the methods? Real 5 10 15 20 Real 10 15 20 Figure: Multivariate
More informationPrincipal Component Analysis (PCA)
Principal Component Analysis (PCA) Additional reading can be found from non-assessed exercises (week 8) in this course unit teaching page. Textbooks: Sect. 6.3 in [1] and Ch. 12 in [2] Outline Introduction
More informationA Tutorial on Data Reduction. Principal Component Analysis Theoretical Discussion. By Shireen Elhabian and Aly Farag
A Tutorial on Data Reduction Principal Component Analysis Theoretical Discussion By Shireen Elhabian and Aly Farag University of Louisville, CVIP Lab November 2008 PCA PCA is A backbone of modern data
More informationSolving Linear Systems
Solving Linear Systems Iterative Solutions Methods Philippe B. Laval KSU Fall 2015 Philippe B. Laval (KSU) Linear Systems Fall 2015 1 / 12 Introduction We continue looking how to solve linear systems of
More information7. Variable extraction and dimensionality reduction
7. Variable extraction and dimensionality reduction The goal of the variable selection in the preceding chapter was to find least useful variables so that it would be possible to reduce the dimensionality
More informationPCA, Kernel PCA, ICA
PCA, Kernel PCA, ICA Learning Representations. Dimensionality Reduction. Maria-Florina Balcan 04/08/2015 Big & High-Dimensional Data High-Dimensions = Lot of Features Document classification Features per
More informationPrincipal Component Analysis -- PCA (also called Karhunen-Loeve transformation)
Principal Component Analysis -- PCA (also called Karhunen-Loeve transformation) PCA transforms the original input space into a lower dimensional space, by constructing dimensions that are linear combinations
More informationFunctions of Several Variables
Functions of Several Variables Extreme Values Philippe B. Laval KSU Today Philippe B. Laval (KSU) Extreme Values Today 1 / 18 Introduction In Calculus I (differential calculus for functions of one variable),
More informationLinear Algebra Methods for Data Mining
Linear Algebra Methods for Data Mining Saara Hyvönen, Saara.Hyvonen@cs.helsinki.fi Spring 2007 Linear Discriminant Analysis Linear Algebra Methods for Data Mining, Spring 2007, University of Helsinki Principal
More informationSingular Value Decomposition
Chapter 6 Singular Value Decomposition In Chapter 5, we derived a number of algorithms for computing the eigenvalues and eigenvectors of matrices A R n n. Having developed this machinery, we complete our
More informationPrincipal Components Analysis (PCA)
Principal Components Analysis (PCA) Principal Components Analysis (PCA) a technique for finding patterns in data of high dimension Outline:. Eigenvectors and eigenvalues. PCA: a) Getting the data b) Centering
More informationB553 Lecture 5: Matrix Algebra Review
B553 Lecture 5: Matrix Algebra Review Kris Hauser January 19, 2012 We have seen in prior lectures how vectors represent points in R n and gradients of functions. Matrices represent linear transformations
More informationMachine Learning (Spring 2012) Principal Component Analysis
1-71 Machine Learning (Spring 1) Principal Component Analysis Yang Xu This note is partly based on Chapter 1.1 in Chris Bishop s book on PRML and the lecture slides on PCA written by Carlos Guestrin in
More informationVectors and Matrices Statistics with Vectors and Matrices
Vectors and Matrices Statistics with Vectors and Matrices Lecture 3 September 7, 005 Analysis Lecture #3-9/7/005 Slide 1 of 55 Today s Lecture Vectors and Matrices (Supplement A - augmented with SAS proc
More informationRepresentation of Functions as Power Series
Representation of Functions as Power Series Philippe B. Laval KSU Today Philippe B. Laval (KSU) Functions as Power Series Today / Introduction In this section and the next, we develop several techniques
More informationUNIT 6: The singular value decomposition.
UNIT 6: The singular value decomposition. María Barbero Liñán Universidad Carlos III de Madrid Bachelor in Statistics and Business Mathematical methods II 2011-2012 A square matrix is symmetric if A T
More informationCS168: The Modern Algorithmic Toolbox Lecture #7: Understanding Principal Component Analysis (PCA)
CS68: The Modern Algorithmic Toolbox Lecture #7: Understanding Principal Component Analysis (PCA) Tim Roughgarden & Gregory Valiant April 0, 05 Introduction. Lecture Goal Principal components analysis
More information14 Singular Value Decomposition
14 Singular Value Decomposition For any high-dimensional data analysis, one s first thought should often be: can I use an SVD? The singular value decomposition is an invaluable analysis tool for dealing
More informationLinear Algebra Methods for Data Mining
Linear Algebra Methods for Data Mining Saara Hyvönen, Saara.Hyvonen@cs.helsinki.fi Spring 2007 The Singular Value Decomposition (SVD) continued Linear Algebra Methods for Data Mining, Spring 2007, University
More informationANOVA: Analysis of Variance - Part I
ANOVA: Analysis of Variance - Part I The purpose of these notes is to discuss the theory behind the analysis of variance. It is a summary of the definitions and results presented in class with a few exercises.
More informationPrincipal Component Analysis
Principal Component Analysis Yingyu Liang yliang@cs.wisc.edu Computer Sciences Department University of Wisconsin, Madison [based on slides from Nina Balcan] slide 1 Goals for the lecture you should understand
More informationAnnouncements (repeat) Principal Components Analysis
4/7/7 Announcements repeat Principal Components Analysis CS 5 Lecture #9 April 4 th, 7 PA4 is due Monday, April 7 th Test # will be Wednesday, April 9 th Test #3 is Monday, May 8 th at 8AM Just hour long
More informationSingular Value Decomposition and Principal Component Analysis (PCA) I
Singular Value Decomposition and Principal Component Analysis (PCA) I Prof Ned Wingreen MOL 40/50 Microarray review Data per array: 0000 genes, I (green) i,i (red) i 000 000+ data points! The expression
More informationMatrix Algebra: Summary
May, 27 Appendix E Matrix Algebra: Summary ontents E. Vectors and Matrtices.......................... 2 E.. Notation.................................. 2 E..2 Special Types of Vectors.........................
More informationPRINCIPAL COMPONENT ANALYSIS
PRINCIPAL COMPONENT ANALYSIS Dimensionality Reduction Tzompanaki Katerina Dimensionality Reduction Unsupervised learning Goal: Find hidden patterns in the data. Used for Visualization Data compression
More informationDimensionality Reduction
394 Chapter 11 Dimensionality Reduction There are many sources of data that can be viewed as a large matrix. We saw in Chapter 5 how the Web can be represented as a transition matrix. In Chapter 9, the
More informationPCA Review. CS 510 February 25 th, 2013
PCA Review CS 510 February 25 th, 2013 Recall the goal: image matching Probe image, registered to gallery Registered Gallery of Images 3/7/13 CS 510, Image Computa5on, Ross Beveridge & Bruce Draper 2 Getting
More informationNumerical Methods I Singular Value Decomposition
Numerical Methods I Singular Value Decomposition Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 MATH-GA 2011.003 / CSCI-GA 2945.003, Fall 2014 October 9th, 2014 A. Donev (Courant Institute)
More informationPrincipal Components Analysis (PCA) and Singular Value Decomposition (SVD) with applications to Microarrays
Principal Components Analysis (PCA) and Singular Value Decomposition (SVD) with applications to Microarrays Prof. Tesler Math 283 Fall 2015 Prof. Tesler Principal Components Analysis Math 283 / Fall 2015
More informationDifferentiation - Important Theorems
Differentiation - Important Theorems Philippe B Laval KSU Spring 2012 Philippe B Laval (KSU) Differentiation - Important Theorems Spring 2012 1 / 10 Introduction We study several important theorems related
More informationCS4495/6495 Introduction to Computer Vision. 8B-L2 Principle Component Analysis (and its use in Computer Vision)
CS4495/6495 Introduction to Computer Vision 8B-L2 Principle Component Analysis (and its use in Computer Vision) Wavelength 2 Wavelength 2 Principal Components Principal components are all about the directions
More informationLinear Algebra Review. Vectors
Linear Algebra Review 9/4/7 Linear Algebra Review By Tim K. Marks UCSD Borrows heavily from: Jana Kosecka http://cs.gmu.edu/~kosecka/cs682.html Virginia de Sa (UCSD) Cogsci 8F Linear Algebra review Vectors
More informationIntroduction to SVD and Applications
Introduction to SVD and Applications Eric Kostelich and Dave Kuhl MSRI Climate Change Summer School July 18, 2008 Introduction The goal of this exercise is to familiarize you with the basics of the singular
More informationNumerical Linear Algebra Homework Assignment - Week 2
Numerical Linear Algebra Homework Assignment - Week 2 Đoàn Trần Nguyên Tùng Student ID: 1411352 8th October 2016 Exercise 2.1: Show that if a matrix A is both triangular and unitary, then it is diagonal.
More informationPrincipal Components Theory Notes
Principal Components Theory Notes Charles J. Geyer August 29, 2007 1 Introduction These are class notes for Stat 5601 (nonparametrics) taught at the University of Minnesota, Spring 2006. This not a theory
More informationHomework 1. Yuan Yao. September 18, 2011
Homework 1 Yuan Yao September 18, 2011 1. Singular Value Decomposition: The goal of this exercise is to refresh your memory about the singular value decomposition and matrix norms. A good reference to
More informationMachine Learning. Principal Components Analysis. Le Song. CSE6740/CS7641/ISYE6740, Fall 2012
Machine Learning CSE6740/CS7641/ISYE6740, Fall 2012 Principal Components Analysis Le Song Lecture 22, Nov 13, 2012 Based on slides from Eric Xing, CMU Reading: Chap 12.1, CB book 1 2 Factor or Component
More informationMore Linear Algebra. Edps/Soc 584, Psych 594. Carolyn J. Anderson
More Linear Algebra Edps/Soc 584, Psych 594 Carolyn J. Anderson Department of Educational Psychology I L L I N O I S university of illinois at urbana-champaign c Board of Trustees, University of Illinois
More informationPrincipal Component Analysis
Principal Component Analysis Yuanzhen Shao MA 26500 Yuanzhen Shao PCA 1 / 13 Data as points in R n Assume that we have a collection of data in R n. x 11 x 21 x 12 S = {X 1 =., X x 22 2 =.,, X x m2 m =.
More informationProbabilistic Latent Semantic Analysis
Probabilistic Latent Semantic Analysis Seungjin Choi Department of Computer Science and Engineering Pohang University of Science and Technology 77 Cheongam-ro, Nam-gu, Pohang 37673, Korea seungjin@postech.ac.kr
More informationMultivariate Statistics (I) 2. Principal Component Analysis (PCA)
Multivariate Statistics (I) 2. Principal Component Analysis (PCA) 2.1 Comprehension of PCA 2.2 Concepts of PCs 2.3 Algebraic derivation of PCs 2.4 Selection and goodness-of-fit of PCs 2.5 Algebraic derivation
More informationNeuroscience Introduction
Neuroscience Introduction The brain As humans, we can identify galaxies light years away, we can study particles smaller than an atom. But we still haven t unlocked the mystery of the three pounds of matter
More informationEigenimaging for Facial Recognition
Eigenimaging for Facial Recognition Aaron Kosmatin, Clayton Broman December 2, 21 Abstract The interest of this paper is Principal Component Analysis, specifically its area of application to facial recognition
More informationMath 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination
Math 0, Winter 07 Final Exam Review Chapter. Matrices and Gaussian Elimination { x + x =,. Different forms of a system of linear equations. Example: The x + 4x = 4. [ ] [ ] [ ] vector form (or the column
More informationIR-MAD Iteratively Re-weighted Multivariate Alteration Detection
IR-MAD Iteratively Re-weighted Multivariate Alteration Detection Nielsen, A. A., Conradsen, K., & Simpson, J. J. (1998). Multivariate Alteration Detection (MAD) and MAF Postprocessing in Multispectral,
More informationCovariance and Principal Components
COMP3204/COMP6223: Computer Vision Covariance and Principal Components Jonathon Hare jsh2@ecs.soton.ac.uk Variance and Covariance Random Variables and Expected Values Mathematicians talk variance (and
More informationImage Registration Lecture 2: Vectors and Matrices
Image Registration Lecture 2: Vectors and Matrices Prof. Charlene Tsai Lecture Overview Vectors Matrices Basics Orthogonal matrices Singular Value Decomposition (SVD) 2 1 Preliminary Comments Some of this
More informationMIT Final Exam Solutions, Spring 2017
MIT 8.6 Final Exam Solutions, Spring 7 Problem : For some real matrix A, the following vectors form a basis for its column space and null space: C(A) = span,, N(A) = span,,. (a) What is the size m n of
More informationCS246: Mining Massive Data Sets Winter Only one late period is allowed for this homework (11:59pm 2/14). General Instructions
CS246: Mining Massive Data Sets Winter 2017 Problem Set 2 Due 11:59pm February 9, 2017 Only one late period is allowed for this homework (11:59pm 2/14). General Instructions Submission instructions: These
More information1 Inner Product and Orthogonality
CSCI 4/Fall 6/Vora/GWU/Orthogonality and Norms Inner Product and Orthogonality Definition : The inner product of two vectors x and y, x x x =.., y =. x n y y... y n is denoted x, y : Note that n x, y =
More informationRecall the convention that, for us, all vectors are column vectors.
Some linear algebra Recall the convention that, for us, all vectors are column vectors. 1. Symmetric matrices Let A be a real matrix. Recall that a complex number λ is an eigenvalue of A if there exists
More information