GEOG 4110/5100 Advanced Remote Sensing Lecture 15

Size: px
Start display at page:

Download "GEOG 4110/5100 Advanced Remote Sensing Lecture 15"

Transcription

1 GEOG 4110/5100 Advanced Remote Sensing Lecture 15 Principal Component Analysis Relevant reading: Richards. Chapters 6.3* *For more information on working with matrices, refer to Richards, Appendix A) GEOG 4110/5100 1

2 Multispectral Transformations of Image Data It is possible to transform brightness data through linear operations on the set of spectral bands Can make image features visible that are not discernable in the original data Can preserve image quality a reduced number of transformed dimensions E.g. for display on a color monitor Principal Component transformation Seeks to minimize correlation in order to minimize redundancy of spectral bands GEOG 4110/5100 2

3 Principal Components Analysis (PCA) PCA is a technique that transforms the original vector image data into smaller set of uncorrelated variables. The variables represent most of the image information and are easier to interpret. Principal components are derived such that the first PC accounts for much of the variation of the original data. The second (vertical) accounts for most of the remaining variation. PCA is useful in reducing the dimensionality (number of bands) that are used for analysis. Minimum noise fraction (MNF) method can be used with hyperspectral data for noise reduction. for a fairly simple explanation: GEOG 4110/5100 3

4 Principal Components Seek new coordinate system in vector space in which data can be represented without correlation Covariance matrix is diagonal y = Gx = D t x GEOG 4110/5100 4

5 Mean Vector and Covariance The mean vector (m) is the vector average of the individual components of a vector The covariance between two real-valued random describes how one variable varies in relation to another. Cov(X,Y) = 1 n 1 n i=1 (X i x )(Y i y ) C = 1 x n 1 n i=1 (X i x)(x i x) T GEOG 4110/5100 5

6 Mean Vector and Covariance The covariance matrix (S x ) is a matrix of covariance values that describes the scatter or spread between variables. Computation of Covariance Matrix (Table 8.1 from Richards and Jia, 2006) x = 1 n n 1 i= 1 ( x i m)( x i m) t m = No Correlation between x 1 and x 2 6

7 Relationship Between x and y Covariance Matrices y = Gx = D t x Each component of y is a linear combination of all of the elements of x; the weighting coefficients are the elements of the matrix G (or D T ) S y = ξ[(y-m y )(y-m y ) t ] m y = ξ[y] = ξ[d t x] = D t ξ [x] = D t m x S y = ξ[(d t x-d t m x )(D t x-d t m x ) t ] S y = D t ξ[(x-m x )(x-m x ) t ]D S y = D t S x D S x is the covariance of the pixel data in x space (S y in y) - m x and m y are the mean vectors in x and y respectively - ξ is the Expected value (here taken as the mean for m) Identifying a y coordinate space in which the pixel data exhibits no correlation requires S y to be a diagonal matrix GEOG 4110/5100 7

8 Eigenvalues and Eigenvectors When we have a transformation matrix operating on a vector, a new vector is produced: Sometimes that new vector is simply the product of a scalar and the original vector Eigenvalue Eigenvector When this is the case, the scalar is referred to as the Eigenvalue, and the vector is referred to as the Eigenvector (simple explanation) (more complex explanation) GEOG 4110/5100 8

9 Eigenvalues and Eigenvectors Eigenvalues (l) and eigenvectors (x) of a Matrix (M) are scalar and vector terms such that the multiplication of x by l has the same result as the matrix transformation of x by matrix M or Mx = lx (i.e. y = lx is equivalent to y = Mx) Mx - lx = 0 à (M-lI)x =0; where I is the identity matrix (x is a vector with n elements, where n = number of bands) For the above to be true, then either x = 0 or M-lI = 0 This is the characteristic equation from which the eigenvalues (l) can be determined When plugged into the equation: (M-lI)x =0, the eigenvectors (x) can be determined GEOG 4110/5100 9

10 Calculating Determinants From:

11 Principal Component Transformation S y = D t S x D - S x is the covariance of the pixel data in x space - D is a matrix of Eigenvectors derived from S x - The covariance matrix in y-space is given by: y $ λ ' & ) 0 λ 2 0 = & ) & ) & ) % 0 0 λ N ( Where N is the dimensionality, and l i represents the eigenvalues in descending order The n th component (n = 1 N) represents z percent of the variance where λ n ζ n = λ 1 + λ λ n - S y is by definition a diagonal covariance matrix with its elements representing the variance in the transformed coordinates - The greatest variance occurs in the first dimension of the transformed coordinate system, the next greatest in the 2 nd, and so-on such that the least variance is found in the n th dimension GEOG 4110/

12 Principal Component Transformation The eigenvectors determine the transformation matrix that produces each principal component The eigenvalue describes the percentage of the variance that is contained within each principal component The higher the eigenvalue as a fraction of the sum of the eigenvalues, the more relative information is contained in the corresponding principal component GEOG 4110/

13 Principal Components Transformation Example in 2 dimensions # = & % ( x $ ' x # & = % ( $ ' GEOG 4110/

14 Principal Components Transformation Example in 2 dimensions x # & = % ( $ ' First we need to find the eigenvalues S x li = λ λ = 0 l 2-3.0l = 0 à l = 2.67 and 0.33 GEOG 4110/

15 Principal Components Transformation Example in 2 dimensions " = $ y # % ' & First component contains 2.67/( ) = 89% of the variance in this example (usually we order the eigenvalues in descending order) Now we seek to find the principal components transformation matrix G = D T Where D T is the transposed matrix of eigenvectors. The first eigenvector (g 1 )corresponds to the first eigenvalue l 1 " [S x li]g 1 = 0 with g 1 = g % 11 t $ ' = d 1 for the two dimensional case # & g 21 Substituting S x and l 1 (2.67) gives the pair of equations: -0.77g g 21 = g g 21 = 0 yields g 11 =1.43g 21 GEOG 4110/

16 Principal Components Transformation Example in 2 dimensions We have the added constraint that the eigenvectors must be normalized (i.e. the G matrix must be orthogonal such that G t = G -1 ) (g 11 ) 2 + (g 21 ) 2 = 1 This produces the following eigenvectors " g 1 = 0.82 % # $ ' g # = 0.57 & % ( & $ 0.82 ' g 2 is the 2 nd eigenvector derived from the 2 nd eigenvalue (replace 2.67 on previous page with 0.33) Which in turn produce the following transformation matrix " G = D t = $ # t % ' & " = $ # % ' & Remember, D is the matrix of eigenvectors GEOG 4110/

17 Principal Components Transformation Example in 2 dimensions GEOG 4110/

18 c (a) Four Landat MSS bans for the region of Andamooka in Central Australia; (b) The four principal components of the image segment; (c) comparison of standard false color composite (R=band 7; G=band 5; B=band 4) with a principal component composite (R, G, B are 1st, 2nd, and 3rd components respectively) a b

19 Highly correlated bands 1, 2, and 3 b a c d e 19

20 a b Bands 4, 3, 2 c PC3, PC2, PC1 d PC4, PC3, PC2 e 20

21 a b c d e 21

22 Principal Component Transformation Steps 1. Compute the covariance matrix of the data set in vector space 2. Calculate the eigenvalues of the covariance matrix 3. The diagonal matrix with the eigenvalues along the diagonal will be the covariance matrix of the transformed axes (principal component axes) 4. Find the matrix of eigenvectors (D i ) for each individual l of interest by solving for [S x l i I]g i = 0. for that l. 5. Transpose the Matrix D to produce principal component transformation matrix (g). The number of rows in g will equal the number of spectral dimensions from which the eigenvalues and eigenvectors were calculated 6. For each g matrix (derived from a given l) the original data values (in original x coordinate system) are multiplied by the rows in g (g 1, g 2, g n where n is the number of dimensions in vector space), to produce coordinates in the transformed dimension (new y coordinate system). Each axis in the original spectral space will be multiplied by its corresponding row in the g matrix to produce the transformed coordinate system (principal component) 7. Steps 4 6 are repeated until the desired number of principal component transformations have been executed.

23 Principal Components Analysis (PCA) PCA is a technique that transforms the original vector image data into smaller set of uncorrelated variables. The variables represent most of the image information and easier to interpret. Principal components are derived such that the first PC accounts for much of the variation of the original data. The second (vertical) accounts for most of the remaining variation. PCA is useful in reducing the dimensionality (number of bands) that used for analysis. Minimum noise fraction (MNF) method can be used with hyperspectral data for noise reduction. GEOG 4110/

24 TM Example for PC Transformation Compute the n-dimensional covariance matrix (7 x 7 for Landsat TM). The variances of the principal components (eigenvalues) contain useful information (e.g. determine the % of total variance explained by each of the principal components) eigenvalue of the p th component sum of the eigenvalue of all components Band Number Variance Table shows the variance of different bands of TM scene. Adapted from Jensen, GEOG 4110/

25 TM Example for PC Transformation Component p (eigenvalues) eigenvalue Table shows the variance of different principal components. Adapted from Jensen, Sum of eigenvalues of all components = % of variance explained by PC1 = ( / )*100 = 84.68% % of variance explained by PC2 = (131.2/ )*100 = % Band Component p Table shows the eigenvectors (coefficients) for each principal component in each column. Adapted from Jensen,

26 calculated to determine which band is associated with each principal TM component. Example This for helps PC in Transformation understanding the information contains by each component. How principal component images are created? Identify the original brightness values of a given pixel (e.g. the first pixel at column 1 and row 1). Obtain the new pixel value by summation of the multiplication of the eigenvector of the component of each band by the original value GEOG 4110/

27 TM Example for PC Transformation Band DN PC1 eigenvector New value = 0.205(20) (30) (50) = 119 This will be carried out for each pixel to produce the PC1 image. If the first three components can explain most of the variation in the data, further analysis can be performed using the transformed images GEOG 4110/

28 TM Example for PC Transformation 7 PC images of TM data. PC1 (bands 4, 5, 7; near infrared and middle infrared). PC2 (bands 1, 2, 3; visible). PC3 (near infrared). PC4 (band 6, thermal). The four PC accounts for % of the variance. From Jensen (2005) GEOG 4110/

29 Back-up GEOG 4110/

GEOG 4110/5100 Advanced Remote Sensing Lecture 12. Classification (Supervised and Unsupervised) Richards: 6.1, ,

GEOG 4110/5100 Advanced Remote Sensing Lecture 12. Classification (Supervised and Unsupervised) Richards: 6.1, , GEOG 4110/5100 Advanced Remote Sensing Lecture 12 Classification (Supervised and Unsupervised) Richards: 6.1, 8.1-8.8.2, 9.1-9.34 GEOG 4110/5100 1 Fourier Transforms Transformations in the Frequency Domain

More information

The Principal Component Analysis

The Principal Component Analysis The Principal Component Analysis Philippe B. Laval KSU Fall 2017 Philippe B. Laval (KSU) PCA Fall 2017 1 / 27 Introduction Every 80 minutes, the two Landsat satellites go around the world, recording images

More information

Principal Components Analysis (PCA)

Principal Components Analysis (PCA) Principal Components Analysis (PCA) Principal Components Analysis (PCA) a technique for finding patterns in data of high dimension Outline:. Eigenvectors and eigenvalues. PCA: a) Getting the data b) Centering

More information

12.2 Dimensionality Reduction

12.2 Dimensionality Reduction 510 Chapter 12 of this dimensionality problem, regularization techniques such as SVD are almost always needed to perform the covariance matrix inversion. Because it appears to be a fundamental property

More information

GNR401 Principles of Satellite Image Processing

GNR401 Principles of Satellite Image Processing Principles of Satellite Image Processing Instructor: Prof. CSRE, IIT Bombay bkmohan@csre.iitb.ac.in Slot 5 Guest Lecture PCT and Band Arithmetic November 07, 2012 9.30 AM 10.55 AM IIT Bombay Slide 1 November

More information

Linear Algebra & Geometry why is linear algebra useful in computer vision?

Linear Algebra & Geometry why is linear algebra useful in computer vision? Linear Algebra & Geometry why is linear algebra useful in computer vision? References: -Any book on linear algebra! -[HZ] chapters 2, 4 Some of the slides in this lecture are courtesy to Prof. Octavia

More information

CHAPTER 4 PRINCIPAL COMPONENT ANALYSIS-BASED FUSION

CHAPTER 4 PRINCIPAL COMPONENT ANALYSIS-BASED FUSION 59 CHAPTER 4 PRINCIPAL COMPONENT ANALYSIS-BASED FUSION 4. INTRODUCTION Weighted average-based fusion algorithms are one of the widely used fusion methods for multi-sensor data integration. These methods

More information

Linear Algebra & Geometry why is linear algebra useful in computer vision?

Linear Algebra & Geometry why is linear algebra useful in computer vision? Linear Algebra & Geometry why is linear algebra useful in computer vision? References: -Any book on linear algebra! -[HZ] chapters 2, 4 Some of the slides in this lecture are courtesy to Prof. Octavia

More information

CS4495/6495 Introduction to Computer Vision. 8B-L2 Principle Component Analysis (and its use in Computer Vision)

CS4495/6495 Introduction to Computer Vision. 8B-L2 Principle Component Analysis (and its use in Computer Vision) CS4495/6495 Introduction to Computer Vision 8B-L2 Principle Component Analysis (and its use in Computer Vision) Wavelength 2 Wavelength 2 Principal Components Principal components are all about the directions

More information

Principle Components Analysis (PCA) Relationship Between a Linear Combination of Variables and Axes Rotation for PCA

Principle Components Analysis (PCA) Relationship Between a Linear Combination of Variables and Axes Rotation for PCA Principle Components Analysis (PCA) Relationship Between a Linear Combination of Variables and Axes Rotation for PCA Principle Components Analysis: Uses one group of variables (we will call this X) In

More information

IR-MAD Iteratively Re-weighted Multivariate Alteration Detection

IR-MAD Iteratively Re-weighted Multivariate Alteration Detection IR-MAD Iteratively Re-weighted Multivariate Alteration Detection Nielsen, A. A., Conradsen, K., & Simpson, J. J. (1998). Multivariate Alteration Detection (MAD) and MAF Postprocessing in Multispectral,

More information

Introduction to Machine Learning

Introduction to Machine Learning 10-701 Introduction to Machine Learning PCA Slides based on 18-661 Fall 2018 PCA Raw data can be Complex, High-dimensional To understand a phenomenon we measure various related quantities If we knew what

More information

Principal Component Analysis-I Geog 210C Introduction to Spatial Data Analysis. Chris Funk. Lecture 17

Principal Component Analysis-I Geog 210C Introduction to Spatial Data Analysis. Chris Funk. Lecture 17 Principal Component Analysis-I Geog 210C Introduction to Spatial Data Analysis Chris Funk Lecture 17 Outline Filters and Rotations Generating co-varying random fields Translating co-varying fields into

More information

VAR Model. (k-variate) VAR(p) model (in the Reduced Form): Y t-2. Y t-1 = A + B 1. Y t + B 2. Y t-p. + ε t. + + B p. where:

VAR Model. (k-variate) VAR(p) model (in the Reduced Form): Y t-2. Y t-1 = A + B 1. Y t + B 2. Y t-p. + ε t. + + B p. where: VAR Model (k-variate VAR(p model (in the Reduced Form: where: Y t = A + B 1 Y t-1 + B 2 Y t-2 + + B p Y t-p + ε t Y t = (y 1t, y 2t,, y kt : a (k x 1 vector of time series variables A: a (k x 1 vector

More information

Lecture: Face Recognition and Feature Reduction

Lecture: Face Recognition and Feature Reduction Lecture: Face Recognition and Feature Reduction Juan Carlos Niebles and Ranjay Krishna Stanford Vision and Learning Lab Lecture 11-1 Recap - Curse of dimensionality Assume 5000 points uniformly distributed

More information

Eigenvalues, Eigenvectors, and an Intro to PCA

Eigenvalues, Eigenvectors, and an Intro to PCA Eigenvalues, Eigenvectors, and an Intro to PCA Eigenvalues, Eigenvectors, and an Intro to PCA Changing Basis We ve talked so far about re-writing our data using a new set of variables, or a new basis.

More information

Eigenvalues, Eigenvectors, and an Intro to PCA

Eigenvalues, Eigenvectors, and an Intro to PCA Eigenvalues, Eigenvectors, and an Intro to PCA Eigenvalues, Eigenvectors, and an Intro to PCA Changing Basis We ve talked so far about re-writing our data using a new set of variables, or a new basis.

More information

Principal Component Analysis

Principal Component Analysis Principal Component Analysis Yingyu Liang yliang@cs.wisc.edu Computer Sciences Department University of Wisconsin, Madison [based on slides from Nina Balcan] slide 1 Goals for the lecture you should understand

More information

Covariance and Principal Components

Covariance and Principal Components COMP3204/COMP6223: Computer Vision Covariance and Principal Components Jonathon Hare jsh2@ecs.soton.ac.uk Variance and Covariance Random Variables and Expected Values Mathematicians talk variance (and

More information

Principal Components Theory Notes

Principal Components Theory Notes Principal Components Theory Notes Charles J. Geyer August 29, 2007 1 Introduction These are class notes for Stat 5601 (nonparametrics) taught at the University of Minnesota, Spring 2006. This not a theory

More information

December 20, MAA704, Multivariate analysis. Christopher Engström. Multivariate. analysis. Principal component analysis

December 20, MAA704, Multivariate analysis. Christopher Engström. Multivariate. analysis. Principal component analysis .. December 20, 2013 Todays lecture. (PCA) (PLS-R) (LDA) . (PCA) is a method often used to reduce the dimension of a large dataset to one of a more manageble size. The new dataset can then be used to make

More information

Lecture: Face Recognition and Feature Reduction

Lecture: Face Recognition and Feature Reduction Lecture: Face Recognition and Feature Reduction Juan Carlos Niebles and Ranjay Krishna Stanford Vision and Learning Lab 1 Recap - Curse of dimensionality Assume 5000 points uniformly distributed in the

More information

PCA, Kernel PCA, ICA

PCA, Kernel PCA, ICA PCA, Kernel PCA, ICA Learning Representations. Dimensionality Reduction. Maria-Florina Balcan 04/08/2015 Big & High-Dimensional Data High-Dimensions = Lot of Features Document classification Features per

More information

PRINCIPAL COMPONENT ANALYSIS

PRINCIPAL COMPONENT ANALYSIS PRINCIPAL COMPONENT ANALYSIS 1 INTRODUCTION One of the main problems inherent in statistics with more than two variables is the issue of visualising or interpreting data. Fortunately, quite often the problem

More information

Chapter 6 Spectral Domain Image Transforms

Chapter 6 Spectral Domain Image Transforms Chapter 6 Spectral Domain Image Transforms 6.1 Introduction Because of the multispectral or vector character of remote sensing image data we can devise transformations that generate new pixel vector descriptions

More information

7. Variable extraction and dimensionality reduction

7. Variable extraction and dimensionality reduction 7. Variable extraction and dimensionality reduction The goal of the variable selection in the preceding chapter was to find least useful variables so that it would be possible to reduce the dimensionality

More information

Mathematical foundations - linear algebra

Mathematical foundations - linear algebra Mathematical foundations - linear algebra Andrea Passerini passerini@disi.unitn.it Machine Learning Vector space Definition (over reals) A set X is called a vector space over IR if addition and scalar

More information

A Tutorial on Data Reduction. Principal Component Analysis Theoretical Discussion. By Shireen Elhabian and Aly Farag

A Tutorial on Data Reduction. Principal Component Analysis Theoretical Discussion. By Shireen Elhabian and Aly Farag A Tutorial on Data Reduction Principal Component Analysis Theoretical Discussion By Shireen Elhabian and Aly Farag University of Louisville, CVIP Lab November 2008 PCA PCA is A backbone of modern data

More information

PRINCIPAL COMPONENT ANALYSIS

PRINCIPAL COMPONENT ANALYSIS PRINCIPAL COMPONENT ANALYSIS Dimensionality Reduction Tzompanaki Katerina Dimensionality Reduction Unsupervised learning Goal: Find hidden patterns in the data. Used for Visualization Data compression

More information

Appendix A: Matrices

Appendix A: Matrices Appendix A: Matrices A matrix is a rectangular array of numbers Such arrays have rows and columns The numbers of rows and columns are referred to as the dimensions of a matrix A matrix with, say, 5 rows

More information

COMPLEX PRINCIPAL COMPONENT SPECTRA EXTRACTION

COMPLEX PRINCIPAL COMPONENT SPECTRA EXTRACTION COMPLEX PRINCIPAL COMPONEN SPECRA EXRACION PROGRAM complex_pca_spectra Computing principal components o begin, click the Formation attributes tab in the AASPI-UIL window and select program complex_pca_spectra:

More information

Covariance to PCA. CS 510 Lecture #8 February 17, 2014

Covariance to PCA. CS 510 Lecture #8 February 17, 2014 Covariance to PCA CS 510 Lecture 8 February 17, 2014 Status Update Programming Assignment 2 is due March 7 th Expect questions about your progress at the start of class I still owe you Assignment 1 back

More information

Data Preprocessing Tasks

Data Preprocessing Tasks Data Tasks 1 2 3 Data Reduction 4 We re here. 1 Dimensionality Reduction Dimensionality reduction is a commonly used approach for generating fewer features. Typically used because too many features can

More information

Coastal Characterization Using EO-1 Hyperion Data

Coastal Characterization Using EO-1 Hyperion Data Coastal Characterization Using EO-1 Hyperion Data Dr. Hsiao-hua K. Burke EO-1 SVT Meeting 18-21 November 2002 Sponsor: NOAA NESDIS GOES 2002-1 Channel Positions of Various Ocean- Color Sensors, 1978-2000*

More information

Covariance to PCA. CS 510 Lecture #14 February 23, 2018

Covariance to PCA. CS 510 Lecture #14 February 23, 2018 Covariance to PCA CS 510 Lecture 14 February 23, 2018 Overview: Goal Assume you have a gallery (database) of images, and a probe (test) image. The goal is to find the database image that is most similar

More information

Face Recognition and Biometric Systems

Face Recognition and Biometric Systems The Eigenfaces method Plan of the lecture Principal Components Analysis main idea Feature extraction by PCA face recognition Eigenfaces training feature extraction Literature M.A.Turk, A.P.Pentland Face

More information

Eigenvalues, Eigenvectors, and an Intro to PCA

Eigenvalues, Eigenvectors, and an Intro to PCA Eigenvalues, Eigenvectors, and an Intro to PCA Eigenvalues, Eigenvectors, and an Intro to PCA Changing Basis We ve talked so far about re-writing our data using a new set of variables, or a new basis.

More information

YORK UNIVERSITY. Faculty of Science Department of Mathematics and Statistics MATH M Test #1. July 11, 2013 Solutions

YORK UNIVERSITY. Faculty of Science Department of Mathematics and Statistics MATH M Test #1. July 11, 2013 Solutions YORK UNIVERSITY Faculty of Science Department of Mathematics and Statistics MATH 222 3. M Test # July, 23 Solutions. For each statement indicate whether it is always TRUE or sometimes FALSE. Note: For

More information

Learning with Singular Vectors

Learning with Singular Vectors Learning with Singular Vectors CIS 520 Lecture 30 October 2015 Barry Slaff Based on: CIS 520 Wiki Materials Slides by Jia Li (PSU) Works cited throughout Overview Linear regression: Given X, Y find w:

More information

Machine Learning (Spring 2012) Principal Component Analysis

Machine Learning (Spring 2012) Principal Component Analysis 1-71 Machine Learning (Spring 1) Principal Component Analysis Yang Xu This note is partly based on Chapter 1.1 in Chris Bishop s book on PRML and the lecture slides on PCA written by Carlos Guestrin in

More information

Dimensionality Reduction: PCA. Nicholas Ruozzi University of Texas at Dallas

Dimensionality Reduction: PCA. Nicholas Ruozzi University of Texas at Dallas Dimensionality Reduction: PCA Nicholas Ruozzi University of Texas at Dallas Eigenvalues λ is an eigenvalue of a matrix A R n n if the linear system Ax = λx has at least one non-zero solution If Ax = λx

More information

MODULE 5 LECTURE NOTES 5 PRINCIPAL COMPONENT ANALYSIS

MODULE 5 LECTURE NOTES 5 PRINCIPAL COMPONENT ANALYSIS MODULE 5 LECTURE NOTES 5 PRINCIPAL COMPONENT ANALYSIS. (PCA) Principal component analysis (PCA), also known as Karhunen-Loeve analysis, transforms the information inherent in multispectral remotely sensed

More information

MultiDimensional Signal Processing Master Degree in Ingegneria delle Telecomunicazioni A.A

MultiDimensional Signal Processing Master Degree in Ingegneria delle Telecomunicazioni A.A MultiDimensional Signal Processing Master Degree in Ingegneria delle Telecomunicazioni A.A. 2017-2018 Pietro Guccione, PhD DEI - DIPARTIMENTO DI INGEGNERIA ELETTRICA E DELL INFORMAZIONE POLITECNICO DI

More information

Chapter 6 Multispectral Transformations of Image Data

Chapter 6 Multispectral Transformations of Image Data Chapter 6 Multispectral Transformations of Image Data The multispectral or vector character of most remote sensing image data renders it amenable to spectral transformations that generate new sets of image

More information

Principal Component Analysis!! Lecture 11!

Principal Component Analysis!! Lecture 11! Principal Component Analysis Lecture 11 1 Eigenvectors and Eigenvalues g Consider this problem of spreading butter on a bread slice 2 Eigenvectors and Eigenvalues g Consider this problem of stretching

More information

Data Mining Lecture 4: Covariance, EVD, PCA & SVD

Data Mining Lecture 4: Covariance, EVD, PCA & SVD Data Mining Lecture 4: Covariance, EVD, PCA & SVD Jo Houghton ECS Southampton February 25, 2019 1 / 28 Variance and Covariance - Expectation A random variable takes on different values due to chance The

More information

1 Principal Components Analysis

1 Principal Components Analysis Lecture 3 and 4 Sept. 18 and Sept.20-2006 Data Visualization STAT 442 / 890, CM 462 Lecture: Ali Ghodsi 1 Principal Components Analysis Principal components analysis (PCA) is a very popular technique for

More information

EECS490: Digital Image Processing. Lecture #26

EECS490: Digital Image Processing. Lecture #26 Lecture #26 Moments; invariant moments Eigenvector, principal component analysis Boundary coding Image primitives Image representation: trees, graphs Object recognition and classes Minimum distance classifiers

More information

LECTURE 16: PCA AND SVD

LECTURE 16: PCA AND SVD Instructor: Sael Lee CS549 Computational Biology LECTURE 16: PCA AND SVD Resource: PCA Slide by Iyad Batal Chapter 12 of PRML Shlens, J. (2003). A tutorial on principal component analysis. CONTENT Principal

More information

Linear Algebra Review. Fei-Fei Li

Linear Algebra Review. Fei-Fei Li Linear Algebra Review Fei-Fei Li 1 / 37 Vectors Vectors and matrices are just collections of ordered numbers that represent something: movements in space, scaling factors, pixel brightnesses, etc. A vector

More information

Vectors and Matrices Statistics with Vectors and Matrices

Vectors and Matrices Statistics with Vectors and Matrices Vectors and Matrices Statistics with Vectors and Matrices Lecture 3 September 7, 005 Analysis Lecture #3-9/7/005 Slide 1 of 55 Today s Lecture Vectors and Matrices (Supplement A - augmented with SAS proc

More information

Computation. For QDA we need to calculate: Lets first consider the case that

Computation. For QDA we need to calculate: Lets first consider the case that Computation For QDA we need to calculate: δ (x) = 1 2 log( Σ ) 1 2 (x µ ) Σ 1 (x µ ) + log(π ) Lets first consider the case that Σ = I,. This is the case where each distribution is spherical, around the

More information

Linear Algebra Methods for Data Mining

Linear Algebra Methods for Data Mining Linear Algebra Methods for Data Mining Saara Hyvönen, Saara.Hyvonen@cs.helsinki.fi Spring 2007 The Singular Value Decomposition (SVD) continued Linear Algebra Methods for Data Mining, Spring 2007, University

More information

Principal Component Analysis and Linear Discriminant Analysis

Principal Component Analysis and Linear Discriminant Analysis Principal Component Analysis and Linear Discriminant Analysis Ying Wu Electrical Engineering and Computer Science Northwestern University Evanston, IL 60208 http://www.eecs.northwestern.edu/~yingwu 1/29

More information

COMP6237 Data Mining Covariance, EVD, PCA & SVD. Jonathon Hare

COMP6237 Data Mining Covariance, EVD, PCA & SVD. Jonathon Hare COMP6237 Data Mining Covariance, EVD, PCA & SVD Jonathon Hare jsh2@ecs.soton.ac.uk Variance and Covariance Random Variables and Expected Values Mathematicians talk variance (and covariance) in terms of

More information

Principal Component Analysis (PCA) Theory, Practice, and Examples

Principal Component Analysis (PCA) Theory, Practice, and Examples Principal Component Analysis (PCA) Theory, Practice, and Examples Data Reduction summarization of data with many (p) variables by a smaller set of (k) derived (synthetic, composite) variables. p k n A

More information

Lecture 13. Principal Component Analysis. Brett Bernstein. April 25, CDS at NYU. Brett Bernstein (CDS at NYU) Lecture 13 April 25, / 26

Lecture 13. Principal Component Analysis. Brett Bernstein. April 25, CDS at NYU. Brett Bernstein (CDS at NYU) Lecture 13 April 25, / 26 Principal Component Analysis Brett Bernstein CDS at NYU April 25, 2017 Brett Bernstein (CDS at NYU) Lecture 13 April 25, 2017 1 / 26 Initial Question Intro Question Question Let S R n n be symmetric. 1

More information

This appendix provides a very basic introduction to linear algebra concepts.

This appendix provides a very basic introduction to linear algebra concepts. APPENDIX Basic Linear Algebra Concepts This appendix provides a very basic introduction to linear algebra concepts. Some of these concepts are intentionally presented here in a somewhat simplified (not

More information

(a)

(a) Chapter 8 Subspace Methods 8. Introduction Principal Component Analysis (PCA) is applied to the analysis of time series data. In this context we discuss measures of complexity and subspace methods for

More information

Econ 204 Supplement to Section 3.6 Diagonalization and Quadratic Forms. 1 Diagonalization and Change of Basis

Econ 204 Supplement to Section 3.6 Diagonalization and Quadratic Forms. 1 Diagonalization and Change of Basis Econ 204 Supplement to Section 3.6 Diagonalization and Quadratic Forms De La Fuente notes that, if an n n matrix has n distinct eigenvalues, it can be diagonalized. In this supplement, we will provide

More information

Independent Component Analysis and Its Application on Accelerator Physics

Independent Component Analysis and Its Application on Accelerator Physics Independent Component Analysis and Its Application on Accelerator Physics Xiaoying Pang LA-UR-12-20069 ICA and PCA Similarities: Blind source separation method (BSS) no model Observed signals are linear

More information

Linear Algebra Methods for Data Mining

Linear Algebra Methods for Data Mining Linear Algebra Methods for Data Mining Saara Hyvönen, Saara.Hyvonen@cs.helsinki.fi Spring 2007 Linear Discriminant Analysis Linear Algebra Methods for Data Mining, Spring 2007, University of Helsinki Principal

More information

Principal Component Analysis

Principal Component Analysis Principal Component Analysis CS5240 Theoretical Foundations in Multimedia Leow Wee Kheng Department of Computer Science School of Computing National University of Singapore Leow Wee Kheng (NUS) Principal

More information

Announcements (repeat) Principal Components Analysis

Announcements (repeat) Principal Components Analysis 4/7/7 Announcements repeat Principal Components Analysis CS 5 Lecture #9 April 4 th, 7 PA4 is due Monday, April 7 th Test # will be Wednesday, April 9 th Test #3 is Monday, May 8 th at 8AM Just hour long

More information

Vectors To begin, let us describe an element of the state space as a point with numerical coordinates, that is x 1. x 2. x =

Vectors To begin, let us describe an element of the state space as a point with numerical coordinates, that is x 1. x 2. x = Linear Algebra Review Vectors To begin, let us describe an element of the state space as a point with numerical coordinates, that is x 1 x x = 2. x n Vectors of up to three dimensions are easy to diagram.

More information

Principal Components Analysis. Sargur Srihari University at Buffalo

Principal Components Analysis. Sargur Srihari University at Buffalo Principal Components Analysis Sargur Srihari University at Buffalo 1 Topics Projection Pursuit Methods Principal Components Examples of using PCA Graphical use of PCA Multidimensional Scaling Srihari 2

More information

Linear Algebra: Matrix Eigenvalue Problems

Linear Algebra: Matrix Eigenvalue Problems CHAPTER8 Linear Algebra: Matrix Eigenvalue Problems Chapter 8 p1 A matrix eigenvalue problem considers the vector equation (1) Ax = λx. 8.0 Linear Algebra: Matrix Eigenvalue Problems Here A is a given

More information

Wavelet Transform And Principal Component Analysis Based Feature Extraction

Wavelet Transform And Principal Component Analysis Based Feature Extraction Wavelet Transform And Principal Component Analysis Based Feature Extraction Keyun Tong June 3, 2010 As the amount of information grows rapidly and widely, feature extraction become an indispensable technique

More information

Principal Component Analysis (PCA) Our starting point consists of T observations from N variables, which will be arranged in an T N matrix R,

Principal Component Analysis (PCA) Our starting point consists of T observations from N variables, which will be arranged in an T N matrix R, Principal Component Analysis (PCA) PCA is a widely used statistical tool for dimension reduction. The objective of PCA is to find common factors, the so called principal components, in form of linear combinations

More information

Machine Learning 2nd Edition

Machine Learning 2nd Edition INTRODUCTION TO Lecture Slides for Machine Learning 2nd Edition ETHEM ALPAYDIN, modified by Leonardo Bobadilla and some parts from http://www.cs.tau.ac.il/~apartzin/machinelearning/ The MIT Press, 2010

More information

15 Singular Value Decomposition

15 Singular Value Decomposition 15 Singular Value Decomposition For any high-dimensional data analysis, one s first thought should often be: can I use an SVD? The singular value decomposition is an invaluable analysis tool for dealing

More information

spring, math 204 (mitchell) list of theorems 1 Linear Systems Linear Transformations Matrix Algebra

spring, math 204 (mitchell) list of theorems 1 Linear Systems Linear Transformations Matrix Algebra spring, 2016. math 204 (mitchell) list of theorems 1 Linear Systems THEOREM 1.0.1 (Theorem 1.1). Uniqueness of Reduced Row-Echelon Form THEOREM 1.0.2 (Theorem 1.2). Existence and Uniqueness Theorem THEOREM

More information

Multivariate Statistical Analysis

Multivariate Statistical Analysis Multivariate Statistical Analysis Fall 2011 C. L. Williams, Ph.D. Lecture 4 for Applied Multivariate Analysis Outline 1 Eigen values and eigen vectors Characteristic equation Some properties of eigendecompositions

More information

Principal Component Analysis (PCA)

Principal Component Analysis (PCA) Principal Component Analysis (PCA) Salvador Dalí, Galatea of the Spheres CSC411/2515: Machine Learning and Data Mining, Winter 2018 Michael Guerzhoy and Lisa Zhang Some slides from Derek Hoiem and Alysha

More information

Principal Component Analysis (PCA) CSC411/2515 Tutorial

Principal Component Analysis (PCA) CSC411/2515 Tutorial Principal Component Analysis (PCA) CSC411/2515 Tutorial Harris Chan Based on previous tutorial slides by Wenjie Luo, Ladislav Rampasek University of Toronto hchan@cs.toronto.edu October 19th, 2017 (UofT)

More information

Principal Component Analysis

Principal Component Analysis Principal Component Analysis Anders Øland David Christiansen 1 Introduction Principal Component Analysis, or PCA, is a commonly used multi-purpose technique in data analysis. It can be used for feature

More information

Econ Slides from Lecture 8

Econ Slides from Lecture 8 Econ 205 Sobel Econ 205 - Slides from Lecture 8 Joel Sobel September 1, 2010 Computational Facts 1. det AB = det BA = det A det B 2. If D is a diagonal matrix, then det D is equal to the product of its

More information

Introduction to Machine Learning. PCA and Spectral Clustering. Introduction to Machine Learning, Slides: Eran Halperin

Introduction to Machine Learning. PCA and Spectral Clustering. Introduction to Machine Learning, Slides: Eran Halperin 1 Introduction to Machine Learning PCA and Spectral Clustering Introduction to Machine Learning, 2013-14 Slides: Eran Halperin Singular Value Decomposition (SVD) The singular value decomposition (SVD)

More information

L3: Review of linear algebra and MATLAB

L3: Review of linear algebra and MATLAB L3: Review of linear algebra and MATLAB Vector and matrix notation Vectors Matrices Vector spaces Linear transformations Eigenvalues and eigenvectors MATLAB primer CSCE 666 Pattern Analysis Ricardo Gutierrez-Osuna

More information

Karhunen-Loève Transform KLT. JanKees van der Poel D.Sc. Student, Mechanical Engineering

Karhunen-Loève Transform KLT. JanKees van der Poel D.Sc. Student, Mechanical Engineering Karhunen-Loève Transform KLT JanKees van der Poel D.Sc. Student, Mechanical Engineering Karhunen-Loève Transform Has many names cited in literature: Karhunen-Loève Transform (KLT); Karhunen-Loève Decomposition

More information

What is a vegetation index?

What is a vegetation index? Vegetation indexes What is a vegetation index? A tool: a vegetation index is a mathematical formula used to estimate the likelihood of vegetation presence in remotely sensed data A product: the result

More information

Principal Component Analysis. Applied Multivariate Statistics Spring 2012

Principal Component Analysis. Applied Multivariate Statistics Spring 2012 Principal Component Analysis Applied Multivariate Statistics Spring 2012 Overview Intuition Four definitions Practical examples Mathematical example Case study 2 PCA: Goals Goal 1: Dimension reduction

More information

5 Linear Algebra and Inverse Problem

5 Linear Algebra and Inverse Problem 5 Linear Algebra and Inverse Problem 5.1 Introduction Direct problem ( Forward problem) is to find field quantities satisfying Governing equations, Boundary conditions, Initial conditions. The direct problem

More information

EECS 275 Matrix Computation

EECS 275 Matrix Computation EECS 275 Matrix Computation Ming-Hsuan Yang Electrical Engineering and Computer Science University of California at Merced Merced, CA 95344 http://faculty.ucmerced.edu/mhyang Lecture 6 1 / 22 Overview

More information

Principal component analysis

Principal component analysis Principal component analysis Angela Montanari 1 Introduction Principal component analysis (PCA) is one of the most popular multivariate statistical methods. It was first introduced by Pearson (1901) and

More information

MATH 423 Linear Algebra II Lecture 33: Diagonalization of normal operators.

MATH 423 Linear Algebra II Lecture 33: Diagonalization of normal operators. MATH 423 Linear Algebra II Lecture 33: Diagonalization of normal operators. Adjoint operator and adjoint matrix Given a linear operator L on an inner product space V, the adjoint of L is a transformation

More information

Digital Image Processing Chapter 11 Representation and Description

Digital Image Processing Chapter 11 Representation and Description Digital Image Processing Chapter 11 Representation and Description Last Updated: July 20, 2003 Preview 11.1 Representation 11.1.1 Chain Codes Chain codes are used to represent a boundary by a connected

More information

Feature Extraction Techniques

Feature Extraction Techniques Feature Extraction Techniques Unsupervised Learning II Feature Extraction Unsupervised ethods can also be used to find features which can be useful for categorization. There are unsupervised ethods that

More information

Singular Value Decomposition and Digital Image Compression

Singular Value Decomposition and Digital Image Compression Singular Value Decomposition and Digital Image Compression Chris Bingham December 1, 016 Page 1 of Abstract The purpose of this document is to be a very basic introduction to the singular value decomposition

More information

1 Linearity and Linear Systems

1 Linearity and Linear Systems Mathematical Tools for Neuroscience (NEU 34) Princeton University, Spring 26 Jonathan Pillow Lecture 7-8 notes: Linear systems & SVD Linearity and Linear Systems Linear system is a kind of mapping f( x)

More information

Linear Dimensionality Reduction

Linear Dimensionality Reduction Outline Hong Chang Institute of Computing Technology, Chinese Academy of Sciences Machine Learning Methods (Fall 2012) Outline Outline I 1 Introduction 2 Principal Component Analysis 3 Factor Analysis

More information

Dimensionality Reduction with Principal Component Analysis

Dimensionality Reduction with Principal Component Analysis 10 Dimensionality Reduction with Principal Component Analysis Working directly with high-dimensional data, such as images, comes with some difficulties: it is hard to analyze, interpretation is difficult,

More information

Linear Algebra Review. Fei-Fei Li

Linear Algebra Review. Fei-Fei Li Linear Algebra Review Fei-Fei Li 1 / 51 Vectors Vectors and matrices are just collections of ordered numbers that represent something: movements in space, scaling factors, pixel brightnesses, etc. A vector

More information

Notes on Implementation of Component Analysis Techniques

Notes on Implementation of Component Analysis Techniques Notes on Implementation of Component Analysis Techniques Dr. Stefanos Zafeiriou January 205 Computing Principal Component Analysis Assume that we have a matrix of centered data observations X = [x µ,...,

More information

Econ Slides from Lecture 7

Econ Slides from Lecture 7 Econ 205 Sobel Econ 205 - Slides from Lecture 7 Joel Sobel August 31, 2010 Linear Algebra: Main Theory A linear combination of a collection of vectors {x 1,..., x k } is a vector of the form k λ ix i for

More information

Digital Change Detection Using Remotely Sensed Data for Monitoring Green Space Destruction in Tabriz

Digital Change Detection Using Remotely Sensed Data for Monitoring Green Space Destruction in Tabriz Int. J. Environ. Res. 1 (1): 35-41, Winter 2007 ISSN:1735-6865 Graduate Faculty of Environment University of Tehran Digital Change Detection Using Remotely Sensed Data for Monitoring Green Space Destruction

More information

2. Matrix Algebra and Random Vectors

2. Matrix Algebra and Random Vectors 2. Matrix Algebra and Random Vectors 2.1 Introduction Multivariate data can be conveniently display as array of numbers. In general, a rectangular array of numbers with, for instance, n rows and p columns

More information

CSC 411 Lecture 12: Principal Component Analysis

CSC 411 Lecture 12: Principal Component Analysis CSC 411 Lecture 12: Principal Component Analysis Roger Grosse, Amir-massoud Farahmand, and Juan Carrasquilla University of Toronto UofT CSC 411: 12-PCA 1 / 23 Overview Today we ll cover the first unsupervised

More information

M.A.P. Matrix Algebra Procedures. by Mary Donovan, Adrienne Copeland, & Patrick Curry

M.A.P. Matrix Algebra Procedures. by Mary Donovan, Adrienne Copeland, & Patrick Curry M.A.P. Matrix Algebra Procedures by Mary Donovan, Adrienne Copeland, & Patrick Curry This document provides an easy to follow background and review of basic matrix definitions and algebra. Because population

More information