CHAPTER 4 PRINCIPAL COMPONENT ANALYSIS-BASED FUSION

Size: px
Start display at page:

Download "CHAPTER 4 PRINCIPAL COMPONENT ANALYSIS-BASED FUSION"

Transcription

1 59 CHAPTER 4 PRINCIPAL COMPONENT ANALYSIS-BASED FUSION 4. INTRODUCTION Weighted average-based fusion algorithms are one of the widely used fusion methods for multi-sensor data integration. These methods involve the selection of appropriate weights to combine the images, so as to reduce the effects of distortion and also give a satisfactory visual image quality. The method of selecting weights based on the energy of the coefficients of the decomposed wavelet coefficients was described in Chapter 3. This chapter outlines a fusion scheme that uses weights based on the statistical nature of the data to be combined. The statistical measure used is the Principal Component Analysis (PCA). Two approaches of using PCA in image fusion have been listed by Genderen and Pohl (998):. PCA of multi-channel images where the first principal component is replaced by different images. This method, known as the Principal Component Substitution (PCS), has been used for remote sensing data fusion by Chavez et al. (99). 2. PCA of all multi-image data channels as outlined by Yesou et al. (993) and whose results were reported by Richards (984).

2 60 In addition to these, a third approach can be included in which the PCA transform has been used to calculate the weights of a linear combination of the input images as shown in the algorithms developed by Das et al. (2000), Haq et al. (2005) and Zheng et al. (2007). Haq et al. and Zheng et al. describe the PCA fusion rule along with multi-resolution image decomposition using the DWT. This research work focuses on the third method detailed above and offers alternative algorithms to determine the weights of fusion. The PCA-based weighted fusion involves separately fusing the high frequency (HF) and the low frequency (LF) parts of an image. The two frequency components are obtained by a filtering mechanism and finally the fused components are added together to get the resultant fused output. This chapter begins with an introduction to the PCA transform from a statistical standpoint; followed by the principal component analysis of images in section 4.3. The newly designed fusion algorithms that make use of the PCA for calculating the weights are described in the next section. This includes a discussion of the PCA Gaussian weighted average fusion method (5) and the efficient PCA Max Fusion scheme (8) in sections and respectively. The performance of the fusion schemes detailed in the previous section is analysed in section 4.5. This chapter concludes with a summary of the proposed techniques. 4.2 PRINCIPAL COMPONENT ANALYSIS (PCA) TRANSFORM 4.2. Introduction Principal Component Analysis is a quantitatively rigorous method for achieving simplification. Often, its operation can be thought of as revealing the internal structure of the data in a way which best explains the variance in the data. The method generates a new set of variables called

3 6 Principal Components (PC). Each principal component is a linear combination of the original variables and all the PCs are orthogonal to each other, and as a whole form an orthogonal basis for the space of the data; thereby removing redundant information, if any. The first principal component is a single axis in space. When each of the observations in the data set is projected on this axis, the resulting values form a new variable and the variance of this variable is the maximum among all possible choices of the first axis. The second principal component is another axis in space, perpendicular to the first. Projecting the observations on this axis generates another new variable, such that the variance of this variable is the maximum among all possible choices of this second axis. The full set of principal components is as large as the original set of variables. In general, the sum of the variances of the first few principal components exceeds 80% of the total variance of the original data. The original data can be recovered from the first few PCs themselves hence the Principal Component Analysis is a method that enables a decrease in the number of channels (or bands) by reducing the inter-channel dependencies. The multidimensional space is mapped into a space of fewer dimensions by transforming the original space using a linear transformation via a principal component analysis. The steps involved in the PCA transform, also called the Karhunen - Loeve (KL) transform, are:. Calculate the covariance matrix or the correlation matrix of the data sets to be transformed. The covariance matrix is used in the case of the unstandardised PCA, while the standardised PCA uses the correlation matrix. 2. Calculate the eigenvalues and the eigenvectors from the correlation / covariance matrix.

4 62 3. Principal Components of the given data set are the eigenvectors of the covariance matrix of the input. If the correlation matrix of the data is constructed and the eigenvectors found and listed in eigenvalue order, then just the first few eigenvectors can be used to reconstruct a large fraction of the variance of the original data. The first few eigenvectors can often be interpreted in terms of the large-scale physical behaviour of the system. The original space has been reduced to the space spanned by a few eigenvectors, with data loss, but retaining the most important variance. A lossless dimensionality reduction is possible if the data in question falls exactly on a smooth, locally flat subspace; however the noisy data prevents such an exact mapping, introducing some loss of information Mathematical Analysis of the Principal Component Analysis Transform x 3 Consider n-dimensional data in the form of M-vectors, x 2, x...x M defined by X as, X = [x,x 2,x 3...x M] (4.) The first principal component is that the n dimensional vector along whose direction, the variance is maximized. The principal component can be easily computed as the eigenvector of the correlation matrix having the largest eigenvalue. The next step is the computation of mean vector of the population, which is defined by the equation, M æ ö m x = ç å x k (4.2) è M ø k=

5 63 where, m x is the mean vector associated with the population of x vectors. For M vector samples from a random population, the covariance matrix can be approximated from the samples by, æ ö C x = ç (4.3) è M ø k= M T T å[ x k x - m xm ] k x where, T indicates vector transposition. C x is covariance matrix of order n x n. Element c ii of Cx is the variance of x i ; the i th component of the x vectors in the population, and element c ij of C x is the covariance between elements x i and x j of these vectors. The matrix C x is real and symmetric. If elements x i and x j are uncorrelated, their covariance is zero and, therefore, c ij =c ji =0. Since C x is real and symmetric, finding a set of n orthogonal eigenvectors always is possible. Let e i and λ i, i =, 2 n, be the eigenvectors and corresponding eigenvalues of C x, arranged in descending order so that l j ³ l j + for j =,2,... n- (4.4) Let A be a matrix whose rows are formed from the eigenvectors of C x ordered so that the first row of A is the eigenvector corresponding to the largest eigenvalue and the last row is the eigenvector corresponding to the

6 64 smallest eigenvalue. The matrix A is called the transformation matrix that maps the vectors x into vectors denoted by y as follows: Y = A x- ( ) m x (4.5) This equation is called the Principal components transform (also called the Hotelling transform). The mean of the y vectors resulting from this transformation is zero; m y = 0 (4.6) the covariance matrix of the y can be obtained in terms of A and C x by C T y = A C X A (4.7) Furthermore, C y is the diagonal matrix whose elements along the main diagonal are Eigen values of C x ; that is, él 0 0 ù C ê ú y = ê 0 l2 0 ú (4.8) êë 0 0 l ú 3 û The off-diagonal elements of the covariance matrix are 0, and hence the elements of the y vectors are uncorrelated. The rows of matrix A are the normalized eigenvectors of C x. Because C x is real and symmetric, these vectors form an orthonormal set, and it follows that the elements along the main diagonal of C y are the eigenvalues of C x. The main diagonal element in the ith row of C y is the variance of vector element y i.

7 65 Because the rows of A are orthonormal, its inverse equals its transpose. Thus, one can recover the vector x by performing the inverse transformation, T x A y+ = m (4.9) x To conclude this section, the principal component analysis is a mathematical way of determining the linear transformation of a sample of points in an N-dimensional space which exhibits the properties of the sample most clearly along the coordinate axes and in this process reduces the interchannel dependencies. The eigenvalues for each principal component correspond to the amount of total variance in the data described by this component. 4.3 PRINCIPAL COMPONENTS ANALYSIS FOR IMAGES 4.3. Basic Steps Involved (Gonzales and Woods 2002) PCA finds applications in image processing, where it has been used for identifying the patterns in data, and expressing the data in such a way as to highlight their similarities and differences. The steps involved in finding the PCA of the given images are the same as followed in the previous section for statistical data sets. However, as the first stage, the given 2 dimensional image is represented as a single vector. In general, an M x N image is represented as an MN x column vector. Then, the covariances of the given images are calculated, followed by the eigenvector and eigenvalue for covariance matrix and finally transforming the images using the eigenvector. This transformed image consists of the principal components of the given images. The principal components of the images are the original data that are represented solely in terms of the eigenvectors.

8 66 This process can be illustrated using an example. Consider two images of the same scene (a road) captured using two different image sensing systems. The first image shown in Figure 4. is taken using an infrared camera, that responds to changes in heat intensity of the source being imaged and the second image is obtained using the conventional CCD camera, shown in Figure 4.2. Figure 4. IR Image Figure 4.2 CCD Image The pixels, which are of two dimensional matrix form M x N, from both the images are vector populated, similar to equation (4.), and shown in Figure 4.5:

9 67 x x2 xm x x2 xm 2x 2x2 2xm 2x 2x2 2xm nx nx2 nxm nx nx2 Nxm Figure 4.3 Matrix representation of Image X Figure 4.4 Matrix representation of Image2 Y st image pixels - X 2 st image pixels - Y x 2x x 2x xm 2x xm 2x Nxm Nxm Figure 4.5 Vector populated matrix The next step is to find the covariance of the vector populated matrix. The covariance matrix gives the relation between the given images. If one is given n images, the covariance matrix will be of n x n dimensions.

10 68 Similar to equation (4.3), for the 2 dimensional case, the covariance can be represented by, n å( Yi - my )( Xi - mx ) i= ( n) C(X, Y) = (4.0) where, m X is the mean of the samples of one image represented by vector X m Y is the mean of the samples of one image represented by vector Y In case of more than 2 source images, more than one covariance measurement is required. For example, from a 3 image (X, Y, Z) vector populated matrix of dimension (No of pixels in a single image x 3), a covariance matrix of dimension 3x3 is obtained, æcov(x, X) cov(x, Y) cov(x, Z) ö ç C(X, Y, Z) = çcov(y, X) cov(y, Y) cov(y, Z) (4.) ç è cov(z,x) cov(z,y) cov(z,z) ø The value of the matrix elements (,), (2,2) and (3,3) gives the variance of images x, y and z, respectively. The element (,2) is the relation between image x and y and similarly for elements (,3) and (2,3). The values of (,2) and (2,) are same. Each entry in the matrix is the result of calculating the covariance between two separate dimensions. The following conclusions can be drawn from the covariance matrix: if the value of covariance is positive, then it indicates that the gray values in both the images increase together; if the value is negative, then it implies a negative correlation, that is, as the gray level in one image increases,

11 69 the value in the other image decreases and finally, the last case of zero covariance indicates that the two images are independent of each other Determination of Eigen Values of the Covariance Matrix Since the covariance matrix is a square matrix, the eigenvectors and eigenvalues for this matrix can be computed, which give useful information about the images. The eigenvector and eigenvalue is in the form as shown below for the images under consideration, 3 æ ö Eigenvalues = ç 3 è ø (4.2) æ ö Eigenvectors = ç è ø (4.3) given in Figure 4.6. The graphical representation of the eigenvectors of the images is Grey Levels of CCD Image Grey Levels of IR Image Figure 4.6 Plot of the images with the eigenvectors of the covariance matrix overlaid on top

12 70 When plotted, both the eigenvectors appear as diagonal dotted lines on the plot and are perpendicular to each other. More importantly, they provide with information about the patterns in the data. It can be seen from the plot that one of the eigenvectors goes through the middle of the points, similar to plotting a line of best fit, and shows how these two data sets are related along that line. The second eigenvector gives the other, less important, pattern in the data; signified by the amount by which all the points following the main line are offset from it. This process of taking the eigenvectors of the covariance matrix enables extraction of the lines that characterise the data. In the above example, the eigenvector with the largest eigenvalue is the one that points down the middle of the data, giving the most significant relationship between the images. The next step in finding the principal components of the given images is to transform the image using the eigenvectors using the equation (4.5), which is expressed below in words; Principal Component = Eigenvector ( Mean subtracted vector populatedmatrix) The eigenvector used above is either one or both of the eigenvectors (if two images are taken). In order that the matrix dimensions agree, either the eigenvector or vector populated matrix has to be transposed. The principal component using the eigenvector with the largest eigenvalue will have more information than that obtained using the eigenvector with the smallest eigenvalue. This can be seen from Figures 4.8 and 4.9 which show the images constructed using the largest eigenvector and the smallest eigenvector, respectively.

13 7 Figure 4.7 Principal components using eigenvector corresponding to the largest eigenvalue Figure 4.8 Principal components using eigenvector corresponding to the smallest eigenvalue From the discussion in this section, it can be summarized that the PCA technique has three effects: it orthogonalizes the components of the input vectors (so that they are uncorrelated with each other); it orders the resulting orthogonal components (principal components) so that those with the largest variation come first; and it eliminates those components that contribute the least to the variation in the data set.

14 72 Gray Levels of CCD Image Gray Levels of IR Image Figure 4.9 Plot of the principal components obtained using both the eigenvectors The PCA process has transformed the data so that it is expressed in terms of the patterns between the images, these patterns being the eigenvectors that most closely describe the relationships between the data. This is helpful because it enables the classification of all the data points (pixels) as a combination of the contributions from each of the eigenvectors. Initially, as in Figure 4.6, the plot has a simple grey level axes, each axis representing grey levels of one image, which does not convey any information on the relationship of the data points with each other. However, after processing the images using the PCA transform, the values of the data points specify exactly where they are present with respect to the trend lines, as seen in Figure 4.7. In the case of the transformation using both the eigenvectors, the data has been altered so that it is in terms of the eigenvectors instead of the usual axes.

15 PCA-BASED WEIGHTED FUSION 4.4. Introduction In this section, two new methods of computing weights for additive fusion of two images are presented. The fusion schemes involve separating the low frequency (LF) and the high frequency (HF) components of the images and detail rules for fusing the frequency components. Consider the multi-sensor input images obtained using an infrared sensor and a visible spectrum CCD sensor as I ir and I vis. The algorithm assumes that the source images are registered with each other and processed to be of the same size. The frequency component separation is achieved by the use of a Gaussian low pass filter used for smoothing the images. A typical smoothing convolution filter is essentially a matrix having an integer value for each row and column, the value chosen depending on the type of filter being used. For the Gaussian low pass filter, the 2-dimensional kernel is given by, h = (4.4) When an image is convolved with this type of filter, the gray value of each pixel is replaced by the average intensity of its eight nearest neighbors and itself. If the gray value of any pixel overlaid by the convolution kernel is dramatically different than that of its neighbors, the averaging effect of the filter will tend to reduce the effect of the noise by distributing it among all of the neighboring pixels. The smoothed images are S = I h (4.5) ir ir * S = I vis h (4.6) vis *

16 74 where * is the convolution operator The images S ir and S vis representing the visible portion of the source images. are the low frequency components The high frequency components are obtained by finding the deviations from the smoothed images. This is given by D - vis = I vis Svis (4.7) D i - r = Iir Sir (4.8) The result of the filtering is shown in Figure 4.0 for the image of a house on a hill captured using a CCD camera and that of the infrared image is given in Figure 4.. Input CCD Image Low Frequency Component Image Result of Low Pass Filtering High Frequency Component Image Figure 4.0 Effect of low pass filtering to separate frequency components of CCD image

17 75 Input IR Image Low Frequency Component Image Result of Low Pass Filtering High Frequency Component Image Figure 4. Effect of low pass filtering to separate frequency components of IR image Then these two frequency components are then fused separately using different fusion rules. The fused components are then added together to get the fused output image. The fusion rules proposed in this thesis are discussed in the following sub-sections PCA Gaussian Fusion Algorithm This fusion rule involves combining the low frequency (LF) components, S ir and S vis, using simple averaging. The high frequency (HF) components, D ir and D vis are combined by weighted addition High Frequency Component Fusion Rule The weights for fusion are calculated from the principal components of the high frequency part of the source images. It involves the computation of principal component of the deviation components as discussed in section Let these components be denoted as PC and PC 2, corresponding to the largest eigenvalue and the smallest eigenvalue, respectively.

18 76 These principal components are used to define the ratios P, P 2 ; which are in turn used to calculate the weights for the fusion. The ratios P and P 2 are obtained as follows: P PC PC + PC = (4.9) 2 P 2 PC PC + PC 2 = (4.20) 2 where PC is principal components corresponding to largest eigen value, and PC 2 is principal components corresponding to smallest eigen value. The weights are obtained by smoothing the ratio of principal components using a similar low pass filter, as in section 4.4. with a kernel defined by equations (4.5) and (4.6). The weights are given by, w * = P h (4.2) w 2 2* = P h (4.22) where * is the convolution operator and h is the Gaussian kernel defined in equation (4.4). Image fusion of the high frequency components is achieved by the weighted, normalized sum of the deviations defined by the fusion rule; D fuse {( D w ) + ( D w )} ir vis 2 = (4.23) [w + w ] 2 The fused output is shown in Figure 4.2

19 77 Figure 4.2 Fusion of HF components Low Frequency Component Fusion Rule The low frequency components are combined by averaging the intensity levels of the two images. This LF image contributes to the background information in the picture and is used to identify the presence of objects in the image. Mathematically, S fuse ( S + S ) ir vis = (4.24) 2

20 78 Figure 4.3 Fusion of LF components The final fused output is obtained by adding the weighted deviations, equations (4.7) and (4.8), to the low frequency background image and is shown below. I + fuse = Sfuse Dfuse (4.25) Figure 4.4 Fused image output using additive weighted fusion

21 Fusion Scheme Using the PCA Max Rule This newly designed scheme overcomes the performance limitation of the weight-based fusion algorithm proposed in section It uses the same technique of separating the image into the low and high frequency components. The low frequency components are combined using the principle of choosing maximum intensity pixels from each LF image. The HF components are fused using the weighted additive fusion rule as used in section Low Frequency Component Fusion Rule The low frequency images S ir and S vis are fused using the Select Max principle as discussed by Zheng et al. (2007). Since the visible information is contained in the low frequency components, fusing the images by selecting the pixel values with the highest intensity gives an output image that has a very high quality as perceived by the human observer. This rule involves choosing the maximum intensity levels of corresponding pixels from each LF image to represent the pixel value in the fused image. However, to enable comparison of the images on a pixel by pixel basis a histogram matching of the two images is performed. This is achieved by matching the histogram of the visible image to that of the IIR image (Gonzales and Woods 2002, Jain 989). The decision map of the low frequency image fusion rule is: ìsir (m, n) if Sir (m, n) > S vis (m, n) Sfuse (m, n) = í (4.26) îs vis (m, n) if Sir (m, n) < S vis (m, n) where, m =, 2, 3,. S m

22 80 n =, 2, 3, S n and S m x S n are the dimensions of the LF component image The fused output is shown in Figure 4.5 Figure 4.5 LF component fusion using choose max rule High Frequency Component Fusion Rule The weights for fusion are calculated from the principal components of the high frequency part of the source images. It involves the computation of principal component of the deviation components as discussed in section Let these components be denoted as PC and PC 2, corresponding to the largest eigenvalue and the smallest eigenvalue, respectively. These principal components are used to define the ratios P, P 2, which are in turn used to calculate the weights for the fusion. The ratios P and P 2 are obtained as follows:

23 8 P PC PC + PC = (4.27) 2 P 2 PC PC + PC 2 = (4.28) 2 where PC is principal components corresponding to largest eigen value, and PC 2 is principal components corresponding to smallest eigen value. The weights are obtained by smoothing the ratio of principal components using a similar low pass filter as in section 4.4. with a kernel defined by equations (4.5) and (4.6). The weights are given by, w * = P h (4.29) w 2 2* = P h (4.30) where * is the convolution operator and h is the Gaussian kernel defined in equation (4.4). Image fusion of the high frequency components is achieved by the weighted, normalized sum of the deviations defined by the fusion rule; D fuse {( D w ) + ( D w )} ir vis 2 = (4.3) [w + w ] 2 The fused output is shown in Figure 4.6,

24 82 Figure 4.6 HF component fusion using weighted average method The final fused output is given by I = S + D (4.32) fuse fuse fuse as shown in Figure 4.7. Figure 4.7 Fused output using the PCA max method

25 RESULTS Experiments were carried out with the newly developed PCA fusion algorithms on different sets of images pertaining to surveillance and night vision applications. The results of the experiments are tabulated in Table 4.. A more detailed comparison of the results is given in Chapter 6. In addition, the average results obtained by conducting experiments on 20 sets of images are shown in Table 4.2. Here the results are compared with the existing DWT-PCA-Max algorithm (Zheng et al. 2007) and the PCA weighted superposition method (Rockinger 999). Based on the experimental work performed using the various quality metrics for the newly designed PCA-based fusion, the results obtained are given in Table 4.. Table 4. Performance metrics for the newly designed PCA-based fusion algorithms Image Fusion Scheme En SSIM MI SD CE Boat Road Scene PCA Max PCA Gaussian PCA Max PCA Gaussian

26 84 (a) PCA Max (b)pca Gaussian Figure 4.8 Fusion output for boat image (c) PCA Max (d) PCA Gaussian Figure 4.9 Fusion output for road scene image Table 4.2 Average value of different set of images Image 40 Sets of images Fusion Scheme PCA Max PCA Gaussian DWT PCA Max PCA fusion CE EN SSIM MI SD

27 CONCLUDING REMARKS This chapter gave a brief introduction to the Principal Component Analysis followed by the application of the PCA for analyzing images. Then the fusion rule that combines the PCA weighted scheme and selecting maximum intensity pixels was presented in this chapter. This algorithm has a very good performance both in terms of the visual quality of the fused image evaluated subjectively, and also in terms of quality metrics. Next the PCA Gaussian fusion scheme was described. This new fusion technique designed makes use of the principal components as the weights for the additive fusion. The performance of these two methods is compared in Table 4.. From the parameters it is seen that the PCA Gaussian method gives a higher SSIM index. Also from the output image it can be observed that this rule gives an output image of better quality when the source images have very high intensity pixels. In case of source images with an evenly distributed histogram the two algorithms give an equivalent performance. These results are presented in section 6.3. A comparative study of the fusion schemes proposed in this thesis along with existing algorithms is presented in chapter 6. Also the average results obtained for a set of 40 images shows a higher value of the SSIM based measure for the two new algorithms compared to the existing techniques. The cross entropy (CE) of the existing methods is better and so is the standard deviation (SD). However, these two metrics determine only the amount of information transferred from the source images to the fused output. The observation of the fused output images correlates the SSIM results on the efficiency of the new PCA algorithms.

7. Variable extraction and dimensionality reduction

7. Variable extraction and dimensionality reduction 7. Variable extraction and dimensionality reduction The goal of the variable selection in the preceding chapter was to find least useful variables so that it would be possible to reduce the dimensionality

More information

A Tutorial on Data Reduction. Principal Component Analysis Theoretical Discussion. By Shireen Elhabian and Aly Farag

A Tutorial on Data Reduction. Principal Component Analysis Theoretical Discussion. By Shireen Elhabian and Aly Farag A Tutorial on Data Reduction Principal Component Analysis Theoretical Discussion By Shireen Elhabian and Aly Farag University of Louisville, CVIP Lab November 2008 PCA PCA is A backbone of modern data

More information

Introduction to Machine Learning

Introduction to Machine Learning 10-701 Introduction to Machine Learning PCA Slides based on 18-661 Fall 2018 PCA Raw data can be Complex, High-dimensional To understand a phenomenon we measure various related quantities If we knew what

More information

Principal Components Analysis (PCA)

Principal Components Analysis (PCA) Principal Components Analysis (PCA) Principal Components Analysis (PCA) a technique for finding patterns in data of high dimension Outline:. Eigenvectors and eigenvalues. PCA: a) Getting the data b) Centering

More information

Mathematical foundations - linear algebra

Mathematical foundations - linear algebra Mathematical foundations - linear algebra Andrea Passerini passerini@disi.unitn.it Machine Learning Vector space Definition (over reals) A set X is called a vector space over IR if addition and scalar

More information

GEOG 4110/5100 Advanced Remote Sensing Lecture 15

GEOG 4110/5100 Advanced Remote Sensing Lecture 15 GEOG 4110/5100 Advanced Remote Sensing Lecture 15 Principal Component Analysis Relevant reading: Richards. Chapters 6.3* http://www.ce.yildiz.edu.tr/personal/songul/file/1097/principal_components.pdf *For

More information

PCA & ICA. CE-717: Machine Learning Sharif University of Technology Spring Soleymani

PCA & ICA. CE-717: Machine Learning Sharif University of Technology Spring Soleymani PCA & ICA CE-717: Machine Learning Sharif University of Technology Spring 2015 Soleymani Dimensionality Reduction: Feature Selection vs. Feature Extraction Feature selection Select a subset of a given

More information

Announcements (repeat) Principal Components Analysis

Announcements (repeat) Principal Components Analysis 4/7/7 Announcements repeat Principal Components Analysis CS 5 Lecture #9 April 4 th, 7 PA4 is due Monday, April 7 th Test # will be Wednesday, April 9 th Test #3 is Monday, May 8 th at 8AM Just hour long

More information

Dimension Reduction Techniques. Presented by Jie (Jerry) Yu

Dimension Reduction Techniques. Presented by Jie (Jerry) Yu Dimension Reduction Techniques Presented by Jie (Jerry) Yu Outline Problem Modeling Review of PCA and MDS Isomap Local Linear Embedding (LLE) Charting Background Advances in data collection and storage

More information

PCA, Kernel PCA, ICA

PCA, Kernel PCA, ICA PCA, Kernel PCA, ICA Learning Representations. Dimensionality Reduction. Maria-Florina Balcan 04/08/2015 Big & High-Dimensional Data High-Dimensions = Lot of Features Document classification Features per

More information

Implementation of Max Principle with PCA in Image Fusion for Surveillance and Navigation Application

Implementation of Max Principle with PCA in Image Fusion for Surveillance and Navigation Application Electronic Letters on Computer Vision and Image Analysis 10(1):1-10, 2011 Implementation of Max Principle with PCA in Image Fusion for Surveillance and Navigation Application Senthil Kumar Sadhasivam *,

More information

1 Principal Components Analysis

1 Principal Components Analysis Lecture 3 and 4 Sept. 18 and Sept.20-2006 Data Visualization STAT 442 / 890, CM 462 Lecture: Ali Ghodsi 1 Principal Components Analysis Principal components analysis (PCA) is a very popular technique for

More information

CS4495/6495 Introduction to Computer Vision. 8B-L2 Principle Component Analysis (and its use in Computer Vision)

CS4495/6495 Introduction to Computer Vision. 8B-L2 Principle Component Analysis (and its use in Computer Vision) CS4495/6495 Introduction to Computer Vision 8B-L2 Principle Component Analysis (and its use in Computer Vision) Wavelength 2 Wavelength 2 Principal Components Principal components are all about the directions

More information

Singular Value Decomposition and Principal Component Analysis (PCA) I

Singular Value Decomposition and Principal Component Analysis (PCA) I Singular Value Decomposition and Principal Component Analysis (PCA) I Prof Ned Wingreen MOL 40/50 Microarray review Data per array: 0000 genes, I (green) i,i (red) i 000 000+ data points! The expression

More information

Pattern Recognition 2

Pattern Recognition 2 Pattern Recognition 2 KNN,, Dr. Terence Sim School of Computing National University of Singapore Outline 1 2 3 4 5 Outline 1 2 3 4 5 The Bayes Classifier is theoretically optimum. That is, prob. of error

More information

Structure in Data. A major objective in data analysis is to identify interesting features or structure in the data.

Structure in Data. A major objective in data analysis is to identify interesting features or structure in the data. Structure in Data A major objective in data analysis is to identify interesting features or structure in the data. The graphical methods are very useful in discovering structure. There are basically two

More information

Data Preprocessing Tasks

Data Preprocessing Tasks Data Tasks 1 2 3 Data Reduction 4 We re here. 1 Dimensionality Reduction Dimensionality reduction is a commonly used approach for generating fewer features. Typically used because too many features can

More information

12.2 Dimensionality Reduction

12.2 Dimensionality Reduction 510 Chapter 12 of this dimensionality problem, regularization techniques such as SVD are almost always needed to perform the covariance matrix inversion. Because it appears to be a fundamental property

More information

GNR401 Principles of Satellite Image Processing

GNR401 Principles of Satellite Image Processing Principles of Satellite Image Processing Instructor: Prof. CSRE, IIT Bombay bkmohan@csre.iitb.ac.in Slot 5 Guest Lecture PCT and Band Arithmetic November 07, 2012 9.30 AM 10.55 AM IIT Bombay Slide 1 November

More information

EECS490: Digital Image Processing. Lecture #26

EECS490: Digital Image Processing. Lecture #26 Lecture #26 Moments; invariant moments Eigenvector, principal component analysis Boundary coding Image primitives Image representation: trees, graphs Object recognition and classes Minimum distance classifiers

More information

System 1 (last lecture) : limited to rigidly structured shapes. System 2 : recognition of a class of varying shapes. Need to:

System 1 (last lecture) : limited to rigidly structured shapes. System 2 : recognition of a class of varying shapes. Need to: System 2 : Modelling & Recognising Modelling and Recognising Classes of Classes of Shapes Shape : PDM & PCA All the same shape? System 1 (last lecture) : limited to rigidly structured shapes System 2 :

More information

Introduction to Machine Learning. PCA and Spectral Clustering. Introduction to Machine Learning, Slides: Eran Halperin

Introduction to Machine Learning. PCA and Spectral Clustering. Introduction to Machine Learning, Slides: Eran Halperin 1 Introduction to Machine Learning PCA and Spectral Clustering Introduction to Machine Learning, 2013-14 Slides: Eran Halperin Singular Value Decomposition (SVD) The singular value decomposition (SVD)

More information

Principal Component Analysis

Principal Component Analysis Principal Component Analysis Yingyu Liang yliang@cs.wisc.edu Computer Sciences Department University of Wisconsin, Madison [based on slides from Nina Balcan] slide 1 Goals for the lecture you should understand

More information

LECTURE 16: PCA AND SVD

LECTURE 16: PCA AND SVD Instructor: Sael Lee CS549 Computational Biology LECTURE 16: PCA AND SVD Resource: PCA Slide by Iyad Batal Chapter 12 of PRML Shlens, J. (2003). A tutorial on principal component analysis. CONTENT Principal

More information

Singular Value Decomposition. 1 Singular Value Decomposition and the Four Fundamental Subspaces

Singular Value Decomposition. 1 Singular Value Decomposition and the Four Fundamental Subspaces Singular Value Decomposition This handout is a review of some basic concepts in linear algebra For a detailed introduction, consult a linear algebra text Linear lgebra and its pplications by Gilbert Strang

More information

Image enhancement. Why image enhancement? Why image enhancement? Why image enhancement? Example of artifacts caused by image encoding

Image enhancement. Why image enhancement? Why image enhancement? Why image enhancement? Example of artifacts caused by image encoding 13 Why image enhancement? Image enhancement Example of artifacts caused by image encoding Computer Vision, Lecture 14 Michael Felsberg Computer Vision Laboratory Department of Electrical Engineering 12

More information

What is Principal Component Analysis?

What is Principal Component Analysis? What is Principal Component Analysis? Principal component analysis (PCA) Reduce the dimensionality of a data set by finding a new set of variables, smaller than the original set of variables Retains most

More information

Dimensionality Reduction with Principal Component Analysis

Dimensionality Reduction with Principal Component Analysis 10 Dimensionality Reduction with Principal Component Analysis Working directly with high-dimensional data, such as images, comes with some difficulties: it is hard to analyze, interpretation is difficult,

More information

HST.582J/6.555J/16.456J

HST.582J/6.555J/16.456J Blind Source Separation: PCA & ICA HST.582J/6.555J/16.456J Gari D. Clifford gari [at] mit. edu http://www.mit.edu/~gari G. D. Clifford 2005-2009 What is BSS? Assume an observation (signal) is a linear

More information

The Principal Component Analysis

The Principal Component Analysis The Principal Component Analysis Philippe B. Laval KSU Fall 2017 Philippe B. Laval (KSU) PCA Fall 2017 1 / 27 Introduction Every 80 minutes, the two Landsat satellites go around the world, recording images

More information

Maximum variance formulation

Maximum variance formulation 12.1. Principal Component Analysis 561 Figure 12.2 Principal component analysis seeks a space of lower dimensionality, known as the principal subspace and denoted by the magenta line, such that the orthogonal

More information

Wavelet Transform And Principal Component Analysis Based Feature Extraction

Wavelet Transform And Principal Component Analysis Based Feature Extraction Wavelet Transform And Principal Component Analysis Based Feature Extraction Keyun Tong June 3, 2010 As the amount of information grows rapidly and widely, feature extraction become an indispensable technique

More information

Principal Component Analysis

Principal Component Analysis Principal Component Analysis Yuanzhen Shao MA 26500 Yuanzhen Shao PCA 1 / 13 Data as points in R n Assume that we have a collection of data in R n. x 11 x 21 x 12 S = {X 1 =., X x 22 2 =.,, X x m2 m =.

More information

Universität Potsdam Institut für Informatik Lehrstuhl Maschinelles Lernen PCA. Tobias Scheffer

Universität Potsdam Institut für Informatik Lehrstuhl Maschinelles Lernen PCA. Tobias Scheffer Universität Potsdam Institut für Informatik Lehrstuhl Maschinelles Lernen PCA Tobias Scheffer Overview Principal Component Analysis (PCA) Kernel-PCA Fisher Linear Discriminant Analysis t-sne 2 PCA: Motivation

More information

Digital Image Processing Chapter 11 Representation and Description

Digital Image Processing Chapter 11 Representation and Description Digital Image Processing Chapter 11 Representation and Description Last Updated: July 20, 2003 Preview 11.1 Representation 11.1.1 Chain Codes Chain codes are used to represent a boundary by a connected

More information

Principal Component Analysis -- PCA (also called Karhunen-Loeve transformation)

Principal Component Analysis -- PCA (also called Karhunen-Loeve transformation) Principal Component Analysis -- PCA (also called Karhunen-Loeve transformation) PCA transforms the original input space into a lower dimensional space, by constructing dimensions that are linear combinations

More information

Overview of clustering analysis. Yuehua Cui

Overview of clustering analysis. Yuehua Cui Overview of clustering analysis Yuehua Cui Email: cuiy@msu.edu http://www.stt.msu.edu/~cui A data set with clear cluster structure How would you design an algorithm for finding the three clusters in this

More information

Principal Component Analysis

Principal Component Analysis Principal Component Analysis CS5240 Theoretical Foundations in Multimedia Leow Wee Kheng Department of Computer Science School of Computing National University of Singapore Leow Wee Kheng (NUS) Principal

More information

Independent Component Analysis and Its Application on Accelerator Physics

Independent Component Analysis and Its Application on Accelerator Physics Independent Component Analysis and Its Application on Accelerator Physics Xiaoying Pang LA-UR-12-20069 ICA and PCA Similarities: Blind source separation method (BSS) no model Observed signals are linear

More information

Lecture: Face Recognition and Feature Reduction

Lecture: Face Recognition and Feature Reduction Lecture: Face Recognition and Feature Reduction Juan Carlos Niebles and Ranjay Krishna Stanford Vision and Learning Lab 1 Recap - Curse of dimensionality Assume 5000 points uniformly distributed in the

More information

Dimensionality Reduction

Dimensionality Reduction Lecture 5 1 Outline 1. Overview a) What is? b) Why? 2. Principal Component Analysis (PCA) a) Objectives b) Explaining variability c) SVD 3. Related approaches a) ICA b) Autoencoders 2 Example 1: Sportsball

More information

Focus was on solving matrix inversion problems Now we look at other properties of matrices Useful when A represents a transformations.

Focus was on solving matrix inversion problems Now we look at other properties of matrices Useful when A represents a transformations. Previously Focus was on solving matrix inversion problems Now we look at other properties of matrices Useful when A represents a transformations y = Ax Or A simply represents data Notion of eigenvectors,

More information

Methods for sparse analysis of high-dimensional data, II

Methods for sparse analysis of high-dimensional data, II Methods for sparse analysis of high-dimensional data, II Rachel Ward May 23, 2011 High dimensional data with low-dimensional structure 300 by 300 pixel images = 90, 000 dimensions 2 / 47 High dimensional

More information

Lecture 24: Principal Component Analysis. Aykut Erdem May 2016 Hacettepe University

Lecture 24: Principal Component Analysis. Aykut Erdem May 2016 Hacettepe University Lecture 4: Principal Component Analysis Aykut Erdem May 016 Hacettepe University This week Motivation PCA algorithms Applications PCA shortcomings Autoencoders Kernel PCA PCA Applications Data Visualization

More information

Lecture Notes 2: Matrices

Lecture Notes 2: Matrices Optimization-based data analysis Fall 2017 Lecture Notes 2: Matrices Matrices are rectangular arrays of numbers, which are extremely useful for data analysis. They can be interpreted as vectors in a vector

More information

Gopalkrishna Veni. Project 4 (Active Shape Models)

Gopalkrishna Veni. Project 4 (Active Shape Models) Gopalkrishna Veni Project 4 (Active Shape Models) Introduction Active shape Model (ASM) is a technique of building a model by learning the variability patterns from training datasets. ASMs try to deform

More information

Lecture: Face Recognition and Feature Reduction

Lecture: Face Recognition and Feature Reduction Lecture: Face Recognition and Feature Reduction Juan Carlos Niebles and Ranjay Krishna Stanford Vision and Learning Lab Lecture 11-1 Recap - Curse of dimensionality Assume 5000 points uniformly distributed

More information

Methods for sparse analysis of high-dimensional data, II

Methods for sparse analysis of high-dimensional data, II Methods for sparse analysis of high-dimensional data, II Rachel Ward May 26, 2011 High dimensional data with low-dimensional structure 300 by 300 pixel images = 90, 000 dimensions 2 / 55 High dimensional

More information

COMPLEX PRINCIPAL COMPONENT SPECTRA EXTRACTION

COMPLEX PRINCIPAL COMPONENT SPECTRA EXTRACTION COMPLEX PRINCIPAL COMPONEN SPECRA EXRACION PROGRAM complex_pca_spectra Computing principal components o begin, click the Formation attributes tab in the AASPI-UIL window and select program complex_pca_spectra:

More information

Derivation of the Kalman Filter

Derivation of the Kalman Filter Derivation of the Kalman Filter Kai Borre Danish GPS Center, Denmark Block Matrix Identities The key formulas give the inverse of a 2 by 2 block matrix, assuming T is invertible: T U 1 L M. (1) V W N P

More information

Principal component analysis

Principal component analysis Principal component analysis Angela Montanari 1 Introduction Principal component analysis (PCA) is one of the most popular multivariate statistical methods. It was first introduced by Pearson (1901) and

More information

Linear Algebra in Computer Vision. Lecture2: Basic Linear Algebra & Probability. Vector. Vector Operations

Linear Algebra in Computer Vision. Lecture2: Basic Linear Algebra & Probability. Vector. Vector Operations Linear Algebra in Computer Vision CSED441:Introduction to Computer Vision (2017F Lecture2: Basic Linear Algebra & Probability Bohyung Han CSE, POSTECH bhhan@postech.ac.kr Mathematics in vector space Linear

More information

Analysis of Spectral Kernel Design based Semi-supervised Learning

Analysis of Spectral Kernel Design based Semi-supervised Learning Analysis of Spectral Kernel Design based Semi-supervised Learning Tong Zhang IBM T. J. Watson Research Center Yorktown Heights, NY 10598 Rie Kubota Ando IBM T. J. Watson Research Center Yorktown Heights,

More information

15 Singular Value Decomposition

15 Singular Value Decomposition 15 Singular Value Decomposition For any high-dimensional data analysis, one s first thought should often be: can I use an SVD? The singular value decomposition is an invaluable analysis tool for dealing

More information

Machine Learning. Principal Components Analysis. Le Song. CSE6740/CS7641/ISYE6740, Fall 2012

Machine Learning. Principal Components Analysis. Le Song. CSE6740/CS7641/ISYE6740, Fall 2012 Machine Learning CSE6740/CS7641/ISYE6740, Fall 2012 Principal Components Analysis Le Song Lecture 22, Nov 13, 2012 Based on slides from Eric Xing, CMU Reading: Chap 12.1, CB book 1 2 Factor or Component

More information

Edge Detection in Computer Vision Systems

Edge Detection in Computer Vision Systems 1 CS332 Visual Processing in Computer and Biological Vision Systems Edge Detection in Computer Vision Systems This handout summarizes much of the material on the detection and description of intensity

More information

Principal Component Analysis

Principal Component Analysis CSci 5525: Machine Learning Dec 3, 2008 The Main Idea Given a dataset X = {x 1,..., x N } The Main Idea Given a dataset X = {x 1,..., x N } Find a low-dimensional linear projection The Main Idea Given

More information

Principal Component Analysis

Principal Component Analysis Principal Component Analysis Anders Øland David Christiansen 1 Introduction Principal Component Analysis, or PCA, is a commonly used multi-purpose technique in data analysis. It can be used for feature

More information

Principal Component Analysis (PCA)

Principal Component Analysis (PCA) Principal Component Analysis (PCA) Additional reading can be found from non-assessed exercises (week 8) in this course unit teaching page. Textbooks: Sect. 6.3 in [1] and Ch. 12 in [2] Outline Introduction

More information

Advanced Introduction to Machine Learning CMU-10715

Advanced Introduction to Machine Learning CMU-10715 Advanced Introduction to Machine Learning CMU-10715 Principal Component Analysis Barnabás Póczos Contents Motivation PCA algorithms Applications Some of these slides are taken from Karl Booksh Research

More information

Principle Components Analysis (PCA) Relationship Between a Linear Combination of Variables and Axes Rotation for PCA

Principle Components Analysis (PCA) Relationship Between a Linear Combination of Variables and Axes Rotation for PCA Principle Components Analysis (PCA) Relationship Between a Linear Combination of Variables and Axes Rotation for PCA Principle Components Analysis: Uses one group of variables (we will call this X) In

More information

Computation. For QDA we need to calculate: Lets first consider the case that

Computation. For QDA we need to calculate: Lets first consider the case that Computation For QDA we need to calculate: δ (x) = 1 2 log( Σ ) 1 2 (x µ ) Σ 1 (x µ ) + log(π ) Lets first consider the case that Σ = I,. This is the case where each distribution is spherical, around the

More information

EECS 275 Matrix Computation

EECS 275 Matrix Computation EECS 275 Matrix Computation Ming-Hsuan Yang Electrical Engineering and Computer Science University of California at Merced Merced, CA 95344 http://faculty.ucmerced.edu/mhyang Lecture 6 1 / 22 Overview

More information

Principal Components Analysis (PCA) and Singular Value Decomposition (SVD) with applications to Microarrays

Principal Components Analysis (PCA) and Singular Value Decomposition (SVD) with applications to Microarrays Principal Components Analysis (PCA) and Singular Value Decomposition (SVD) with applications to Microarrays Prof. Tesler Math 283 Fall 2015 Prof. Tesler Principal Components Analysis Math 283 / Fall 2015

More information

Basic Principles of Video Coding

Basic Principles of Video Coding Basic Principles of Video Coding Introduction Categories of Video Coding Schemes Information Theory Overview of Video Coding Techniques Predictive coding Transform coding Quantization Entropy coding Motion

More information

Robot Image Credit: Viktoriya Sukhanova 123RF.com. Dimensionality Reduction

Robot Image Credit: Viktoriya Sukhanova 123RF.com. Dimensionality Reduction Robot Image Credit: Viktoriya Sukhanova 13RF.com Dimensionality Reduction Feature Selection vs. Dimensionality Reduction Feature Selection (last time) Select a subset of features. When classifying novel

More information

Data Mining. Dimensionality reduction. Hamid Beigy. Sharif University of Technology. Fall 1395

Data Mining. Dimensionality reduction. Hamid Beigy. Sharif University of Technology. Fall 1395 Data Mining Dimensionality reduction Hamid Beigy Sharif University of Technology Fall 1395 Hamid Beigy (Sharif University of Technology) Data Mining Fall 1395 1 / 42 Outline 1 Introduction 2 Feature selection

More information

IMAGE COMPRESSION-II. Week IX. 03/6/2003 Image Compression-II 1

IMAGE COMPRESSION-II. Week IX. 03/6/2003 Image Compression-II 1 IMAGE COMPRESSION-II Week IX 3/6/23 Image Compression-II 1 IMAGE COMPRESSION Data redundancy Self-information and Entropy Error-free and lossy compression Huffman coding Predictive coding Transform coding

More information

Waveform-Based Coding: Outline

Waveform-Based Coding: Outline Waveform-Based Coding: Transform and Predictive Coding Yao Wang Polytechnic University, Brooklyn, NY11201 http://eeweb.poly.edu/~yao Based on: Y. Wang, J. Ostermann, and Y.-Q. Zhang, Video Processing and

More information

Unsupervised Machine Learning and Data Mining. DS 5230 / DS Fall Lecture 7. Jan-Willem van de Meent

Unsupervised Machine Learning and Data Mining. DS 5230 / DS Fall Lecture 7. Jan-Willem van de Meent Unsupervised Machine Learning and Data Mining DS 5230 / DS 4420 - Fall 2018 Lecture 7 Jan-Willem van de Meent DIMENSIONALITY REDUCTION Borrowing from: Percy Liang (Stanford) Dimensionality Reduction Goal:

More information

Expectation Maximization

Expectation Maximization Expectation Maximization Machine Learning CSE546 Carlos Guestrin University of Washington November 13, 2014 1 E.M.: The General Case E.M. widely used beyond mixtures of Gaussians The recipe is the same

More information

Least Squares Optimization

Least Squares Optimization Least Squares Optimization The following is a brief review of least squares optimization and constrained optimization techniques. Broadly, these techniques can be used in data analysis and visualization

More information

Eigenimaging for Facial Recognition

Eigenimaging for Facial Recognition Eigenimaging for Facial Recognition Aaron Kosmatin, Clayton Broman December 2, 21 Abstract The interest of this paper is Principal Component Analysis, specifically its area of application to facial recognition

More information

Statistical Machine Learning

Statistical Machine Learning Statistical Machine Learning Christoph Lampert Spring Semester 2015/2016 // Lecture 12 1 / 36 Unsupervised Learning Dimensionality Reduction 2 / 36 Dimensionality Reduction Given: data X = {x 1,..., x

More information

Computer Assisted Image Analysis

Computer Assisted Image Analysis Computer Assisted Image Analysis Lecture 0 - Object Descriptors II Amin Allalou amin@cb.uu.se Centre for Image Analysis Uppsala University 2009-04-27 A. Allalou (Uppsala University) Object Descriptors

More information

Face Recognition. Face Recognition. Subspace-Based Face Recognition Algorithms. Application of Face Recognition

Face Recognition. Face Recognition. Subspace-Based Face Recognition Algorithms. Application of Face Recognition ace Recognition Identify person based on the appearance of face CSED441:Introduction to Computer Vision (2017) Lecture10: Subspace Methods and ace Recognition Bohyung Han CSE, POSTECH bhhan@postech.ac.kr

More information

MACHINE LEARNING. Methods for feature extraction and reduction of dimensionality: Probabilistic PCA and kernel PCA

MACHINE LEARNING. Methods for feature extraction and reduction of dimensionality: Probabilistic PCA and kernel PCA 1 MACHINE LEARNING Methods for feature extraction and reduction of dimensionality: Probabilistic PCA and kernel PCA 2 Practicals Next Week Next Week, Practical Session on Computer Takes Place in Room GR

More information

Karhunen-Loève Transform KLT. JanKees van der Poel D.Sc. Student, Mechanical Engineering

Karhunen-Loève Transform KLT. JanKees van der Poel D.Sc. Student, Mechanical Engineering Karhunen-Loève Transform KLT JanKees van der Poel D.Sc. Student, Mechanical Engineering Karhunen-Loève Transform Has many names cited in literature: Karhunen-Loève Transform (KLT); Karhunen-Loève Decomposition

More information

2. LINEAR ALGEBRA. 1. Definitions. 2. Linear least squares problem. 3. QR factorization. 4. Singular value decomposition (SVD) 5.

2. LINEAR ALGEBRA. 1. Definitions. 2. Linear least squares problem. 3. QR factorization. 4. Singular value decomposition (SVD) 5. 2. LINEAR ALGEBRA Outline 1. Definitions 2. Linear least squares problem 3. QR factorization 4. Singular value decomposition (SVD) 5. Pseudo-inverse 6. Eigenvalue decomposition (EVD) 1 Definitions Vector

More information

1 Inner Product and Orthogonality

1 Inner Product and Orthogonality CSCI 4/Fall 6/Vora/GWU/Orthogonality and Norms Inner Product and Orthogonality Definition : The inner product of two vectors x and y, x x x =.., y =. x n y y... y n is denoted x, y : Note that n x, y =

More information

Signal Analysis. Principal Component Analysis

Signal Analysis. Principal Component Analysis Multi dimensional Signal Analysis Lecture 2E Principal Component Analysis Subspace representation Note! Given avector space V of dimension N a scalar product defined by G 0 a subspace U of dimension M

More information

Machine Learning. Dimensionality reduction. Hamid Beigy. Sharif University of Technology. Fall 1395

Machine Learning. Dimensionality reduction. Hamid Beigy. Sharif University of Technology. Fall 1395 Machine Learning Dimensionality reduction Hamid Beigy Sharif University of Technology Fall 1395 Hamid Beigy (Sharif University of Technology) Machine Learning Fall 1395 1 / 47 Table of contents 1 Introduction

More information

Principal Component Analysis

Principal Component Analysis B: Chapter 1 HTF: Chapter 1.5 Principal Component Analysis Barnabás Póczos University of Alberta Nov, 009 Contents Motivation PCA algorithms Applications Face recognition Facial expression recognition

More information

Basic Concepts of. Feature Selection

Basic Concepts of. Feature Selection Basic Concepts of Pattern Recognition and Feature Selection Xiaojun Qi -- REU Site Program in CVMA (2011 Summer) 1 Outline Pattern Recognition Pattern vs. Features Pattern Classes Classification Feature

More information

COMS 4721: Machine Learning for Data Science Lecture 19, 4/6/2017

COMS 4721: Machine Learning for Data Science Lecture 19, 4/6/2017 COMS 4721: Machine Learning for Data Science Lecture 19, 4/6/2017 Prof. John Paisley Department of Electrical Engineering & Data Science Institute Columbia University PRINCIPAL COMPONENT ANALYSIS DIMENSIONALITY

More information

Linear Subspace Models

Linear Subspace Models Linear Subspace Models Goal: Explore linear models of a data set. Motivation: A central question in vision concerns how we represent a collection of data vectors. The data vectors may be rasterized images,

More information

UCLA STAT 233 Statistical Methods in Biomedical Imaging

UCLA STAT 233 Statistical Methods in Biomedical Imaging UCLA STAT 233 Statistical Methods in Biomedical Imaging Instructor: Ivo Dinov, Asst. Prof. In Statistics and Neurology University of California, Los Angeles, Spring 2004 http://www.stat.ucla.edu/~dinov/

More information

Tutorial on Principal Component Analysis

Tutorial on Principal Component Analysis Tutorial on Principal Component Analysis Copyright c 1997, 2003 Javier R. Movellan. This is an open source document. Permission is granted to copy, distribute and/or modify this document under the terms

More information

LECTURE NOTE #11 PROF. ALAN YUILLE

LECTURE NOTE #11 PROF. ALAN YUILLE LECTURE NOTE #11 PROF. ALAN YUILLE 1. NonLinear Dimension Reduction Spectral Methods. The basic idea is to assume that the data lies on a manifold/surface in D-dimensional space, see figure (1) Perform

More information

Covariance and Principal Components

Covariance and Principal Components COMP3204/COMP6223: Computer Vision Covariance and Principal Components Jonathon Hare jsh2@ecs.soton.ac.uk Variance and Covariance Random Variables and Expected Values Mathematicians talk variance (and

More information

Conceptual Questions for Review

Conceptual Questions for Review Conceptual Questions for Review Chapter 1 1.1 Which vectors are linear combinations of v = (3, 1) and w = (4, 3)? 1.2 Compare the dot product of v = (3, 1) and w = (4, 3) to the product of their lengths.

More information

Principal Component Analysis vs. Independent Component Analysis for Damage Detection

Principal Component Analysis vs. Independent Component Analysis for Damage Detection 6th European Workshop on Structural Health Monitoring - Fr..D.4 Principal Component Analysis vs. Independent Component Analysis for Damage Detection D. A. TIBADUIZA, L. E. MUJICA, M. ANAYA, J. RODELLAR

More information

Lecture Notes 1: Vector spaces

Lecture Notes 1: Vector spaces Optimization-based data analysis Fall 2017 Lecture Notes 1: Vector spaces In this chapter we review certain basic concepts of linear algebra, highlighting their application to signal processing. 1 Vector

More information

Data Mining and Analysis: Fundamental Concepts and Algorithms

Data Mining and Analysis: Fundamental Concepts and Algorithms Data Mining and Analysis: Fundamental Concepts and Algorithms dataminingbook.info Mohammed J. Zaki 1 Wagner Meira Jr. 2 1 Department of Computer Science Rensselaer Polytechnic Institute, Troy, NY, USA

More information

Computational paradigms for the measurement signals processing. Metodologies for the development of classification algorithms.

Computational paradigms for the measurement signals processing. Metodologies for the development of classification algorithms. Computational paradigms for the measurement signals processing. Metodologies for the development of classification algorithms. January 5, 25 Outline Methodologies for the development of classification

More information

Principal Components Analysis. Sargur Srihari University at Buffalo

Principal Components Analysis. Sargur Srihari University at Buffalo Principal Components Analysis Sargur Srihari University at Buffalo 1 Topics Projection Pursuit Methods Principal Components Examples of using PCA Graphical use of PCA Multidimensional Scaling Srihari 2

More information

PRINCIPAL COMPONENT ANALYSIS

PRINCIPAL COMPONENT ANALYSIS PRINCIPAL COMPONENT ANALYSIS 1 INTRODUCTION One of the main problems inherent in statistics with more than two variables is the issue of visualising or interpreting data. Fortunately, quite often the problem

More information

GI07/COMPM012: Mathematical Programming and Research Methods (Part 2) 2. Least Squares and Principal Components Analysis. Massimiliano Pontil

GI07/COMPM012: Mathematical Programming and Research Methods (Part 2) 2. Least Squares and Principal Components Analysis. Massimiliano Pontil GI07/COMPM012: Mathematical Programming and Research Methods (Part 2) 2. Least Squares and Principal Components Analysis Massimiliano Pontil 1 Today s plan SVD and principal component analysis (PCA) Connection

More information

Linear Algebra & Geometry why is linear algebra useful in computer vision?

Linear Algebra & Geometry why is linear algebra useful in computer vision? Linear Algebra & Geometry why is linear algebra useful in computer vision? References: -Any book on linear algebra! -[HZ] chapters 2, 4 Some of the slides in this lecture are courtesy to Prof. Octavia

More information

Principal Component Analysis (PCA)

Principal Component Analysis (PCA) Principal Component Analysis (PCA) Salvador Dalí, Galatea of the Spheres CSC411/2515: Machine Learning and Data Mining, Winter 2018 Michael Guerzhoy and Lisa Zhang Some slides from Derek Hoiem and Alysha

More information