Manifold Learning and it s application

Similar documents
Focus was on solving matrix inversion problems Now we look at other properties of matrices Useful when A represents a transformations.

Non-linear Dimensionality Reduction

Nonlinear Dimensionality Reduction. Jose A. Costa

Nonlinear Manifold Learning Summary

Machine Learning. Data visualization and dimensionality reduction. Eric Xing. Lecture 7, August 13, Eric Xing Eric CMU,

L26: Advanced dimensionality reduction

Lecture 10: Dimension Reduction Techniques

Dimension Reduction and Low-dimensional Embedding

Dimension Reduction Techniques. Presented by Jie (Jerry) Yu

Nonlinear Dimensionality Reduction

Nonlinear Dimensionality Reduction

Unsupervised dimensionality reduction

Global (ISOMAP) versus Local (LLE) Methods in Nonlinear Dimensionality Reduction

Face Recognition Using Laplacianfaces He et al. (IEEE Trans PAMI, 2005) presented by Hassan A. Kingravi

Apprentissage non supervisée

Manifold Learning: Theory and Applications to HRI

Nonlinear Methods. Data often lies on or near a nonlinear low-dimensional curve aka manifold.

ISSN: (Online) Volume 3, Issue 5, May 2015 International Journal of Advance Research in Computer Science and Management Studies

CSE 291. Assignment Spectral clustering versus k-means. Out: Wed May 23 Due: Wed Jun 13

Statistical Pattern Recognition

Intrinsic Structure Study on Whale Vocalizations

Dimensionality Reduction AShortTutorial

LECTURE NOTE #11 PROF. ALAN YUILLE

(Non-linear) dimensionality reduction. Department of Computer Science, Czech Technical University in Prague

Locality Preserving Projections

Laplacian Eigenmaps for Dimensionality Reduction and Data Representation

Laplacian Eigenmaps for Dimensionality Reduction and Data Representation

Statistical Machine Learning

Distance Metric Learning in Data Mining (Part II) Fei Wang and Jimeng Sun IBM TJ Watson Research Center

Dimensionality Reduction: A Comparative Review

Advanced Machine Learning & Perception

Statistical and Computational Analysis of Locality Preserving Projection

Dimensionality Reduc1on

ISOMAP TRACKING WITH PARTICLE FILTER

Learning Eigenfunctions: Links with Spectral Clustering and Kernel PCA

Robust Laplacian Eigenmaps Using Global Information

Machine Learning. CUNY Graduate Center, Spring Lectures 11-12: Unsupervised Learning 1. Professor Liang Huang.

EECS 275 Matrix Computation

DIMENSION REDUCTION. min. j=1

Learning a Kernel Matrix for Nonlinear Dimensionality Reduction

Dimensionality Reduction: A Comparative Review

Manifold Learning: From Linear to nonlinear. Presenter: Wei-Lun (Harry) Chao Date: April 26 and May 3, 2012 At: AMMAI 2012

Principal Component Analysis -- PCA (also called Karhunen-Loeve transformation)

Discriminant Uncorrelated Neighborhood Preserving Projections

Learning a kernel matrix for nonlinear dimensionality reduction

Dimensionality Reduction:

Riemannian Manifold Learning for Nonlinear Dimensionality Reduction

Lecture: Some Practical Considerations (3 of 4)

Unsupervised Learning Techniques Class 07, 1 March 2006 Andrea Caponnetto

A Duality View of Spectral Methods for Dimensionality Reduction

Large-Scale Manifold Learning

A Duality View of Spectral Methods for Dimensionality Reduction

An Empirical Comparison of Dimensionality Reduction Methods for Classifying Gene and Protein Expression Datasets

Data-dependent representations: Laplacian Eigenmaps

Dimensionality Reduction

Founda'ons of Large- Scale Mul'media Informa'on Management and Retrieval. Lecture #4 Similarity. Edward Chang

Smart PCA. Yi Zhang Machine Learning Department Carnegie Mellon University

Data Mining. Dimensionality reduction. Hamid Beigy. Sharif University of Technology. Fall 1395

Data Mining II. Prof. Dr. Karsten Borgwardt, Department Biosystems, ETH Zürich. Basel, Spring Semester 2016 D-BSSE

Performance Evaluation of Nonlinear Dimensionality Reduction Methods on the BANCA Database

Graphs, Geometry and Semi-supervised Learning

Connection of Local Linear Embedding, ISOMAP, and Kernel Principal Component Analysis

Gaussian Process Latent Random Field

Nonlinear Learning using Local Coordinate Coding

DIDELĖS APIMTIES DUOMENŲ VIZUALI ANALIZĖ

Permutation-invariant regularization of large covariance matrices. Liza Levina

PARAMETERIZATION OF NON-LINEAR MANIFOLDS

Introduction to Machine Learning

Maximum likelihood estimation of Riemannian metrics from Euclidean data

Diffusion Geometries, Diffusion Wavelets and Harmonic Analysis of large data sets.

Graph Metrics and Dimension Reduction

Statistical Learning. Dong Liu. Dept. EEIS, USTC

Preprocessing & dimensionality reduction

Spectral Dimensionality Reduction via Maximum Entropy

Classification of handwritten digits using supervised locally linear embedding algorithm and support vector machine

Supplemental Materials for. Local Multidimensional Scaling for. Nonlinear Dimension Reduction, Graph Drawing. and Proximity Analysis

Lecture 3: Compressive Classification

Part I Generalized Principal Component Analysis

Improved Local Coordinate Coding using Local Tangents

Face Recognition. Face Recognition. Subspace-Based Face Recognition Algorithms. Application of Face Recognition

Metric Learning on Manifolds

A Tour of Unsupervised Learning Part I Graphical models and dimension reduction

Data dependent operators for the spatial-spectral fusion problem

Multivariate Statistics Random Projections and Johnson-Lindenstrauss Lemma

Machine Learning. Dimensionality reduction. Hamid Beigy. Sharif University of Technology. Fall 1395

Spherical Euclidean Distance Embedding of a Graph

TANGENT SPACE INTRINSIC MANIFOLD REGULARIZATION FOR DATA REPRESENTATION. Shiliang Sun

ECE 521. Lecture 11 (not on midterm material) 13 February K-means clustering, Dimensionality reduction

Is Manifold Learning for Toy Data only?

Methods for sparse analysis of high-dimensional data, II

Machine Learning. Chao Lan

Nonlinear Dimensionality Reduction by Semidefinite Programming and Kernel Matrix Factorization

Uncorrelated Multilinear Principal Component Analysis through Successive Variance Maximization

Dimensionality Reduction. CS57300 Data Mining Fall Instructor: Bruno Ribeiro

Machine Learning (BSMC-GA 4439) Wenke Liu

Regression on Manifolds Using Kernel Dimension Reduction

Eigenfaces. Face Recognition Using Principal Components Analysis

Dimensionality Estimation, Manifold Learning and Function Approximation using Tensor Voting

Dimensionality Reduction

Random projection trees and low dimensional manifolds. Sanjoy Dasgupta and Yoav Freund University of California, San Diego

Transcription:

Manifold Learning and it s application Nandan Dubey SE367

Outline 1 Introduction Manifold Examples image as vector Importance Dimension Reduction Techniques 2 Linear Methods PCA Example MDS Perception 3 Non-Linear Techniques 4 LLE 5 Isomap 6 References Further Reading

Manifold What is Manifold? Definition Manifold is a mathematical space that on a small enough scale resembles the Euclidean space of a specific dimension. Hence a line and a circle are one-dimensional manifolds, a plane and sphere (the surface of a ball) are two-dimensional manifolds.

Examples Introduction contd. S -shape Swiss roll

Examples Introduction contd. S -shape Swiss roll Both are two-dim. data embedded in 3D

image as vector Images as Vectors

Importance Why reduce Dimension? Curse of Dimensionality. Image data(each pixel), spectral coefficients, Text categorization(frequencies of phrases in a document), genes, many more. Practical data lie in subspace which is low dimensional.

Dimension Reduction Techniques Dimension reduction techniques Linear methods Principal Component Analysis(PCA) Multidimensional Scaling(MDS) Non-linear Methods Locally Linear Embedding (Roweis and Saul) IsoMap (Tenenbaum, de Silva, and Langford)

PCA PCA-Principal Component Analysis PCA Transforms possibly correlated variables into a smaller dimensional uncorrelated variables called principal components. Find linear subspace projection P which preserves the data locations (under quadratic error) Algorithm Evaluate Covariance matrix C = (x µ x )(x µ x ) T Simple eigenvector(v q ) solution Top q eigenvectors of XCC T X T V q =[v 1,... v q ] is a basis for the q-dim subspace Locations given by V T q XC

Example

Example PCA - Algorithm n is number of data points, D is dimension of input, d is output dimension. Data = {X 1, X 2,..., X n }, X i ɛr D, i = 1,... n goal is y i ɛr d, i =1,..., n d << D minimize reconstruction error min PɛR mxm n i=1 X i PX i 2 where P = U U T UɛR Dxd maximize the variance max y i y j 2 Then y i = U T X i

MDS Multidimensional Scaling Given pre-distances D Find Euclidean q-dim space which preserves these distances Solution is given by the eigenstructure of Top q eigenvectors This is exactly the same solution as PCA

MDS Deficiencies of Linear Methods Data may not be best summarized by linear combination of features Example: PCA cannot discover 1D structure of a helix

Perception How brain store these? Every pixel? Or perceptually meaningful structure?

Perception How brain store these? Every pixel? Or perceptually meaningful structure? : Up-down pose Left-right pose Lighting direction So, our brain successfully reduced the high-dimensional inputs to an intrinsically 3-dimensional manifold!

LLE and Isomap Local relationships: Two solutions which preserve local structure: LLE Locally Linear Embedding (LLE) : Change to a local representation (at each point) Base the local rep. on position of neighboring points Isomap IsoMap: Estimate actual (geodesic) distances in p-dim. space Find q-dim representation preserving those distances

Both rely on the locally flat nature of the manifold How do we find a locality in which this is true?

Both rely on the locally flat nature of the manifold How do we find a locality in which this is true? k-nearest-neighbors

Both rely on the locally flat nature of the manifold How do we find a locality in which this is true? k-nearest-neighbors epsilon-ball

Local Linear Embedding Select a local neighborhood Change each point into a coordinate system based on its neighbors Find new (q-dim) coordinates which reproduce these local relationships

LLE by image

Locally Linear Embedding Find new (q-dim) coordinates which reproduce these local coordinates. This can be solved using the eigenstructure as well: We want the min. variance

Locally Linear Embedding minimize n i=1 X i k j=1 W ijx j 2 W is n x n sparse matrix minimize n i=1 y i k j=1 W ijy j 2 or n n i=1 j=1 M ijyi T y i M = n x n matrix M = (I - W) T (I W )

Isomap Distance Matrix Build a graph with K-nearest neighbor or ɛ nearest neighbor. Geodesic Distance Infer other interpoint distances by finding shortest paths on the graph (Dijkstra s algorithm) MDS Apply MDS to embed the nearest neighbors.

an Application

an Application Linear Interpolation

an Application Interpolation By Manifold

another Application(Isomap)

other Applications Example Pose estimation image denoising missing data interpolation

Image References Learning-based Methods in Vision Alexei Efros, CMU, Spring 2007 Nonlinear Dimensionality Reduction by Locally Linear Embedding Sam T. Roweis and Lawrence K. Saul

Further Reading Further Reading MDS : Encyclopedia of Cognitive Science Multidimensional Scaling Mark Steyvers PCA: Jonathon Shlens, A Tutorial on Principal Component Analysis or refer Wikipedia ISOMAP: A Global Geometric Framework for Nonlinear Dimensionality Reduction Joshua B. Tenenbaum, Vin de Silva, John C. Langford LLE: Nonlinear Dimensionality Reduction by Locally Linear Embedding Sam T. Roweis and Lawrence K. Saul

Further Reading Further Reading Thank you! http://home.iitk.ac.in/ ndubey/thesis