Data-dependent representations: Laplacian Eigenmaps

Similar documents
Data dependent operators for the spatial-spectral fusion problem

Laplacian Eigenmaps for Dimensionality Reduction and Data Representation

Nonlinear Dimensionality Reduction. Jose A. Costa

Nonlinear Methods. Data often lies on or near a nonlinear low-dimensional curve aka manifold.

Learning Eigenfunctions: Links with Spectral Clustering and Kernel PCA

Nonlinear Dimensionality Reduction

Data Analysis and Manifold Learning Lecture 3: Graphs, Graph Matrices, and Graph Embeddings

Nonlinear Dimensionality Reduction

Laplacian Eigenmaps for Dimensionality Reduction and Data Representation

Unsupervised dimensionality reduction

Graph Metrics and Dimension Reduction

Non-linear Dimensionality Reduction

Intrinsic Structure Study on Whale Vocalizations

Unsupervised Learning Techniques Class 07, 1 March 2006 Andrea Caponnetto

Data Analysis and Manifold Learning Lecture 7: Spectral Clustering

CSE 291. Assignment Spectral clustering versus k-means. Out: Wed May 23 Due: Wed Jun 13

Beyond Scalar Affinities for Network Analysis or Vector Diffusion Maps and the Connection Laplacian

Dimensionality Reduction: A Comparative Review

Robust Laplacian Eigenmaps Using Global Information

Dimensionality Reduc1on

Learning a Kernel Matrix for Nonlinear Dimensionality Reduction

EECS 275 Matrix Computation

Locality Preserving Projections

Manifold Learning: Theory and Applications to HRI

Diffusion Geometries, Diffusion Wavelets and Harmonic Analysis of large data sets.

Contribution from: Springer Verlag Berlin Heidelberg 2005 ISBN

Connection of Local Linear Embedding, ISOMAP, and Kernel Principal Component Analysis

Learning gradients: prescriptive models

Multiscale Analysis and Diffusion Semigroups With Applications

Graphs, Geometry and Semi-supervised Learning

Lecture: Some Practical Considerations (3 of 4)

Lecture 10: Dimension Reduction Techniques

Learning on Graphs and Manifolds. CMPSCI 689 Sridhar Mahadevan U.Mass Amherst

Learning a kernel matrix for nonlinear dimensionality reduction

Dimensionality Reduction: A Comparative Review

LECTURE NOTE #11 PROF. ALAN YUILLE

Diffusion Wavelets and Applications

March 13, Paper: R.R. Coifman, S. Lafon, Diffusion maps ([Coifman06]) Seminar: Learning with Graphs, Prof. Hein, Saarland University

Face Recognition Using Laplacianfaces He et al. (IEEE Trans PAMI, 2005) presented by Hassan A. Kingravi

Apprentissage non supervisée

DIMENSION REDUCTION. min. j=1

Mid-year Report Linear and Non-linear Dimentionality. Reduction. applied to gene expression data of cancer tissue samples

A Duality View of Spectral Methods for Dimensionality Reduction

Spectral Dimensionality Reduction

EE Technion, Spring then. can be isometrically embedded into can be realized as a Gram matrix of rank, Properties:

A Duality View of Spectral Methods for Dimensionality Reduction

Dimension Reduction Techniques. Presented by Jie (Jerry) Yu

Dimensionality Reduction AShortTutorial

Focus was on solving matrix inversion problems Now we look at other properties of matrices Useful when A represents a transformations.

Certifying the Global Optimality of Graph Cuts via Semidefinite Programming: A Theoretic Guarantee for Spectral Clustering

Global vs. Multiscale Approaches

Advances in Manifold Learning Presented by: Naku Nak l Verm r a June 10, 2008

Manifold Regularization

Multiscale Manifold Learning

Statistical and Computational Analysis of Locality Preserving Projection

Introduction to Machine Learning

Linear and Non-linear Dimension Reduction Applied to Gene Expression Data of Cancer Tissue Samples

Analysis of Spectral Kernel Design based Semi-supervised Learning

PARAMETERIZATION OF NON-LINEAR MANIFOLDS

Advanced Machine Learning & Perception

TANGENT SPACE INTRINSIC MANIFOLD REGULARIZATION FOR DATA REPRESENTATION. Shiliang Sun

An Algorithmist s Toolkit September 10, Lecture 1

Distance Metric Learning in Data Mining (Part II) Fei Wang and Jimeng Sun IBM TJ Watson Research Center

Learning Eigenfunctions Links Spectral Embedding

Machine Learning for Data Science (CS4786) Lecture 11

Self-Tuning Semantic Image Segmentation

Informative Laplacian Projection

Functional Maps ( ) Dr. Emanuele Rodolà Room , Informatik IX

Nonlinear Dimensionality Reduction by Semidefinite Programming and Kernel Matrix Factorization

Machine Learning - MT Clustering

Manifold Alignment using Procrustes Analysis

The Nyström Extension and Spectral Methods in Learning

Large-Scale Manifold Learning

Diffusion Geometries, Global and Multiscale

Is Manifold Learning for Toy Data only?

Dimension Reduction and Low-dimensional Embedding

Definition and basic properties of heat kernels I, An introduction

SPECTRAL CLUSTERING AND KERNEL PRINCIPAL COMPONENT ANALYSIS ARE PURSUING GOOD PROJECTIONS

Manifold Learning: From Linear to nonlinear. Presenter: Wei-Lun (Harry) Chao Date: April 26 and May 3, 2012 At: AMMAI 2012

Dimensionality Reduction:

Statistical Learning. Dong Liu. Dept. EEIS, USTC

Hot Spots Conjecture and Its Application to Modeling Tubular Structures

Basic Calculus Review

Spectral Techniques for Clustering

HYPERGRAPH BASED SEMI-SUPERVISED LEARNING ALGORITHMS APPLIED TO SPEECH RECOGNITION PROBLEM: A NOVEL APPROACH

Machine Learning. Data visualization and dimensionality reduction. Eric Xing. Lecture 7, August 13, Eric Xing Eric CMU,

Directed Graph Embedding: an Algorithm based on Continuous Limits of Laplacian-type Operators

Bi-stochastic kernels via asymmetric affinity functions

Diffusion Maps, Spectral Clustering and Eigenfunctions of Fokker-Planck Operators

Metric Learning on Manifolds

EE16B Designing Information Devices and Systems II

Nearly Isometric Embedding by Relaxation

Statistical Pattern Recognition

Graphs in Machine Learning

High-Dimensional Pattern Recognition using Low-Dimensional Embedding and Earth Mover s Distance

STA141C: Big Data & High Performance Statistical Computing

Manifold Learning and it s application

Image Analysis & Retrieval Lec 13 - Feature Dimension Reduction

Clustering in kernel embedding spaces and organization of documents

Nonlinear Dimensionality Reduction II: Diffusion Maps

Transcription:

Data-dependent representations: Laplacian Eigenmaps November 4, 2015

Data Organization and Manifold Learning There are many techniques for Data Organization and Manifold Learning, e.g., Principal Component Analysis (PCA), Locally Linear Embedding (LLE), Isomap, genetic algorithms, and neural networks. We are interested in a subfamily of these techniques known as Kernel Eigenmap Methods. These include Kernel PCA, LLE, Hessian LLE (HLLE), and Laplacian Eigenmaps. Kernel eigenmap methods require two steps. Given data space X of N vectors in R D. 1 Construction of an N N symmetric, positive semi-definite kernel, K, from these N data points in R D. 2 Diagonalization of K, and then choosing d D significant eigenmaps of K. These become our new coordinates, and accomplish dimensionality reduction. We are particularly interested in diffusion kernels K, which are defined by means of transition matrices.

Kernel Eigenmap Methods for Dimension Reduction - Kernel Construction Kernel eigenmap methods were introduced to address complexities not resolvable by linear methods. The idea behind kernel methods is to express correlations or similarities between vectors in the data space X in terms of a symmetric, positive semi-definite kernel function K : X X R. Generally, there exists a Hilbert space K and a mapping Φ : X K such that K (x, y) = Φ(x), Φ(y). Then, diagonalize by the spectral theorem and choose significant eigenmaps to obtain dimensionality reduction. Kernels can be constructed by many kernel eigenmap methods. These include Kernel PCA, LLE, HLLE, and Laplacian Eigenmaps.

Kernel Eigenmap Methods for Dimension Reduction - Kernel Diagonalization The second step in kernel eigenmap methods is the diagonalization of the kernel. Let e j, j = 1,..., N, be the set of eigenvectors of the kernel matrix K, with eigenvalues λ j. Order the eigenvalues monotonically. Choose the top d << D significant eigenvectors to map the original data points x i R D to (e 1 (i),..., e d (i)) R d, i = 1,..., N.

Data Organization X Y 1 K 2 There are other alternative interpretations for the steps of our diagram: 1 Constructions of kernels K may be independent from data and based on principles. 2 Redundant representations, such as frames, can be used to replace orthonormal eigendecompositions. We need not select the target dimensionality to be lower than the dimension of the input. This leads, to data expansion, or data organization, rather then dimensionality reduction.

Operator Theory on Graphs Presented approach leads to analysis of operators on data-dependent structures, such as graphs or manifolds. Locally Linear Embedding, Diffusion Maps, Diffusion Wavelets, Laplacian Eigenmaps, Schroedinger Eigenmaps Mathematical core: Pick a positive semidefinite bounded operator A as the infinitesimal generator of a semigroup of operators, e ta, t > 0. The semigroup can be identified with the Markov processes of diffusion or random walks, as is the case, e.g., with Diffusion Maps and Diffusion Wavelets The infinitesimal generator and the semigroup share the common representation, e.g., eigenbasis

Laplacian Eigenmaps M. Belkin and P. Niyogi, 2003. Points close on the manifold should remain close in R d. Use Laplace-Beltrami operator M to control the embedding. Use discrete approximations for practical problems. Proven convergence (Belkin and Niyogi, 2003 2008). Gave rise to Diffusion Maps and Diffusion Wavelets, among others.

Laplacian Eigenmaps - Implementation 1 Put an edge between nodes i and j if x i and x j are close. Precisely, given a parameter k N, put an edge between nodes i and j if x i is among the k nearest neighbors of x j or vice versa. 2 Given a parameter t > 0, if nodes i and j are connected, set W i,j = e x i x j 2 t. 3 Set D i,i = j W i,j, and let L = D W. Solve Lf = λdf, under the constraint y Dy = Id. Let f 0, f 1,..., f d be d + 1 eigenvector solutions corresponding to the first eigenvalues 0 = λ 0 λ 1 λ d. Discard f 0 and use the next d eigenvectors to embed in d-dimensional Euclidean space using the map x i (f 1 (i), f 2 (i),..., f d (i)).

Swiss Roll Figure : a) Original, b) LE

From eigenproblems to optimization Consider the following minimization problem, y R d, 1 min y 2 y i y j 2 W i,j = min tr(y Ly). Dy=Id y Dy=E i,j Its solution is given by the d minimal non-zero eigenvalue solutions of Lf = λdf under the constraint y Dy = Id.

The Matlab Toolbox for Dimensionality Reduction, developed by Laurens van der Maaten, contains Matlab implementations of 34 techniques for dimensionality reduction and metric learning, including Laplacian Eigenmaps (LE). It can be found at and downloaded form his website: http://lvdmaaten.github.io/drtoolbox/ Please note that LE solves a generalized eigenvalue problem, and does not simply take eigenvectors of Graph Laplacian.