Graph Metrics and Dimension Reduction

Size: px
Start display at page:

Download "Graph Metrics and Dimension Reduction"

Transcription

1 Graph Metrics and Dimension Reduction Minh Tang 1 Michael Trosset 2 1 Applied Mathematics and Statistics The Johns Hopkins University 2 Department of Statistics Indiana University, Bloomington November 11, 2010 Tang & Trosset Graph Metrics and Dimension Reduction November 11,

2 Outline 1 Problem Description 2 Preliminaries 3 Distances on Undirected Graphs 4 From Distances to Embeddings 5 Distances on Directed Graphs Tang & Trosset Graph Metrics and Dimension Reduction November 11,

3 Benefits of Dimension Reduction High-dimensional data, i.e., a great many measurements taken on each member of a set of objects are now ubiquitous. Bellman s curse of dimensionality [Bellman(1957)] refers to the problems caused by the exponential increase in volume of a mathematical space as additional dimensions are added. A list of problems might include, Slow convergence of statistical estimators, e.g., density estimators. Overfitting of models to noise. Difficulties in performing exploratory data analysis. Nearest-neighbour searches are inefficient. Dimension reduction is the process of replacing a multivariate data set with a data set of lower dimension. Classical approaches to dimension reduction involve the use of principal component analysis (PCA) [Pearson(1901), Hotelling(1933)] or classical multidimensional scaling [Torgersen(1952), Gower(1966)]. Tang & Trosset Graph Metrics and Dimension Reduction November 11,

4 Problem Description We consider the problem of constructing a low-dimensional Euclidean representation of data described by pairwise similarities. The low-dimensional representation can serve as the basis for other exploitation tasks, e.g., visualization, clustering, or classification. Our basic strategy is: 1 Transform the similarities into some notion of dissimilarities; 2 Embed the derived dissimilarities. Our concerns are closely related to the concerns of manifold learning. Various manifold learning techniques can be interpreted as transformations from similarities to dissimilarities. Tang & Trosset Graph Metrics and Dimension Reduction November 11,

5 A Motivating Example: MNIST Dataset 1200 images of digits (4 or 5) from MNIST [LeCun et al.(1998)]. Each image is viewed as a point in R How should we measure the proximity of a pair of images? δ ij = x i x j γ ij = exp( x i x j 2 σ 2 ) PC Digit 4 Digit Digit 4 Digit PC1 Tang & Trosset Graph Metrics and Dimension Reduction November 11,

6 From Similarities to Graphs Isomap [Tenenbaum et al.(2000)] employed two commonly used approach to graph construction. In both approaches, vertices correspond to feature vectors. 1 ɛ-neighborhood approach. v i v j iff x i x j < ɛ. 2 K-NN approach. Connect v i v j iff x j is a K nearest neighbor of x i. This graph is directed, but in practice it is often symmetrized. Tang & Trosset Graph Metrics and Dimension Reduction November 11,

7 Outline 1 Problem Description 2 Preliminaries 3 Distances on Undirected Graphs 4 From Distances to Embeddings 5 Distances on Directed Graphs Tang & Trosset Graph Metrics and Dimension Reduction November 11,

8 Distances on Graphs Given a n n similarity matrix Γ = (γ ij ): 1 Transform the similarities to dissimilarities. (Isomap starts with distances in an ambient input space.) (a) Construct a weighted graph G = (V, E, ω) with n vertices and edge weights ω ij = γ ij. (b) Choose a suitable measure of dissimilarity (typically a distance) on G. Let denote the pairwise dissimilarities. 2 Embed. Several popular approaches that transform similarity to distance rely on the concept of a random walk. Tang & Trosset Graph Metrics and Dimension Reduction November 11,

9 Random Walks on Graphs Let G = (V, E, ω) be an undirected graph. We define the transition matrix P = (p uv ) of a Markov chain with state space V as p uv = { ω({u,v}) deg(u) if u v 0 otherwise (1) Suppose that G is connected. Then the stationary distribution π of P exists and is unique. Furthermore, if G is connected and not bipartite, then lim k Pk = 1π T := Q (2) Tang & Trosset Graph Metrics and Dimension Reduction November 11,

10 Distance Geometry Definition (Euclidean Distance Matrix) Let = (δ ij ) be a n n dissimilarity matrix. is a Type-2 Euclidean distance matrix (EDM-2) if there exists n points x 1, x 2,..., x n R p for some p such that δ ij = x i x j 2. Let A and B be n n matrices. Define two linear transforms τ and κ by τ(a) = 1 2 (I 11T /n)a(i 11 T /n). κ(b) = B dg 11 T B B T + 11 T B dg There is an equivalence between EDM-2 and p.s.d matrices. Theorem ([Schoenberg(1935), Young and Householder(1938)]) is EDM-2 iff τ( ) is p.s.d. As a corollary, if B is p.s.d, then = κ(b) is EDM-2. Tang & Trosset Graph Metrics and Dimension Reduction November 11,

11 Outline 1 Problem Description 2 Preliminaries 3 Distances on Undirected Graphs 4 From Distances to Embeddings 5 Distances on Directed Graphs Tang & Trosset Graph Metrics and Dimension Reduction November 11,

12 Expected Commute Time Following [Kemeny and Snell(1960)], let Π = diag(π) and Z = (I P + Q) 1. The expected first passage times are given by and the expected commute times are M = (11 T Z dg Z)Π 1 ect = M + M T = κ(zπ 1 ) It turns out that ZΠ 1 0; hence ect is EDM-2. This result is well known; see [Trosset and Tang(2010)] for an elementary proof. Tang & Trosset Graph Metrics and Dimension Reduction November 11,

13 Diffusion Distances Let e i and e j denote point masses at vertices v i and v j. After r time steps, under the random walk model with transition matrix P, these distributions have diffused to e T i P r and e T j P r. The diffusion distance [Coifman and Lafon(2006)] at time r between v i and v j is ρ r (v i, v j ) = e T i P r e T j P r 1/π where the inner product, 1/π is defined as u, v 1/π = k u(k)v(k)/π(k) It turns out that ρ 2 r = κ(p 2r Π 1 ) and P 2r Π 1 0; hence, ρ 2 r is EDM-2. Tang & Trosset Graph Metrics and Dimension Reduction November 11,

14 Some Remarks on ECT and Diffusion Distances 1 ect can be written as ( ect = κ(zπ 1 ) = κ (P Q) t Π 1). Even though ( κ((p Q) t Π 1 ) = κ(p t Π 1 ) for any t, ) ect κ t=0 Pt Π 1 because t=0 Pt Π 1 does not necessarily converge. 2 ρ 2 t can be written as t=0 ρ 2 t = κ(p 2t Π 1 ) = κ ( (P Q) 2t Π 1) 3 Diffusion distance between v i and v j at time t takes into account only paths of length 2t while expected commute time takes into account paths of all lengths. In fact, expected commute time with respect to P 2 is the sum of diffusion distances for t = 0, 1,... with respect to P. Tang & Trosset Graph Metrics and Dimension Reduction November 11,

15 General Framework for Euclidean Distances on Graphs We introduce a general family of Euclidean distances constructed from random walks on graphs. Let f be a real-valued function with a series expansion f (x) = a 0 + a 1 x + a 2 x 2 + and radius of convergence R 1. For a square matrix X, define f (X) by Theorem ([Tang(2010)]) f (X) = a 0 I + a 1 X + a 2 X 2 + Assume that P is irreducible and aperiodic. If f (x) 0 for x ( 1, 1), then ( = κ(f (P Q)Π 1 ) = κ (a 0 I+a 1 (P Q)+a 2 (P Q) 2 + )Π 1) (3) is well-defined and EDM-2. Tang & Trosset Graph Metrics and Dimension Reduction November 11,

16 Euclidean Distances on Graphs: Some Examples = κ(f (P Q)Π 1 ) = κ((a 0 I + (a 1 P Q) + a 2 (P Q) 2 + )Π 1 ) f (x) Comments 1 trivial notion of distance x 2r diffusion distance at time r 1/(1 x) expected commute time 1/(1 x) k for k 2 longer paths have higher weights log (1 x 2 ) longer paths have lower weights exp(x) heavily weights paths of short lengths Tang & Trosset Graph Metrics and Dimension Reduction November 11,

17 Outline 1 Problem Description 2 Preliminaries 3 Distances on Undirected Graphs 4 From Distances to Embeddings 5 Distances on Directed Graphs Tang & Trosset Graph Metrics and Dimension Reduction November 11,

18 Embedding = κ(f (P Q)Π 1 ) in R d : Method 1 Embed by CMDS. 1 Compute B = τ( ) = (I 11 T /n)f (P Q)Π 1 (I 11 T /n) 2 Let λ 1 λ 2 λ n denote the eigenvalues of B and let v 1, v 2,..., v n denote the corresponding set of orthonormal eigenvectors. Then [ X d = λ1 v 1 λ 2 v 2 ] λ d v d produces a configuration of points in R d whose interpoint distances approximate. Tang & Trosset Graph Metrics and Dimension Reduction November 11,

19 Example: Embedding ect by CMDS Let G be an undirected graph and L be the combinatorial Laplacian of G. It turns out that L, the Moore-Penrose pseudoinverse of L, is related to ZΠ 1 by L = c(i 11 T /n)zπ 1 (I 11 T /n) (4) where c is a constant. Therefore, ect = cκ(l ). Furthermore, τ( ect ) = cl. The d-dimensional embedding of ect is thus given by [ X d = c λ1 ν 1 λ 2 ν 2... ] λ d ν d (5) where λ 1 λ 2... and ν 1, ν 2,... are the eigenvalues and corresponding eigenvectors of L. The embedding in Eq. (5) is called a combinatorial Laplacian eigenmap. Tang & Trosset Graph Metrics and Dimension Reduction November 11,

20 Embedding = κ(f (P Q)Π 1 ) in R d : Method 2 Embed by the eigenvalues and eigenvectors of P. 1 Let µ 1, µ 2,..., µ n 1 < 1 = µ n denote the eigenvalues of P, sorted so that f (µ i ) f (µ i+1 ). Let u 1, u 2,..., u n denote the corresponding set of eigenvectors, orthonormal with respect to the inner product u, v π = k u(k)v(k)π(k) 2 Then [ X d = f (µ1 )u 1 f (µ 2 )u 2 ] f (µ d )u d produces a configuration of points in R d whose interpoint distances approximate. By rescaling individual coordinates, the embedding from any one f can be transformed to the embedding for any other f. Tang & Trosset Graph Metrics and Dimension Reduction November 11,

21 Comparing the Embeddings Method 1: Classical MDS 1 [ The embedding X = λ1 v 1 ] λ n 1 v n 1 recovers completely. 2 The embedding dimension of is almost surely n 1. 3 The best (least squares) d-dim[ representation of X is λ1 X d = v 1 ] λ d v d. Method 2: Eigensystem of P 1 The [ embedding X = f (µ1 )u 1 ] f (µ n 1 )u n 1 recovers completely. 2 The embedding dimension of is almost surely n 1. 3 The best d-dim representation of X is (usually) [ not f X d = (µ1 )u 1 ] f (µ d )u d Tang & Trosset Graph Metrics and Dimension Reduction November 11,

22 Normalized Laplacian Eigenmaps 1 Construct a graph G = (V, E, ω) with V = X. 2 Compute the eigenvalues λ and eigenvectors f of the generalized eigenvalue problem Lf = λdf (6) 3 Let λ 0 λ 1 λ n 1 be the eigenvalues of Eq. (6) and f 0, f 1,..., f n 1 be the corresponding eigenvectors. ( ) 4 Embed into R d by x i 1 1 λ1 f 1 (i), 1 λ2, f 2 (i),..., λd f d (i). Under our framework, steps 2 to 4 are equivalent to embedding ect using the eigenvalues and eigenvectors of P (Method 2). This is not equivalent to embedding ect using the eigensystem of L (Method 1). Tang & Trosset Graph Metrics and Dimension Reduction November 11,

23 Diffusion Maps [Coifman and Lafon(2006)] 1 Construct a graph G = (V, E, ω) with V = X. 2 Generate the transition matrix P of G. 3 Let λ 0 λ 1 λ n 1 be the eigenvalues of P and f 0, f 1,..., f n 1 be the corresponding eigenvectors. 4 Embed into R d by x i ( λ t 1 f 1(i), λ t 2 f 2(i),..., λ t mf d (i)). Recall that ρ 2 t = κ((p Q) 2t Π 1 ) is the matrix of diffusion distances. Under our framework, steps 2 to 4 are equivalent to embedding ρ 2 t using the eigenvalues and eigenvectors of P (Method 2). Normalized Laplacian eigenmaps and diffusion maps are thus coordinate rescalings of one another. Tang & Trosset Graph Metrics and Dimension Reduction November 11,

24 Coordinates rescaling of embeddings Digit 4 Digit Digit 4 Digit (a) Normalized Laplacian eigenmap (b) Diffusion map at time t = 10 Tang & Trosset Graph Metrics and Dimension Reduction November 11,

25 Paths of Even Length & Diffusion Distances Y Y X X (a)original Data (b)diffusion map at time t = 5 Tang & Trosset Graph Metrics and Dimension Reduction November 11,

26 Outline 1 Problem Description 2 Preliminaries 3 Distances on Undirected Graphs 4 From Distances to Embeddings 5 Distances on Directed Graphs Tang & Trosset Graph Metrics and Dimension Reduction November 11,

27 Expected Commute Time for Directed Graphs Analogous to the case of expected commute time on undirected graphs, let Π = diag(π) and Z = (I P + Q) 1. The expected first passage times for directed graphs are also given by M = (11 T diag(z) Z)Π 1 and the expected commute times are ect = M + M T = κ(zπ 1 ) = κ(h(zπ 1 )) where H(A) = 1 2 (A + AT ) is the Hermitian part of A. It turns out that H(ZΠ 1 ) 0; hence, ect for directed graphs is EDM-2. Tang & Trosset Graph Metrics and Dimension Reduction November 11,

28 Diffusion Distances for Directed Graphs Let e i and e j denote point masses at vertices v i and v j. Analogous to the case of diffusion distance on undirected graphs, after r time steps, under the random walk model with transition matrix P, these distributions had diffused to e T i P r and e T j P r. The diffusion distance on directed graph at time r between v i and v j is ρ r (v i, v j ) = e T i P r e T j P r 1/π where the inner product, 1/π is defined as u, v 1/π = k u(k)v(k)/π(k) ρ 2 r = κ(p r Π 1 (P r ) T ); hence ρ 2 r for directed graphs is EDM-2. Tang & Trosset Graph Metrics and Dimension Reduction November 11,

29 Distances on Directed Graphs: Some Comments 1 It is harder to derive a framework for Euclidean distances on directed graphs. If G is a directed graph and P is the transition matrix on G, then there exists a k 2 such that = κ((i P + Q) k Π 1 ) is not EDM-2. 2 Expected commute time under the random walk model with transition matrix P 2 is no longer the sum of the squared diffusion distances through all time scales. We can interpret this as saying that the symmetrization performed in constructing expected commute time is incompatible with the symmetrization performed in constructing diffusion distances. Tang & Trosset Graph Metrics and Dimension Reduction November 11,

30 From Distances to Embeddings Let G be a directed graph with transition matrix P. Suppose that is a dissimilarity/distance matrix constructed by considering random walks on G. Consider the problem of embedding into Euclidean space. Embedding using CMDS is straightforward. Because the eigenvalues and eigenvectors of P are possibly complex-valued, embedding using the eigenvalues and eigenvectors of P might not be possible. The notions of the combinatorial Laplacian and normalized Laplacian need to be extended to directed graphs. This is usually accomplished by symmetrization. However, the decision of how to symmetrize is not obvious. For example, the following definition of the combinatorial Laplacian for directed graphs ([Chung(2005)]) L = Π ΠP + PT Π 2 does not generate expected commute time for directed graphs. (7) Tang & Trosset Graph Metrics and Dimension Reduction November 11,

31 Embedding Distances for Directed Graphs We construct the directed K-NN graph with K = 100 for our motivating example. We compute the matrix of expected commute times and the matrix of diffusion distances at time t = 30. We then embed the resulting matrices using CMDS Digit 4 Digit Digit 4 Digit (a) ect (b) ρ 2 t Tang & Trosset Graph Metrics and Dimension Reduction November 11,

32 An Auxiliary Result A small class of Euclidean distances on directed graphs can be established by considering relaxed random walks. Let G be a directed graph with transition matrix P. A relaxed random walk on G is a random walk with P α = αi + (1 α)p for some α (0, 1). Let f be a real-valued function with series expansion and radius of convergence R 1. f (x) = a 0 + a 1 x + a 2 x 2 + If a 0 > 0 and a i 0 for all i 1, then for any irreducible and aperiodic P, there exists a α [0, 1) such that is well-defined and EDM-2. = κ(f (P α Q)Π 1 ) Tang & Trosset Graph Metrics and Dimension Reduction November 11,

33 PC Digit 4 Digit PC1 Tang & Trosset Graph Metrics and Dimension Reduction November 11,

34 Digit 4 Digit Tang & Trosset Graph Metrics and Dimension Reduction November 11,

35 Digit 4 Digit Tang & Trosset Graph Metrics and Dimension Reduction November 11,

36 Digit 4 Digit Tang & Trosset Graph Metrics and Dimension Reduction November 11,

37 M. Belkin and P. Niyogi. Laplacian eigenmaps for dimensionality reduction and data representation. Neural Computation, 15: , R. E. Bellman. Dynamic programming. Princeton University Press, F. Chung. Laplacians and the Cheeger inequality for directed graphs. Annals of Combinatorics, 9:1 19, R. Coifman and S. Lafon. Diffusion maps. Applied and Computational Harmonic Analysis, 21:5 30, J. C. Gower. Some distance properties of latent root and vector methods used in multivariate analysis. Biometrika, 53: , Tang & Trosset Graph Metrics and Dimension Reduction November 11,

38 H. Hotelling. Analysis of a complex of statistical variables into principle components. Journal of Educational Psychology, 24: , J. G. Kemeny and J. L. Snell. Finite Markov Chains. Springer, Y. LeCun et al. Gradient-based learning applied to document recognition. In Proceedings of the IEEE, volume 86, pages , K. Pearson. On lines and planes of closest fit to a system of points in space. Philosophical Magazine, 2: , M. Saerens et al. The principal components analysis of a graph and its relationships to spectral clustering. Tang & Trosset Graph Metrics and Dimension Reduction November 11,

39 Proceedings of the fifteenth European conference on machine learning, I. J. Schoenberg. Remarks to maurice frechet s article sur la definition axiomatique d une classe d espace distances vectoriellement applicable sur l espace de hilbert. The Annals of Mathematics, 36(3):pp , J. Shi and J. Malik. Normalized cuts and image segmentation. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages , A. Smola and R. Kondor. Kernels and regularization on graphs. In Conference on Learning Theory, M. Tang. Graph metrics and dimensionality reduction. PhD thesis, Indiana University, Bloomington, Tang & Trosset Graph Metrics and Dimension Reduction November 11,

40 J. B. Tenenbaum et al. A global geometric framework for nonlinear dimensionality reduction. Science, 290: , December W. S. Torgersen. Multidimensional scaling: I. Theory and method. Psychometrika, 17: , M. W. Trosset and C. E. Priebe. The out-of-sample problem for classical multidimensional scaling. Computational Statistics and Data Analysis, 52: , M. W. Trosset and M. Tang. On combinatorial Laplacian eigenmaps. Technical report, Indiana University, Bloomington, G. Young and A. S. Householder. Discussion of a set of points in terms of their mutual distances. Psychometrika, 3:19 22, Tang & Trosset Graph Metrics and Dimension Reduction November 11,

Data-dependent representations: Laplacian Eigenmaps

Data-dependent representations: Laplacian Eigenmaps Data-dependent representations: Laplacian Eigenmaps November 4, 2015 Data Organization and Manifold Learning There are many techniques for Data Organization and Manifold Learning, e.g., Principal Component

More information

Data Analysis and Manifold Learning Lecture 7: Spectral Clustering

Data Analysis and Manifold Learning Lecture 7: Spectral Clustering Data Analysis and Manifold Learning Lecture 7: Spectral Clustering Radu Horaud INRIA Grenoble Rhone-Alpes, France Radu.Horaud@inrialpes.fr http://perception.inrialpes.fr/ Outline of Lecture 7 What is spectral

More information

Unsupervised dimensionality reduction

Unsupervised dimensionality reduction Unsupervised dimensionality reduction Guillaume Obozinski Ecole des Ponts - ParisTech SOCN course 2014 Guillaume Obozinski Unsupervised dimensionality reduction 1/30 Outline 1 PCA 2 Kernel PCA 3 Multidimensional

More information

Laplacian Eigenmaps for Dimensionality Reduction and Data Representation

Laplacian Eigenmaps for Dimensionality Reduction and Data Representation Laplacian Eigenmaps for Dimensionality Reduction and Data Representation Neural Computation, June 2003; 15 (6):1373-1396 Presentation for CSE291 sp07 M. Belkin 1 P. Niyogi 2 1 University of Chicago, Department

More information

The Out-of-Sample Problem for Classical Multidimensional Scaling

The Out-of-Sample Problem for Classical Multidimensional Scaling The Out-of-Sample Problem for Classical Multidimensional Scaling Michael W. Trosset a and Carey E. Priebe b December 21, 2006; revised February 10, 2008 Technical Report 06-04 Department of Statistics

More information

Non-linear Dimensionality Reduction

Non-linear Dimensionality Reduction Non-linear Dimensionality Reduction CE-725: Statistical Pattern Recognition Sharif University of Technology Spring 2013 Soleymani Outline Introduction Laplacian Eigenmaps Locally Linear Embedding (LLE)

More information

Nonlinear Methods. Data often lies on or near a nonlinear low-dimensional curve aka manifold.

Nonlinear Methods. Data often lies on or near a nonlinear low-dimensional curve aka manifold. Nonlinear Methods Data often lies on or near a nonlinear low-dimensional curve aka manifold. 27 Laplacian Eigenmaps Linear methods Lower-dimensional linear projection that preserves distances between all

More information

Apprentissage non supervisée

Apprentissage non supervisée Apprentissage non supervisée Cours 3 Higher dimensions Jairo Cugliari Master ECD 2015-2016 From low to high dimension Density estimation Histograms and KDE Calibration can be done automacally But! Let

More information

Nonlinear Dimensionality Reduction. Jose A. Costa

Nonlinear Dimensionality Reduction. Jose A. Costa Nonlinear Dimensionality Reduction Jose A. Costa Mathematics of Information Seminar, Dec. Motivation Many useful of signals such as: Image databases; Gene expression microarrays; Internet traffic time

More information

Data dependent operators for the spatial-spectral fusion problem

Data dependent operators for the spatial-spectral fusion problem Data dependent operators for the spatial-spectral fusion problem Wien, December 3, 2012 Joint work with: University of Maryland: J. J. Benedetto, J. A. Dobrosotskaya, T. Doster, K. W. Duke, M. Ehler, A.

More information

Learning Eigenfunctions: Links with Spectral Clustering and Kernel PCA

Learning Eigenfunctions: Links with Spectral Clustering and Kernel PCA Learning Eigenfunctions: Links with Spectral Clustering and Kernel PCA Yoshua Bengio Pascal Vincent Jean-François Paiement University of Montreal April 2, Snowbird Learning 2003 Learning Modal Structures

More information

Unsupervised Learning Techniques Class 07, 1 March 2006 Andrea Caponnetto

Unsupervised Learning Techniques Class 07, 1 March 2006 Andrea Caponnetto Unsupervised Learning Techniques 9.520 Class 07, 1 March 2006 Andrea Caponnetto About this class Goal To introduce some methods for unsupervised learning: Gaussian Mixtures, K-Means, ISOMAP, HLLE, Laplacian

More information

Lecture: Some Practical Considerations (3 of 4)

Lecture: Some Practical Considerations (3 of 4) Stat260/CS294: Spectral Graph Methods Lecture 14-03/10/2015 Lecture: Some Practical Considerations (3 of 4) Lecturer: Michael Mahoney Scribe: Michael Mahoney Warning: these notes are still very rough.

More information

Focus was on solving matrix inversion problems Now we look at other properties of matrices Useful when A represents a transformations.

Focus was on solving matrix inversion problems Now we look at other properties of matrices Useful when A represents a transformations. Previously Focus was on solving matrix inversion problems Now we look at other properties of matrices Useful when A represents a transformations y = Ax Or A simply represents data Notion of eigenvectors,

More information

Intrinsic Structure Study on Whale Vocalizations

Intrinsic Structure Study on Whale Vocalizations 1 2015 DCLDE Conference Intrinsic Structure Study on Whale Vocalizations Yin Xian 1, Xiaobai Sun 2, Yuan Zhang 3, Wenjing Liao 3 Doug Nowacek 1,4, Loren Nolte 1, Robert Calderbank 1,2,3 1 Department of

More information

Data Analysis and Manifold Learning Lecture 3: Graphs, Graph Matrices, and Graph Embeddings

Data Analysis and Manifold Learning Lecture 3: Graphs, Graph Matrices, and Graph Embeddings Data Analysis and Manifold Learning Lecture 3: Graphs, Graph Matrices, and Graph Embeddings Radu Horaud INRIA Grenoble Rhone-Alpes, France Radu.Horaud@inrialpes.fr http://perception.inrialpes.fr/ Outline

More information

Beyond Scalar Affinities for Network Analysis or Vector Diffusion Maps and the Connection Laplacian

Beyond Scalar Affinities for Network Analysis or Vector Diffusion Maps and the Connection Laplacian Beyond Scalar Affinities for Network Analysis or Vector Diffusion Maps and the Connection Laplacian Amit Singer Princeton University Department of Mathematics and Program in Applied and Computational Mathematics

More information

Laplacian Eigenmaps for Dimensionality Reduction and Data Representation

Laplacian Eigenmaps for Dimensionality Reduction and Data Representation Introduction and Data Representation Mikhail Belkin & Partha Niyogi Department of Electrical Engieering University of Minnesota Mar 21, 2017 1/22 Outline Introduction 1 Introduction 2 3 4 Connections to

More information

Nonlinear Dimensionality Reduction

Nonlinear Dimensionality Reduction Outline Hong Chang Institute of Computing Technology, Chinese Academy of Sciences Machine Learning Methods (Fall 2012) Outline Outline I 1 Kernel PCA 2 Isomap 3 Locally Linear Embedding 4 Laplacian Eigenmap

More information

March 13, Paper: R.R. Coifman, S. Lafon, Diffusion maps ([Coifman06]) Seminar: Learning with Graphs, Prof. Hein, Saarland University

March 13, Paper: R.R. Coifman, S. Lafon, Diffusion maps ([Coifman06]) Seminar: Learning with Graphs, Prof. Hein, Saarland University Kernels March 13, 2008 Paper: R.R. Coifman, S. Lafon, maps ([Coifman06]) Seminar: Learning with Graphs, Prof. Hein, Saarland University Kernels Figure: Example Application from [LafonWWW] meaningful geometric

More information

How to learn from very few examples?

How to learn from very few examples? How to learn from very few examples? Dengyong Zhou Department of Empirical Inference Max Planck Institute for Biological Cybernetics Spemannstr. 38, 72076 Tuebingen, Germany Outline Introduction Part A

More information

Diffusion Geometries, Diffusion Wavelets and Harmonic Analysis of large data sets.

Diffusion Geometries, Diffusion Wavelets and Harmonic Analysis of large data sets. Diffusion Geometries, Diffusion Wavelets and Harmonic Analysis of large data sets. R.R. Coifman, S. Lafon, MM Mathematics Department Program of Applied Mathematics. Yale University Motivations The main

More information

Robust Laplacian Eigenmaps Using Global Information

Robust Laplacian Eigenmaps Using Global Information Manifold Learning and its Applications: Papers from the AAAI Fall Symposium (FS-9-) Robust Laplacian Eigenmaps Using Global Information Shounak Roychowdhury ECE University of Texas at Austin, Austin, TX

More information

ISSN: (Online) Volume 3, Issue 5, May 2015 International Journal of Advance Research in Computer Science and Management Studies

ISSN: (Online) Volume 3, Issue 5, May 2015 International Journal of Advance Research in Computer Science and Management Studies ISSN: 2321-7782 (Online) Volume 3, Issue 5, May 2015 International Journal of Advance Research in Computer Science and Management Studies Research Article / Survey Paper / Case Study Available online at:

More information

Advances in Manifold Learning Presented by: Naku Nak l Verm r a June 10, 2008

Advances in Manifold Learning Presented by: Naku Nak l Verm r a June 10, 2008 Advances in Manifold Learning Presented by: Nakul Verma June 10, 008 Outline Motivation Manifolds Manifold Learning Random projection of manifolds for dimension reduction Introduction to random projections

More information

Dimensionality Reduction AShortTutorial

Dimensionality Reduction AShortTutorial Dimensionality Reduction AShortTutorial Ali Ghodsi Department of Statistics and Actuarial Science University of Waterloo Waterloo, Ontario, Canada, 2006 c Ali Ghodsi, 2006 Contents 1 An Introduction to

More information

Face Recognition Using Laplacianfaces He et al. (IEEE Trans PAMI, 2005) presented by Hassan A. Kingravi

Face Recognition Using Laplacianfaces He et al. (IEEE Trans PAMI, 2005) presented by Hassan A. Kingravi Face Recognition Using Laplacianfaces He et al. (IEEE Trans PAMI, 2005) presented by Hassan A. Kingravi Overview Introduction Linear Methods for Dimensionality Reduction Nonlinear Methods and Manifold

More information

Spherical Euclidean Distance Embedding of a Graph

Spherical Euclidean Distance Embedding of a Graph Spherical Euclidean Distance Embedding of a Graph Hou-Duo Qi University of Southampton Presented at Isaac Newton Institute Polynomial Optimization August 9, 2013 Spherical Embedding Problem The Problem:

More information

Nonlinear Dimensionality Reduction

Nonlinear Dimensionality Reduction Nonlinear Dimensionality Reduction Piyush Rai CS5350/6350: Machine Learning October 25, 2011 Recap: Linear Dimensionality Reduction Linear Dimensionality Reduction: Based on a linear projection of the

More information

EECS 275 Matrix Computation

EECS 275 Matrix Computation EECS 275 Matrix Computation Ming-Hsuan Yang Electrical Engineering and Computer Science University of California at Merced Merced, CA 95344 http://faculty.ucmerced.edu/mhyang Lecture 23 1 / 27 Overview

More information

Statistical Machine Learning

Statistical Machine Learning Statistical Machine Learning Christoph Lampert Spring Semester 2015/2016 // Lecture 12 1 / 36 Unsupervised Learning Dimensionality Reduction 2 / 36 Dimensionality Reduction Given: data X = {x 1,..., x

More information

THE HIDDEN CONVEXITY OF SPECTRAL CLUSTERING

THE HIDDEN CONVEXITY OF SPECTRAL CLUSTERING THE HIDDEN CONVEXITY OF SPECTRAL CLUSTERING Luis Rademacher, Ohio State University, Computer Science and Engineering. Joint work with Mikhail Belkin and James Voss This talk A new approach to multi-way

More information

CSE 291. Assignment Spectral clustering versus k-means. Out: Wed May 23 Due: Wed Jun 13

CSE 291. Assignment Spectral clustering versus k-means. Out: Wed May 23 Due: Wed Jun 13 CSE 291. Assignment 3 Out: Wed May 23 Due: Wed Jun 13 3.1 Spectral clustering versus k-means Download the rings data set for this problem from the course web site. The data is stored in MATLAB format as

More information

Learning from Labeled and Unlabeled Data: Semi-supervised Learning and Ranking p. 1/31

Learning from Labeled and Unlabeled Data: Semi-supervised Learning and Ranking p. 1/31 Learning from Labeled and Unlabeled Data: Semi-supervised Learning and Ranking Dengyong Zhou zhou@tuebingen.mpg.de Dept. Schölkopf, Max Planck Institute for Biological Cybernetics, Germany Learning from

More information

Graphs, Geometry and Semi-supervised Learning

Graphs, Geometry and Semi-supervised Learning Graphs, Geometry and Semi-supervised Learning Mikhail Belkin The Ohio State University, Dept of Computer Science and Engineering and Dept of Statistics Collaborators: Partha Niyogi, Vikas Sindhwani In

More information

Graph-Laplacian PCA: Closed-form Solution and Robustness

Graph-Laplacian PCA: Closed-form Solution and Robustness 2013 IEEE Conference on Computer Vision and Pattern Recognition Graph-Laplacian PCA: Closed-form Solution and Robustness Bo Jiang a, Chris Ding b,a, Bin Luo a, Jin Tang a a School of Computer Science and

More information

Locality Preserving Projections

Locality Preserving Projections Locality Preserving Projections Xiaofei He Department of Computer Science The University of Chicago Chicago, IL 60637 xiaofei@cs.uchicago.edu Partha Niyogi Department of Computer Science The University

More information

Learning gradients: prescriptive models

Learning gradients: prescriptive models Department of Statistical Science Institute for Genome Sciences & Policy Department of Computer Science Duke University May 11, 2007 Relevant papers Learning Coordinate Covariances via Gradients. Sayan

More information

Linearly-solvable Markov decision problems

Linearly-solvable Markov decision problems Advances in Neural Information Processing Systems 2 Linearly-solvable Markov decision problems Emanuel Todorov Department of Cognitive Science University of California San Diego todorov@cogsci.ucsd.edu

More information

Markov Chains, Random Walks on Graphs, and the Laplacian

Markov Chains, Random Walks on Graphs, and the Laplacian Markov Chains, Random Walks on Graphs, and the Laplacian CMPSCI 791BB: Advanced ML Sridhar Mahadevan Random Walks! There is significant interest in the problem of random walks! Markov chain analysis! Computer

More information

Spectral Techniques for Clustering

Spectral Techniques for Clustering Nicola Rebagliati 1/54 Spectral Techniques for Clustering Nicola Rebagliati 29 April, 2010 Nicola Rebagliati 2/54 Thesis Outline 1 2 Data Representation for Clustering Setting Data Representation and Methods

More information

Spectral Clustering on Handwritten Digits Database

Spectral Clustering on Handwritten Digits Database University of Maryland-College Park Advance Scientific Computing I,II Spectral Clustering on Handwritten Digits Database Author: Danielle Middlebrooks Dmiddle1@math.umd.edu Second year AMSC Student Advisor:

More information

Dimensionality Reduc1on

Dimensionality Reduc1on Dimensionality Reduc1on contd Aarti Singh Machine Learning 10-601 Nov 10, 2011 Slides Courtesy: Tom Mitchell, Eric Xing, Lawrence Saul 1 Principal Component Analysis (PCA) Principal Components are the

More information

Lecture: Modeling graphs with electrical networks

Lecture: Modeling graphs with electrical networks Stat260/CS294: Spectral Graph Methods Lecture 16-03/17/2015 Lecture: Modeling graphs with electrical networks Lecturer: Michael Mahoney Scribe: Michael Mahoney Warning: these notes are still very rough.

More information

Learning on Graphs and Manifolds. CMPSCI 689 Sridhar Mahadevan U.Mass Amherst

Learning on Graphs and Manifolds. CMPSCI 689 Sridhar Mahadevan U.Mass Amherst Learning on Graphs and Manifolds CMPSCI 689 Sridhar Mahadevan U.Mass Amherst Outline Manifold learning is a relatively new area of machine learning (2000-now). Main idea Model the underlying geometry of

More information

Dimension Reduction and Low-dimensional Embedding

Dimension Reduction and Low-dimensional Embedding Dimension Reduction and Low-dimensional Embedding Ying Wu Electrical Engineering and Computer Science Northwestern University Evanston, IL 60208 http://www.eecs.northwestern.edu/~yingwu 1/26 Dimension

More information

SPECTRAL CLUSTERING AND KERNEL PRINCIPAL COMPONENT ANALYSIS ARE PURSUING GOOD PROJECTIONS

SPECTRAL CLUSTERING AND KERNEL PRINCIPAL COMPONENT ANALYSIS ARE PURSUING GOOD PROJECTIONS SPECTRAL CLUSTERING AND KERNEL PRINCIPAL COMPONENT ANALYSIS ARE PURSUING GOOD PROJECTIONS VIKAS CHANDRAKANT RAYKAR DECEMBER 5, 24 Abstract. We interpret spectral clustering algorithms in the light of unsupervised

More information

Contribution from: Springer Verlag Berlin Heidelberg 2005 ISBN

Contribution from: Springer Verlag Berlin Heidelberg 2005 ISBN Contribution from: Mathematical Physics Studies Vol. 7 Perspectives in Analysis Essays in Honor of Lennart Carleson s 75th Birthday Michael Benedicks, Peter W. Jones, Stanislav Smirnov (Eds.) Springer

More information

Spectral Clustering. Zitao Liu

Spectral Clustering. Zitao Liu Spectral Clustering Zitao Liu Agenda Brief Clustering Review Similarity Graph Graph Laplacian Spectral Clustering Algorithm Graph Cut Point of View Random Walk Point of View Perturbation Theory Point of

More information

Bi-stochastic kernels via asymmetric affinity functions

Bi-stochastic kernels via asymmetric affinity functions Bi-stochastic kernels via asymmetric affinity functions Ronald R. Coifman, Matthew J. Hirn Yale University Department of Mathematics P.O. Box 208283 New Haven, Connecticut 06520-8283 USA ariv:1209.0237v4

More information

Clustering in kernel embedding spaces and organization of documents

Clustering in kernel embedding spaces and organization of documents Clustering in kernel embedding spaces and organization of documents Stéphane Lafon Collaborators: Raphy Coifman (Yale), Yosi Keller (Yale), Ioannis G. Kevrekidis (Princeton), Ann B. Lee (CMU), Boaz Nadler

More information

Manifold Learning and it s application

Manifold Learning and it s application Manifold Learning and it s application Nandan Dubey SE367 Outline 1 Introduction Manifold Examples image as vector Importance Dimension Reduction Techniques 2 Linear Methods PCA Example MDS Perception

More information

Manifold Learning: Theory and Applications to HRI

Manifold Learning: Theory and Applications to HRI Manifold Learning: Theory and Applications to HRI Seungjin Choi Department of Computer Science Pohang University of Science and Technology, Korea seungjin@postech.ac.kr August 19, 2008 1 / 46 Greek Philosopher

More information

High-Dimensional Pattern Recognition using Low-Dimensional Embedding and Earth Mover s Distance

High-Dimensional Pattern Recognition using Low-Dimensional Embedding and Earth Mover s Distance High-Dimensional Pattern Recognition using Low-Dimensional Embedding and Earth Mover s Distance Linh Lieu,a, Naoki Saito a a Department of Mathematics, University of California, Davis, CA 95616, USA Abstract

More information

Statistical and Computational Analysis of Locality Preserving Projection

Statistical and Computational Analysis of Locality Preserving Projection Statistical and Computational Analysis of Locality Preserving Projection Xiaofei He xiaofei@cs.uchicago.edu Department of Computer Science, University of Chicago, 00 East 58th Street, Chicago, IL 60637

More information

Nonlinear Manifold Learning Summary

Nonlinear Manifold Learning Summary Nonlinear Manifold Learning 6.454 Summary Alexander Ihler ihler@mit.edu October 6, 2003 Abstract Manifold learning is the process of estimating a low-dimensional structure which underlies a collection

More information

On the eigenvalues of Euclidean distance matrices

On the eigenvalues of Euclidean distance matrices Volume 27, N. 3, pp. 237 250, 2008 Copyright 2008 SBMAC ISSN 00-8205 www.scielo.br/cam On the eigenvalues of Euclidean distance matrices A.Y. ALFAKIH Department of Mathematics and Statistics University

More information

Lecture 10: Dimension Reduction Techniques

Lecture 10: Dimension Reduction Techniques Lecture 10: Dimension Reduction Techniques Radu Balan Department of Mathematics, AMSC, CSCAMM and NWC University of Maryland, College Park, MD April 17, 2018 Input Data It is assumed that there is a set

More information

DIMENSION REDUCTION. min. j=1

DIMENSION REDUCTION. min. j=1 DIMENSION REDUCTION 1 Principal Component Analysis (PCA) Principal components analysis (PCA) finds low dimensional approximations to the data by projecting the data onto linear subspaces. Let X R d and

More information

MACHINE LEARNING. Methods for feature extraction and reduction of dimensionality: Probabilistic PCA and kernel PCA

MACHINE LEARNING. Methods for feature extraction and reduction of dimensionality: Probabilistic PCA and kernel PCA 1 MACHINE LEARNING Methods for feature extraction and reduction of dimensionality: Probabilistic PCA and kernel PCA 2 Practicals Next Week Next Week, Practical Session on Computer Takes Place in Room GR

More information

Distance Metric Learning in Data Mining (Part II) Fei Wang and Jimeng Sun IBM TJ Watson Research Center

Distance Metric Learning in Data Mining (Part II) Fei Wang and Jimeng Sun IBM TJ Watson Research Center Distance Metric Learning in Data Mining (Part II) Fei Wang and Jimeng Sun IBM TJ Watson Research Center 1 Outline Part I - Applications Motivation and Introduction Patient similarity application Part II

More information

Francois Fouss, Alain Pirotte, Jean-Michel Renders & Marco Saerens. January 31, 2006

Francois Fouss, Alain Pirotte, Jean-Michel Renders & Marco Saerens. January 31, 2006 A novel way of computing dissimilarities between nodes of a graph, with application to collaborative filtering and subspace projection of the graph nodes Francois Fouss, Alain Pirotte, Jean-Michel Renders

More information

PARAMETERIZATION OF NON-LINEAR MANIFOLDS

PARAMETERIZATION OF NON-LINEAR MANIFOLDS PARAMETERIZATION OF NON-LINEAR MANIFOLDS C. W. GEAR DEPARTMENT OF CHEMICAL AND BIOLOGICAL ENGINEERING PRINCETON UNIVERSITY, PRINCETON, NJ E-MAIL:WGEAR@PRINCETON.EDU Abstract. In this report we consider

More information

Dimension Reduction Techniques. Presented by Jie (Jerry) Yu

Dimension Reduction Techniques. Presented by Jie (Jerry) Yu Dimension Reduction Techniques Presented by Jie (Jerry) Yu Outline Problem Modeling Review of PCA and MDS Isomap Local Linear Embedding (LLE) Charting Background Advances in data collection and storage

More information

Diffusion Wavelets and Applications

Diffusion Wavelets and Applications Diffusion Wavelets and Applications J.C. Bremer, R.R. Coifman, P.W. Jones, S. Lafon, M. Mohlenkamp, MM, R. Schul, A.D. Szlam Demos, web pages and preprints available at: S.Lafon: www.math.yale.edu/~sl349

More information

L26: Advanced dimensionality reduction

L26: Advanced dimensionality reduction L26: Advanced dimensionality reduction The snapshot CA approach Oriented rincipal Components Analysis Non-linear dimensionality reduction (manifold learning) ISOMA Locally Linear Embedding CSCE 666 attern

More information

Multiscale Manifold Learning

Multiscale Manifold Learning Multiscale Manifold Learning Chang Wang IBM T J Watson Research Lab Kitchawan Rd Yorktown Heights, New York 598 wangchan@usibmcom Sridhar Mahadevan Computer Science Department University of Massachusetts

More information

Manifold Regularization

Manifold Regularization 9.520: Statistical Learning Theory and Applications arch 3rd, 200 anifold Regularization Lecturer: Lorenzo Rosasco Scribe: Hooyoung Chung Introduction In this lecture we introduce a class of learning algorithms,

More information

Certifying the Global Optimality of Graph Cuts via Semidefinite Programming: A Theoretic Guarantee for Spectral Clustering

Certifying the Global Optimality of Graph Cuts via Semidefinite Programming: A Theoretic Guarantee for Spectral Clustering Certifying the Global Optimality of Graph Cuts via Semidefinite Programming: A Theoretic Guarantee for Spectral Clustering Shuyang Ling Courant Institute of Mathematical Sciences, NYU Aug 13, 2018 Joint

More information

Linear and Non-Linear Dimensionality Reduction

Linear and Non-Linear Dimensionality Reduction Linear and Non-Linear Dimensionality Reduction Alexander Schulz aschulz(at)techfak.uni-bielefeld.de University of Pisa, Pisa 4.5.215 and 7.5.215 Overview Dimensionality Reduction Motivation Linear Projections

More information

Statistical Pattern Recognition

Statistical Pattern Recognition Statistical Pattern Recognition Feature Extraction Hamid R. Rabiee Jafar Muhammadi, Alireza Ghasemi, Payam Siyari Spring 2014 http://ce.sharif.edu/courses/92-93/2/ce725-2/ Agenda Dimensionality Reduction

More information

Connection of Local Linear Embedding, ISOMAP, and Kernel Principal Component Analysis

Connection of Local Linear Embedding, ISOMAP, and Kernel Principal Component Analysis Connection of Local Linear Embedding, ISOMAP, and Kernel Principal Component Analysis Alvina Goh Vision Reading Group 13 October 2005 Connection of Local Linear Embedding, ISOMAP, and Kernel Principal

More information

Diffusion Maps, Spectral Clustering and Eigenfunctions of Fokker-Planck Operators

Diffusion Maps, Spectral Clustering and Eigenfunctions of Fokker-Planck Operators Diffusion Maps, Spectral Clustering and Eigenfunctions of Fokker-Planck Operators Boaz Nadler Stéphane Lafon Ronald R. Coifman Department of Mathematics, Yale University, New Haven, CT 652. {boaz.nadler,stephane.lafon,ronald.coifman}@yale.edu

More information

Graph Partitioning Using Random Walks

Graph Partitioning Using Random Walks Graph Partitioning Using Random Walks A Convex Optimization Perspective Lorenzo Orecchia Computer Science Why Spectral Algorithms for Graph Problems in practice? Simple to implement Can exploit very efficient

More information

Diffusion Maps, Spectral Clustering and Eigenfunctions of Fokker-Planck Operators

Diffusion Maps, Spectral Clustering and Eigenfunctions of Fokker-Planck Operators Diffusion Maps, Spectral Clustering and Eigenfunctions of Fokker-Planck Operators Boaz Nadler Stéphane Lafon Ronald R. Coifman Department of Mathematics, Yale University, New Haven, CT 652. {boaz.nadler,stephane.lafon,ronald.coifman}@yale.edu

More information

Spectral Clustering. by HU Pili. June 16, 2013

Spectral Clustering. by HU Pili. June 16, 2013 Spectral Clustering by HU Pili June 16, 2013 Outline Clustering Problem Spectral Clustering Demo Preliminaries Clustering: K-means Algorithm Dimensionality Reduction: PCA, KPCA. Spectral Clustering Framework

More information

Spectral Clustering. Spectral Clustering? Two Moons Data. Spectral Clustering Algorithm: Bipartioning. Spectral methods

Spectral Clustering. Spectral Clustering? Two Moons Data. Spectral Clustering Algorithm: Bipartioning. Spectral methods Spectral Clustering Seungjin Choi Department of Computer Science POSTECH, Korea seungjin@postech.ac.kr 1 Spectral methods Spectral Clustering? Methods using eigenvectors of some matrices Involve eigen-decomposition

More information

Semi-supervised Eigenvectors for Locally-biased Learning

Semi-supervised Eigenvectors for Locally-biased Learning Semi-supervised Eigenvectors for Locally-biased Learning Toke Jansen Hansen Section for Cognitive Systems DTU Informatics Technical University of Denmark tjha@imm.dtu.dk Michael W. Mahoney Department of

More information

MATH 829: Introduction to Data Mining and Analysis Clustering II

MATH 829: Introduction to Data Mining and Analysis Clustering II his lecture is based on U. von Luxburg, A Tutorial on Spectral Clustering, Statistics and Computing, 17 (4), 2007. MATH 829: Introduction to Data Mining and Analysis Clustering II Dominique Guillot Departments

More information

Graph Partitioning Algorithms and Laplacian Eigenvalues

Graph Partitioning Algorithms and Laplacian Eigenvalues Graph Partitioning Algorithms and Laplacian Eigenvalues Luca Trevisan Stanford Based on work with Tsz Chiu Kwok, Lap Chi Lau, James Lee, Yin Tat Lee, and Shayan Oveis Gharan spectral graph theory Use linear

More information

Learning a Kernel Matrix for Nonlinear Dimensionality Reduction

Learning a Kernel Matrix for Nonlinear Dimensionality Reduction Learning a Kernel Matrix for Nonlinear Dimensionality Reduction Kilian Q. Weinberger kilianw@cis.upenn.edu Fei Sha feisha@cis.upenn.edu Lawrence K. Saul lsaul@cis.upenn.edu Department of Computer and Information

More information

Manifold Coarse Graining for Online Semi-supervised Learning

Manifold Coarse Graining for Online Semi-supervised Learning for Online Semi-supervised Learning Mehrdad Farajtabar, Amirreza Shaban, Hamid R. Rabiee, Mohammad H. Rohban Digital Media Lab, Department of Computer Engineering, Sharif University of Technology, Tehran,

More information

Data Analysis and Manifold Learning Lecture 9: Diffusion on Manifolds and on Graphs

Data Analysis and Manifold Learning Lecture 9: Diffusion on Manifolds and on Graphs Data Analysis and Manifold Learning Lecture 9: Diffusion on Manifolds and on Graphs Radu Horaud INRIA Grenoble Rhone-Alpes, France Radu.Horaud@inrialpes.fr http://perception.inrialpes.fr/ Outline of Lecture

More information

Smart PCA. Yi Zhang Machine Learning Department Carnegie Mellon University

Smart PCA. Yi Zhang Machine Learning Department Carnegie Mellon University Smart PCA Yi Zhang Machine Learning Department Carnegie Mellon University yizhang1@cs.cmu.edu Abstract PCA can be smarter and makes more sensible projections. In this paper, we propose smart PCA, an extension

More information

Exploiting Sparse Non-Linear Structure in Astronomical Data

Exploiting Sparse Non-Linear Structure in Astronomical Data Exploiting Sparse Non-Linear Structure in Astronomical Data Ann B. Lee Department of Statistics and Department of Machine Learning, Carnegie Mellon University Joint work with P. Freeman, C. Schafer, and

More information

Beyond the Point Cloud: From Transductive to Semi-Supervised Learning

Beyond the Point Cloud: From Transductive to Semi-Supervised Learning Beyond the Point Cloud: From Transductive to Semi-Supervised Learning Vikas Sindhwani, Partha Niyogi, Mikhail Belkin Andrew B. Goldberg goldberg@cs.wisc.edu Department of Computer Sciences University of

More information

Global (ISOMAP) versus Local (LLE) Methods in Nonlinear Dimensionality Reduction

Global (ISOMAP) versus Local (LLE) Methods in Nonlinear Dimensionality Reduction Global (ISOMAP) versus Local (LLE) Methods in Nonlinear Dimensionality Reduction A presentation by Evan Ettinger on a Paper by Vin de Silva and Joshua B. Tenenbaum May 12, 2005 Outline Introduction The

More information

Diffusion and random walks on graphs

Diffusion and random walks on graphs Diffusion and random walks on graphs Leonid E. Zhukov School of Data Analysis and Artificial Intelligence Department of Computer Science National Research University Higher School of Economics Structural

More information

Spectral approximations in machine learning

Spectral approximations in machine learning Spectral approximations in machine learning Darren Homrighausen Department of Statistics Carnegie Mellon University Pittsburgh, PA 15213 dhomrigh@stat.cmu.edu Daniel J. McDonald Department of Statistics

More information

Machine Learning. B. Unsupervised Learning B.2 Dimensionality Reduction. Lars Schmidt-Thieme, Nicolas Schilling

Machine Learning. B. Unsupervised Learning B.2 Dimensionality Reduction. Lars Schmidt-Thieme, Nicolas Schilling Machine Learning B. Unsupervised Learning B.2 Dimensionality Reduction Lars Schmidt-Thieme, Nicolas Schilling Information Systems and Machine Learning Lab (ISMLL) Institute for Computer Science University

More information

Spectral Clustering on Handwritten Digits Database Mid-Year Pr

Spectral Clustering on Handwritten Digits Database Mid-Year Pr Spectral Clustering on Handwritten Digits Database Mid-Year Presentation Danielle dmiddle1@math.umd.edu Advisor: Kasso Okoudjou kasso@umd.edu Department of Mathematics University of Maryland- College Park

More information

Proximity data visualization with h-plots

Proximity data visualization with h-plots The fifth international conference user! 2009 Proximity data visualization with h-plots Irene Epifanio Dpt. Matemàtiques, Univ. Jaume I (SPAIN) epifanio@uji.es; http://www3.uji.es/~epifanio Outline Motivating

More information

Spectral Clustering. Guokun Lai 2016/10

Spectral Clustering. Guokun Lai 2016/10 Spectral Clustering Guokun Lai 2016/10 1 / 37 Organization Graph Cut Fundamental Limitations of Spectral Clustering Ng 2002 paper (if we have time) 2 / 37 Notation We define a undirected weighted graph

More information

Signal processing methods have significantly changed. Diffusion Maps. Signal Processing

Signal processing methods have significantly changed. Diffusion Maps. Signal Processing [ Ronen Talmon, Israel Cohen, Sharon Gannot, and Ronald R. Coifman ] Diffusion Maps for Signal Processing Advances in Kernel-based Learning for Signal Processing istockphoto.com/ aleksandar velasevic [

More information

LECTURE NOTE #11 PROF. ALAN YUILLE

LECTURE NOTE #11 PROF. ALAN YUILLE LECTURE NOTE #11 PROF. ALAN YUILLE 1. NonLinear Dimension Reduction Spectral Methods. The basic idea is to assume that the data lies on a manifold/surface in D-dimensional space, see figure (1) Perform

More information

Multiscale Wavelets on Trees, Graphs and High Dimensional Data

Multiscale Wavelets on Trees, Graphs and High Dimensional Data Multiscale Wavelets on Trees, Graphs and High Dimensional Data ICML 2010, Haifa Matan Gavish (Weizmann/Stanford) Boaz Nadler (Weizmann) Ronald Coifman (Yale) Boaz Nadler Ronald Coifman Motto... the relationships

More information

Linear algebra and applications to graphs Part 1

Linear algebra and applications to graphs Part 1 Linear algebra and applications to graphs Part 1 Written up by Mikhail Belkin and Moon Duchin Instructor: Laszlo Babai June 17, 2001 1 Basic Linear Algebra Exercise 1.1 Let V and W be linear subspaces

More information

Linear Spectral Hashing

Linear Spectral Hashing Linear Spectral Hashing Zalán Bodó and Lehel Csató Babeş Bolyai University - Faculty of Mathematics and Computer Science Kogălniceanu 1., 484 Cluj-Napoca - Romania Abstract. assigns binary hash keys to

More information

215 Problem 1. (a) Define the total variation distance µ ν tv for probability distributions µ, ν on a finite set S. Show that

215 Problem 1. (a) Define the total variation distance µ ν tv for probability distributions µ, ν on a finite set S. Show that 15 Problem 1. (a) Define the total variation distance µ ν tv for probability distributions µ, ν on a finite set S. Show that µ ν tv = (1/) x S µ(x) ν(x) = x S(µ(x) ν(x)) + where a + = max(a, 0). Show that

More information

Spectral Feature Vectors for Graph Clustering

Spectral Feature Vectors for Graph Clustering Spectral Feature Vectors for Graph Clustering Bin Luo,, Richard C. Wilson,andEdwinR.Hancock Department of Computer Science, University of York York YO DD, UK Anhui University, P.R. China {luo,wilson,erh}@cs.york.ac.uk

More information