Beyond Scalar Affinities for Network Analysis or Vector Diffusion Maps and the Connection Laplacian
|
|
- Karin Edwards
- 5 years ago
- Views:
Transcription
1 Beyond Scalar Affinities for Network Analysis or Vector Diffusion Maps and the Connection Laplacian Amit Singer Princeton University Department of Mathematics and Program in Applied and Computational Mathematics May 28, 2015 Amit Singer (Princeton University) May / 27
2 Single Particle Reconstruction using cryo-em Schematic drawing of the imaging process: The cryo-em problem: Amit Singer (Princeton University) May / 27
3 Main Algorithmic Challenges 1 Orientation assignment 2 Heterogeneity (resolving structural variability) 3 2D Class averaging (de-noising) 4 Symmetry detection 5 Motion correction 6 Particle picking Amit Singer (Princeton University) May / 27
4 Class Averaging in Cryo-EM: Improve SNR Amit Singer (Princeton University) May / 27
5 Image denoising by vector diffusion maps Generalization of Laplacian Eigenmaps (Belkin, Niyogi 2003) and Diffusion Maps (Coifman, Lafon 2006) Introduced the graph Connection Laplacian S, Zhao, Shkolnisky, Hadani (SIIMS 2011) Hadani, S (FoCM 2011) S, Wu (Comm. Pure Appl. Math 2012) Zhao, S (J Struct. Bio. 2014) Experimental images (70S) courtesy of Dr. Joachim Frank (Columbia) Class averages by vector diffusion maps (averaging with 20 nearest neighbors) Amit Singer (Princeton University) May / 27
6 Rotation Invariant Distances Projection images I 1,I 2,...,I n with unknown rotations R 1,R 2,...,R n SO(3) Rotationally Invariant Distances (RID) Cluster the images using K-means. d RID (i,j) = min O SO(2) I i O I j Images are not centered; also possible to include translations and to optimize over the special Euclidean group. Problem with this approach: outliers. At low SNR images with completely different viewing directions may have relatively small d RID (noise aligns well, instead of underlying signal). Amit Singer (Princeton University) May / 27
7 Outliers: Small World Graph on S 2 Define graph G = (V,E) by {i,j} E d RID (i,j) ε. 3 R i 2 R i 1 R i Optimal rotation angles O ij = argmin I i O I j, i,j = 1,...,n. O SO(2) Triplet consistency relation good triangles O ij O jk O ki I 2 2. How to use information of optimal rotations in a systematic way? Vector Diffusion Maps Non-local means with rotations Amit Singer (Princeton University) May / 27
8 Vector Diffusion Maps: Setup j w ij O ij i In VDM, the relationships between data points (e.g., cryo-em images) are represented as a weighted graph, where the weights w ij describing affinities between data points are accompanied by linear orthogonal transformations O ij. Amit Singer (Princeton University) May / 27
9 Manifold Learning: Point cloud in R p x 1,x 2,...,x n R p. Manifold assumption: x 1,...,x n M d, with d p. Local Principal Component Analysis (PCA) gives an approximate orthonormal basis O i for the tangent space T xi M. O i is a p d matrix with orthonormal columns: O T i O i = I d d. Alignment: O ij = argmin O O(d) O O T i O j HS (computed through the singular value decomposition of O T i O j ) Amit Singer (Princeton University) May / 27
10 Parallel Transport O ij approximates the parallel transport operator P xi,x j : T xj M T xi M Amit Singer (Princeton University) May / 27
11 Laplacian Eigenmap (Belkin and Niyogi 2003) and Diffusion Map (Coifman and Lafon 2006) Symmetric n n matrix W 0 : W 0 (i,j) = { wij (i,j) E, 0 (i,j) / E. Diagonal matrix D 0 of the same size: D 0 (i,i) = deg(i) = j:(i,j) E w ij. Graph Laplacian, Normalized graph Laplacian and the random walk matrix: L 0 = D 0 W 0, L 0 = I D 1/2 0 W 0 D 1/2 0, A 0 = D 1 0 W 0 The diffusion map Φ t is defined in terms of the eigenvectors of A 0 : A 0 φ l = λ l φ l, l = 1,...,n Φ t : i (λ t l φ l(i)) n l=1. Amit Singer (Princeton University) May / 27
12 Vector diffusion mapping: W 1 and D 1 Symmetric nd nd matrix W 1 : { wij O W 1 (i,j) = ij (i,j) E, 0 d d (i,j) / E. n n blocks, each of which is of size d d. Diagonal matrix D 1 of the same size, where the diagonal d d blocks are scalar matrices with the weighted degrees: D 1 (i,i) = deg(i)i d d, and deg(i) = j:(i,j) E w ij Amit Singer (Princeton University) May / 27
13 A 1 = D 1 1 W 1 is an averaging operator for vector fields The matrix A 1 can be applied to vectors v of length nd, which we regard as n vectors of length d, such that v(i) is a vector in R d viewed as a vector in T xi M. The matrix A 1 = D1 1 W 1 is an averaging operator for vector fields, since (A 1 v)(i) = 1 deg(i) j:(i,j) E w ij O ij v(j). This implies that the operator A 1 transport vectors from the tangent spaces T xj M (that are nearby to T xi M) to T xi M and then averages the transported vectors in T xi M. Amit Singer (Princeton University) May / 27
14 Affinity between nodes based on consistency of transformations In the VDM framework, we define the affinity between i and j by considering all paths of length t connecting them, but instead of just summing the weights of all paths, we sum the transformations. Every path from j to i may result in a different transformation (like parallel transport due to curvature). When adding transformations of different paths, cancelations may happen. We define the affinity between i and j as the consistency between these transformations. A 1 = D 1 1 W 1 is similar to the symmetric matrix W 1 W 1 = D 1/2 1 W 1 D 1/2 1 We define the affinity between i and j as W 2t 1 (i,j) 2 HS = deg(i) deg(j) (D 1 1 W 1) 2t (i,j) 2 HS. Amit Singer (Princeton University) May / 27
15 Embedding into a Hilbert Space Since W 1 is symmetric, it has a complete set of eigenvectors {v l } nd l=1 and eigenvalues {λ l } nd l=1 (ordered as λ 1 λ 2... λ nd ). Spectral decompositions of W 1 and W 2t W 1 (i,j) = nd l=1 1 : λ l v l (i)v l (j) T, and W2t 1 (i,j) = nd l=1 where v l (i) R d for i = 1,...,n and l = 1,...,nd. The HS norm of W2t 1 (i,j) is calculated using the trace: W 2t λ 2t l v l (i)v l (j) T, nd 1 (i,j) 2 HS = (λ l λ r ) 2t v l (i),v r (i) v l (j),v r (j). l,r=1 The affinity W 2t 1 (i,j) 2 HS = V t(i),v t (j) is an inner product for the finite dimensional Hilbert space R (nd)2 via the mapping V t : V t : i ( (λ l λ r ) t v l (i),v r (i) ) nd l,r=1. Amit Singer (Princeton University) May / 27
16 Vector Diffusion Distance The vector diffusion mapping is defined as V t : i ( (λ l λ r ) t v l (i),v r (i) ) nd l,r=1. The vector diffusion distance between nodes i and j is denoted d VDM,t (i,j) and is defined as d 2 VDM,t (i,j) = V t(i),v t (i) + V t (j),v t (j) 2 V t (i),v t (j). Other normalizations of the matrix W 1 are possible and lead to slightly different embeddings and distances (similar to diffusion maps). The matrices I W 1 and I + W 1 are positive semidefinite, because v T (I ±D 1/2 1 W 1 D 1/2 1 )v = (i,j) E v(i) ± w ijo ij v(j) 2 0, deg(i) deg(j) for any v R nd. Therefore, λ l [ 1,1]. As a result, the vector diffusion mapping and distances can be well approximated by using only the few largest eigenvalues and their corresponding eigenvectors. Amit Singer (Princeton University) May / 27
17 Application to the class averaging problem in Cryo-EM (S, Zhao, Shkolnisky, Hadani 2011) 14 x x degrees degrees (a) Neighbors are identified using d RID (b) Neighbors are identified using d VDM,t=2 Figure : SNR=1/64: Histogram of the angles (x-axis, in degrees) between the viewing directions of each image (out of 40000) and it 40 neighboring images. Left: neighbors are identified using the original rotationally invariant distances d RID. Right: neighbors are post identified using vector diffusion distances. Amit Singer (Princeton University) May / 27
18 Computational Aspects Zhao, S J. Struct. Biol Naïve implementation requires O(n 2 ) rotational alignments of images Rotational invariant representation of images: bispectrum Dimensionality reduction using a randomized algorithm for PCA (Rokhlin, Liberty, Tygert, Martinsson, Halko, Tropp, Szlam,...) Randomized approximated nearest neighbors search in nearly linear time (Jones, Osipov, Rokhlin 2011) Amit Singer (Princeton University) May / 27
19 The Hairy Ball Theorem There is no non-vanishing continuous tangent vector field on the sphere. Cannot find O i such that O ij = O i O 1 j. No global rotational alignment of all images. Amit Singer (Princeton University) May / 27
20 Convergence Theorem to the Connection-Laplacian (S, Wu) Let ι : M R p be a smooth d-dim closed Riemannian manifold embedded in R p, with metric g induced from the canonical metric on R p, and the data set {x i } i=1,...,n is independently uniformly distributed over M. Let K C 2 (R + ) be a positive kernel function decaying exponentially, that is, there exist T > 0 and C > 0 such that K(t) Ce t when t > T. ( ι(xi ) ι(x j ) R p For ǫ > 0, let K ǫ (x i,x j ) = K ǫ ). Then, for X C 3 (TM) and for all x i almost surely we have [ n lim 1 j=1 K ] ǫ(x i,x j )O ij X j ǫ 0n ǫ n j=1 K X i = m 2 ( ι 2 X(x i ),e l ) d ǫ(x i,x j ) 2dm l=1, 0 where 2 is the connection Laplacian, X i ( ι X(x i ),e l ) d l=1 Rd for all i, {e l (x i )} l=1,...,d is an orthonormal basis of ι T xi M, m l = R d x l K( x )dx, and O ij is the optimal orthogonal transformation determined by the algorithm in the alignment step. Amit Singer (Princeton University) May / 27
21 Example: Connection-Laplacian for S d embedded in R d+1 The connection-laplacian commutes with rotations and the eigenvalues and eigen-vector-fields are calculated using representation theory: S 2 : 6,10,14,... S 3 : 4,6,9,16,16,... S 4 : 5,10,14,... S 5 : 6,15,20, (a) S 2 (b) S 3 (c) S 4 (d) S 5 Figure : Bar plots of the largest 30 eigenvalues of A 1 for n = 8000 points uniformly distributed over spheres of different dimensions. Amit Singer (Princeton University) May / 27
22 Spectral graph theory with vector fields The Graph Connection Laplacian L 1 = D 1 W 1 The Normalized Graph Connection Laplacian L 1 = I D 1/2 1 W 1 D 1/2 = I W 1 Averaging operator / random walk matrix for vector diffusion: A 1 = D 1 1 W 1 A Cheeger inequality for the graph connection Laplacian (Bandeira, S, Spielman SIAM Matrix Analysis 2013) Amit Singer (Princeton University) May / 27
23 More applications of VDM: Orientability from a point cloud Encode the information about reflections in a symmetric n n matrix Z with entries { detoij (i,j) E, Z ij = 0 (i,j) / E. That is, Z ij = 1 if no reflection is needed, Z ij = 1 if a reflection is needed, and Z ij = 0 if the points are not nearby. Normalize Z by the node degrees (a) S 2 (b) Klein bottle (c) RP 2 Figure : Histogram of the values of the top eigenvector of D 1 0 Z. Amit Singer (Princeton University) May / 27
24 Orientable Double Covering Embedding obtained using the eigenvectors of the (normalized) matrix [ ] ( ) Z Z 1 1 = Z, Z Z 1 1 Figure : Left: the orientable double covering of RP(2), which is S 2 ; Middle: the orientable double covering of the Klein bottle, which is T 2 ; Right: the orientable double covering of the Möbius strip, which is a cylinder. Amit Singer (Princeton University) May / 27
25 Registration and Ordering of Drosophila Embryogenesis Images Dsilva, Lim, Lu, S, Kevrekidis, Shvartsman, Development 2015 Amit Singer (Princeton University) May / 27
26 Registered and Ordered using VDM Amit Singer (Princeton University) May / 27
27 References A. Singer, Z. Zhao, Y. Shkolnisky, R. Hadani, Viewing Angle Classification of Cryo-Electron Microscopy Images using Eigenvectors, SIAM Journal on Imaging Sciences, 4 (2), pp (2011). A. Singer, H.-T. Wu, Orientability and Diffusion Maps, Applied and Computational Harmonic Analysis, 31 (1), pp (2011). R. Hadani, A. Singer, Representation theoretic patterns in three dimensional Cryo-Electron Microscopy II The class averaging problem, Foundations of Computational Mathematics, 11 (5), pp (2011). A. Singer, H.-T. Wu, Vector Diffusion Maps and the Connection Laplacian, Communications on Pure and Applied Mathematics, 65 (8), pp (2012). A. S. Bandeira, A. Singer, D. A. Spielman, A Cheeger inequality for the graph connection Laplacian, SIAM Journal on Matrix Analysis and Applications, 34 (4), pp (2013). Z. Zhao, A. Singer, Rotationally Invariant Image Representation for Viewing Direction Classification in Cryo-EM, Journal of Structural Biology, 186 (1), pp (2014). C. J. Dsilva, B. Lim, H. Lu, A. Singer, I. G. Kevrekidis, S. Y. Shvartsman, Temporal ordering and registration of images in studies of developmental dynamics, Development, 142, (2015). Amit Singer (Princeton University) May / 27
Class Averaging in Cryo-Electron Microscopy
Class Averaging in Cryo-Electron Microscopy Zhizhen Jane Zhao Courant Institute of Mathematical Sciences, NYU Mathematics in Data Sciences ICERM, Brown University July 30 2015 Single Particle Reconstruction
More informationIntroduction to Three Dimensional Structure Determination of Macromolecules by Cryo-Electron Microscopy
Introduction to Three Dimensional Structure Determination of Macromolecules by Cryo-Electron Microscopy Amit Singer Princeton University, Department of Mathematics and PACM July 23, 2014 Amit Singer (Princeton
More informationData-dependent representations: Laplacian Eigenmaps
Data-dependent representations: Laplacian Eigenmaps November 4, 2015 Data Organization and Manifold Learning There are many techniques for Data Organization and Manifold Learning, e.g., Principal Component
More information3D Structure Determination using Cryo-Electron Microscopy Computational Challenges
3D Structure Determination using Cryo-Electron Microscopy Computational Challenges Amit Singer Princeton University Department of Mathematics and Program in Applied and Computational Mathematics July 28,
More informationSome Optimization and Statistical Learning Problems in Structural Biology
Some Optimization and Statistical Learning Problems in Structural Biology Amit Singer Princeton University, Department of Mathematics and PACM January 8, 2013 Amit Singer (Princeton University) January
More informationData dependent operators for the spatial-spectral fusion problem
Data dependent operators for the spatial-spectral fusion problem Wien, December 3, 2012 Joint work with: University of Maryland: J. J. Benedetto, J. A. Dobrosotskaya, T. Doster, K. W. Duke, M. Ehler, A.
More informationLearning Eigenfunctions: Links with Spectral Clustering and Kernel PCA
Learning Eigenfunctions: Links with Spectral Clustering and Kernel PCA Yoshua Bengio Pascal Vincent Jean-François Paiement University of Montreal April 2, Snowbird Learning 2003 Learning Modal Structures
More informationNon-linear Dimensionality Reduction
Non-linear Dimensionality Reduction CE-725: Statistical Pattern Recognition Sharif University of Technology Spring 2013 Soleymani Outline Introduction Laplacian Eigenmaps Locally Linear Embedding (LLE)
More informationAnalysis of Spectral Kernel Design based Semi-supervised Learning
Analysis of Spectral Kernel Design based Semi-supervised Learning Tong Zhang IBM T. J. Watson Research Center Yorktown Heights, NY 10598 Rie Kubota Ando IBM T. J. Watson Research Center Yorktown Heights,
More informationPARAMETERIZATION OF NON-LINEAR MANIFOLDS
PARAMETERIZATION OF NON-LINEAR MANIFOLDS C. W. GEAR DEPARTMENT OF CHEMICAL AND BIOLOGICAL ENGINEERING PRINCETON UNIVERSITY, PRINCETON, NJ E-MAIL:WGEAR@PRINCETON.EDU Abstract. In this report we consider
More informationNonlinear Methods. Data often lies on or near a nonlinear low-dimensional curve aka manifold.
Nonlinear Methods Data often lies on or near a nonlinear low-dimensional curve aka manifold. 27 Laplacian Eigenmaps Linear methods Lower-dimensional linear projection that preserves distances between all
More informationContribution from: Springer Verlag Berlin Heidelberg 2005 ISBN
Contribution from: Mathematical Physics Studies Vol. 7 Perspectives in Analysis Essays in Honor of Lennart Carleson s 75th Birthday Michael Benedicks, Peter W. Jones, Stanislav Smirnov (Eds.) Springer
More informationUnsupervised Learning Techniques Class 07, 1 March 2006 Andrea Caponnetto
Unsupervised Learning Techniques 9.520 Class 07, 1 March 2006 Andrea Caponnetto About this class Goal To introduce some methods for unsupervised learning: Gaussian Mixtures, K-Means, ISOMAP, HLLE, Laplacian
More informationFrom graph to manifold Laplacian: The convergence rate
Appl. Comput. Harmon. Anal. 2 (2006) 28 34 www.elsevier.com/locate/acha Letter to the Editor From graph to manifold Laplacian: The convergence rate A. Singer Department of athematics, Yale University,
More informationMarch 13, Paper: R.R. Coifman, S. Lafon, Diffusion maps ([Coifman06]) Seminar: Learning with Graphs, Prof. Hein, Saarland University
Kernels March 13, 2008 Paper: R.R. Coifman, S. Lafon, maps ([Coifman06]) Seminar: Learning with Graphs, Prof. Hein, Saarland University Kernels Figure: Example Application from [LafonWWW] meaningful geometric
More informationGraph Metrics and Dimension Reduction
Graph Metrics and Dimension Reduction Minh Tang 1 Michael Trosset 2 1 Applied Mathematics and Statistics The Johns Hopkins University 2 Department of Statistics Indiana University, Bloomington November
More informationVector spaces. DS-GA 1013 / MATH-GA 2824 Optimization-based Data Analysis.
Vector spaces DS-GA 1013 / MATH-GA 2824 Optimization-based Data Analysis http://www.cims.nyu.edu/~cfgranda/pages/obda_fall17/index.html Carlos Fernandez-Granda Vector space Consists of: A set V A scalar
More informationGlobal Positioning from Local Distances
Global Positioning from Local Distances Amit Singer Yale University, Department of Mathematics, Program in Applied Mathematics SIAM Annual Meeting, July 2008 Amit Singer (Yale University) San Diego 1 /
More informationGeometry on Probability Spaces
Geometry on Probability Spaces Steve Smale Toyota Technological Institute at Chicago 427 East 60th Street, Chicago, IL 60637, USA E-mail: smale@math.berkeley.edu Ding-Xuan Zhou Department of Mathematics,
More informationThree-dimensional structure determination of molecules without crystallization
Three-dimensional structure determination of molecules without crystallization Amit Singer Princeton University, Department of Mathematics and PACM June 4, 214 Amit Singer (Princeton University) June 214
More informationMultiscale Analysis and Diffusion Semigroups With Applications
Multiscale Analysis and Diffusion Semigroups With Applications Karamatou Yacoubou Djima Advisor: Wojciech Czaja Norbert Wiener Center Department of Mathematics University of Maryland, College Park http://www.norbertwiener.umd.edu
More informationUnsupervised dimensionality reduction
Unsupervised dimensionality reduction Guillaume Obozinski Ecole des Ponts - ParisTech SOCN course 2014 Guillaume Obozinski Unsupervised dimensionality reduction 1/30 Outline 1 PCA 2 Kernel PCA 3 Multidimensional
More informationGraphs, Geometry and Semi-supervised Learning
Graphs, Geometry and Semi-supervised Learning Mikhail Belkin The Ohio State University, Dept of Computer Science and Engineering and Dept of Statistics Collaborators: Partha Niyogi, Vikas Sindhwani In
More informationOrientation Assignment in Cryo-EM
Orientation Assignment in Cryo-EM Amit Singer Princeton University Department of Mathematics and Program in Applied and Computational Mathematics July 24, 2014 Amit Singer (Princeton University) July 2014
More informationSpectral Processing. Misha Kazhdan
Spectral Processing Misha Kazhdan [Taubin, 1995] A Signal Processing Approach to Fair Surface Design [Desbrun, et al., 1999] Implicit Fairing of Arbitrary Meshes [Vallet and Levy, 2008] Spectral Geometry
More informationAdvances in Manifold Learning Presented by: Naku Nak l Verm r a June 10, 2008
Advances in Manifold Learning Presented by: Nakul Verma June 10, 008 Outline Motivation Manifolds Manifold Learning Random projection of manifolds for dimension reduction Introduction to random projections
More informationApplications of Randomized Methods for Decomposing and Simulating from Large Covariance Matrices
Applications of Randomized Methods for Decomposing and Simulating from Large Covariance Matrices Vahid Dehdari and Clayton V. Deutsch Geostatistical modeling involves many variables and many locations.
More informationData Analysis and Manifold Learning Lecture 9: Diffusion on Manifolds and on Graphs
Data Analysis and Manifold Learning Lecture 9: Diffusion on Manifolds and on Graphs Radu Horaud INRIA Grenoble Rhone-Alpes, France Radu.Horaud@inrialpes.fr http://perception.inrialpes.fr/ Outline of Lecture
More informationDiffusion Geometries, Diffusion Wavelets and Harmonic Analysis of large data sets.
Diffusion Geometries, Diffusion Wavelets and Harmonic Analysis of large data sets. R.R. Coifman, S. Lafon, MM Mathematics Department Program of Applied Mathematics. Yale University Motivations The main
More informationData Analysis and Manifold Learning Lecture 3: Graphs, Graph Matrices, and Graph Embeddings
Data Analysis and Manifold Learning Lecture 3: Graphs, Graph Matrices, and Graph Embeddings Radu Horaud INRIA Grenoble Rhone-Alpes, France Radu.Horaud@inrialpes.fr http://perception.inrialpes.fr/ Outline
More informationStatistical Pattern Recognition
Statistical Pattern Recognition Feature Extraction Hamid R. Rabiee Jafar Muhammadi, Alireza Ghasemi, Payam Siyari Spring 2014 http://ce.sharif.edu/courses/92-93/2/ce725-2/ Agenda Dimensionality Reduction
More information8.1 Concentration inequality for Gaussian random matrix (cont d)
MGMT 69: Topics in High-dimensional Data Analysis Falll 26 Lecture 8: Spectral clustering and Laplacian matrices Lecturer: Jiaming Xu Scribe: Hyun-Ju Oh and Taotao He, October 4, 26 Outline Concentration
More informationDimension Reduction Techniques. Presented by Jie (Jerry) Yu
Dimension Reduction Techniques Presented by Jie (Jerry) Yu Outline Problem Modeling Review of PCA and MDS Isomap Local Linear Embedding (LLE) Charting Background Advances in data collection and storage
More informationIntroduction to Machine Learning. PCA and Spectral Clustering. Introduction to Machine Learning, Slides: Eran Halperin
1 Introduction to Machine Learning PCA and Spectral Clustering Introduction to Machine Learning, 2013-14 Slides: Eran Halperin Singular Value Decomposition (SVD) The singular value decomposition (SVD)
More informationThe crucial role of statistics in manifold learning
The crucial role of statistics in manifold learning Tyrus Berry Postdoc, Dept. of Mathematical Sciences, GMU Statistics Seminar GMU Feb., 26 Postdoctoral position supported by NSF ANALYSIS OF POINT CLOUDS
More informationRobust Laplacian Eigenmaps Using Global Information
Manifold Learning and its Applications: Papers from the AAAI Fall Symposium (FS-9-) Robust Laplacian Eigenmaps Using Global Information Shounak Roychowdhury ECE University of Texas at Austin, Austin, TX
More informationSpectral Algorithms I. Slides based on Spectral Mesh Processing Siggraph 2010 course
Spectral Algorithms I Slides based on Spectral Mesh Processing Siggraph 2010 course Why Spectral? A different way to look at functions on a domain Why Spectral? Better representations lead to simpler solutions
More informationData Analysis and Manifold Learning Lecture 7: Spectral Clustering
Data Analysis and Manifold Learning Lecture 7: Spectral Clustering Radu Horaud INRIA Grenoble Rhone-Alpes, France Radu.Horaud@inrialpes.fr http://perception.inrialpes.fr/ Outline of Lecture 7 What is spectral
More informationDimensionality Reduc1on
Dimensionality Reduc1on contd Aarti Singh Machine Learning 10-601 Nov 10, 2011 Slides Courtesy: Tom Mitchell, Eric Xing, Lawrence Saul 1 Principal Component Analysis (PCA) Principal Components are the
More informationLecture Notes 1: Vector spaces
Optimization-based data analysis Fall 2017 Lecture Notes 1: Vector spaces In this chapter we review certain basic concepts of linear algebra, highlighting their application to signal processing. 1 Vector
More informationConnection of Local Linear Embedding, ISOMAP, and Kernel Principal Component Analysis
Connection of Local Linear Embedding, ISOMAP, and Kernel Principal Component Analysis Alvina Goh Vision Reading Group 13 October 2005 Connection of Local Linear Embedding, ISOMAP, and Kernel Principal
More informationFiltering via a Reference Set. A.Haddad, D. Kushnir, R.R. Coifman Technical Report YALEU/DCS/TR-1441 February 21, 2011
Patch-based de-noising algorithms and patch manifold smoothing have emerged as efficient de-noising methods. This paper provides a new insight on these methods, such as the Non Local Means or the image
More informationL26: Advanced dimensionality reduction
L26: Advanced dimensionality reduction The snapshot CA approach Oriented rincipal Components Analysis Non-linear dimensionality reduction (manifold learning) ISOMA Locally Linear Embedding CSCE 666 attern
More informationApprentissage non supervisée
Apprentissage non supervisée Cours 3 Higher dimensions Jairo Cugliari Master ECD 2015-2016 From low to high dimension Density estimation Histograms and KDE Calibration can be done automacally But! Let
More informationDiffusion Wavelets and Applications
Diffusion Wavelets and Applications J.C. Bremer, R.R. Coifman, P.W. Jones, S. Lafon, M. Mohlenkamp, MM, R. Schul, A.D. Szlam Demos, web pages and preprints available at: S.Lafon: www.math.yale.edu/~sl349
More informationNonlinear Dimensionality Reduction. Jose A. Costa
Nonlinear Dimensionality Reduction Jose A. Costa Mathematics of Information Seminar, Dec. Motivation Many useful of signals such as: Image databases; Gene expression microarrays; Internet traffic time
More informationFast Angular Synchronization for Phase Retrieval via Incomplete Information
Fast Angular Synchronization for Phase Retrieval via Incomplete Information Aditya Viswanathan a and Mark Iwen b a Department of Mathematics, Michigan State University; b Department of Mathematics & Department
More informationStatistical Machine Learning
Statistical Machine Learning Christoph Lampert Spring Semester 2015/2016 // Lecture 12 1 / 36 Unsupervised Learning Dimensionality Reduction 2 / 36 Dimensionality Reduction Given: data X = {x 1,..., x
More informationCOMS 4721: Machine Learning for Data Science Lecture 19, 4/6/2017
COMS 4721: Machine Learning for Data Science Lecture 19, 4/6/2017 Prof. John Paisley Department of Electrical Engineering & Data Science Institute Columbia University PRINCIPAL COMPONENT ANALYSIS DIMENSIONALITY
More informationMath 307 Learning Goals. March 23, 2010
Math 307 Learning Goals March 23, 2010 Course Description The course presents core concepts of linear algebra by focusing on applications in Science and Engineering. Examples of applications from recent
More informationIntrinsic Structure Study on Whale Vocalizations
1 2015 DCLDE Conference Intrinsic Structure Study on Whale Vocalizations Yin Xian 1, Xiaobai Sun 2, Yuan Zhang 3, Wenjing Liao 3 Doug Nowacek 1,4, Loren Nolte 1, Robert Calderbank 1,2,3 1 Department of
More informationNonlinear Dimensionality Reduction
Nonlinear Dimensionality Reduction Piyush Rai CS5350/6350: Machine Learning October 25, 2011 Recap: Linear Dimensionality Reduction Linear Dimensionality Reduction: Based on a linear projection of the
More informationEECS 275 Matrix Computation
EECS 275 Matrix Computation Ming-Hsuan Yang Electrical Engineering and Computer Science University of California at Merced Merced, CA 95344 http://faculty.ucmerced.edu/mhyang Lecture 23 1 / 27 Overview
More informationIFT LAPLACIAN APPLICATIONS. Mikhail Bessmeltsev
IFT 6112 09 LAPLACIAN APPLICATIONS http://www-labs.iro.umontreal.ca/~bmpix/teaching/6112/2018/ Mikhail Bessmeltsev Rough Intuition http://pngimg.com/upload/hammer_png3886.png You can learn a lot about
More informationNormalized power iterations for the computation of SVD
Normalized power iterations for the computation of SVD Per-Gunnar Martinsson Department of Applied Mathematics University of Colorado Boulder, Co. Per-gunnar.Martinsson@Colorado.edu Arthur Szlam Courant
More informationFunctional Analysis Review
Outline 9.520: Statistical Learning Theory and Applications February 8, 2010 Outline 1 2 3 4 Vector Space Outline A vector space is a set V with binary operations +: V V V and : R V V such that for all
More informationLaplacian Eigenmaps for Dimensionality Reduction and Data Representation
Introduction and Data Representation Mikhail Belkin & Partha Niyogi Department of Electrical Engieering University of Minnesota Mar 21, 2017 1/22 Outline Introduction 1 Introduction 2 3 4 Connections to
More informationLaplacian Eigenmaps for Dimensionality Reduction and Data Representation
Laplacian Eigenmaps for Dimensionality Reduction and Data Representation Neural Computation, June 2003; 15 (6):1373-1396 Presentation for CSE291 sp07 M. Belkin 1 P. Niyogi 2 1 University of Chicago, Department
More informationBeyond the Point Cloud: From Transductive to Semi-Supervised Learning
Beyond the Point Cloud: From Transductive to Semi-Supervised Learning Vikas Sindhwani, Partha Niyogi, Mikhail Belkin Andrew B. Goldberg goldberg@cs.wisc.edu Department of Computer Sciences University of
More informationLearning gradients: prescriptive models
Department of Statistical Science Institute for Genome Sciences & Policy Department of Computer Science Duke University May 11, 2007 Relevant papers Learning Coordinate Covariances via Gradients. Sayan
More informationBasic Calculus Review
Basic Calculus Review Lorenzo Rosasco ISML Mod. 2 - Machine Learning Vector Spaces Functionals and Operators (Matrices) Vector Space A vector space is a set V with binary operations +: V V V and : R V
More information1 Singular Value Decomposition and Principal Component
Singular Value Decomposition and Principal Component Analysis In these lectures we discuss the SVD and the PCA, two of the most widely used tools in machine learning. Principal Component Analysis (PCA)
More informationLECTURE NOTE #11 PROF. ALAN YUILLE
LECTURE NOTE #11 PROF. ALAN YUILLE 1. NonLinear Dimension Reduction Spectral Methods. The basic idea is to assume that the data lies on a manifold/surface in D-dimensional space, see figure (1) Perform
More informationSpectral Clustering. by HU Pili. June 16, 2013
Spectral Clustering by HU Pili June 16, 2013 Outline Clustering Problem Spectral Clustering Demo Preliminaries Clustering: K-means Algorithm Dimensionality Reduction: PCA, KPCA. Spectral Clustering Framework
More informationCS168: The Modern Algorithmic Toolbox Lectures #11 and #12: Spectral Graph Theory
CS168: The Modern Algorithmic Toolbox Lectures #11 and #12: Spectral Graph Theory Tim Roughgarden & Gregory Valiant May 2, 2016 Spectral graph theory is the powerful and beautiful theory that arises from
More informationDiffusion and random walks on graphs
Diffusion and random walks on graphs Leonid E. Zhukov School of Data Analysis and Artificial Intelligence Department of Computer Science National Research University Higher School of Economics Structural
More informationDiffusion/Inference geometries of data features, situational awareness and visualization. Ronald R Coifman Mathematics Yale University
Diffusion/Inference geometries of data features, situational awareness and visualization Ronald R Coifman Mathematics Yale University Digital data is generally converted to point clouds in high dimensional
More informationA Statistical Look at Spectral Graph Analysis. Deep Mukhopadhyay
A Statistical Look at Spectral Graph Analysis Deep Mukhopadhyay Department of Statistics, Temple University Office: Speakman 335 deep@temple.edu http://sites.temple.edu/deepstat/ Graph Signal Processing
More information8 Eigenvectors and the Anisotropic Multivariate Gaussian Distribution
Eigenvectors and the Anisotropic Multivariate Gaussian Distribution Eigenvectors and the Anisotropic Multivariate Gaussian Distribution EIGENVECTORS [I don t know if you were properly taught about eigenvectors
More informationMultiscale bi-harmonic Analysis of Digital Data Bases and Earth moving distances.
Multiscale bi-harmonic Analysis of Digital Data Bases and Earth moving distances. R. Coifman, Department of Mathematics, program of Applied Mathematics Yale University Joint work with M. Gavish and W.
More informationMachine Learning for Data Science (CS4786) Lecture 11
Machine Learning for Data Science (CS4786) Lecture 11 Spectral clustering Course Webpage : http://www.cs.cornell.edu/courses/cs4786/2016sp/ ANNOUNCEMENT 1 Assignment P1 the Diagnostic assignment 1 will
More informationLecture: Some Practical Considerations (3 of 4)
Stat260/CS294: Spectral Graph Methods Lecture 14-03/10/2015 Lecture: Some Practical Considerations (3 of 4) Lecturer: Michael Mahoney Scribe: Michael Mahoney Warning: these notes are still very rough.
More informationPCA from noisy linearly reduced measurements
PCA from noisy linearly reduced measurements Amit Singer Princeton University Department of Mathematics and Program in Applied and Computational Mathematics September 27, 2016 Amit Singer (Princeton University)
More informationSynchronization over Cartan motion groups via contraction
Synchronization over Cartan motion groups via contraction Onur Özyeşil, Nir Sharon 1, and Amit Singer 1, 1 Program in Applied and Computational Mathematics, Princeton University Department of Mathematics,
More informationExercises * on Principal Component Analysis
Exercises * on Principal Component Analysis Laurenz Wiskott Institut für Neuroinformatik Ruhr-Universität Bochum, Germany, EU 4 February 207 Contents Intuition 3. Problem statement..........................................
More informationLecture 3: Review of Linear Algebra
ECE 83 Fall 2 Statistical Signal Processing instructor: R Nowak Lecture 3: Review of Linear Algebra Very often in this course we will represent signals as vectors and operators (eg, filters, transforms,
More informationA graph based approach to semi-supervised learning
A graph based approach to semi-supervised learning 1 Feb 2011 Two papers M. Belkin, P. Niyogi, and V Sindhwani. Manifold regularization: a geometric framework for learning from labeled and unlabeled examples.
More informationJustin Solomon MIT, Spring 2017
Justin Solomon MIT, Spring 2017 http://pngimg.com/upload/hammer_png3886.png You can learn a lot about a shape by hitting it (lightly) with a hammer! What can you learn about its shape from vibration frequencies
More informationNonlinear Dimensionality Reduction
Outline Hong Chang Institute of Computing Technology, Chinese Academy of Sciences Machine Learning Methods (Fall 2012) Outline Outline I 1 Kernel PCA 2 Isomap 3 Locally Linear Embedding 4 Laplacian Eigenmap
More informationLecture 3: Review of Linear Algebra
ECE 83 Fall 2 Statistical Signal Processing instructor: R Nowak, scribe: R Nowak Lecture 3: Review of Linear Algebra Very often in this course we will represent signals as vectors and operators (eg, filters,
More informationDirectional Field. Xiao-Ming Fu
Directional Field Xiao-Ming Fu Outlines Introduction Discretization Representation Objectives and Constraints Outlines Introduction Discretization Representation Objectives and Constraints Definition Spatially-varying
More informationIntroduction to Machine Learning
10-701 Introduction to Machine Learning PCA Slides based on 18-661 Fall 2018 PCA Raw data can be Complex, High-dimensional To understand a phenomenon we measure various related quantities If we knew what
More informationSpectral Graph Theory and its Applications. Daniel A. Spielman Dept. of Computer Science Program in Applied Mathematics Yale Unviersity
Spectral Graph Theory and its Applications Daniel A. Spielman Dept. of Computer Science Program in Applied Mathematics Yale Unviersity Outline Adjacency matrix and Laplacian Intuition, spectral graph drawing
More informationStatistical and Computational Analysis of Locality Preserving Projection
Statistical and Computational Analysis of Locality Preserving Projection Xiaofei He xiaofei@cs.uchicago.edu Department of Computer Science, University of Chicago, 00 East 58th Street, Chicago, IL 60637
More informationLocality Preserving Projections
Locality Preserving Projections Xiaofei He Department of Computer Science The University of Chicago Chicago, IL 60637 xiaofei@cs.uchicago.edu Partha Niyogi Department of Computer Science The University
More informationRandomized Algorithms
Randomized Algorithms Saniv Kumar, Google Research, NY EECS-6898, Columbia University - Fall, 010 Saniv Kumar 9/13/010 EECS6898 Large Scale Machine Learning 1 Curse of Dimensionality Gaussian Mixture Models
More informationAPPENDIX A. Background Mathematics. A.1 Linear Algebra. Vector algebra. Let x denote the n-dimensional column vector with components x 1 x 2.
APPENDIX A Background Mathematics A. Linear Algebra A.. Vector algebra Let x denote the n-dimensional column vector with components 0 x x 2 B C @. A x n Definition 6 (scalar product). The scalar product
More informationCSE 554 Lecture 7: Alignment
CSE 554 Lecture 7: Alignment Fall 2012 CSE554 Alignment Slide 1 Review Fairing (smoothing) Relocating vertices to achieve a smoother appearance Method: centroid averaging Simplification Reducing vertex
More informationSection 3.9. Matrix Norm
3.9. Matrix Norm 1 Section 3.9. Matrix Norm Note. We define several matrix norms, some similar to vector norms and some reflecting how multiplication by a matrix affects the norm of a vector. We use matrix
More informationDiffusion Maps, Spectral Clustering and Eigenfunctions of Fokker-Planck Operators
Diffusion Maps, Spectral Clustering and Eigenfunctions of Fokker-Planck Operators Boaz Nadler Stéphane Lafon Ronald R. Coifman Department of Mathematics, Yale University, New Haven, CT 652. {boaz.nadler,stephane.lafon,ronald.coifman}@yale.edu
More informationUnderstanding graphs through spectral densities
Understanding graphs through spectral densities David Bindel Department of Computer Science Cornell University SCAN seminar, 29 Feb 216 SCAN 1 / 1 Can One Hear the Shape of a Drum? 2 u = λu on Ω, u = on
More informationAn Algorithmist s Toolkit September 10, Lecture 1
18.409 An Algorithmist s Toolkit September 10, 2009 Lecture 1 Lecturer: Jonathan Kelner Scribe: Jesse Geneson (2009) 1 Overview The class s goals, requirements, and policies were introduced, and topics
More informationPrincipal Component Analysis
Machine Learning Michaelmas 2017 James Worrell Principal Component Analysis 1 Introduction 1.1 Goals of PCA Principal components analysis (PCA) is a dimensionality reduction technique that can be used
More informationDiffusion Geometries, Global and Multiscale
Diffusion Geometries, Global and Multiscale R.R. Coifman, S. Lafon, MM, J.C. Bremer Jr., A.D. Szlam, P.W. Jones, R.Schul Papers, talks, other materials available at: www.math.yale.edu/~mmm82 Data and functions
More informationUnsupervised Machine Learning and Data Mining. DS 5230 / DS Fall Lecture 7. Jan-Willem van de Meent
Unsupervised Machine Learning and Data Mining DS 5230 / DS 4420 - Fall 2018 Lecture 7 Jan-Willem van de Meent DIMENSIONALITY REDUCTION Borrowing from: Percy Liang (Stanford) Dimensionality Reduction Goal:
More informationMachine Learning - MT & 14. PCA and MDS
Machine Learning - MT 2016 13 & 14. PCA and MDS Varun Kanade University of Oxford November 21 & 23, 2016 Announcements Sheet 4 due this Friday by noon Practical 3 this week (continue next week if necessary)
More informationMultiscale Wavelets on Trees, Graphs and High Dimensional Data
Multiscale Wavelets on Trees, Graphs and High Dimensional Data ICML 2010, Haifa Matan Gavish (Weizmann/Stanford) Boaz Nadler (Weizmann) Ronald Coifman (Yale) Boaz Nadler Ronald Coifman Motto... the relationships
More informationAdvanced Machine Learning & Perception
Advanced Machine Learning & Perception Instructor: Tony Jebara Topic 2 Nonlinear Manifold Learning Multidimensional Scaling (MDS) Locally Linear Embedding (LLE) Beyond Principal Components Analysis (PCA)
More informationDIMENSION REDUCTION. min. j=1
DIMENSION REDUCTION 1 Principal Component Analysis (PCA) Principal components analysis (PCA) finds low dimensional approximations to the data by projecting the data onto linear subspaces. Let X R d and
More informationCMPSCI 791BB: Advanced ML: Laplacian Learning
CMPSCI 791BB: Advanced ML: Laplacian Learning Sridhar Mahadevan Outline! Spectral graph operators! Combinatorial graph Laplacian! Normalized graph Laplacian! Random walks! Machine learning on graphs! Clustering!
More information