Multiscale Wavelets on Trees, Graphs and High Dimensional Data

Size: px
Start display at page:

Download "Multiscale Wavelets on Trees, Graphs and High Dimensional Data"

Transcription

1 Multiscale Wavelets on Trees, Graphs and High Dimensional Data ICML 2010, Haifa Matan Gavish (Weizmann/Stanford) Boaz Nadler (Weizmann) Ronald Coifman (Yale)

2 Boaz Nadler Ronald Coifman

3 Motto... the relationships between smoothness and frequency forming the core ideas of Euclidean harmonic analysis are remarkably resilient, persisting in very general geometries. - Szlam, Maggioni, Coifman (2008)

4 Problem setup: Processing functions on a dataset Given a dataset X = {x 1,..., x N } with similarity matrix W i,j (or X = {x 1,..., x N } R d ) Nonparametric inference of f : X R Denoise: observe g = f + ε, recover f SSL / classification: extend f from X X to X

5 Problem setup: Processing functions on a dataset Given a dataset X = {x 1,..., x N } with similarity matrix W i,j (or X = {x 1,..., x N } R d ) Nonparametric inference of f : X R Denoise: observe g = f + ε, recover f SSL / classification: extend f from X X to X

6 Problem setup: Processing functions on a dataset Given a dataset X = {x 1,..., x N } with similarity matrix W i,j (or X = {x 1,..., x N } R d ) Nonparametric inference of f : X R Denoise: observe g = f + ε, recover f SSL / classification: extend f from X X to X

7 Problem setup: Processing functions on a dataset Given a dataset X = {x 1,..., x N } with similarity matrix W i,j (or X = {x 1,..., x N } R d ) Nonparametric inference of f : X R Denoise: observe g = f + ε, recover f SSL / classification: extend f from X X to X

8 Problem setup: Processing functions on a dataset Given a dataset X = {x 1,..., x N } with similarity matrix W i,j (or X = {x 1,..., x N } R d ) Nonparametric inference of f : X R Denoise: observe g = f + ε, recover f SSL / classification: extend f from X X to X

9 Problem setup: Processing functions on a dataset Given a dataset X = {x 1,..., x N } with similarity matrix W i,j (or X = {x 1,..., x N } R d ) Nonparametric inference of f : X R Denoise: observe g = f + ε, recover f SSL / classification: extend f from X X to X

10 Problem setup: Processing functions on a dataset Given a dataset X = {x 1,..., x N } with similarity matrix W i,j (or X = {x 1,..., x N } R d ) Nonparametric inference of f : X R Denoise: observe g = f + ε, recover f SSL / classification: extend f from X X to X

11 Problem setup: Processing functions on a dataset Given a dataset X = {x 1,..., x N } with similarity matrix W i,j (or X = {x 1,..., x N } R d ) Nonparametric inference of f : X R Denoise: observe g = f + ε, recover f SSL / classification: extend f from X X to X

12 Problem setup: Data adaptive orthobasis Can use local geometry W, but why reinvent the wheel? Enter Euclid Harmonic analysis wisdom in low dim Euclidean space: use orthobasis {ψ i } for space of functions f : X R Popular bases: Fourier, wavelet Process f in coefficient domain e.g. estimate, thereshold Exit Euclid We want to build {ψ i } according to graph W

13 Problem setup: Data adaptive orthobasis Can use local geometry W, but why reinvent the wheel? Enter Euclid Harmonic analysis wisdom in low dim Euclidean space: use orthobasis {ψ i } for space of functions f : X R Popular bases: Fourier, wavelet Process f in coefficient domain e.g. estimate, thereshold Exit Euclid We want to build {ψ i } according to graph W

14 Problem setup: Data adaptive orthobasis Can use local geometry W, but why reinvent the wheel? Enter Euclid Harmonic analysis wisdom in low dim Euclidean space: use orthobasis {ψ i } for space of functions f : X R Popular bases: Fourier, wavelet Process f in coefficient domain e.g. estimate, thereshold Exit Euclid We want to build {ψ i } according to graph W

15 Problem setup: Data adaptive orthobasis Can use local geometry W, but why reinvent the wheel? Enter Euclid Harmonic analysis wisdom in low dim Euclidean space: use orthobasis {ψ i } for space of functions f : X R Popular bases: Fourier, wavelet Process f in coefficient domain e.g. estimate, thereshold Exit Euclid We want to build {ψ i } according to graph W

16 Problem setup: Data adaptive orthobasis Can use local geometry W, but why reinvent the wheel? Enter Euclid Harmonic analysis wisdom in low dim Euclidean space: use orthobasis {ψ i } for space of functions f : X R Popular bases: Fourier, wavelet Process f in coefficient domain e.g. estimate, thereshold Exit Euclid We want to build {ψ i } according to graph W

17 Problem setup: Data adaptive orthobasis Can use local geometry W, but why reinvent the wheel? Enter Euclid Harmonic analysis wisdom in low dim Euclidean space: use orthobasis {ψ i } for space of functions f : X R Popular bases: Fourier, wavelet Process f in coefficient domain e.g. estimate, thereshold Exit Euclid We want to build {ψ i } according to graph W

18 Problem setup: Data adaptive orthobasis Can use local geometry W, but why reinvent the wheel? Enter Euclid Harmonic analysis wisdom in low dim Euclidean space: use orthobasis {ψ i } for space of functions f : X R Popular bases: Fourier, wavelet Process f in coefficient domain e.g. estimate, thereshold Exit Euclid We want to build {ψ i } according to graph W

19 Toy example Chapelle, Scholkopf and Zien, Semi-supervised learning, 2006 Example: USPS benchmark X is USPS (ML benchmark) as 1500 vectors in R = R 256 Affinity W i,j = exp ( x i x j 2) f : X {1, 1} is the class label.

20 Toy example Chapelle, Scholkopf and Zien, Semi-supervised learning, 2006 Example: USPS benchmark X is USPS (ML benchmark) as 1500 vectors in R = R 256 Affinity W i,j = exp ( x i x j 2) f : X {1, 1} is the class label.

21 Toy example Chapelle, Scholkopf and Zien, Semi-supervised learning, 2006 Example: USPS benchmark X is USPS (ML benchmark) as 1500 vectors in R = R 256 Affinity W i,j = exp ( x i x j 2) f : X {1, 1} is the class label.

22 Toy example Chapelle, Scholkopf and Zien, Semi-supervised learning, 2006 Example: USPS benchmark X is USPS (ML benchmark) as 1500 vectors in R = R 256 Affinity W i,j = exp ( x i x j 2) f : X {1, 1} is the class label.

23 Toy example: visualization by kernel PCA

24 Fourier Basis for {f : X R} Belkin and Niyogi, Using manifold structure for partially labelled classification, 2003 Generalizing Fourier: The Graph Laplacian eigenbasis Take (W D)ψ i = λ i ψ i where D i,i = j W i,j

25 Cons of Laplacian Eigenbasis Laplacian ( Graph Fourier ) basis Oscillatory, nonlocalized Uninterpretable Scalability challenging No theoretical bound on f, ψ i Empirically slow coefficient decay No fast transform

26 Toy example: Graph Laplacian Eigenbasis

27 Toy example: Graph Laplacian Eigenbasis

28 Cons of Laplacian Eigenbasis Laplacian eigenbasis Oscillatory, nonlocalized Uninterpretable Scalability Challenging No theory bound on f, ψ i Empirically slow decay No fast transform Dream Basis Localized Interpretable Computation scalable Provably fast coef. decay Empricially fast decay Fast transform

29 On Euclidean space, Wavelet basis solves this Localized Interpretable - scale/shift of same function Fundamental wavelet property on R - coeffs decay: If ψ is a regular wavelet and 0 < α < 1, then f (x) f (y) C x y α f, ψ l,k C 2 l(α+ 1 2) Fast transform

30 Wavelet basis for {f : X R}? Prior Art Diffusion wavelets (Coifman, Maggioni) Anisotropic Haar bases (Donoho) Treelets (Nadler, Lee, Wasserman)

31 Main Message Any Balanced Partition Tree whose metric preserves smoothness in W yields an extremely simple Wavelet Dream Basis

32

33 Cons of Laplacian Eigenbasis Laplacian eigenbasis Oscillatory, nonlocalized Uninterpretable Scalability Challenging No theory bound on f, ψ i Empirically slow decay No fast transform Haar-like Basis Localized Easily interpretable Computation scalable f smooth f, ψ l,k c l Empricially fast decay Fast transform

34 Toy example: Haar-like coeffs decay

35 Toy example: Haar-like coeffs decay

36 Toy example: Haar-like coeffs decay

37 Toy example: Haar-like coeffs decay

38 Eigenfunctions are oscillatory

39 Toy example: Haar-like basis function

40 Main Message Any Balanced Partition Tree, whose metric preserves smoothness in W,yields an extremely simple Basis

41 A Partition Tree on the nodes

42 A Partition Tree on the nodes

43 A Partition Tree on the nodes

44 A Partition Tree on the nodes

45 A Partition Tree on the nodes

46 Partition Tree (Dendrogram)

47 Partition Tree (Dendrogram)

48 Partition Tree (Dendrogram)

49 Partition Tree (Dendrogram)

50 Partition Tree (Dendrogram)

51 Partition Tree (Dendrogram)

52 Partition Tree (Dendrogram)

53 The Haar Basis on [0, 1]

54 The Haar Basis on [0, 1]

55 The Haar Basis on [0, 1]

56 The Haar Basis on [0, 1]

57 The Haar Basis on [0, 1]

58 The Haar Basis on [0, 1]

59 The Haar Basis on [0, 1]

60 The Haar Basis on [0, 1]

61 The Haar Basis on [0, 1]

62 The Haar Basis on [0, 1]

63 The Haar Basis on [0, 1]

64 Partition Tree (Dendrogram)

65 Partition Tree (Dendrogram)

66 Partition Tree (Dendrogram)

67 Partition Tree (Dendrogram)

68 Partition Tree (Dendrogram)

69 Partition Tree (Dendrogram)

70 Partition Tree (Dendrogram)

71 Partition Tree Haar-like basis

72 Partition Tree Haar-like basis

73 Partition Tree Haar-like basis

74 Partition Tree Haar-like basis

75 Partition Tree Haar-like basis

76 Partition Tree Haar-like basis

77 Partition Tree Haar-like basis

78 Partition Tree Haar-like basis

79 Partition Tree Haar-like basis

80 Partition Tree Haar-like basis

81 Partition Tree Haar-like basis

82

83 Main Message Any Balanced Partition Tree,whose metric preserves smoothness in W, yields an extremely simple WaveletBasis

84 f smooth in tree metric coefs decay How to define smoothness Partition tree T induces natrual tree (ultra-) metric d Measure smoothness of f : X R w.r.t d Theorem Let f : X R. Then f (x) f (y) C d (x, y) α f, ψ l,k C supp (ψ l,k ) (α+ 1 2) for any Haar-like basis {ψ l.k } based on the tree T. If the tree is balanced offspring folder q parent folder Then f (x) f (y) C d (x, y) α f, ψ l,k C q l(α+ 1 2) (for classical Haar, q = 1 2 ).

85 f smooth in tree metric coefs decay How to define smoothness Partition tree T induces natrual tree (ultra-) metric d Measure smoothness of f : X R w.r.t d Theorem Let f : X R. Then f (x) f (y) C d (x, y) α f, ψ l,k C supp (ψ l,k ) (α+ 1 2) for any Haar-like basis {ψ l.k } based on the tree T. If the tree is balanced offspring folder q parent folder Then f (x) f (y) C d (x, y) α f, ψ l,k C q l(α+ 1 2) (for classical Haar, q = 1 2 ).

86 f smooth in tree metric coefs decay How to define smoothness Partition tree T induces natrual tree (ultra-) metric d Measure smoothness of f : X R w.r.t d Theorem Let f : X R. Then f (x) f (y) C d (x, y) α f, ψ l,k C supp (ψ l,k ) (α+ 1 2) for any Haar-like basis {ψ l.k } based on the tree T. If the tree is balanced offspring folder q parent folder Then f (x) f (y) C d (x, y) α f, ψ l,k C q l(α+ 1 2) (for classical Haar, q = 1 2 ).

87 f smooth in tree metric coefs decay How to define smoothness Partition tree T induces natrual tree (ultra-) metric d Measure smoothness of f : X R w.r.t d Theorem Let f : X R. Then f (x) f (y) C d (x, y) α f, ψ l,k C supp (ψ l,k ) (α+ 1 2) for any Haar-like basis {ψ l.k } based on the tree T. If the tree is balanced offspring folder q parent folder Then f (x) f (y) C d (x, y) α f, ψ l,k C q l(α+ 1 2) (for classical Haar, q = 1 2 ).

88 f smooth in tree metric coefs decay How to define smoothness Partition tree T induces natrual tree (ultra-) metric d Measure smoothness of f : X R w.r.t d Theorem Let f : X R. Then f (x) f (y) C d (x, y) α f, ψ l,k C supp (ψ l,k ) (α+ 1 2) for any Haar-like basis {ψ l.k } based on the tree T. If the tree is balanced offspring folder q parent folder Then f (x) f (y) C d (x, y) α f, ψ l,k C q l(α+ 1 2) (for classical Haar, q = 1 2 ).

89 Results Results 1 Any partition tree on X induces wavelet Haar-like bases 2 Balanced tree f smooth equals fast coefficient decay 3 Application to semi-supervised learning 4 Beyond basics: Comparing trees, Tensor product of Haar-like bases

90 Application: Semi supervised learning Classification/Regression with Haar-like basis Task: Given values of smooth f on X X, extend f to X. Step 1: Build a partition tree s.t. f is smooth w.r.t tree metric Step 2: Construct a Haar-like basis {ψ l,i } Step 3: Estimate ˆf = f, ψl,i ψ l,i Control over coefficient decay non-parametric risk analysis Bound on E f ˆf 2 depends only on smoothness α of the target f and # labeled points

91 Application: Semi supervised learning Classification/Regression with Haar-like basis Task: Given values of smooth f on X X, extend f to X. Step 1: Build a partition tree s.t. f is smooth w.r.t tree metric Step 2: Construct a Haar-like basis {ψ l,i } Step 3: Estimate ˆf = f, ψl,i ψ l,i Control over coefficient decay non-parametric risk analysis Bound on E f ˆf 2 depends only on smoothness α of the target f and # labeled points

92 Application: Semi supervised learning Classification/Regression with Haar-like basis Task: Given values of smooth f on X X, extend f to X. Step 1: Build a partition tree s.t. f is smooth w.r.t tree metric Step 2: Construct a Haar-like basis {ψ l,i } Step 3: Estimate ˆf = f, ψl,i ψ l,i Control over coefficient decay non-parametric risk analysis Bound on E f ˆf 2 depends only on smoothness α of the target f and # labeled points

93 Application: Semi supervised learning Classification/Regression with Haar-like basis Task: Given values of smooth f on X X, extend f to X. Step 1: Build a partition tree s.t. f is smooth w.r.t tree metric Step 2: Construct a Haar-like basis {ψ l,i } Step 3: Estimate ˆf = f, ψl,i ψ l,i Control over coefficient decay non-parametric risk analysis Bound on E f ˆf 2 depends only on smoothness α of the target f and # labeled points

94 Application: Semi supervised learning Classification/Regression with Haar-like basis Task: Given values of smooth f on X X, extend f to X. Step 1: Build a partition tree s.t. f is smooth w.r.t tree metric Step 2: Construct a Haar-like basis {ψ l,i } Step 3: Estimate ˆf = f, ψl,i ψ l,i Control over coefficient decay non-parametric risk analysis Bound on E f ˆf 2 depends only on smoothness α of the target f and # labeled points

95 Application: Semi supervised learning Classification/Regression with Haar-like basis Task: Given values of smooth f on X X, extend f to X. Step 1: Build a partition tree s.t. f is smooth w.r.t tree metric Step 2: Construct a Haar-like basis {ψ l,i } Step 3: Estimate ˆf = f, ψl,i ψ l,i Control over coefficient decay non-parametric risk analysis Bound on E f ˆf 2 depends only on smoothness α of the target f and # labeled points

96 Application: Semi supervised learning Classification/Regression with Haar-like basis Task: Given values of smooth f on X X, extend f to X. Step 1: Build a partition tree s.t. f is smooth w.r.t tree metric Step 2: Construct a Haar-like basis {ψ l,i } Step 3: Estimate ˆf = f, ψl,i ψ l,i Control over coefficient decay non-parametric risk analysis Bound on E f ˆf 2 depends only on smoothness α of the target f and # labeled points

97 Toy Example benchmark Test Error (%) Laplacian Eigenmaps Laplacian Reg. Adaptive Threshold Haar like basis State of the art # of labeled points (out of 1500)

98 MNIST Digits 8 vs. {3,4,5,7} Laplacian Eigenmaps Laplacian Reg. Adaptive Threshold Haar like basis Test Error (%) # of labeled points (out of 1000)

99 Results Results 1 Any partition tree on X induces wavelet Haar-like bases 2 Balanced tree f smooth equals fast coefficient decay 3 Application to semi-supervised learning 4 Beyond basics: Tensor product of Haar-like bases, Coefficient thresholding

100 Tensor product of Haar-like bases Coifman and G, Harmonic Analysis of Digital Data Bases (2010)

101 Tensor product of Haar-like bases Coifman and G, Harmonic Analysis of Digital Data Bases (2010)

102 Tensor product of Haar-like bases Coifman and G, Harmonic Analysis of Digital Data Bases (2010)

103 Tensor product of Haar-like bases Coifman and G, Harmonic Analysis of Digital Data Bases (2010)

104 Tensor product of Haar-like bases Coifman and G, Harmonic Analysis of Digital Data Bases (2010)

105 Tensor product of Haar-like bases Coifman and G, Harmonic Analysis of Digital Data Bases (2010)

106 Tensor product of Haar-like bases Coifman and G, Harmonic Analysis of Digital Data Bases (2010)

107 Tensor product of Haar-like bases Coifman and G, Harmonic Analysis of Digital Data Bases (2010)

108 Tensor product of Haar-like bases Coifman and G, Harmonic Analysis of Digital Data Bases (2010)

109 Tensor product of Haar-like bases Coifman and G, Harmonic Analysis of Digital Data Bases (2010)

110 Tensor product of Haar-like bases Coifman and G, Harmonic Analysis of Digital Data Bases (2010)

111 Tensor product of Haar-like bases Coifman and G, Harmonic Analysis of Digital Data Bases (2010)

112 Tensor product of Haar-like bases Coifman and G, Harmonic Analysis of Digital Data Bases (2010)

113 Tensor product of Haar-like bases Coifman and G, Harmonic Analysis of Digital Data Bases (2010)

114 Tensor product of Haar-like bases

115 The Big Picture Coifman and Weiss, Extensions of Hardy spaces and their use in analysis, 1979 Geometric tools for Data analysis widely recognized Analysis tools (e.g. function spaces, wavelet theory) in graph or general geometries valuable and largely unexplored in ML context Deep theory, long tradition: geometry of X bases for {f : X R} ( Spaces of Homogeneous Type )

116 The Big Picture Coifman and Weiss, Extensions of Hardy spaces and their use in analysis, 1979 Geometric tools for Data analysis widely recognized Analysis tools (e.g. function spaces, wavelet theory) in graph or general geometries valuable and largely unexplored in ML context Deep theory, long tradition: geometry of X bases for {f : X R} ( Spaces of Homogeneous Type )

117 The Big Picture Coifman and Weiss, Extensions of Hardy spaces and their use in analysis, 1979 Geometric tools for Data analysis widely recognized Analysis tools (e.g. function spaces, wavelet theory) in graph or general geometries valuable and largely unexplored in ML context Deep theory, long tradition: geometry of X bases for {f : X R} ( Spaces of Homogeneous Type )

118 The Big Picture Coifman and Weiss, Extensions of Hardy spaces and their use in analysis, 1979 Geometric tools for Data analysis widely recognized Analysis tools (e.g. function spaces, wavelet theory) in graph or general geometries valuable and largely unexplored in ML context Deep theory, long tradition: geometry of X bases for {f : X R} ( Spaces of Homogeneous Type )

119 Summary Motto... the relationships between smoothness and frequency forming the core ideas of Euclidean harmonic analysis are remarkably resilient, persisting in very general geometries. - Szlam, Maggioni, Coifman (2008) Main message Any Balanced Partition Tree whose metric preserves smoothness in W yields an extremely simple Dream Wavelet Basis Fascinating open question Which graphs admit Balanced Partition Trees, whose metric preserves smoothness in W?

120 Supporting Information - proofs & code: gavish ; nadler/ G, Nadler and Coifman, Multiscale wavelets on trees, graphs and high dimensional data, Proceedings of ICML 2010 G, Nadler and Coifman, Inference using Haar-like wavelets, preprint (2010) Coifman and G, Harmonic analysis of digital data bases, to appear in Wavelets: Old and New Perspectives, Springer (2010) Coifman and Weiss, Extensions of Hardy spaces and their use in analysis, Bul. Of the AMS, 83 #4, pp (1977) Singh, Nowak and Calderbank, Detecting weak but hierarchically-structured patterns in networks, Proceedings of AISTATS 2010

Global vs. Multiscale Approaches

Global vs. Multiscale Approaches Harmonic Analysis on Graphs Global vs. Multiscale Approaches Weizmann Institute of Science, Rehovot, Israel July 2011 Joint work with Matan Gavish (WIS/Stanford), Ronald Coifman (Yale), ICML 10' Challenge:

More information

Harmonic Analysis and Geometries of Digital Data Bases

Harmonic Analysis and Geometries of Digital Data Bases Harmonic Analysis and Geometries of Digital Data Bases AMS Session Special Sesson on the Mathematics of Information and Knowledge, Ronald Coifman (Yale) and Matan Gavish (Stanford, Yale) January 14, 2010

More information

Multiscale Wavelets on Trees, Graphs and High Dimensional Data: Theory and Applications to Semi Supervised Learning

Multiscale Wavelets on Trees, Graphs and High Dimensional Data: Theory and Applications to Semi Supervised Learning Multiscale Wavelets on Trees, Graphs and High Dimensional Data: Theory and Applications to Semi Supervised Learning Matan Gavish 1 Boaz Nadler Weizmann Institute of Science, P.O. Box 26, Rehovot, 76100,

More information

Multiscale bi-harmonic Analysis of Digital Data Bases and Earth moving distances.

Multiscale bi-harmonic Analysis of Digital Data Bases and Earth moving distances. Multiscale bi-harmonic Analysis of Digital Data Bases and Earth moving distances. R. Coifman, Department of Mathematics, program of Applied Mathematics Yale University Joint work with M. Gavish and W.

More information

Graphs, Geometry and Semi-supervised Learning

Graphs, Geometry and Semi-supervised Learning Graphs, Geometry and Semi-supervised Learning Mikhail Belkin The Ohio State University, Dept of Computer Science and Engineering and Dept of Statistics Collaborators: Partha Niyogi, Vikas Sindhwani In

More information

Diffusion Wavelets and Applications

Diffusion Wavelets and Applications Diffusion Wavelets and Applications J.C. Bremer, R.R. Coifman, P.W. Jones, S. Lafon, M. Mohlenkamp, MM, R. Schul, A.D. Szlam Demos, web pages and preprints available at: S.Lafon: www.math.yale.edu/~sl349

More information

Diffusion Geometries, Global and Multiscale

Diffusion Geometries, Global and Multiscale Diffusion Geometries, Global and Multiscale R.R. Coifman, S. Lafon, MM, J.C. Bremer Jr., A.D. Szlam, P.W. Jones, R.Schul Papers, talks, other materials available at: www.math.yale.edu/~mmm82 Data and functions

More information

Learning gradients: prescriptive models

Learning gradients: prescriptive models Department of Statistical Science Institute for Genome Sciences & Policy Department of Computer Science Duke University May 11, 2007 Relevant papers Learning Coordinate Covariances via Gradients. Sayan

More information

Conference in Honor of Aline Bonami Orleans, June 2014

Conference in Honor of Aline Bonami Orleans, June 2014 Conference in Honor of Aline Bonami Orleans, June 2014 Harmonic Analysis and functional duality, as a tool for organization of information, and learning. R. Coifman Department of Mathematics, program of

More information

Learning on Graphs and Manifolds. CMPSCI 689 Sridhar Mahadevan U.Mass Amherst

Learning on Graphs and Manifolds. CMPSCI 689 Sridhar Mahadevan U.Mass Amherst Learning on Graphs and Manifolds CMPSCI 689 Sridhar Mahadevan U.Mass Amherst Outline Manifold learning is a relatively new area of machine learning (2000-now). Main idea Model the underlying geometry of

More information

Learning Eigenfunctions: Links with Spectral Clustering and Kernel PCA

Learning Eigenfunctions: Links with Spectral Clustering and Kernel PCA Learning Eigenfunctions: Links with Spectral Clustering and Kernel PCA Yoshua Bengio Pascal Vincent Jean-François Paiement University of Montreal April 2, Snowbird Learning 2003 Learning Modal Structures

More information

From graph to manifold Laplacian: The convergence rate

From graph to manifold Laplacian: The convergence rate Appl. Comput. Harmon. Anal. 2 (2006) 28 34 www.elsevier.com/locate/acha Letter to the Editor From graph to manifold Laplacian: The convergence rate A. Singer Department of athematics, Yale University,

More information

Diffusion/Inference geometries of data features, situational awareness and visualization. Ronald R Coifman Mathematics Yale University

Diffusion/Inference geometries of data features, situational awareness and visualization. Ronald R Coifman Mathematics Yale University Diffusion/Inference geometries of data features, situational awareness and visualization Ronald R Coifman Mathematics Yale University Digital data is generally converted to point clouds in high dimensional

More information

Contribution from: Springer Verlag Berlin Heidelberg 2005 ISBN

Contribution from: Springer Verlag Berlin Heidelberg 2005 ISBN Contribution from: Mathematical Physics Studies Vol. 7 Perspectives in Analysis Essays in Honor of Lennart Carleson s 75th Birthday Michael Benedicks, Peter W. Jones, Stanislav Smirnov (Eds.) Springer

More information

Semi-Supervised Learning with the Graph Laplacian: The Limit of Infinite Unlabelled Data

Semi-Supervised Learning with the Graph Laplacian: The Limit of Infinite Unlabelled Data Semi-Supervised Learning with the Graph Laplacian: The Limit of Infinite Unlabelled Data Boaz Nadler Dept. of Computer Science and Applied Mathematics Weizmann Institute of Science Rehovot, Israel 76 boaz.nadler@weizmann.ac.il

More information

Semi-Supervised Learning in Gigantic Image Collections. Rob Fergus (New York University) Yair Weiss (Hebrew University) Antonio Torralba (MIT)

Semi-Supervised Learning in Gigantic Image Collections. Rob Fergus (New York University) Yair Weiss (Hebrew University) Antonio Torralba (MIT) Semi-Supervised Learning in Gigantic Image Collections Rob Fergus (New York University) Yair Weiss (Hebrew University) Antonio Torralba (MIT) Gigantic Image Collections What does the world look like? High

More information

Data-dependent representations: Laplacian Eigenmaps

Data-dependent representations: Laplacian Eigenmaps Data-dependent representations: Laplacian Eigenmaps November 4, 2015 Data Organization and Manifold Learning There are many techniques for Data Organization and Manifold Learning, e.g., Principal Component

More information

Diffusion Geometries, Diffusion Wavelets and Harmonic Analysis of large data sets.

Diffusion Geometries, Diffusion Wavelets and Harmonic Analysis of large data sets. Diffusion Geometries, Diffusion Wavelets and Harmonic Analysis of large data sets. R.R. Coifman, S. Lafon, MM Mathematics Department Program of Applied Mathematics. Yale University Motivations The main

More information

Nonlinear Dimensionality Reduction. Jose A. Costa

Nonlinear Dimensionality Reduction. Jose A. Costa Nonlinear Dimensionality Reduction Jose A. Costa Mathematics of Information Seminar, Dec. Motivation Many useful of signals such as: Image databases; Gene expression microarrays; Internet traffic time

More information

Clustering in kernel embedding spaces and organization of documents

Clustering in kernel embedding spaces and organization of documents Clustering in kernel embedding spaces and organization of documents Stéphane Lafon Collaborators: Raphy Coifman (Yale), Yosi Keller (Yale), Ioannis G. Kevrekidis (Princeton), Ann B. Lee (CMU), Boaz Nadler

More information

March 13, Paper: R.R. Coifman, S. Lafon, Diffusion maps ([Coifman06]) Seminar: Learning with Graphs, Prof. Hein, Saarland University

March 13, Paper: R.R. Coifman, S. Lafon, Diffusion maps ([Coifman06]) Seminar: Learning with Graphs, Prof. Hein, Saarland University Kernels March 13, 2008 Paper: R.R. Coifman, S. Lafon, maps ([Coifman06]) Seminar: Learning with Graphs, Prof. Hein, Saarland University Kernels Figure: Example Application from [LafonWWW] meaningful geometric

More information

Stable MMPI-2 Scoring: Introduction to Kernel. Extension Techniques

Stable MMPI-2 Scoring: Introduction to Kernel. Extension Techniques 1 Stable MMPI-2 Scoring: Introduction to Kernel Extension Techniques Liberty,E., Almagor,M., Zucker,S., Keller,Y., and Coifman,R.R. Abstract The current study introduces a new technique called Geometric

More information

Multiscale analysis on graphs

Multiscale analysis on graphs Mathematics and Computer Science Duke University I.P.A.M. 11/17/08 In collaboration with R.R. Coifman, P.W. Jones, Y-M. Jung, R. Schul, A.D. Szlam & J.C. Bremer Jr. Funding: NSF/DHS-FODAVA, DMS, IIS, CCF;

More information

Learning from Labeled and Unlabeled Data: Semi-supervised Learning and Ranking p. 1/31

Learning from Labeled and Unlabeled Data: Semi-supervised Learning and Ranking p. 1/31 Learning from Labeled and Unlabeled Data: Semi-supervised Learning and Ranking Dengyong Zhou zhou@tuebingen.mpg.de Dept. Schölkopf, Max Planck Institute for Biological Cybernetics, Germany Learning from

More information

Beyond Scalar Affinities for Network Analysis or Vector Diffusion Maps and the Connection Laplacian

Beyond Scalar Affinities for Network Analysis or Vector Diffusion Maps and the Connection Laplacian Beyond Scalar Affinities for Network Analysis or Vector Diffusion Maps and the Connection Laplacian Amit Singer Princeton University Department of Mathematics and Program in Applied and Computational Mathematics

More information

Diffusion Maps, Spectral Clustering and Eigenfunctions of Fokker-Planck Operators

Diffusion Maps, Spectral Clustering and Eigenfunctions of Fokker-Planck Operators Diffusion Maps, Spectral Clustering and Eigenfunctions of Fokker-Planck Operators Boaz Nadler Stéphane Lafon Ronald R. Coifman Department of Mathematics, Yale University, New Haven, CT 652. {boaz.nadler,stephane.lafon,ronald.coifman}@yale.edu

More information

Semi-Supervised Learning by Multi-Manifold Separation

Semi-Supervised Learning by Multi-Manifold Separation Semi-Supervised Learning by Multi-Manifold Separation Xiaojin (Jerry) Zhu Department of Computer Sciences University of Wisconsin Madison Joint work with Andrew Goldberg, Zhiting Xu, Aarti Singh, and Rob

More information

Graph-Based Semi-Supervised Learning

Graph-Based Semi-Supervised Learning Graph-Based Semi-Supervised Learning Olivier Delalleau, Yoshua Bengio and Nicolas Le Roux Université de Montréal CIAR Workshop - April 26th, 2005 Graph-Based Semi-Supervised Learning Yoshua Bengio, Olivier

More information

Beyond the Point Cloud: From Transductive to Semi-Supervised Learning

Beyond the Point Cloud: From Transductive to Semi-Supervised Learning Beyond the Point Cloud: From Transductive to Semi-Supervised Learning Vikas Sindhwani, Partha Niyogi, Mikhail Belkin Andrew B. Goldberg goldberg@cs.wisc.edu Department of Computer Sciences University of

More information

Analysis of Spectral Kernel Design based Semi-supervised Learning

Analysis of Spectral Kernel Design based Semi-supervised Learning Analysis of Spectral Kernel Design based Semi-supervised Learning Tong Zhang IBM T. J. Watson Research Center Yorktown Heights, NY 10598 Rie Kubota Ando IBM T. J. Watson Research Center Yorktown Heights,

More information

Self-Tuning Semantic Image Segmentation

Self-Tuning Semantic Image Segmentation Self-Tuning Semantic Image Segmentation Sergey Milyaev 1,2, Olga Barinova 2 1 Voronezh State University sergey.milyaev@gmail.com 2 Lomonosov Moscow State University obarinova@graphics.cs.msu.su Abstract.

More information

On the Phase Transition Phenomenon of Graph Laplacian Eigenfunctions on Trees

On the Phase Transition Phenomenon of Graph Laplacian Eigenfunctions on Trees On the Phase Transition Phenomenon of Graph Laplacian Eigenfunctions on Trees Naoki Saito and Ernest Woei Department of Mathematics University of California Davis, CA 9566 USA Email: saito@math.ucdavis.edu;

More information

Manifold Coarse Graining for Online Semi-supervised Learning

Manifold Coarse Graining for Online Semi-supervised Learning for Online Semi-supervised Learning Mehrdad Farajtabar, Amirreza Shaban, Hamid R. Rabiee, Mohammad H. Rohban Digital Media Lab, Department of Computer Engineering, Sharif University of Technology, Tehran,

More information

How to learn from very few examples?

How to learn from very few examples? How to learn from very few examples? Dengyong Zhou Department of Empirical Inference Max Planck Institute for Biological Cybernetics Spemannstr. 38, 72076 Tuebingen, Germany Outline Introduction Part A

More information

CSE 291. Assignment Spectral clustering versus k-means. Out: Wed May 23 Due: Wed Jun 13

CSE 291. Assignment Spectral clustering versus k-means. Out: Wed May 23 Due: Wed Jun 13 CSE 291. Assignment 3 Out: Wed May 23 Due: Wed Jun 13 3.1 Spectral clustering versus k-means Download the rings data set for this problem from the course web site. The data is stored in MATLAB format as

More information

Justin Solomon MIT, Spring 2017

Justin Solomon MIT, Spring 2017 Justin Solomon MIT, Spring 2017 http://pngimg.com/upload/hammer_png3886.png You can learn a lot about a shape by hitting it (lightly) with a hammer! What can you learn about its shape from vibration frequencies

More information

Discrete vs. Continuous: Two Sides of Machine Learning

Discrete vs. Continuous: Two Sides of Machine Learning Discrete vs. Continuous: Two Sides of Machine Learning Dengyong Zhou Department of Empirical Inference Max Planck Institute for Biological Cybernetics Spemannstr. 38, 72076 Tuebingen, Germany Oct. 18,

More information

Semi-supervised Eigenvectors for Locally-biased Learning

Semi-supervised Eigenvectors for Locally-biased Learning Semi-supervised Eigenvectors for Locally-biased Learning Toke Jansen Hansen Section for Cognitive Systems DTU Informatics Technical University of Denmark tjha@imm.dtu.dk Michael W. Mahoney Department of

More information

Sparse Approximation: from Image Restoration to High Dimensional Classification

Sparse Approximation: from Image Restoration to High Dimensional Classification Sparse Approximation: from Image Restoration to High Dimensional Classification Bin Dong Beijing International Center for Mathematical Research Beijing Institute of Big Data Research Peking University

More information

Bi-stochastic kernels via asymmetric affinity functions

Bi-stochastic kernels via asymmetric affinity functions Bi-stochastic kernels via asymmetric affinity functions Ronald R. Coifman, Matthew J. Hirn Yale University Department of Mathematics P.O. Box 208283 New Haven, Connecticut 06520-8283 USA ariv:1209.0237v4

More information

Predicting Graph Labels using Perceptron. Shuang Song

Predicting Graph Labels using Perceptron. Shuang Song Predicting Graph Labels using Perceptron Shuang Song shs037@eng.ucsd.edu Online learning over graphs M. Herbster, M. Pontil, and L. Wainer, Proc. 22nd Int. Conf. Machine Learning (ICML'05), 2005 Prediction

More information

A graph based approach to semi-supervised learning

A graph based approach to semi-supervised learning A graph based approach to semi-supervised learning 1 Feb 2011 Two papers M. Belkin, P. Niyogi, and V Sindhwani. Manifold regularization: a geometric framework for learning from labeled and unlabeled examples.

More information

Semi-Supervised Classification with Universum

Semi-Supervised Classification with Universum Semi-Supervised Classification with Universum Dan Zhang 1, Jingdong Wang 2, Fei Wang 3, Changshui Zhang 4 1,3,4 State Key Laboratory on Intelligent Technology and Systems, Tsinghua National Laboratory

More information

Diffusion Maps, Spectral Clustering and Eigenfunctions of Fokker-Planck Operators

Diffusion Maps, Spectral Clustering and Eigenfunctions of Fokker-Planck Operators Diffusion Maps, Spectral Clustering and Eigenfunctions of Fokker-Planck Operators Boaz Nadler Stéphane Lafon Ronald R. Coifman Department of Mathematics, Yale University, New Haven, CT 652. {boaz.nadler,stephane.lafon,ronald.coifman}@yale.edu

More information

Semi-Supervised Learning of Speech Sounds

Semi-Supervised Learning of Speech Sounds Aren Jansen Partha Niyogi Department of Computer Science Interspeech 2007 Objectives 1 Present a manifold learning algorithm based on locality preserving projections for semi-supervised phone classification

More information

What is semi-supervised learning?

What is semi-supervised learning? What is semi-supervised learning? In many practical learning domains, there is a large supply of unlabeled data but limited labeled data, which can be expensive to generate text processing, video-indexing,

More information

Nonlinear Methods. Data often lies on or near a nonlinear low-dimensional curve aka manifold.

Nonlinear Methods. Data often lies on or near a nonlinear low-dimensional curve aka manifold. Nonlinear Methods Data often lies on or near a nonlinear low-dimensional curve aka manifold. 27 Laplacian Eigenmaps Linear methods Lower-dimensional linear projection that preserves distances between all

More information

Semi-Supervised Learning

Semi-Supervised Learning Semi-Supervised Learning getting more for less in natural language processing and beyond Xiaojin (Jerry) Zhu School of Computer Science Carnegie Mellon University 1 Semi-supervised Learning many human

More information

Persistent homology and nonparametric regression

Persistent homology and nonparametric regression Cleveland State University March 10, 2009, BIRS: Data Analysis using Computational Topology and Geometric Statistics joint work with Gunnar Carlsson (Stanford), Moo Chung (Wisconsin Madison), Peter Kim

More information

Diffusion Wavelets for multiscale analysis on manifolds and graphs: constructions and applications

Diffusion Wavelets for multiscale analysis on manifolds and graphs: constructions and applications Diffusion Wavelets for multiscale analysis on manifolds and graphs: constructions and applications Mauro Maggioni EPFL, Lausanne Dec. 19th, 5 EPFL Multiscale Analysis and Diffusion Wavelets - Mauro Maggioni,

More information

Multiscale Analysis and Diffusion Semigroups With Applications

Multiscale Analysis and Diffusion Semigroups With Applications Multiscale Analysis and Diffusion Semigroups With Applications Karamatou Yacoubou Djima Advisor: Wojciech Czaja Norbert Wiener Center Department of Mathematics University of Maryland, College Park http://www.norbertwiener.umd.edu

More information

Limits of Spectral Clustering

Limits of Spectral Clustering Limits of Spectral Clustering Ulrike von Luxburg and Olivier Bousquet Max Planck Institute for Biological Cybernetics Spemannstr. 38, 72076 Tübingen, Germany {ulrike.luxburg,olivier.bousquet}@tuebingen.mpg.de

More information

THE HIDDEN CONVEXITY OF SPECTRAL CLUSTERING

THE HIDDEN CONVEXITY OF SPECTRAL CLUSTERING THE HIDDEN CONVEXITY OF SPECTRAL CLUSTERING Luis Rademacher, Ohio State University, Computer Science and Engineering. Joint work with Mikhail Belkin and James Voss This talk A new approach to multi-way

More information

Geometry on Probability Spaces

Geometry on Probability Spaces Geometry on Probability Spaces Steve Smale Toyota Technological Institute at Chicago 427 East 60th Street, Chicago, IL 60637, USA E-mail: smale@math.berkeley.edu Ding-Xuan Zhou Department of Mathematics,

More information

Face Recognition Using Laplacianfaces He et al. (IEEE Trans PAMI, 2005) presented by Hassan A. Kingravi

Face Recognition Using Laplacianfaces He et al. (IEEE Trans PAMI, 2005) presented by Hassan A. Kingravi Face Recognition Using Laplacianfaces He et al. (IEEE Trans PAMI, 2005) presented by Hassan A. Kingravi Overview Introduction Linear Methods for Dimensionality Reduction Nonlinear Methods and Manifold

More information

Multiscale Manifold Learning

Multiscale Manifold Learning Multiscale Manifold Learning Chang Wang IBM T J Watson Research Lab Kitchawan Rd Yorktown Heights, New York 598 wangchan@usibmcom Sridhar Mahadevan Computer Science Department University of Massachusetts

More information

Unsupervised Learning Techniques Class 07, 1 March 2006 Andrea Caponnetto

Unsupervised Learning Techniques Class 07, 1 March 2006 Andrea Caponnetto Unsupervised Learning Techniques 9.520 Class 07, 1 March 2006 Andrea Caponnetto About this class Goal To introduce some methods for unsupervised learning: Gaussian Mixtures, K-Means, ISOMAP, HLLE, Laplacian

More information

Fast Direct Policy Evaluation using Multiscale Analysis of Markov Diffusion Processes

Fast Direct Policy Evaluation using Multiscale Analysis of Markov Diffusion Processes Fast Direct Policy Evaluation using Multiscale Analysis of Markov Diffusion Processes Mauro Maggioni mauro.maggioni@yale.edu Department of Mathematics, Yale University, P.O. Box 88, New Haven, CT,, U.S.A.

More information

Efficient Inference in Fully Connected CRFs with Gaussian Edge Potentials

Efficient Inference in Fully Connected CRFs with Gaussian Edge Potentials Efficient Inference in Fully Connected CRFs with Gaussian Edge Potentials Philipp Krähenbühl and Vladlen Koltun Stanford University Presenter: Yuan-Ting Hu 1 Conditional Random Field (CRF) E x I = φ u

More information

Unlabeled Data: Now It Helps, Now It Doesn t

Unlabeled Data: Now It Helps, Now It Doesn t institution-logo-filena A. Singh, R. D. Nowak, and X. Zhu. In NIPS, 2008. 1 Courant Institute, NYU April 21, 2015 Outline institution-logo-filena 1 Conflicting Views in Semi-supervised Learning The Cluster

More information

Directed Graph Embedding: an Algorithm based on Continuous Limits of Laplacian-type Operators

Directed Graph Embedding: an Algorithm based on Continuous Limits of Laplacian-type Operators Directed Graph Embedding: an Algorithm based on Continuous Limits of Laplacian-type Operators Dominique C. Perrault-Joncas Department of Statistics University of Washington Seattle, WA 98195 dcpj@stat.washington.edu

More information

Manifold Learning: Theory and Applications to HRI

Manifold Learning: Theory and Applications to HRI Manifold Learning: Theory and Applications to HRI Seungjin Choi Department of Computer Science Pohang University of Science and Technology, Korea seungjin@postech.ac.kr August 19, 2008 1 / 46 Greek Philosopher

More information

The Generalized Haar-Walsh Transform (GHWT) for Data Analysis on Graphs and Networks

The Generalized Haar-Walsh Transform (GHWT) for Data Analysis on Graphs and Networks The Generalized Haar-Walsh Transform (GHWT) for Data Analysis on Graphs and Networks Jeff Irion & Naoki Saito Department of Mathematics University of California, Davis SIAM Annual Meeting 2014 Chicago,

More information

Manifold Regularization

Manifold Regularization 9.520: Statistical Learning Theory and Applications arch 3rd, 200 anifold Regularization Lecturer: Lorenzo Rosasco Scribe: Hooyoung Chung Introduction In this lecture we introduce a class of learning algorithms,

More information

Local regression, intrinsic dimension, and nonparametric sparsity

Local regression, intrinsic dimension, and nonparametric sparsity Local regression, intrinsic dimension, and nonparametric sparsity Samory Kpotufe Toyota Technological Institute - Chicago and Max Planck Institute for Intelligent Systems I. Local regression and (local)

More information

Nonlinear Dimensionality Reduction

Nonlinear Dimensionality Reduction Outline Hong Chang Institute of Computing Technology, Chinese Academy of Sciences Machine Learning Methods (Fall 2012) Outline Outline I 1 Kernel PCA 2 Isomap 3 Locally Linear Embedding 4 Laplacian Eigenmap

More information

Linear Regression. Aarti Singh. Machine Learning / Sept 27, 2010

Linear Regression. Aarti Singh. Machine Learning / Sept 27, 2010 Linear Regression Aarti Singh Machine Learning 10-701/15-781 Sept 27, 2010 Discrete to Continuous Labels Classification Sports Science News Anemic cell Healthy cell Regression X = Document Y = Topic X

More information

IFT LAPLACIAN APPLICATIONS. Mikhail Bessmeltsev

IFT LAPLACIAN APPLICATIONS.   Mikhail Bessmeltsev IFT 6112 09 LAPLACIAN APPLICATIONS http://www-labs.iro.umontreal.ca/~bmpix/teaching/6112/2018/ Mikhail Bessmeltsev Rough Intuition http://pngimg.com/upload/hammer_png3886.png You can learn a lot about

More information

Hou, Ch. et al. IEEE Transactions on Neural Networks March 2011

Hou, Ch. et al. IEEE Transactions on Neural Networks March 2011 Hou, Ch. et al. IEEE Transactions on Neural Networks March 2011 Semi-supervised approach which attempts to incorporate partial information from unlabeled data points Semi-supervised approach which attempts

More information

Certifying the Global Optimality of Graph Cuts via Semidefinite Programming: A Theoretic Guarantee for Spectral Clustering

Certifying the Global Optimality of Graph Cuts via Semidefinite Programming: A Theoretic Guarantee for Spectral Clustering Certifying the Global Optimality of Graph Cuts via Semidefinite Programming: A Theoretic Guarantee for Spectral Clustering Shuyang Ling Courant Institute of Mathematical Sciences, NYU Aug 13, 2018 Joint

More information

Advances in Manifold Learning Presented by: Naku Nak l Verm r a June 10, 2008

Advances in Manifold Learning Presented by: Naku Nak l Verm r a June 10, 2008 Advances in Manifold Learning Presented by: Nakul Verma June 10, 008 Outline Motivation Manifolds Manifold Learning Random projection of manifolds for dimension reduction Introduction to random projections

More information

DIMENSIONALITY REDUCTION METHODS IN INDEPENDENT SUBSPACE ANALYSIS FOR SIGNAL DETECTION. Mijail Guillemard, Armin Iske, Sara Krause-Solberg

DIMENSIONALITY REDUCTION METHODS IN INDEPENDENT SUBSPACE ANALYSIS FOR SIGNAL DETECTION. Mijail Guillemard, Armin Iske, Sara Krause-Solberg DIMENSIONALIY EDUCION MEHODS IN INDEPENDEN SUBSPACE ANALYSIS FO SIGNAL DEECION Mijail Guillemard, Armin Iske, Sara Krause-Solberg Department of Mathematics, University of Hamburg, {guillemard, iske, krause-solberg}@math.uni-hamburg.de

More information

Non-linear Dimensionality Reduction

Non-linear Dimensionality Reduction Non-linear Dimensionality Reduction CE-725: Statistical Pattern Recognition Sharif University of Technology Spring 2013 Soleymani Outline Introduction Laplacian Eigenmaps Locally Linear Embedding (LLE)

More information

EECS 275 Matrix Computation

EECS 275 Matrix Computation EECS 275 Matrix Computation Ming-Hsuan Yang Electrical Engineering and Computer Science University of California at Merced Merced, CA 95344 http://faculty.ucmerced.edu/mhyang Lecture 23 1 / 27 Overview

More information

A Statistical Look at Spectral Graph Analysis. Deep Mukhopadhyay

A Statistical Look at Spectral Graph Analysis. Deep Mukhopadhyay A Statistical Look at Spectral Graph Analysis Deep Mukhopadhyay Department of Statistics, Temple University Office: Speakman 335 deep@temple.edu http://sites.temple.edu/deepstat/ Graph Signal Processing

More information

Graphs in Machine Learning

Graphs in Machine Learning Graphs in Machine Learning Michal Valko Inria Lille - Nord Europe, France TA: Pierre Perrault Partially based on material by: Mikhail Belkin, Jerry Zhu, Olivier Chapelle, Branislav Kveton October 30, 2017

More information

Laplacian Eigenmaps for Dimensionality Reduction and Data Representation

Laplacian Eigenmaps for Dimensionality Reduction and Data Representation Laplacian Eigenmaps for Dimensionality Reduction and Data Representation Neural Computation, June 2003; 15 (6):1373-1396 Presentation for CSE291 sp07 M. Belkin 1 P. Niyogi 2 1 University of Chicago, Department

More information

Statistical learning on graphs

Statistical learning on graphs Statistical learning on graphs Jean-Philippe Vert Jean-Philippe.Vert@ensmp.fr ParisTech, Ecole des Mines de Paris Institut Curie INSERM U900 Seminar of probabilities, Institut Joseph Fourier, Grenoble,

More information

Mathematical Methods in Machine Learning

Mathematical Methods in Machine Learning UMD, Spring 2016 Outline Lecture 2: Role of Directionality 1 Lecture 2: Role of Directionality Anisotropic Harmonic Analysis Harmonic analysis decomposes signals into simpler elements called analyzing

More information

Manifold Denoising. Matthias Hein Markus Maier Max Planck Institute for Biological Cybernetics Tübingen, Germany

Manifold Denoising. Matthias Hein Markus Maier Max Planck Institute for Biological Cybernetics Tübingen, Germany Manifold Denoising Matthias Hein Markus Maier Max Planck Institute for Biological Cybernetics Tübingen, Germany {first.last}@tuebingen.mpg.de Abstract We consider the problem of denoising a noisily sampled

More information

Improved Graph Laplacian via Geometric Consistency

Improved Graph Laplacian via Geometric Consistency Improved Graph Laplacian via Geometric Consistency Dominique C. Perrault-Joncas Google, Inc. dominiquep@google.com Marina Meilă Department of Statistics University of Washington mmp2@uw.edu James McQueen

More information

Machine Learning with Quantum-Inspired Tensor Networks

Machine Learning with Quantum-Inspired Tensor Networks Machine Learning with Quantum-Inspired Tensor Networks E.M. Stoudenmire and David J. Schwab Advances in Neural Information Processing 29 arxiv:1605.05775 RIKEN AICS - Mar 2017 Collaboration with David

More information

LOCALITY PRESERVING HASHING. Electrical Engineering and Computer Science University of California, Merced Merced, CA 95344, USA

LOCALITY PRESERVING HASHING. Electrical Engineering and Computer Science University of California, Merced Merced, CA 95344, USA LOCALITY PRESERVING HASHING Yi-Hsuan Tsai Ming-Hsuan Yang Electrical Engineering and Computer Science University of California, Merced Merced, CA 95344, USA ABSTRACT The spectral hashing algorithm relaxes

More information

Data dependent operators for the spatial-spectral fusion problem

Data dependent operators for the spatial-spectral fusion problem Data dependent operators for the spatial-spectral fusion problem Wien, December 3, 2012 Joint work with: University of Maryland: J. J. Benedetto, J. A. Dobrosotskaya, T. Doster, K. W. Duke, M. Ehler, A.

More information

Diffusion on a Tensor Product Graph for Semi-Supervised Learning and Interactive Image Segmentation

Diffusion on a Tensor Product Graph for Semi-Supervised Learning and Interactive Image Segmentation Diffusion on a Tensor Product Graph for Semi-Supervised Learning and Interactive Image Segmentation Xingwei Yang Department of Computer and Information Sciences Temple University Philadelphia, PA 19122,

More information

A Multiscale Framework for Markov Decision Processes using Diffusion Wavelets

A Multiscale Framework for Markov Decision Processes using Diffusion Wavelets A Multiscale Framework for Markov Decision Processes using Diffusion Wavelets Mauro Maggioni Program in Applied Mathematics Department of Mathematics Yale University New Haven, CT 6 mauro.maggioni@yale.edu

More information

Risi Kondor, The University of Chicago. Nedelina Teneva UChicago. Vikas K Garg TTI-C, MIT

Risi Kondor, The University of Chicago. Nedelina Teneva UChicago. Vikas K Garg TTI-C, MIT Risi Kondor, The University of Chicago Nedelina Teneva UChicago Vikas K Garg TTI-C, MIT Risi Kondor, The University of Chicago Nedelina Teneva UChicago Vikas K Garg TTI-C, MIT {(x 1, y 1 ), (x 2, y 2 ),,

More information

Semi-Supervised Learning with Graphs. Xiaojin (Jerry) Zhu School of Computer Science Carnegie Mellon University

Semi-Supervised Learning with Graphs. Xiaojin (Jerry) Zhu School of Computer Science Carnegie Mellon University Semi-Supervised Learning with Graphs Xiaojin (Jerry) Zhu School of Computer Science Carnegie Mellon University 1 Semi-supervised Learning classification classifiers need labeled data to train labeled data

More information

Spectral Hashing: Learning to Leverage 80 Million Images

Spectral Hashing: Learning to Leverage 80 Million Images Spectral Hashing: Learning to Leverage 80 Million Images Yair Weiss, Antonio Torralba, Rob Fergus Hebrew University, MIT, NYU Outline Motivation: Brute Force Computer Vision. Semantic Hashing. Spectral

More information

Graph Metrics and Dimension Reduction

Graph Metrics and Dimension Reduction Graph Metrics and Dimension Reduction Minh Tang 1 Michael Trosset 2 1 Applied Mathematics and Statistics The Johns Hopkins University 2 Department of Statistics Indiana University, Bloomington November

More information

Distance Metric Learning in Data Mining (Part II) Fei Wang and Jimeng Sun IBM TJ Watson Research Center

Distance Metric Learning in Data Mining (Part II) Fei Wang and Jimeng Sun IBM TJ Watson Research Center Distance Metric Learning in Data Mining (Part II) Fei Wang and Jimeng Sun IBM TJ Watson Research Center 1 Outline Part I - Applications Motivation and Introduction Patient similarity application Part II

More information

An Iterated Graph Laplacian Approach for Ranking on Manifolds

An Iterated Graph Laplacian Approach for Ranking on Manifolds An Iterated Graph Laplacian Approach for Ranking on Manifolds Xueyuan Zhou Department of Computer Science University of Chicago Chicago, IL zhouxy@cs.uchicago.edu Mikhail Belkin Department of Computer

More information

Is Manifold Learning for Toy Data only?

Is Manifold Learning for Toy Data only? s Manifold Learning for Toy Data only? Marina Meilă University of Washington mmp@stat.washington.edu MMDS Workshop 2016 Outline What is non-linear dimension reduction? Metric Manifold Learning Estimating

More information

Filtering via a Reference Set. A.Haddad, D. Kushnir, R.R. Coifman Technical Report YALEU/DCS/TR-1441 February 21, 2011

Filtering via a Reference Set. A.Haddad, D. Kushnir, R.R. Coifman Technical Report YALEU/DCS/TR-1441 February 21, 2011 Patch-based de-noising algorithms and patch manifold smoothing have emerged as efficient de-noising methods. This paper provides a new insight on these methods, such as the Non Local Means or the image

More information

Overview of Statistical Tools. Statistical Inference. Bayesian Framework. Modeling. Very simple case. Things are usually more complicated

Overview of Statistical Tools. Statistical Inference. Bayesian Framework. Modeling. Very simple case. Things are usually more complicated Fall 3 Computer Vision Overview of Statistical Tools Statistical Inference Haibin Ling Observation inference Decision Prior knowledge http://www.dabi.temple.edu/~hbling/teaching/3f_5543/index.html Bayesian

More information

Integrating Global and Local Structures: A Least Squares Framework for Dimensionality Reduction

Integrating Global and Local Structures: A Least Squares Framework for Dimensionality Reduction Integrating Global and Local Structures: A Least Squares Framework for Dimensionality Reduction Jianhui Chen, Jieping Ye Computer Science and Engineering Department Arizona State University {jianhui.chen,

More information

Detecting Sparse Structures in Data in Sub-Linear Time: A group testing approach

Detecting Sparse Structures in Data in Sub-Linear Time: A group testing approach Detecting Sparse Structures in Data in Sub-Linear Time: A group testing approach Boaz Nadler The Weizmann Institute of Science Israel Joint works with Inbal Horev, Ronen Basri, Meirav Galun and Ery Arias-Castro

More information

Scalable machine learning for massive datasets: Fast summation algorithms

Scalable machine learning for massive datasets: Fast summation algorithms Scalable machine learning for massive datasets: Fast summation algorithms Getting good enough solutions as fast as possible Vikas Chandrakant Raykar vikas@cs.umd.edu University of Maryland, CollegePark

More information

Deep Learning: Approximation of Functions by Composition

Deep Learning: Approximation of Functions by Composition Deep Learning: Approximation of Functions by Composition Zuowei Shen Department of Mathematics National University of Singapore Outline 1 A brief introduction of approximation theory 2 Deep learning: approximation

More information

Improved Local Coordinate Coding using Local Tangents

Improved Local Coordinate Coding using Local Tangents Improved Local Coordinate Coding using Local Tangents Kai Yu NEC Laboratories America, 10081 N. Wolfe Road, Cupertino, CA 95129 Tong Zhang Rutgers University, 110 Frelinghuysen Road, Piscataway, NJ 08854

More information