MATH 829: Introduction to Data Mining and Analysis Clustering II
|
|
- Verity Parsons
- 6 years ago
- Views:
Transcription
1 his lecture is based on U. von Luxburg, A Tutorial on Spectral Clustering, Statistics and Computing, 17 (4), MATH 829: Introduction to Data Mining and Analysis Clustering II Dominique Guillot Departments of Mathematical Sciences University of Delaware April 27, 2016
2 Spectral clustering: overview 2/11 In the previous lecture, we discussed how K-means can be used to cluster points in R p. Spectral clustering: Very popular clustering method. Often outperforms other methods such as K-means. Can be used for various types of data (not only points in R p ). Easy to implement. Only uses basic linear algebra.
3 Spectral clustering: overview 2/11 In the previous lecture, we discussed how K-means can be used to cluster points in R p. Spectral clustering: Very popular clustering method. Often outperforms other methods such as K-means. Can be used for various types of data (not only points in R p ). Easy to implement. Only uses basic linear algebra. Overview of spectral clustering:
4 Spectral clustering: overview 2/11 In the previous lecture, we discussed how K-means can be used to cluster points in R p. Spectral clustering: Very popular clustering method. Often outperforms other methods such as K-means. Can be used for various types of data (not only points in R p ). Easy to implement. Only uses basic linear algebra. Overview of spectral clustering: 1 Construct a similarity matrix measuring the similarity of pairs of objects.
5 Spectral clustering: overview 2/11 In the previous lecture, we discussed how K-means can be used to cluster points in R p. Spectral clustering: Very popular clustering method. Often outperforms other methods such as K-means. Can be used for various types of data (not only points in R p ). Easy to implement. Only uses basic linear algebra. Overview of spectral clustering: 1 Construct a similarity matrix measuring the similarity of pairs of objects. 2 Use the similarity matrix to construct a (weighted or unweighted) graph.
6 Spectral clustering: overview 2/11 In the previous lecture, we discussed how K-means can be used to cluster points in R p. Spectral clustering: Very popular clustering method. Often outperforms other methods such as K-means. Can be used for various types of data (not only points in R p ). Easy to implement. Only uses basic linear algebra. Overview of spectral clustering: 1 Construct a similarity matrix measuring the similarity of pairs of objects. 2 Use the similarity matrix to construct a (weighted or unweighted) graph. 3 Compute eigenvectors of the graph Laplacian.
7 Spectral clustering: overview 2/11 In the previous lecture, we discussed how K-means can be used to cluster points in R p. Spectral clustering: Very popular clustering method. Often outperforms other methods such as K-means. Can be used for various types of data (not only points in R p ). Easy to implement. Only uses basic linear algebra. Overview of spectral clustering: 1 Construct a similarity matrix measuring the similarity of pairs of objects. 2 Use the similarity matrix to construct a (weighted or unweighted) graph. 3 Compute eigenvectors of the graph Laplacian. 4 Cluster the graph using the eigenvectors of the graph Laplacian using the K-means algorithm.
8 Notation 3/11 We will use the following notation/conventions: G = (V, E) a graph with vertex set V = {v 1,..., v n } and edge set E V V.
9 Notation 3/11 We will use the following notation/conventions: G = (V, E) a graph with vertex set V = {v 1,..., v n } and edge set E V V. Each edge carries a weight w ij 0.
10 Notation 3/11 We will use the following notation/conventions: G = (V, E) a graph with vertex set V = {v 1,..., v n } and edge set E V V. Each edge carries a weight w ij 0. The adjacency matrix of G is W = W G = (w ij ) n i,j=1. We will assume W is symmetric (undirected graphs).
11 Notation 3/11 We will use the following notation/conventions: G = (V, E) a graph with vertex set V = {v 1,..., v n } and edge set E V V. Each edge carries a weight w ij 0. The adjacency matrix of G is W = W G = (w ij ) n i,j=1. We will assume W is symmetric (undirected graphs). The degree of v i is d i := n w ij. j=1
12 Notation 3/11 We will use the following notation/conventions: G = (V, E) a graph with vertex set V = {v 1,..., v n } and edge set E V V. Each edge carries a weight w ij 0. The adjacency matrix of G is W = W G = (w ij ) n i,j=1. We will assume W is symmetric (undirected graphs). The degree of v i is d i := n w ij. j=1 The degree matrix of G is D := diag(d 1,..., d n ).
13 Notation 3/11 We will use the following notation/conventions: G = (V, E) a graph with vertex set V = {v 1,..., v n } and edge set E V V. Each edge carries a weight w ij 0. The adjacency matrix of G is W = W G = (w ij ) n i,j=1. We will assume W is symmetric (undirected graphs). The degree of v i is d i := n w ij. j=1 The degree matrix of G is D := diag(d 1,..., d n ). We denote the complement of A V by A.
14 Notation 3/11 We will use the following notation/conventions: G = (V, E) a graph with vertex set V = {v 1,..., v n } and edge set E V V. Each edge carries a weight w ij 0. The adjacency matrix of G is W = W G = (w ij ) n i,j=1. We will assume W is symmetric (undirected graphs). The degree of v i is d i := n w ij. j=1 The degree matrix of G is D := diag(d 1,..., d n ). We denote the complement of A V by A. If A V, then we let 1 A = (f 1,..., f n ) T R n, where f i = 1 if v i A and 0 otherwise.
15 Similarity graphs 4/11 We assume we are given a measure of similarity s between data points x 1,..., x n X : s : X X [0, ).
16 Similarity graphs 4/11 We assume we are given a measure of similarity s between data points x 1,..., x n X : s : X X [0, ). We denote by s ij := s(x i, x j ) the measure of similarity between x i and x j.
17 Similarity graphs 4/11 We assume we are given a measure of similarity s between data points x 1,..., x n X : s : X X [0, ). We denote by s ij := s(x i, x j ) the measure of similarity between x i and x j. Equivalently, we may assume we have a measure of distance between data points (e.g. (X, d) is a metric space).
18 Similarity graphs 4/11 We assume we are given a measure of similarity s between data points x 1,..., x n X : s : X X [0, ). We denote by s ij := s(x i, x j ) the measure of similarity between x i and x j. Equivalently, we may assume we have a measure of distance between data points (e.g. (X, d) is a metric space). Let d ij := d(x i, x j ), the distance between x i and x j.
19 Similarity graphs 4/11 We assume we are given a measure of similarity s between data points x 1,..., x n X : s : X X [0, ). We denote by s ij := s(x i, x j ) the measure of similarity between x i and x j. Equivalently, we may assume we have a measure of distance between data points (e.g. (X, d) is a metric space). Let d ij := d(x i, x j ), the distance between x i and x j. From d ij (or s ij ), we naturally build a similarity graph.
20 Similarity graphs 4/11 We assume we are given a measure of similarity s between data points x 1,..., x n X : s : X X [0, ). We denote by s ij := s(x i, x j ) the measure of similarity between x i and x j. Equivalently, we may assume we have a measure of distance between data points (e.g. (X, d) is a metric space). Let d ij := d(x i, x j ), the distance between x i and x j. From d ij (or s ij ), we naturally build a similarity graph. We will discuss 3 popular ways of building a similarity graph.
21 Similarity graphs (cont.) 5/11 Vertex set = {v 1,..., v n } where n is the number of data points.
22 Similarity graphs (cont.) 5/11 Vertex set = {v 1,..., v n } where n is the number of data points. 1 The ɛ-neighborhood graph: Connect all points whose pairwise distances are smaller than some ɛ > 0. We usually don't weight the edges. The graph is thus a simple graph (unweighted, undirected graph containing no loops or multiple edges).
23 5/11 Similarity graphs (cont.) Vertex set = {v 1,..., v n } where n is the number of data points. 1 The ɛ-neighborhood graph: Connect all points whose pairwise distances are smaller than some ɛ > 0. We usually don't weight the edges. The graph is thus a simple graph (unweighted, undirected graph containing no loops or multiple edges). 2 The k-nearest neighbor graph: The goal is to connect v i to v j if x j is among the k nearest neighbords of x i. However, this leads to a directed graph. We therefore dene:
24 5/11 Similarity graphs (cont.) Vertex set = {v 1,..., v n } where n is the number of data points. 1 The ɛ-neighborhood graph: Connect all points whose pairwise distances are smaller than some ɛ > 0. We usually don't weight the edges. The graph is thus a simple graph (unweighted, undirected graph containing no loops or multiple edges). 2 The k-nearest neighbor graph: The goal is to connect v i to v j if x j is among the k nearest neighbords of x i. However, this leads to a directed graph. We therefore dene: the k-nearest neighbor graph: v i is adjacent to v j i x j is among the k nearest neighbords of x i OR x i is among the k nearest neighbords of x j.
25 5/11 Similarity graphs (cont.) Vertex set = {v 1,..., v n } where n is the number of data points. 1 The ɛ-neighborhood graph: Connect all points whose pairwise distances are smaller than some ɛ > 0. We usually don't weight the edges. The graph is thus a simple graph (unweighted, undirected graph containing no loops or multiple edges). 2 The k-nearest neighbor graph: The goal is to connect v i to v j if x j is among the k nearest neighbords of x i. However, this leads to a directed graph. We therefore dene: the k-nearest neighbor graph: v i is adjacent to v j i x j is among the k nearest neighbords of x i OR x i is among the k nearest neighbords of x j. the mutual k-nearest neighbor graph: v i is adjacent to v j i x j is among the k nearest neighbords of x i AND x i is among the k nearest neighbors of x j.
26 5/11 Similarity graphs (cont.) Vertex set = {v 1,..., v n } where n is the number of data points. 1 The ɛ-neighborhood graph: Connect all points whose pairwise distances are smaller than some ɛ > 0. We usually don't weight the edges. The graph is thus a simple graph (unweighted, undirected graph containing no loops or multiple edges). 2 The k-nearest neighbor graph: The goal is to connect v i to v j if x j is among the k nearest neighbords of x i. However, this leads to a directed graph. We therefore dene: the k-nearest neighbor graph: v i is adjacent to v j i x j is among the k nearest neighbords of x i OR x i is among the k nearest neighbords of x j. the mutual k-nearest neighbor graph: v i is adjacent to v j i x j is among the k nearest neighbords of x i AND x i is among the k nearest neighbors of x j. We weight the edges by the similarity of their endpoints.
27 Similarity graphs (cont.) 6/11 3 The fully connected graph: Connect all points with edge weights s ij.
28 Similarity graphs (cont.) 6/11 3 The fully connected graph: Connect all points with edge weights s ij. For example, one could use the Gaussian similarity function to represent a local neighborhood relationships: s ij = s(x i, x j ) = exp( x i x j 2 /(2σ 2 )) (σ 2 > 0). Note: σ 2 controls the width of the neighborhoods.
29 Similarity graphs (cont.) 6/11 3 The fully connected graph: Connect all points with edge weights s ij. For example, one could use the Gaussian similarity function to represent a local neighborhood relationships: s ij = s(x i, x j ) = exp( x i x j 2 /(2σ 2 )) (σ 2 > 0). Note: σ 2 controls the width of the neighborhoods. All graphs mentioned above are regularly used in spectral clustering.
30 Graph Laplacians 7/11 There are three commonly used denitions of the graph Laplacian: 1 The unnormalized Laplacian is L := D W.
31 Graph Laplacians 7/11 There are three commonly used denitions of the graph Laplacian: 1 The unnormalized Laplacian is L := D W. 2 The normalized symmetric Laplacian is L sym := D 1/2 LD 1/2 = I D 1/2 W D 1/2.
32 Graph Laplacians 7/11 There are three commonly used denitions of the graph Laplacian: 1 The unnormalized Laplacian is L := D W. 2 The normalized symmetric Laplacian is L sym := D 1/2 LD 1/2 = I D 1/2 W D 1/2. 3 The normalized random walk Laplacian is L rw := D 1 L = I D 1 W.
33 Graph Laplacians 7/11 There are three commonly used denitions of the graph Laplacian: 1 The unnormalized Laplacian is L := D W. 2 The normalized symmetric Laplacian is L sym := D 1/2 LD 1/2 = I D 1/2 W D 1/2. 3 The normalized random walk Laplacian is L rw := D 1 L = I D 1 W. We begin by studying properties of the unnormalized Laplacian.
34 The unnormalized Laplacian 8/11 Proposition: The matrix L satises the following properties:
35 The unnormalized Laplacian 8/11 Proposition: The matrix L satises the following properties: 1 For any f R n : f T Lf = 1 n w ij (f i f j ) 2. 2 i,j=1
36 The unnormalized Laplacian 8/11 Proposition: The matrix L satises the following properties: 1 For any f R n : f T Lf = 1 n w ij (f i f j ) 2. 2 i,j=1 2 L is symmetric and positive semidenite.
37 The unnormalized Laplacian 8/11 Proposition: The matrix L satises the following properties: 1 For any f R n : f T Lf = 1 n w ij (f i f j ) 2. 2 i,j=1 2 L is symmetric and positive semidenite. 3 0 is an eigenvalue of L with associated constant eigenvector 1.
38 The unnormalized Laplacian 8/11 Proposition: The matrix L satises the following properties: 1 For any f R n : f T Lf = 1 n w ij (f i f j ) 2. 2 i,j=1 2 L is symmetric and positive semidenite. 3 0 is an eigenvalue of L with associated constant eigenvector 1. Proof:
39 The unnormalized Laplacian (2) follows from (1). (3) is easy. 8/11 Proposition: The matrix L satises the following properties: 1 For any f R n : f T Lf = 1 n w ij (f i f j ) 2. 2 i,j=1 2 L is symmetric and positive semidenite. 3 0 is an eigenvalue of L with associated constant eigenvector 1. Proof: To prove (1), f T Lf = f T Df f T W f = n d i fi 2 i=1 n w ij f i f j i,j=1 = 1 n d i fi = 1 2 i=1 n w ij f i f j + i,j=1 n w ij (f i f j ) 2. i,j=1 n d j fj 2 j=1
40 The unnormalized Laplacian (cont.) 9/11 Proposition: Let G be an undirected graph with non-negative weights. Then:
41 The unnormalized Laplacian (cont.) 9/11 Proposition: Let G be an undirected graph with non-negative weights. Then: 1 The multiplicity k of the eigenvalue 0 of L equals the number of connected components A 1,..., A k in the graph.
42 The unnormalized Laplacian (cont.) 9/11 Proposition: Let G be an undirected graph with non-negative weights. Then: 1 The multiplicity k of the eigenvalue 0 of L equals the number of connected components A 1,..., A k in the graph. 2 The eigenspace of eigenvalue 0 is spanned by the indicator vectors 1 A1,..., 1 Ak of those components.
43 The unnormalized Laplacian (cont.) 9/11 Proposition: Let G be an undirected graph with non-negative weights. Then: 1 The multiplicity k of the eigenvalue 0 of L equals the number of connected components A 1,..., A k in the graph. 2 The eigenspace of eigenvalue 0 is spanned by the indicator vectors 1 A1,..., 1 Ak of those components. Proof:
44 The unnormalized Laplacian (cont.) 9/11 Proposition: Let G be an undirected graph with non-negative weights. Then: 1 The multiplicity k of the eigenvalue 0 of L equals the number of connected components A 1,..., A k in the graph. 2 The eigenspace of eigenvalue 0 is spanned by the indicator vectors 1 A1,..., 1 Ak of those components. Proof: If f is an eigenvector associate to λ = 0, then 0 = f T Lf = n w ij (f i f j ) 2. i,j=1 It follows that f i = f j whenever w ij > 0. Thus f is constant on the connected components of G. We conclude that the eigenspace of 0 is contained in span(1 A1,..., 1 Ak ). Conversely, it is not hard to see that each 1 Ai is an eigenvector associated to 0 (write L in block diagonal form).
45 The normalized Laplacians 10/11 Proposition: The normalized Laplacians satisfy the following properties: 1 For every f R n, we have f T L sym f = 1 2 ( ) 2 n f i w ij f j. di dj i,j=1 2 λ is an eigenvalue of L rw with eigenvector u if and only if λ is an eigenvalue of L sym with eigenvector w = D 1/2 u. 3 λ is an eigenvalue of L rw with eigenvector u if and only if λ and u solve the generalized eigenproblem Lu = λdu. Proof: The proof of (1) is similar to the proof of the analogous result for the unnormalized Laplacian. (2) and (3) follow easily by using appropriate rescalings.
46 The normalized Laplacians (cont.) 11/11 Proposition: Let G be an undirected graph with non-negative weights. Then: 1 The multiplicity k of the eigenvalue 0 of both L sym and L rw equals the number of connected components A 1,..., A k in the graph. 2 For L rw, the eigenspace of eigenvalue 0 is spanned by the indicator vectors 1 Ai, i = 1,..., k. 3 For L sym, the eigenspace of eigenvalue 0 is spanned by the vectors D 1/2 1 Ai, i = 1,..., k. Proof: Similar to the proof of the analogous result for the unnormalized Laplacian.
MATH 567: Mathematical Techniques in Data Science Clustering II
This lecture is based on U. von Luxburg, A Tutorial on Spectral Clustering, Statistics and Computing, 17 (4), 2007. MATH 567: Mathematical Techniques in Data Science Clustering II Dominique Guillot Departments
More informationMATH 567: Mathematical Techniques in Data Science Clustering II
Spectral clustering: overview MATH 567: Mathematical Techniques in Data Science Clustering II Dominique uillot Departments of Mathematical Sciences University of Delaware Overview of spectral clustering:
More informationSpectral Clustering. Spectral Clustering? Two Moons Data. Spectral Clustering Algorithm: Bipartioning. Spectral methods
Spectral Clustering Seungjin Choi Department of Computer Science POSTECH, Korea seungjin@postech.ac.kr 1 Spectral methods Spectral Clustering? Methods using eigenvectors of some matrices Involve eigen-decomposition
More informationSpectral Clustering. Zitao Liu
Spectral Clustering Zitao Liu Agenda Brief Clustering Review Similarity Graph Graph Laplacian Spectral Clustering Algorithm Graph Cut Point of View Random Walk Point of View Perturbation Theory Point of
More informationData Analysis and Manifold Learning Lecture 3: Graphs, Graph Matrices, and Graph Embeddings
Data Analysis and Manifold Learning Lecture 3: Graphs, Graph Matrices, and Graph Embeddings Radu Horaud INRIA Grenoble Rhone-Alpes, France Radu.Horaud@inrialpes.fr http://perception.inrialpes.fr/ Outline
More informationAn Algorithmist s Toolkit September 10, Lecture 1
18.409 An Algorithmist s Toolkit September 10, 2009 Lecture 1 Lecturer: Jonathan Kelner Scribe: Jesse Geneson (2009) 1 Overview The class s goals, requirements, and policies were introduced, and topics
More informationMLCC Clustering. Lorenzo Rosasco UNIGE-MIT-IIT
MLCC 2018 - Clustering Lorenzo Rosasco UNIGE-MIT-IIT About this class We will consider an unsupervised setting, and in particular the problem of clustering unlabeled data into coherent groups. MLCC 2018
More informationData Analysis and Manifold Learning Lecture 7: Spectral Clustering
Data Analysis and Manifold Learning Lecture 7: Spectral Clustering Radu Horaud INRIA Grenoble Rhone-Alpes, France Radu.Horaud@inrialpes.fr http://perception.inrialpes.fr/ Outline of Lecture 7 What is spectral
More informationMachine Learning for Data Science (CS4786) Lecture 11
Machine Learning for Data Science (CS4786) Lecture 11 Spectral clustering Course Webpage : http://www.cs.cornell.edu/courses/cs4786/2016sp/ ANNOUNCEMENT 1 Assignment P1 the Diagnostic assignment 1 will
More informationSpectral clustering. Two ideal clusters, with two points each. Spectral clustering algorithms
A simple example Two ideal clusters, with two points each Spectral clustering Lecture 2 Spectral clustering algorithms 4 2 3 A = Ideally permuted Ideal affinities 2 Indicator vectors Each cluster has an
More information8.1 Concentration inequality for Gaussian random matrix (cont d)
MGMT 69: Topics in High-dimensional Data Analysis Falll 26 Lecture 8: Spectral clustering and Laplacian matrices Lecturer: Jiaming Xu Scribe: Hyun-Ju Oh and Taotao He, October 4, 26 Outline Concentration
More informationSpectral Clustering. Guokun Lai 2016/10
Spectral Clustering Guokun Lai 2016/10 1 / 37 Organization Graph Cut Fundamental Limitations of Spectral Clustering Ng 2002 paper (if we have time) 2 / 37 Notation We define a undirected weighted graph
More informationSpectral Clustering on Handwritten Digits Database
University of Maryland-College Park Advance Scientific Computing I,II Spectral Clustering on Handwritten Digits Database Author: Danielle Middlebrooks Dmiddle1@math.umd.edu Second year AMSC Student Advisor:
More informationSpectral Clustering on Handwritten Digits Database Mid-Year Pr
Spectral Clustering on Handwritten Digits Database Mid-Year Presentation Danielle dmiddle1@math.umd.edu Advisor: Kasso Okoudjou kasso@umd.edu Department of Mathematics University of Maryland- College Park
More informationLecture 9: Laplacian Eigenmaps
Lecture 9: Radu Balan Department of Mathematics, AMSC, CSCAMM and NWC University of Maryland, College Park, MD April 18, 2017 Optimization Criteria Assume G = (V, W ) is a undirected weighted graph with
More information17.1 Directed Graphs, Undirected Graphs, Incidence Matrices, Adjacency Matrices, Weighted Graphs
Chapter 17 Graphs and Graph Laplacians 17.1 Directed Graphs, Undirected Graphs, Incidence Matrices, Adjacency Matrices, Weighted Graphs Definition 17.1. A directed graph isa pairg =(V,E), where V = {v
More informationMATH 829: Introduction to Data Mining and Analysis Graphical Models I
MATH 829: Introduction to Data Mining and Analysis Graphical Models I Dominique Guillot Departments of Mathematical Sciences University of Delaware May 2, 2016 1/12 Independence and conditional independence:
More informationSpectral Theory of Unsigned and Signed Graphs Applications to Graph Clustering. Some Slides
Spectral Theory of Unsigned and Signed Graphs Applications to Graph Clustering Some Slides Jean Gallier Department of Computer and Information Science University of Pennsylvania Philadelphia, PA 19104,
More informationLecture 12 : Graph Laplacians and Cheeger s Inequality
CPS290: Algorithmic Foundations of Data Science March 7, 2017 Lecture 12 : Graph Laplacians and Cheeger s Inequality Lecturer: Kamesh Munagala Scribe: Kamesh Munagala Graph Laplacian Maybe the most beautiful
More informationLecture 12: Introduction to Spectral Graph Theory, Cheeger s inequality
CSE 521: Design and Analysis of Algorithms I Spring 2016 Lecture 12: Introduction to Spectral Graph Theory, Cheeger s inequality Lecturer: Shayan Oveis Gharan May 4th Scribe: Gabriel Cadamuro Disclaimer:
More informationLecture 13: Spectral Graph Theory
CSE 521: Design and Analysis of Algorithms I Winter 2017 Lecture 13: Spectral Graph Theory Lecturer: Shayan Oveis Gharan 11/14/18 Disclaimer: These notes have not been subjected to the usual scrutiny reserved
More informationLecture 1: Graphs, Adjacency Matrices, Graph Laplacian
Lecture 1: Graphs, Adjacency Matrices, Graph Laplacian Radu Balan January 31, 2017 G = (V, E) An undirected graph G is given by two pieces of information: a set of vertices V and a set of edges E, G =
More informationNotes on Elementary Spectral Graph Theory Applications to Graph Clustering
Notes on Elementary Spectral Graph Theory Applications to Graph Clustering Jean Gallier Department of Computer and Information Science University of Pennsylvania Philadelphia, PA 19104, USA e-mail: jean@cis.upenn.edu
More informationCertifying the Global Optimality of Graph Cuts via Semidefinite Programming: A Theoretic Guarantee for Spectral Clustering
Certifying the Global Optimality of Graph Cuts via Semidefinite Programming: A Theoretic Guarantee for Spectral Clustering Shuyang Ling Courant Institute of Mathematical Sciences, NYU Aug 13, 2018 Joint
More information1 Matrix notation and preliminaries from spectral graph theory
Graph clustering (or community detection or graph partitioning) is one of the most studied problems in network analysis. One reason for this is that there are a variety of ways to define a cluster or community.
More information1 Matrix notation and preliminaries from spectral graph theory
Graph clustering (or community detection or graph partitioning) is one of the most studied problems in network analysis. One reason for this is that there are a variety of ways to define a cluster or community.
More informationNotes on Elementary Spectral Graph Theory Applications to Graph Clustering Using Normalized Cuts
University of Pennsylvania ScholarlyCommons Technical Reports (CIS) Department of Computer & Information Science 1-1-2013 Notes on Elementary Spectral Graph Theory Applications to Graph Clustering Using
More informationMATH 829: Introduction to Data Mining and Analysis Support vector machines and kernels
1/12 MATH 829: Introduction to Data Mining and Analysis Support vector machines and kernels Dominique Guillot Departments of Mathematical Sciences University of Delaware March 14, 2016 Separating sets:
More informationA lower bound for the Laplacian eigenvalues of a graph proof of a conjecture by Guo
A lower bound for the Laplacian eigenvalues of a graph proof of a conjecture by Guo A. E. Brouwer & W. H. Haemers 2008-02-28 Abstract We show that if µ j is the j-th largest Laplacian eigenvalue, and d
More information18.312: Algebraic Combinatorics Lionel Levine. Lecture 19
832: Algebraic Combinatorics Lionel Levine Lecture date: April 2, 20 Lecture 9 Notes by: David Witmer Matrix-Tree Theorem Undirected Graphs Let G = (V, E) be a connected, undirected graph with n vertices,
More informationMarch 13, Paper: R.R. Coifman, S. Lafon, Diffusion maps ([Coifman06]) Seminar: Learning with Graphs, Prof. Hein, Saarland University
Kernels March 13, 2008 Paper: R.R. Coifman, S. Lafon, maps ([Coifman06]) Seminar: Learning with Graphs, Prof. Hein, Saarland University Kernels Figure: Example Application from [LafonWWW] meaningful geometric
More informationSpectral Theory of Unsigned and Signed Graphs Applications to Graph Clustering: a Survey
Spectral Theory of Unsigned and Signed Graphs Applications to Graph Clustering: a Survey Jean Gallier Department of Computer and Information Science University of Pennsylvania Philadelphia, PA 19104, USA
More informationLearning from Sensor Data: Set II. Behnaam Aazhang J.S. Abercombie Professor Electrical and Computer Engineering Rice University
Learning from Sensor Data: Set II Behnaam Aazhang J.S. Abercombie Professor Electrical and Computer Engineering Rice University 1 6. Data Representation The approach for learning from data Probabilistic
More informationClustering using Mixture Models
Clustering using Mixture Models The full posterior of the Gaussian Mixture Model is p(x, Z, µ,, ) =p(x Z, µ, )p(z )p( )p(µ, ) data likelihood (Gaussian) correspondence prob. (Multinomial) mixture prior
More informationAn Introduction to Spectral Graph Theory
An Introduction to Spectral Graph Theory Mackenzie Wheeler Supervisor: Dr. Gary MacGillivray University of Victoria WheelerM@uvic.ca Outline Outline 1. How many walks are there from vertices v i to v j
More informationCS168: The Modern Algorithmic Toolbox Lectures #11 and #12: Spectral Graph Theory
CS168: The Modern Algorithmic Toolbox Lectures #11 and #12: Spectral Graph Theory Tim Roughgarden & Gregory Valiant May 2, 2016 Spectral graph theory is the powerful and beautiful theory that arises from
More informationDiffusion and random walks on graphs
Diffusion and random walks on graphs Leonid E. Zhukov School of Data Analysis and Artificial Intelligence Department of Computer Science National Research University Higher School of Economics Structural
More informationLaplacian Eigenmaps for Dimensionality Reduction and Data Representation
Laplacian Eigenmaps for Dimensionality Reduction and Data Representation Neural Computation, June 2003; 15 (6):1373-1396 Presentation for CSE291 sp07 M. Belkin 1 P. Niyogi 2 1 University of Chicago, Department
More informationNonlinear Methods. Data often lies on or near a nonlinear low-dimensional curve aka manifold.
Nonlinear Methods Data often lies on or near a nonlinear low-dimensional curve aka manifold. 27 Laplacian Eigenmaps Linear methods Lower-dimensional linear projection that preserves distances between all
More informationLecture 10: Dimension Reduction Techniques
Lecture 10: Dimension Reduction Techniques Radu Balan Department of Mathematics, AMSC, CSCAMM and NWC University of Maryland, College Park, MD April 17, 2018 Input Data It is assumed that there is a set
More informationComputer Vision Group Prof. Daniel Cremers. 14. Clustering
Group Prof. Daniel Cremers 14. Clustering Motivation Supervised learning is good for interaction with humans, but labels from a supervisor are hard to obtain Clustering is unsupervised learning, i.e. it
More informationStatistical Learning Notes III-Section 2.4.3
Statistical Learning Notes III-Section 2.4.3 "Graphical" Spectral Features Stephen Vardeman Analytics Iowa LLC January 2019 Stephen Vardeman (Analytics Iowa LLC) Statistical Learning Notes III-Section
More informationTwo Laplacians for the distance matrix of a graph
Two Laplacians for the distance matrix of a graph Mustapha Aouchiche and Pierre Hansen GERAD and HEC Montreal, Canada CanaDAM 2013, Memorial University, June 1013 Aouchiche & Hansen CanaDAM 2013, Memorial
More informationLecture 13 Spectral Graph Algorithms
COMS 995-3: Advanced Algorithms March 6, 7 Lecture 3 Spectral Graph Algorithms Instructor: Alex Andoni Scribe: Srikar Varadaraj Introduction Today s topics: Finish proof from last lecture Example of random
More informationGraph fundamentals. Matrices associated with a graph
Graph fundamentals Matrices associated with a graph Drawing a picture of a graph is one way to represent it. Another type of representation is via a matrix. Let G be a graph with V (G) ={v 1,v,...,v n
More informationDiffuse interface methods on graphs: Data clustering and Gamma-limits
Diffuse interface methods on graphs: Data clustering and Gamma-limits Yves van Gennip joint work with Andrea Bertozzi, Jeff Brantingham, Blake Hunter Department of Mathematics, UCLA Research made possible
More informationMATH 829: Introduction to Data Mining and Analysis Computing the lasso solution
1/16 MATH 829: Introduction to Data Mining and Analysis Computing the lasso solution Dominique Guillot Departments of Mathematical Sciences University of Delaware February 26, 2016 Computing the lasso
More informationUnsupervised dimensionality reduction
Unsupervised dimensionality reduction Guillaume Obozinski Ecole des Ponts - ParisTech SOCN course 2014 Guillaume Obozinski Unsupervised dimensionality reduction 1/30 Outline 1 PCA 2 Kernel PCA 3 Multidimensional
More informationData Mining and Analysis: Fundamental Concepts and Algorithms
: Fundamental Concepts and Algorithms dataminingbook.info Mohammed J. Zaki 1 Wagner Meira Jr. 2 1 Department of Computer Science Rensselaer Polytechnic Institute, Troy, NY, USA 2 Department of Computer
More informationPhysical Metaphors for Graphs
Graphs and Networks Lecture 3 Physical Metaphors for Graphs Daniel A. Spielman October 9, 203 3. Overview We will examine physical metaphors for graphs. We begin with a spring model and then discuss resistor
More informationMATH 829: Introduction to Data Mining and Analysis Graphical Models II - Gaussian Graphical Models
1/13 MATH 829: Introduction to Data Mining and Analysis Graphical Models II - Gaussian Graphical Models Dominique Guillot Departments of Mathematical Sciences University of Delaware May 4, 2016 Recall
More informationGraphs in Machine Learning
Graphs in Machine Learning Michal Valko Inria Lille - Nord Europe, France TA: Pierre Perrault Partially based on material by: Ulrike von Luxburg, Gary Miller, Doyle & Schnell, Daniel Spielman October 10th,
More informationCMPSCI 791BB: Advanced ML: Laplacian Learning
CMPSCI 791BB: Advanced ML: Laplacian Learning Sridhar Mahadevan Outline! Spectral graph operators! Combinatorial graph Laplacian! Normalized graph Laplacian! Random walks! Machine learning on graphs! Clustering!
More informationMath 126 Lecture 8. Summary of results Application to the Laplacian of the cube
Math 126 Lecture 8 Summary of results Application to the Laplacian of the cube Summary of key results Every representation is determined by its character. If j is the character of a representation and
More informationSome notes on SVD, dimensionality reduction, and clustering. Guangliang Chen
Some notes on SVD, dimensionality reduction, and clustering Guangliang Chen Contents 1. Introduction 4 2. Review of necessary linear algebra 4 3. Matrix SVD and its applications 8 Practice problems set
More informationNetworks and Their Spectra
Networks and Their Spectra Victor Amelkin University of California, Santa Barbara Department of Computer Science victor@cs.ucsb.edu December 4, 2017 1 / 18 Introduction Networks (= graphs) are everywhere.
More informationEECS 275 Matrix Computation
EECS 275 Matrix Computation Ming-Hsuan Yang Electrical Engineering and Computer Science University of California at Merced Merced, CA 95344 http://faculty.ucmerced.edu/mhyang Lecture 23 1 / 27 Overview
More informationSpectral Graph Theory
Spectral Graph Theory Aaron Mishtal April 27, 2016 1 / 36 Outline Overview Linear Algebra Primer History Theory Applications Open Problems Homework Problems References 2 / 36 Outline Overview Linear Algebra
More informationStatistical Machine Learning
Statistical Machine Learning Christoph Lampert Spring Semester 2015/2016 // Lecture 12 1 / 36 Unsupervised Learning Dimensionality Reduction 2 / 36 Dimensionality Reduction Given: data X = {x 1,..., x
More informationOn the inverse matrix of the Laplacian and all ones matrix
On the inverse matrix of the Laplacian and all ones matrix Sho Suda (Joint work with Michio Seto and Tetsuji Taniguchi) International Christian University JSPS Research Fellow PD November 21, 2012 Sho
More informationGraphs in Machine Learning
Graphs in Machine Learning Michal Valko INRIA Lille - Nord Europe, France Partially based on material by: Ulrike von Luxburg, Gary Miller, Doyle & Schnell, Daniel Spielman January 27, 2015 MVA 2014/2015
More information1 Review: symmetric matrices, their eigenvalues and eigenvectors
Cornell University, Fall 2012 Lecture notes on spectral methods in algorithm design CS 6820: Algorithms Studying the eigenvalues and eigenvectors of matrices has powerful consequences for at least three
More informationHYPERGRAPH BASED SEMI-SUPERVISED LEARNING ALGORITHMS APPLIED TO SPEECH RECOGNITION PROBLEM: A NOVEL APPROACH
HYPERGRAPH BASED SEMI-SUPERVISED LEARNING ALGORITHMS APPLIED TO SPEECH RECOGNITION PROBLEM: A NOVEL APPROACH Hoang Trang 1, Tran Hoang Loc 1 1 Ho Chi Minh City University of Technology-VNU HCM, Ho Chi
More informationCOMPSCI 514: Algorithms for Data Science
COMPSCI 514: Algorithms for Data Science Arya Mazumdar University of Massachusetts at Amherst Fall 2018 Lecture 8 Spectral Clustering Spectral clustering Curse of dimensionality Dimensionality Reduction
More informationLecture: Some Practical Considerations (3 of 4)
Stat260/CS294: Spectral Graph Methods Lecture 14-03/10/2015 Lecture: Some Practical Considerations (3 of 4) Lecturer: Michael Mahoney Scribe: Michael Mahoney Warning: these notes are still very rough.
More informationUnsupervised Learning Techniques Class 07, 1 March 2006 Andrea Caponnetto
Unsupervised Learning Techniques 9.520 Class 07, 1 March 2006 Andrea Caponnetto About this class Goal To introduce some methods for unsupervised learning: Gaussian Mixtures, K-Means, ISOMAP, HLLE, Laplacian
More informationInformation Propagation Analysis of Social Network Using the Universality of Random Matrix
Information Propagation Analysis of Social Network Using the Universality of Random Matrix Yusuke Sakumoto, Tsukasa Kameyama, Chisa Takano and Masaki Aida Tokyo Metropolitan University, 6-6 Asahigaoka,
More informationMath 3191 Applied Linear Algebra
Math 9 Applied Linear Algebra Lecture 9: Diagonalization Stephen Billups University of Colorado at Denver Math 9Applied Linear Algebra p./9 Section. Diagonalization The goal here is to develop a useful
More informationData-dependent representations: Laplacian Eigenmaps
Data-dependent representations: Laplacian Eigenmaps November 4, 2015 Data Organization and Manifold Learning There are many techniques for Data Organization and Manifold Learning, e.g., Principal Component
More informationGraphic sequences, adjacency matrix
Chapter 2 Graphic sequences, adjacency matrix Definition 2.1. A sequence of integers (d 1,..., d n ) is called graphic if it is the degree sequence of a graph. Example 2.2. 1. (1, 2, 2, 3) is graphic:
More informationIntroduction to Spectral Graph Theory and Graph Clustering
Introduction to Spectral Graph Theory and Graph Clustering Chengming Jiang ECS 231 Spring 2016 University of California, Davis 1 / 40 Motivation Image partitioning in computer vision 2 / 40 Motivation
More informationRelations Between Adjacency And Modularity Graph Partitioning: Principal Component Analysis vs. Modularity Component Analysis
Relations Between Adjacency And Modularity Graph Partitioning: Principal Component Analysis vs. Modularity Component Analysis Hansi Jiang Carl Meyer North Carolina State University October 27, 2015 1 /
More informationCommunities, Spectral Clustering, and Random Walks
Communities, Spectral Clustering, and Random Walks David Bindel Department of Computer Science Cornell University 26 Sep 2011 20 21 19 16 22 28 17 18 29 26 27 30 23 1 25 5 8 24 2 4 14 3 9 13 15 11 10 12
More informationThe f-adjusted Graph Laplacian: a Diagonal Modification with a Geometric Interpretation
The f-adjusted Graph Laplacian: a Diagonal Modification with a Geometric Interpretation Sven Kurras Ulrike von Luxburg Department of Computer Science, University of Hamburg, Germany Gilles Blanchard Department
More informationMath 443/543 Graph Theory Notes 5: Graphs as matrices, spectral graph theory, and PageRank
Math 443/543 Graph Theory Notes 5: Graphs as matrices, spectral graph theory, and PageRank David Glickenstein November 3, 4 Representing graphs as matrices It will sometimes be useful to represent graphs
More informationLaplacian Eigenmaps for Dimensionality Reduction and Data Representation
Introduction and Data Representation Mikhail Belkin & Partha Niyogi Department of Electrical Engieering University of Minnesota Mar 21, 2017 1/22 Outline Introduction 1 Introduction 2 3 4 Connections to
More informationA New Spectral Technique Using Normalized Adjacency Matrices for Graph Matching 1
CHAPTER-3 A New Spectral Technique Using Normalized Adjacency Matrices for Graph Matching Graph matching problem has found many applications in areas as diverse as chemical structure analysis, pattern
More informationLecture 1: Systems of linear equations and their solutions
Lecture 1: Systems of linear equations and their solutions Course overview Topics to be covered this semester: Systems of linear equations and Gaussian elimination: Solving linear equations and applications
More informationWalks, Springs, and Resistor Networks
Spectral Graph Theory Lecture 12 Walks, Springs, and Resistor Networks Daniel A. Spielman October 8, 2018 12.1 Overview In this lecture we will see how the analysis of random walks, spring networks, and
More informationConsistency of Spectral Clustering
Max Planck Institut für biologische Kybernetik Max Planck Institute for Biological Cybernetics Technical Report No. TR-34 Consistency of Spectral Clustering Ulrike von Luxburg, Mikhail Belkin 2, Olivier
More informationLecture 7: Spectral Graph Theory II
A Theorist s Toolkit (CMU 18-859T, Fall 2013) Lecture 7: Spectral Graph Theory II September 30, 2013 Lecturer: Ryan O Donnell Scribe: Christian Tjandraatmadja 1 Review We spend a page to review the previous
More informationLinear Spectral Hashing
Linear Spectral Hashing Zalán Bodó and Lehel Csató Babeş Bolyai University - Faculty of Mathematics and Computer Science Kogălniceanu 1., 484 Cluj-Napoca - Romania Abstract. assigns binary hash keys to
More informationConsistent Manifold Representation for Topological Data Analysis
Consistent Manifold Representation for Topological Data Analysis Tyrus Berry (tberry@gmu.edu) and Timothy Sauer Dept. of Mathematical Sciences, George Mason University, Fairfax, VA 3 Abstract For data
More informationORIE 4741: Learning with Big Messy Data. Spectral Graph Theory
ORIE 4741: Learning with Big Messy Data Spectral Graph Theory Mika Sumida Operations Research and Information Engineering Cornell September 15, 2017 1 / 32 Outline Graph Theory Spectral Graph Theory Laplacian
More informationEnergy of Graphs. Sivaram K. Narayan Central Michigan University. Presented at CMU on October 10, 2015
Energy of Graphs Sivaram K. Narayan Central Michigan University Presented at CMU on October 10, 2015 1 / 32 Graphs We will consider simple graphs (no loops, no multiple edges). Let V = {v 1, v 2,..., v
More information22m:033 Notes: 7.1 Diagonalization of Symmetric Matrices
m:33 Notes: 7. Diagonalization of Symmetric Matrices Dennis Roseman University of Iowa Iowa City, IA http://www.math.uiowa.edu/ roseman May 3, Symmetric matrices Definition. A symmetric matrix is a matrix
More informationMATH 829: Introduction to Data Mining and Analysis Consistency of Linear Regression
1/9 MATH 829: Introduction to Data Mining and Analysis Consistency of Linear Regression Dominique Guillot Deartments of Mathematical Sciences University of Delaware February 15, 2016 Distribution of regression
More informationLecture: Some Statistical Inference Issues (2 of 3)
Stat260/CS294: Spectral Graph Methods Lecture 23-04/16/2015 Lecture: Some Statistical Inference Issues (2 of 3) Lecturer: Michael Mahoney Scribe: Michael Mahoney Warning: these notes are still very rough.
More informationThe University of Texas at Austin Department of Electrical and Computer Engineering. EE381V: Large Scale Learning Spring 2013.
The University of Texas at Austin Department of Electrical and Computer Engineering EE381V: Large Scale Learning Spring 2013 Assignment 1 Caramanis/Sanghavi Due: Thursday, Feb. 7, 2013. (Problems 1 and
More informationTutorial: Gaussian conditional independence and graphical models. Thomas Kahle Otto-von-Guericke Universität Magdeburg
Tutorial: Gaussian conditional independence and graphical models Thomas Kahle Otto-von-Guericke Universität Magdeburg The central dogma of algebraic statistics Statistical models are varieties The central
More informationThe Simplest Construction of Expanders
Spectral Graph Theory Lecture 14 The Simplest Construction of Expanders Daniel A. Spielman October 16, 2009 14.1 Overview I am going to present the simplest construction of expanders that I have been able
More informationDamped random walks and the characteristic polynomial of the weighted Laplacian on a graph
Damped random walks and the characteristic polynomial of the weighted Laplacian on a graph arxiv:mathpr/0506460 v1 Jun 005 MADHAV P DESAI and HARIHARAN NARAYANAN September 11, 006 Abstract For λ > 0, we
More informationMATH 304 Linear Algebra Lecture 33: Bases of eigenvectors. Diagonalization.
MATH 304 Linear Algebra Lecture 33: Bases of eigenvectors. Diagonalization. Eigenvalues and eigenvectors of an operator Definition. Let V be a vector space and L : V V be a linear operator. A number λ
More informationDimensionality Reduc1on
Dimensionality Reduc1on contd Aarti Singh Machine Learning 10-601 Nov 10, 2011 Slides Courtesy: Tom Mitchell, Eric Xing, Lawrence Saul 1 Principal Component Analysis (PCA) Principal Components are the
More informationLecture 5: Random Walks and Markov Chain
Spectral Graph Theory and Applications WS 20/202 Lecture 5: Random Walks and Markov Chain Lecturer: Thomas Sauerwald & He Sun Introduction to Markov Chains Definition 5.. A sequence of random variables
More informationAN EXAMPLE OF USING STAR COMPLEMENTS IN CLASSIFYING STRONGLY REGULAR GRAPHS
Faculty of Sciences and Mathematics, University of Niš, Serbia Available at: http://www.pmf.ni.ac.yu/filomat Filomat 22:2 (2008), 53 57 AN EXAMPLE OF USING STAR COMPLEMENTS IN CLASSIFYING STRONGLY REGULAR
More informationLecture 1 and 2: Introduction and Graph theory basics. Spring EE 194, Networked estimation and control (Prof. Khan) January 23, 2012
Lecture 1 and 2: Introduction and Graph theory basics Spring 2012 - EE 194, Networked estimation and control (Prof. Khan) January 23, 2012 Spring 2012: EE-194-02 Networked estimation and control Schedule
More informationMath 18, Linear Algebra, Lecture C00, Spring 2017 Review and Practice Problems for Final Exam
Math 8, Linear Algebra, Lecture C, Spring 7 Review and Practice Problems for Final Exam. The augmentedmatrix of a linear system has been transformed by row operations into 5 4 8. Determine if the system
More informationKernels of Directed Graph Laplacians. J. S. Caughman and J.J.P. Veerman
Kernels of Directed Graph Laplacians J. S. Caughman and J.J.P. Veerman Department of Mathematics and Statistics Portland State University PO Box 751, Portland, OR 97207. caughman@pdx.edu, veerman@pdx.edu
More informationMATH 304 Linear Algebra Lecture 34: Review for Test 2.
MATH 304 Linear Algebra Lecture 34: Review for Test 2. Topics for Test 2 Linear transformations (Leon 4.1 4.3) Matrix transformations Matrix of a linear mapping Similar matrices Orthogonality (Leon 5.1
More information