MATH 829: Introduction to Data Mining and Analysis Clustering II

Size: px
Start display at page:

Download "MATH 829: Introduction to Data Mining and Analysis Clustering II"

Transcription

1 his lecture is based on U. von Luxburg, A Tutorial on Spectral Clustering, Statistics and Computing, 17 (4), MATH 829: Introduction to Data Mining and Analysis Clustering II Dominique Guillot Departments of Mathematical Sciences University of Delaware April 27, 2016

2 Spectral clustering: overview 2/11 In the previous lecture, we discussed how K-means can be used to cluster points in R p. Spectral clustering: Very popular clustering method. Often outperforms other methods such as K-means. Can be used for various types of data (not only points in R p ). Easy to implement. Only uses basic linear algebra.

3 Spectral clustering: overview 2/11 In the previous lecture, we discussed how K-means can be used to cluster points in R p. Spectral clustering: Very popular clustering method. Often outperforms other methods such as K-means. Can be used for various types of data (not only points in R p ). Easy to implement. Only uses basic linear algebra. Overview of spectral clustering:

4 Spectral clustering: overview 2/11 In the previous lecture, we discussed how K-means can be used to cluster points in R p. Spectral clustering: Very popular clustering method. Often outperforms other methods such as K-means. Can be used for various types of data (not only points in R p ). Easy to implement. Only uses basic linear algebra. Overview of spectral clustering: 1 Construct a similarity matrix measuring the similarity of pairs of objects.

5 Spectral clustering: overview 2/11 In the previous lecture, we discussed how K-means can be used to cluster points in R p. Spectral clustering: Very popular clustering method. Often outperforms other methods such as K-means. Can be used for various types of data (not only points in R p ). Easy to implement. Only uses basic linear algebra. Overview of spectral clustering: 1 Construct a similarity matrix measuring the similarity of pairs of objects. 2 Use the similarity matrix to construct a (weighted or unweighted) graph.

6 Spectral clustering: overview 2/11 In the previous lecture, we discussed how K-means can be used to cluster points in R p. Spectral clustering: Very popular clustering method. Often outperforms other methods such as K-means. Can be used for various types of data (not only points in R p ). Easy to implement. Only uses basic linear algebra. Overview of spectral clustering: 1 Construct a similarity matrix measuring the similarity of pairs of objects. 2 Use the similarity matrix to construct a (weighted or unweighted) graph. 3 Compute eigenvectors of the graph Laplacian.

7 Spectral clustering: overview 2/11 In the previous lecture, we discussed how K-means can be used to cluster points in R p. Spectral clustering: Very popular clustering method. Often outperforms other methods such as K-means. Can be used for various types of data (not only points in R p ). Easy to implement. Only uses basic linear algebra. Overview of spectral clustering: 1 Construct a similarity matrix measuring the similarity of pairs of objects. 2 Use the similarity matrix to construct a (weighted or unweighted) graph. 3 Compute eigenvectors of the graph Laplacian. 4 Cluster the graph using the eigenvectors of the graph Laplacian using the K-means algorithm.

8 Notation 3/11 We will use the following notation/conventions: G = (V, E) a graph with vertex set V = {v 1,..., v n } and edge set E V V.

9 Notation 3/11 We will use the following notation/conventions: G = (V, E) a graph with vertex set V = {v 1,..., v n } and edge set E V V. Each edge carries a weight w ij 0.

10 Notation 3/11 We will use the following notation/conventions: G = (V, E) a graph with vertex set V = {v 1,..., v n } and edge set E V V. Each edge carries a weight w ij 0. The adjacency matrix of G is W = W G = (w ij ) n i,j=1. We will assume W is symmetric (undirected graphs).

11 Notation 3/11 We will use the following notation/conventions: G = (V, E) a graph with vertex set V = {v 1,..., v n } and edge set E V V. Each edge carries a weight w ij 0. The adjacency matrix of G is W = W G = (w ij ) n i,j=1. We will assume W is symmetric (undirected graphs). The degree of v i is d i := n w ij. j=1

12 Notation 3/11 We will use the following notation/conventions: G = (V, E) a graph with vertex set V = {v 1,..., v n } and edge set E V V. Each edge carries a weight w ij 0. The adjacency matrix of G is W = W G = (w ij ) n i,j=1. We will assume W is symmetric (undirected graphs). The degree of v i is d i := n w ij. j=1 The degree matrix of G is D := diag(d 1,..., d n ).

13 Notation 3/11 We will use the following notation/conventions: G = (V, E) a graph with vertex set V = {v 1,..., v n } and edge set E V V. Each edge carries a weight w ij 0. The adjacency matrix of G is W = W G = (w ij ) n i,j=1. We will assume W is symmetric (undirected graphs). The degree of v i is d i := n w ij. j=1 The degree matrix of G is D := diag(d 1,..., d n ). We denote the complement of A V by A.

14 Notation 3/11 We will use the following notation/conventions: G = (V, E) a graph with vertex set V = {v 1,..., v n } and edge set E V V. Each edge carries a weight w ij 0. The adjacency matrix of G is W = W G = (w ij ) n i,j=1. We will assume W is symmetric (undirected graphs). The degree of v i is d i := n w ij. j=1 The degree matrix of G is D := diag(d 1,..., d n ). We denote the complement of A V by A. If A V, then we let 1 A = (f 1,..., f n ) T R n, where f i = 1 if v i A and 0 otherwise.

15 Similarity graphs 4/11 We assume we are given a measure of similarity s between data points x 1,..., x n X : s : X X [0, ).

16 Similarity graphs 4/11 We assume we are given a measure of similarity s between data points x 1,..., x n X : s : X X [0, ). We denote by s ij := s(x i, x j ) the measure of similarity between x i and x j.

17 Similarity graphs 4/11 We assume we are given a measure of similarity s between data points x 1,..., x n X : s : X X [0, ). We denote by s ij := s(x i, x j ) the measure of similarity between x i and x j. Equivalently, we may assume we have a measure of distance between data points (e.g. (X, d) is a metric space).

18 Similarity graphs 4/11 We assume we are given a measure of similarity s between data points x 1,..., x n X : s : X X [0, ). We denote by s ij := s(x i, x j ) the measure of similarity between x i and x j. Equivalently, we may assume we have a measure of distance between data points (e.g. (X, d) is a metric space). Let d ij := d(x i, x j ), the distance between x i and x j.

19 Similarity graphs 4/11 We assume we are given a measure of similarity s between data points x 1,..., x n X : s : X X [0, ). We denote by s ij := s(x i, x j ) the measure of similarity between x i and x j. Equivalently, we may assume we have a measure of distance between data points (e.g. (X, d) is a metric space). Let d ij := d(x i, x j ), the distance between x i and x j. From d ij (or s ij ), we naturally build a similarity graph.

20 Similarity graphs 4/11 We assume we are given a measure of similarity s between data points x 1,..., x n X : s : X X [0, ). We denote by s ij := s(x i, x j ) the measure of similarity between x i and x j. Equivalently, we may assume we have a measure of distance between data points (e.g. (X, d) is a metric space). Let d ij := d(x i, x j ), the distance between x i and x j. From d ij (or s ij ), we naturally build a similarity graph. We will discuss 3 popular ways of building a similarity graph.

21 Similarity graphs (cont.) 5/11 Vertex set = {v 1,..., v n } where n is the number of data points.

22 Similarity graphs (cont.) 5/11 Vertex set = {v 1,..., v n } where n is the number of data points. 1 The ɛ-neighborhood graph: Connect all points whose pairwise distances are smaller than some ɛ > 0. We usually don't weight the edges. The graph is thus a simple graph (unweighted, undirected graph containing no loops or multiple edges).

23 5/11 Similarity graphs (cont.) Vertex set = {v 1,..., v n } where n is the number of data points. 1 The ɛ-neighborhood graph: Connect all points whose pairwise distances are smaller than some ɛ > 0. We usually don't weight the edges. The graph is thus a simple graph (unweighted, undirected graph containing no loops or multiple edges). 2 The k-nearest neighbor graph: The goal is to connect v i to v j if x j is among the k nearest neighbords of x i. However, this leads to a directed graph. We therefore dene:

24 5/11 Similarity graphs (cont.) Vertex set = {v 1,..., v n } where n is the number of data points. 1 The ɛ-neighborhood graph: Connect all points whose pairwise distances are smaller than some ɛ > 0. We usually don't weight the edges. The graph is thus a simple graph (unweighted, undirected graph containing no loops or multiple edges). 2 The k-nearest neighbor graph: The goal is to connect v i to v j if x j is among the k nearest neighbords of x i. However, this leads to a directed graph. We therefore dene: the k-nearest neighbor graph: v i is adjacent to v j i x j is among the k nearest neighbords of x i OR x i is among the k nearest neighbords of x j.

25 5/11 Similarity graphs (cont.) Vertex set = {v 1,..., v n } where n is the number of data points. 1 The ɛ-neighborhood graph: Connect all points whose pairwise distances are smaller than some ɛ > 0. We usually don't weight the edges. The graph is thus a simple graph (unweighted, undirected graph containing no loops or multiple edges). 2 The k-nearest neighbor graph: The goal is to connect v i to v j if x j is among the k nearest neighbords of x i. However, this leads to a directed graph. We therefore dene: the k-nearest neighbor graph: v i is adjacent to v j i x j is among the k nearest neighbords of x i OR x i is among the k nearest neighbords of x j. the mutual k-nearest neighbor graph: v i is adjacent to v j i x j is among the k nearest neighbords of x i AND x i is among the k nearest neighbors of x j.

26 5/11 Similarity graphs (cont.) Vertex set = {v 1,..., v n } where n is the number of data points. 1 The ɛ-neighborhood graph: Connect all points whose pairwise distances are smaller than some ɛ > 0. We usually don't weight the edges. The graph is thus a simple graph (unweighted, undirected graph containing no loops or multiple edges). 2 The k-nearest neighbor graph: The goal is to connect v i to v j if x j is among the k nearest neighbords of x i. However, this leads to a directed graph. We therefore dene: the k-nearest neighbor graph: v i is adjacent to v j i x j is among the k nearest neighbords of x i OR x i is among the k nearest neighbords of x j. the mutual k-nearest neighbor graph: v i is adjacent to v j i x j is among the k nearest neighbords of x i AND x i is among the k nearest neighbors of x j. We weight the edges by the similarity of their endpoints.

27 Similarity graphs (cont.) 6/11 3 The fully connected graph: Connect all points with edge weights s ij.

28 Similarity graphs (cont.) 6/11 3 The fully connected graph: Connect all points with edge weights s ij. For example, one could use the Gaussian similarity function to represent a local neighborhood relationships: s ij = s(x i, x j ) = exp( x i x j 2 /(2σ 2 )) (σ 2 > 0). Note: σ 2 controls the width of the neighborhoods.

29 Similarity graphs (cont.) 6/11 3 The fully connected graph: Connect all points with edge weights s ij. For example, one could use the Gaussian similarity function to represent a local neighborhood relationships: s ij = s(x i, x j ) = exp( x i x j 2 /(2σ 2 )) (σ 2 > 0). Note: σ 2 controls the width of the neighborhoods. All graphs mentioned above are regularly used in spectral clustering.

30 Graph Laplacians 7/11 There are three commonly used denitions of the graph Laplacian: 1 The unnormalized Laplacian is L := D W.

31 Graph Laplacians 7/11 There are three commonly used denitions of the graph Laplacian: 1 The unnormalized Laplacian is L := D W. 2 The normalized symmetric Laplacian is L sym := D 1/2 LD 1/2 = I D 1/2 W D 1/2.

32 Graph Laplacians 7/11 There are three commonly used denitions of the graph Laplacian: 1 The unnormalized Laplacian is L := D W. 2 The normalized symmetric Laplacian is L sym := D 1/2 LD 1/2 = I D 1/2 W D 1/2. 3 The normalized random walk Laplacian is L rw := D 1 L = I D 1 W.

33 Graph Laplacians 7/11 There are three commonly used denitions of the graph Laplacian: 1 The unnormalized Laplacian is L := D W. 2 The normalized symmetric Laplacian is L sym := D 1/2 LD 1/2 = I D 1/2 W D 1/2. 3 The normalized random walk Laplacian is L rw := D 1 L = I D 1 W. We begin by studying properties of the unnormalized Laplacian.

34 The unnormalized Laplacian 8/11 Proposition: The matrix L satises the following properties:

35 The unnormalized Laplacian 8/11 Proposition: The matrix L satises the following properties: 1 For any f R n : f T Lf = 1 n w ij (f i f j ) 2. 2 i,j=1

36 The unnormalized Laplacian 8/11 Proposition: The matrix L satises the following properties: 1 For any f R n : f T Lf = 1 n w ij (f i f j ) 2. 2 i,j=1 2 L is symmetric and positive semidenite.

37 The unnormalized Laplacian 8/11 Proposition: The matrix L satises the following properties: 1 For any f R n : f T Lf = 1 n w ij (f i f j ) 2. 2 i,j=1 2 L is symmetric and positive semidenite. 3 0 is an eigenvalue of L with associated constant eigenvector 1.

38 The unnormalized Laplacian 8/11 Proposition: The matrix L satises the following properties: 1 For any f R n : f T Lf = 1 n w ij (f i f j ) 2. 2 i,j=1 2 L is symmetric and positive semidenite. 3 0 is an eigenvalue of L with associated constant eigenvector 1. Proof:

39 The unnormalized Laplacian (2) follows from (1). (3) is easy. 8/11 Proposition: The matrix L satises the following properties: 1 For any f R n : f T Lf = 1 n w ij (f i f j ) 2. 2 i,j=1 2 L is symmetric and positive semidenite. 3 0 is an eigenvalue of L with associated constant eigenvector 1. Proof: To prove (1), f T Lf = f T Df f T W f = n d i fi 2 i=1 n w ij f i f j i,j=1 = 1 n d i fi = 1 2 i=1 n w ij f i f j + i,j=1 n w ij (f i f j ) 2. i,j=1 n d j fj 2 j=1

40 The unnormalized Laplacian (cont.) 9/11 Proposition: Let G be an undirected graph with non-negative weights. Then:

41 The unnormalized Laplacian (cont.) 9/11 Proposition: Let G be an undirected graph with non-negative weights. Then: 1 The multiplicity k of the eigenvalue 0 of L equals the number of connected components A 1,..., A k in the graph.

42 The unnormalized Laplacian (cont.) 9/11 Proposition: Let G be an undirected graph with non-negative weights. Then: 1 The multiplicity k of the eigenvalue 0 of L equals the number of connected components A 1,..., A k in the graph. 2 The eigenspace of eigenvalue 0 is spanned by the indicator vectors 1 A1,..., 1 Ak of those components.

43 The unnormalized Laplacian (cont.) 9/11 Proposition: Let G be an undirected graph with non-negative weights. Then: 1 The multiplicity k of the eigenvalue 0 of L equals the number of connected components A 1,..., A k in the graph. 2 The eigenspace of eigenvalue 0 is spanned by the indicator vectors 1 A1,..., 1 Ak of those components. Proof:

44 The unnormalized Laplacian (cont.) 9/11 Proposition: Let G be an undirected graph with non-negative weights. Then: 1 The multiplicity k of the eigenvalue 0 of L equals the number of connected components A 1,..., A k in the graph. 2 The eigenspace of eigenvalue 0 is spanned by the indicator vectors 1 A1,..., 1 Ak of those components. Proof: If f is an eigenvector associate to λ = 0, then 0 = f T Lf = n w ij (f i f j ) 2. i,j=1 It follows that f i = f j whenever w ij > 0. Thus f is constant on the connected components of G. We conclude that the eigenspace of 0 is contained in span(1 A1,..., 1 Ak ). Conversely, it is not hard to see that each 1 Ai is an eigenvector associated to 0 (write L in block diagonal form).

45 The normalized Laplacians 10/11 Proposition: The normalized Laplacians satisfy the following properties: 1 For every f R n, we have f T L sym f = 1 2 ( ) 2 n f i w ij f j. di dj i,j=1 2 λ is an eigenvalue of L rw with eigenvector u if and only if λ is an eigenvalue of L sym with eigenvector w = D 1/2 u. 3 λ is an eigenvalue of L rw with eigenvector u if and only if λ and u solve the generalized eigenproblem Lu = λdu. Proof: The proof of (1) is similar to the proof of the analogous result for the unnormalized Laplacian. (2) and (3) follow easily by using appropriate rescalings.

46 The normalized Laplacians (cont.) 11/11 Proposition: Let G be an undirected graph with non-negative weights. Then: 1 The multiplicity k of the eigenvalue 0 of both L sym and L rw equals the number of connected components A 1,..., A k in the graph. 2 For L rw, the eigenspace of eigenvalue 0 is spanned by the indicator vectors 1 Ai, i = 1,..., k. 3 For L sym, the eigenspace of eigenvalue 0 is spanned by the vectors D 1/2 1 Ai, i = 1,..., k. Proof: Similar to the proof of the analogous result for the unnormalized Laplacian.

MATH 567: Mathematical Techniques in Data Science Clustering II

MATH 567: Mathematical Techniques in Data Science Clustering II This lecture is based on U. von Luxburg, A Tutorial on Spectral Clustering, Statistics and Computing, 17 (4), 2007. MATH 567: Mathematical Techniques in Data Science Clustering II Dominique Guillot Departments

More information

MATH 567: Mathematical Techniques in Data Science Clustering II

MATH 567: Mathematical Techniques in Data Science Clustering II Spectral clustering: overview MATH 567: Mathematical Techniques in Data Science Clustering II Dominique uillot Departments of Mathematical Sciences University of Delaware Overview of spectral clustering:

More information

Spectral Clustering. Spectral Clustering? Two Moons Data. Spectral Clustering Algorithm: Bipartioning. Spectral methods

Spectral Clustering. Spectral Clustering? Two Moons Data. Spectral Clustering Algorithm: Bipartioning. Spectral methods Spectral Clustering Seungjin Choi Department of Computer Science POSTECH, Korea seungjin@postech.ac.kr 1 Spectral methods Spectral Clustering? Methods using eigenvectors of some matrices Involve eigen-decomposition

More information

Spectral Clustering. Zitao Liu

Spectral Clustering. Zitao Liu Spectral Clustering Zitao Liu Agenda Brief Clustering Review Similarity Graph Graph Laplacian Spectral Clustering Algorithm Graph Cut Point of View Random Walk Point of View Perturbation Theory Point of

More information

Data Analysis and Manifold Learning Lecture 3: Graphs, Graph Matrices, and Graph Embeddings

Data Analysis and Manifold Learning Lecture 3: Graphs, Graph Matrices, and Graph Embeddings Data Analysis and Manifold Learning Lecture 3: Graphs, Graph Matrices, and Graph Embeddings Radu Horaud INRIA Grenoble Rhone-Alpes, France Radu.Horaud@inrialpes.fr http://perception.inrialpes.fr/ Outline

More information

An Algorithmist s Toolkit September 10, Lecture 1

An Algorithmist s Toolkit September 10, Lecture 1 18.409 An Algorithmist s Toolkit September 10, 2009 Lecture 1 Lecturer: Jonathan Kelner Scribe: Jesse Geneson (2009) 1 Overview The class s goals, requirements, and policies were introduced, and topics

More information

MLCC Clustering. Lorenzo Rosasco UNIGE-MIT-IIT

MLCC Clustering. Lorenzo Rosasco UNIGE-MIT-IIT MLCC 2018 - Clustering Lorenzo Rosasco UNIGE-MIT-IIT About this class We will consider an unsupervised setting, and in particular the problem of clustering unlabeled data into coherent groups. MLCC 2018

More information

Data Analysis and Manifold Learning Lecture 7: Spectral Clustering

Data Analysis and Manifold Learning Lecture 7: Spectral Clustering Data Analysis and Manifold Learning Lecture 7: Spectral Clustering Radu Horaud INRIA Grenoble Rhone-Alpes, France Radu.Horaud@inrialpes.fr http://perception.inrialpes.fr/ Outline of Lecture 7 What is spectral

More information

Machine Learning for Data Science (CS4786) Lecture 11

Machine Learning for Data Science (CS4786) Lecture 11 Machine Learning for Data Science (CS4786) Lecture 11 Spectral clustering Course Webpage : http://www.cs.cornell.edu/courses/cs4786/2016sp/ ANNOUNCEMENT 1 Assignment P1 the Diagnostic assignment 1 will

More information

Spectral clustering. Two ideal clusters, with two points each. Spectral clustering algorithms

Spectral clustering. Two ideal clusters, with two points each. Spectral clustering algorithms A simple example Two ideal clusters, with two points each Spectral clustering Lecture 2 Spectral clustering algorithms 4 2 3 A = Ideally permuted Ideal affinities 2 Indicator vectors Each cluster has an

More information

8.1 Concentration inequality for Gaussian random matrix (cont d)

8.1 Concentration inequality for Gaussian random matrix (cont d) MGMT 69: Topics in High-dimensional Data Analysis Falll 26 Lecture 8: Spectral clustering and Laplacian matrices Lecturer: Jiaming Xu Scribe: Hyun-Ju Oh and Taotao He, October 4, 26 Outline Concentration

More information

Spectral Clustering. Guokun Lai 2016/10

Spectral Clustering. Guokun Lai 2016/10 Spectral Clustering Guokun Lai 2016/10 1 / 37 Organization Graph Cut Fundamental Limitations of Spectral Clustering Ng 2002 paper (if we have time) 2 / 37 Notation We define a undirected weighted graph

More information

Spectral Clustering on Handwritten Digits Database

Spectral Clustering on Handwritten Digits Database University of Maryland-College Park Advance Scientific Computing I,II Spectral Clustering on Handwritten Digits Database Author: Danielle Middlebrooks Dmiddle1@math.umd.edu Second year AMSC Student Advisor:

More information

Spectral Clustering on Handwritten Digits Database Mid-Year Pr

Spectral Clustering on Handwritten Digits Database Mid-Year Pr Spectral Clustering on Handwritten Digits Database Mid-Year Presentation Danielle dmiddle1@math.umd.edu Advisor: Kasso Okoudjou kasso@umd.edu Department of Mathematics University of Maryland- College Park

More information

Lecture 9: Laplacian Eigenmaps

Lecture 9: Laplacian Eigenmaps Lecture 9: Radu Balan Department of Mathematics, AMSC, CSCAMM and NWC University of Maryland, College Park, MD April 18, 2017 Optimization Criteria Assume G = (V, W ) is a undirected weighted graph with

More information

17.1 Directed Graphs, Undirected Graphs, Incidence Matrices, Adjacency Matrices, Weighted Graphs

17.1 Directed Graphs, Undirected Graphs, Incidence Matrices, Adjacency Matrices, Weighted Graphs Chapter 17 Graphs and Graph Laplacians 17.1 Directed Graphs, Undirected Graphs, Incidence Matrices, Adjacency Matrices, Weighted Graphs Definition 17.1. A directed graph isa pairg =(V,E), where V = {v

More information

MATH 829: Introduction to Data Mining and Analysis Graphical Models I

MATH 829: Introduction to Data Mining and Analysis Graphical Models I MATH 829: Introduction to Data Mining and Analysis Graphical Models I Dominique Guillot Departments of Mathematical Sciences University of Delaware May 2, 2016 1/12 Independence and conditional independence:

More information

Spectral Theory of Unsigned and Signed Graphs Applications to Graph Clustering. Some Slides

Spectral Theory of Unsigned and Signed Graphs Applications to Graph Clustering. Some Slides Spectral Theory of Unsigned and Signed Graphs Applications to Graph Clustering Some Slides Jean Gallier Department of Computer and Information Science University of Pennsylvania Philadelphia, PA 19104,

More information

Lecture 12 : Graph Laplacians and Cheeger s Inequality

Lecture 12 : Graph Laplacians and Cheeger s Inequality CPS290: Algorithmic Foundations of Data Science March 7, 2017 Lecture 12 : Graph Laplacians and Cheeger s Inequality Lecturer: Kamesh Munagala Scribe: Kamesh Munagala Graph Laplacian Maybe the most beautiful

More information

Lecture 12: Introduction to Spectral Graph Theory, Cheeger s inequality

Lecture 12: Introduction to Spectral Graph Theory, Cheeger s inequality CSE 521: Design and Analysis of Algorithms I Spring 2016 Lecture 12: Introduction to Spectral Graph Theory, Cheeger s inequality Lecturer: Shayan Oveis Gharan May 4th Scribe: Gabriel Cadamuro Disclaimer:

More information

Lecture 13: Spectral Graph Theory

Lecture 13: Spectral Graph Theory CSE 521: Design and Analysis of Algorithms I Winter 2017 Lecture 13: Spectral Graph Theory Lecturer: Shayan Oveis Gharan 11/14/18 Disclaimer: These notes have not been subjected to the usual scrutiny reserved

More information

Lecture 1: Graphs, Adjacency Matrices, Graph Laplacian

Lecture 1: Graphs, Adjacency Matrices, Graph Laplacian Lecture 1: Graphs, Adjacency Matrices, Graph Laplacian Radu Balan January 31, 2017 G = (V, E) An undirected graph G is given by two pieces of information: a set of vertices V and a set of edges E, G =

More information

Notes on Elementary Spectral Graph Theory Applications to Graph Clustering

Notes on Elementary Spectral Graph Theory Applications to Graph Clustering Notes on Elementary Spectral Graph Theory Applications to Graph Clustering Jean Gallier Department of Computer and Information Science University of Pennsylvania Philadelphia, PA 19104, USA e-mail: jean@cis.upenn.edu

More information

Certifying the Global Optimality of Graph Cuts via Semidefinite Programming: A Theoretic Guarantee for Spectral Clustering

Certifying the Global Optimality of Graph Cuts via Semidefinite Programming: A Theoretic Guarantee for Spectral Clustering Certifying the Global Optimality of Graph Cuts via Semidefinite Programming: A Theoretic Guarantee for Spectral Clustering Shuyang Ling Courant Institute of Mathematical Sciences, NYU Aug 13, 2018 Joint

More information

1 Matrix notation and preliminaries from spectral graph theory

1 Matrix notation and preliminaries from spectral graph theory Graph clustering (or community detection or graph partitioning) is one of the most studied problems in network analysis. One reason for this is that there are a variety of ways to define a cluster or community.

More information

1 Matrix notation and preliminaries from spectral graph theory

1 Matrix notation and preliminaries from spectral graph theory Graph clustering (or community detection or graph partitioning) is one of the most studied problems in network analysis. One reason for this is that there are a variety of ways to define a cluster or community.

More information

Notes on Elementary Spectral Graph Theory Applications to Graph Clustering Using Normalized Cuts

Notes on Elementary Spectral Graph Theory Applications to Graph Clustering Using Normalized Cuts University of Pennsylvania ScholarlyCommons Technical Reports (CIS) Department of Computer & Information Science 1-1-2013 Notes on Elementary Spectral Graph Theory Applications to Graph Clustering Using

More information

MATH 829: Introduction to Data Mining and Analysis Support vector machines and kernels

MATH 829: Introduction to Data Mining and Analysis Support vector machines and kernels 1/12 MATH 829: Introduction to Data Mining and Analysis Support vector machines and kernels Dominique Guillot Departments of Mathematical Sciences University of Delaware March 14, 2016 Separating sets:

More information

A lower bound for the Laplacian eigenvalues of a graph proof of a conjecture by Guo

A lower bound for the Laplacian eigenvalues of a graph proof of a conjecture by Guo A lower bound for the Laplacian eigenvalues of a graph proof of a conjecture by Guo A. E. Brouwer & W. H. Haemers 2008-02-28 Abstract We show that if µ j is the j-th largest Laplacian eigenvalue, and d

More information

18.312: Algebraic Combinatorics Lionel Levine. Lecture 19

18.312: Algebraic Combinatorics Lionel Levine. Lecture 19 832: Algebraic Combinatorics Lionel Levine Lecture date: April 2, 20 Lecture 9 Notes by: David Witmer Matrix-Tree Theorem Undirected Graphs Let G = (V, E) be a connected, undirected graph with n vertices,

More information

March 13, Paper: R.R. Coifman, S. Lafon, Diffusion maps ([Coifman06]) Seminar: Learning with Graphs, Prof. Hein, Saarland University

March 13, Paper: R.R. Coifman, S. Lafon, Diffusion maps ([Coifman06]) Seminar: Learning with Graphs, Prof. Hein, Saarland University Kernels March 13, 2008 Paper: R.R. Coifman, S. Lafon, maps ([Coifman06]) Seminar: Learning with Graphs, Prof. Hein, Saarland University Kernels Figure: Example Application from [LafonWWW] meaningful geometric

More information

Spectral Theory of Unsigned and Signed Graphs Applications to Graph Clustering: a Survey

Spectral Theory of Unsigned and Signed Graphs Applications to Graph Clustering: a Survey Spectral Theory of Unsigned and Signed Graphs Applications to Graph Clustering: a Survey Jean Gallier Department of Computer and Information Science University of Pennsylvania Philadelphia, PA 19104, USA

More information

Learning from Sensor Data: Set II. Behnaam Aazhang J.S. Abercombie Professor Electrical and Computer Engineering Rice University

Learning from Sensor Data: Set II. Behnaam Aazhang J.S. Abercombie Professor Electrical and Computer Engineering Rice University Learning from Sensor Data: Set II Behnaam Aazhang J.S. Abercombie Professor Electrical and Computer Engineering Rice University 1 6. Data Representation The approach for learning from data Probabilistic

More information

Clustering using Mixture Models

Clustering using Mixture Models Clustering using Mixture Models The full posterior of the Gaussian Mixture Model is p(x, Z, µ,, ) =p(x Z, µ, )p(z )p( )p(µ, ) data likelihood (Gaussian) correspondence prob. (Multinomial) mixture prior

More information

An Introduction to Spectral Graph Theory

An Introduction to Spectral Graph Theory An Introduction to Spectral Graph Theory Mackenzie Wheeler Supervisor: Dr. Gary MacGillivray University of Victoria WheelerM@uvic.ca Outline Outline 1. How many walks are there from vertices v i to v j

More information

CS168: The Modern Algorithmic Toolbox Lectures #11 and #12: Spectral Graph Theory

CS168: The Modern Algorithmic Toolbox Lectures #11 and #12: Spectral Graph Theory CS168: The Modern Algorithmic Toolbox Lectures #11 and #12: Spectral Graph Theory Tim Roughgarden & Gregory Valiant May 2, 2016 Spectral graph theory is the powerful and beautiful theory that arises from

More information

Diffusion and random walks on graphs

Diffusion and random walks on graphs Diffusion and random walks on graphs Leonid E. Zhukov School of Data Analysis and Artificial Intelligence Department of Computer Science National Research University Higher School of Economics Structural

More information

Laplacian Eigenmaps for Dimensionality Reduction and Data Representation

Laplacian Eigenmaps for Dimensionality Reduction and Data Representation Laplacian Eigenmaps for Dimensionality Reduction and Data Representation Neural Computation, June 2003; 15 (6):1373-1396 Presentation for CSE291 sp07 M. Belkin 1 P. Niyogi 2 1 University of Chicago, Department

More information

Nonlinear Methods. Data often lies on or near a nonlinear low-dimensional curve aka manifold.

Nonlinear Methods. Data often lies on or near a nonlinear low-dimensional curve aka manifold. Nonlinear Methods Data often lies on or near a nonlinear low-dimensional curve aka manifold. 27 Laplacian Eigenmaps Linear methods Lower-dimensional linear projection that preserves distances between all

More information

Lecture 10: Dimension Reduction Techniques

Lecture 10: Dimension Reduction Techniques Lecture 10: Dimension Reduction Techniques Radu Balan Department of Mathematics, AMSC, CSCAMM and NWC University of Maryland, College Park, MD April 17, 2018 Input Data It is assumed that there is a set

More information

Computer Vision Group Prof. Daniel Cremers. 14. Clustering

Computer Vision Group Prof. Daniel Cremers. 14. Clustering Group Prof. Daniel Cremers 14. Clustering Motivation Supervised learning is good for interaction with humans, but labels from a supervisor are hard to obtain Clustering is unsupervised learning, i.e. it

More information

Statistical Learning Notes III-Section 2.4.3

Statistical Learning Notes III-Section 2.4.3 Statistical Learning Notes III-Section 2.4.3 "Graphical" Spectral Features Stephen Vardeman Analytics Iowa LLC January 2019 Stephen Vardeman (Analytics Iowa LLC) Statistical Learning Notes III-Section

More information

Two Laplacians for the distance matrix of a graph

Two Laplacians for the distance matrix of a graph Two Laplacians for the distance matrix of a graph Mustapha Aouchiche and Pierre Hansen GERAD and HEC Montreal, Canada CanaDAM 2013, Memorial University, June 1013 Aouchiche & Hansen CanaDAM 2013, Memorial

More information

Lecture 13 Spectral Graph Algorithms

Lecture 13 Spectral Graph Algorithms COMS 995-3: Advanced Algorithms March 6, 7 Lecture 3 Spectral Graph Algorithms Instructor: Alex Andoni Scribe: Srikar Varadaraj Introduction Today s topics: Finish proof from last lecture Example of random

More information

Graph fundamentals. Matrices associated with a graph

Graph fundamentals. Matrices associated with a graph Graph fundamentals Matrices associated with a graph Drawing a picture of a graph is one way to represent it. Another type of representation is via a matrix. Let G be a graph with V (G) ={v 1,v,...,v n

More information

Diffuse interface methods on graphs: Data clustering and Gamma-limits

Diffuse interface methods on graphs: Data clustering and Gamma-limits Diffuse interface methods on graphs: Data clustering and Gamma-limits Yves van Gennip joint work with Andrea Bertozzi, Jeff Brantingham, Blake Hunter Department of Mathematics, UCLA Research made possible

More information

MATH 829: Introduction to Data Mining and Analysis Computing the lasso solution

MATH 829: Introduction to Data Mining and Analysis Computing the lasso solution 1/16 MATH 829: Introduction to Data Mining and Analysis Computing the lasso solution Dominique Guillot Departments of Mathematical Sciences University of Delaware February 26, 2016 Computing the lasso

More information

Unsupervised dimensionality reduction

Unsupervised dimensionality reduction Unsupervised dimensionality reduction Guillaume Obozinski Ecole des Ponts - ParisTech SOCN course 2014 Guillaume Obozinski Unsupervised dimensionality reduction 1/30 Outline 1 PCA 2 Kernel PCA 3 Multidimensional

More information

Data Mining and Analysis: Fundamental Concepts and Algorithms

Data Mining and Analysis: Fundamental Concepts and Algorithms : Fundamental Concepts and Algorithms dataminingbook.info Mohammed J. Zaki 1 Wagner Meira Jr. 2 1 Department of Computer Science Rensselaer Polytechnic Institute, Troy, NY, USA 2 Department of Computer

More information

Physical Metaphors for Graphs

Physical Metaphors for Graphs Graphs and Networks Lecture 3 Physical Metaphors for Graphs Daniel A. Spielman October 9, 203 3. Overview We will examine physical metaphors for graphs. We begin with a spring model and then discuss resistor

More information

MATH 829: Introduction to Data Mining and Analysis Graphical Models II - Gaussian Graphical Models

MATH 829: Introduction to Data Mining and Analysis Graphical Models II - Gaussian Graphical Models 1/13 MATH 829: Introduction to Data Mining and Analysis Graphical Models II - Gaussian Graphical Models Dominique Guillot Departments of Mathematical Sciences University of Delaware May 4, 2016 Recall

More information

Graphs in Machine Learning

Graphs in Machine Learning Graphs in Machine Learning Michal Valko Inria Lille - Nord Europe, France TA: Pierre Perrault Partially based on material by: Ulrike von Luxburg, Gary Miller, Doyle & Schnell, Daniel Spielman October 10th,

More information

CMPSCI 791BB: Advanced ML: Laplacian Learning

CMPSCI 791BB: Advanced ML: Laplacian Learning CMPSCI 791BB: Advanced ML: Laplacian Learning Sridhar Mahadevan Outline! Spectral graph operators! Combinatorial graph Laplacian! Normalized graph Laplacian! Random walks! Machine learning on graphs! Clustering!

More information

Math 126 Lecture 8. Summary of results Application to the Laplacian of the cube

Math 126 Lecture 8. Summary of results Application to the Laplacian of the cube Math 126 Lecture 8 Summary of results Application to the Laplacian of the cube Summary of key results Every representation is determined by its character. If j is the character of a representation and

More information

Some notes on SVD, dimensionality reduction, and clustering. Guangliang Chen

Some notes on SVD, dimensionality reduction, and clustering. Guangliang Chen Some notes on SVD, dimensionality reduction, and clustering Guangliang Chen Contents 1. Introduction 4 2. Review of necessary linear algebra 4 3. Matrix SVD and its applications 8 Practice problems set

More information

Networks and Their Spectra

Networks and Their Spectra Networks and Their Spectra Victor Amelkin University of California, Santa Barbara Department of Computer Science victor@cs.ucsb.edu December 4, 2017 1 / 18 Introduction Networks (= graphs) are everywhere.

More information

EECS 275 Matrix Computation

EECS 275 Matrix Computation EECS 275 Matrix Computation Ming-Hsuan Yang Electrical Engineering and Computer Science University of California at Merced Merced, CA 95344 http://faculty.ucmerced.edu/mhyang Lecture 23 1 / 27 Overview

More information

Spectral Graph Theory

Spectral Graph Theory Spectral Graph Theory Aaron Mishtal April 27, 2016 1 / 36 Outline Overview Linear Algebra Primer History Theory Applications Open Problems Homework Problems References 2 / 36 Outline Overview Linear Algebra

More information

Statistical Machine Learning

Statistical Machine Learning Statistical Machine Learning Christoph Lampert Spring Semester 2015/2016 // Lecture 12 1 / 36 Unsupervised Learning Dimensionality Reduction 2 / 36 Dimensionality Reduction Given: data X = {x 1,..., x

More information

On the inverse matrix of the Laplacian and all ones matrix

On the inverse matrix of the Laplacian and all ones matrix On the inverse matrix of the Laplacian and all ones matrix Sho Suda (Joint work with Michio Seto and Tetsuji Taniguchi) International Christian University JSPS Research Fellow PD November 21, 2012 Sho

More information

Graphs in Machine Learning

Graphs in Machine Learning Graphs in Machine Learning Michal Valko INRIA Lille - Nord Europe, France Partially based on material by: Ulrike von Luxburg, Gary Miller, Doyle & Schnell, Daniel Spielman January 27, 2015 MVA 2014/2015

More information

1 Review: symmetric matrices, their eigenvalues and eigenvectors

1 Review: symmetric matrices, their eigenvalues and eigenvectors Cornell University, Fall 2012 Lecture notes on spectral methods in algorithm design CS 6820: Algorithms Studying the eigenvalues and eigenvectors of matrices has powerful consequences for at least three

More information

HYPERGRAPH BASED SEMI-SUPERVISED LEARNING ALGORITHMS APPLIED TO SPEECH RECOGNITION PROBLEM: A NOVEL APPROACH

HYPERGRAPH BASED SEMI-SUPERVISED LEARNING ALGORITHMS APPLIED TO SPEECH RECOGNITION PROBLEM: A NOVEL APPROACH HYPERGRAPH BASED SEMI-SUPERVISED LEARNING ALGORITHMS APPLIED TO SPEECH RECOGNITION PROBLEM: A NOVEL APPROACH Hoang Trang 1, Tran Hoang Loc 1 1 Ho Chi Minh City University of Technology-VNU HCM, Ho Chi

More information

COMPSCI 514: Algorithms for Data Science

COMPSCI 514: Algorithms for Data Science COMPSCI 514: Algorithms for Data Science Arya Mazumdar University of Massachusetts at Amherst Fall 2018 Lecture 8 Spectral Clustering Spectral clustering Curse of dimensionality Dimensionality Reduction

More information

Lecture: Some Practical Considerations (3 of 4)

Lecture: Some Practical Considerations (3 of 4) Stat260/CS294: Spectral Graph Methods Lecture 14-03/10/2015 Lecture: Some Practical Considerations (3 of 4) Lecturer: Michael Mahoney Scribe: Michael Mahoney Warning: these notes are still very rough.

More information

Unsupervised Learning Techniques Class 07, 1 March 2006 Andrea Caponnetto

Unsupervised Learning Techniques Class 07, 1 March 2006 Andrea Caponnetto Unsupervised Learning Techniques 9.520 Class 07, 1 March 2006 Andrea Caponnetto About this class Goal To introduce some methods for unsupervised learning: Gaussian Mixtures, K-Means, ISOMAP, HLLE, Laplacian

More information

Information Propagation Analysis of Social Network Using the Universality of Random Matrix

Information Propagation Analysis of Social Network Using the Universality of Random Matrix Information Propagation Analysis of Social Network Using the Universality of Random Matrix Yusuke Sakumoto, Tsukasa Kameyama, Chisa Takano and Masaki Aida Tokyo Metropolitan University, 6-6 Asahigaoka,

More information

Math 3191 Applied Linear Algebra

Math 3191 Applied Linear Algebra Math 9 Applied Linear Algebra Lecture 9: Diagonalization Stephen Billups University of Colorado at Denver Math 9Applied Linear Algebra p./9 Section. Diagonalization The goal here is to develop a useful

More information

Data-dependent representations: Laplacian Eigenmaps

Data-dependent representations: Laplacian Eigenmaps Data-dependent representations: Laplacian Eigenmaps November 4, 2015 Data Organization and Manifold Learning There are many techniques for Data Organization and Manifold Learning, e.g., Principal Component

More information

Graphic sequences, adjacency matrix

Graphic sequences, adjacency matrix Chapter 2 Graphic sequences, adjacency matrix Definition 2.1. A sequence of integers (d 1,..., d n ) is called graphic if it is the degree sequence of a graph. Example 2.2. 1. (1, 2, 2, 3) is graphic:

More information

Introduction to Spectral Graph Theory and Graph Clustering

Introduction to Spectral Graph Theory and Graph Clustering Introduction to Spectral Graph Theory and Graph Clustering Chengming Jiang ECS 231 Spring 2016 University of California, Davis 1 / 40 Motivation Image partitioning in computer vision 2 / 40 Motivation

More information

Relations Between Adjacency And Modularity Graph Partitioning: Principal Component Analysis vs. Modularity Component Analysis

Relations Between Adjacency And Modularity Graph Partitioning: Principal Component Analysis vs. Modularity Component Analysis Relations Between Adjacency And Modularity Graph Partitioning: Principal Component Analysis vs. Modularity Component Analysis Hansi Jiang Carl Meyer North Carolina State University October 27, 2015 1 /

More information

Communities, Spectral Clustering, and Random Walks

Communities, Spectral Clustering, and Random Walks Communities, Spectral Clustering, and Random Walks David Bindel Department of Computer Science Cornell University 26 Sep 2011 20 21 19 16 22 28 17 18 29 26 27 30 23 1 25 5 8 24 2 4 14 3 9 13 15 11 10 12

More information

The f-adjusted Graph Laplacian: a Diagonal Modification with a Geometric Interpretation

The f-adjusted Graph Laplacian: a Diagonal Modification with a Geometric Interpretation The f-adjusted Graph Laplacian: a Diagonal Modification with a Geometric Interpretation Sven Kurras Ulrike von Luxburg Department of Computer Science, University of Hamburg, Germany Gilles Blanchard Department

More information

Math 443/543 Graph Theory Notes 5: Graphs as matrices, spectral graph theory, and PageRank

Math 443/543 Graph Theory Notes 5: Graphs as matrices, spectral graph theory, and PageRank Math 443/543 Graph Theory Notes 5: Graphs as matrices, spectral graph theory, and PageRank David Glickenstein November 3, 4 Representing graphs as matrices It will sometimes be useful to represent graphs

More information

Laplacian Eigenmaps for Dimensionality Reduction and Data Representation

Laplacian Eigenmaps for Dimensionality Reduction and Data Representation Introduction and Data Representation Mikhail Belkin & Partha Niyogi Department of Electrical Engieering University of Minnesota Mar 21, 2017 1/22 Outline Introduction 1 Introduction 2 3 4 Connections to

More information

A New Spectral Technique Using Normalized Adjacency Matrices for Graph Matching 1

A New Spectral Technique Using Normalized Adjacency Matrices for Graph Matching 1 CHAPTER-3 A New Spectral Technique Using Normalized Adjacency Matrices for Graph Matching Graph matching problem has found many applications in areas as diverse as chemical structure analysis, pattern

More information

Lecture 1: Systems of linear equations and their solutions

Lecture 1: Systems of linear equations and their solutions Lecture 1: Systems of linear equations and their solutions Course overview Topics to be covered this semester: Systems of linear equations and Gaussian elimination: Solving linear equations and applications

More information

Walks, Springs, and Resistor Networks

Walks, Springs, and Resistor Networks Spectral Graph Theory Lecture 12 Walks, Springs, and Resistor Networks Daniel A. Spielman October 8, 2018 12.1 Overview In this lecture we will see how the analysis of random walks, spring networks, and

More information

Consistency of Spectral Clustering

Consistency of Spectral Clustering Max Planck Institut für biologische Kybernetik Max Planck Institute for Biological Cybernetics Technical Report No. TR-34 Consistency of Spectral Clustering Ulrike von Luxburg, Mikhail Belkin 2, Olivier

More information

Lecture 7: Spectral Graph Theory II

Lecture 7: Spectral Graph Theory II A Theorist s Toolkit (CMU 18-859T, Fall 2013) Lecture 7: Spectral Graph Theory II September 30, 2013 Lecturer: Ryan O Donnell Scribe: Christian Tjandraatmadja 1 Review We spend a page to review the previous

More information

Linear Spectral Hashing

Linear Spectral Hashing Linear Spectral Hashing Zalán Bodó and Lehel Csató Babeş Bolyai University - Faculty of Mathematics and Computer Science Kogălniceanu 1., 484 Cluj-Napoca - Romania Abstract. assigns binary hash keys to

More information

Consistent Manifold Representation for Topological Data Analysis

Consistent Manifold Representation for Topological Data Analysis Consistent Manifold Representation for Topological Data Analysis Tyrus Berry (tberry@gmu.edu) and Timothy Sauer Dept. of Mathematical Sciences, George Mason University, Fairfax, VA 3 Abstract For data

More information

ORIE 4741: Learning with Big Messy Data. Spectral Graph Theory

ORIE 4741: Learning with Big Messy Data. Spectral Graph Theory ORIE 4741: Learning with Big Messy Data Spectral Graph Theory Mika Sumida Operations Research and Information Engineering Cornell September 15, 2017 1 / 32 Outline Graph Theory Spectral Graph Theory Laplacian

More information

Energy of Graphs. Sivaram K. Narayan Central Michigan University. Presented at CMU on October 10, 2015

Energy of Graphs. Sivaram K. Narayan Central Michigan University. Presented at CMU on October 10, 2015 Energy of Graphs Sivaram K. Narayan Central Michigan University Presented at CMU on October 10, 2015 1 / 32 Graphs We will consider simple graphs (no loops, no multiple edges). Let V = {v 1, v 2,..., v

More information

22m:033 Notes: 7.1 Diagonalization of Symmetric Matrices

22m:033 Notes: 7.1 Diagonalization of Symmetric Matrices m:33 Notes: 7. Diagonalization of Symmetric Matrices Dennis Roseman University of Iowa Iowa City, IA http://www.math.uiowa.edu/ roseman May 3, Symmetric matrices Definition. A symmetric matrix is a matrix

More information

MATH 829: Introduction to Data Mining and Analysis Consistency of Linear Regression

MATH 829: Introduction to Data Mining and Analysis Consistency of Linear Regression 1/9 MATH 829: Introduction to Data Mining and Analysis Consistency of Linear Regression Dominique Guillot Deartments of Mathematical Sciences University of Delaware February 15, 2016 Distribution of regression

More information

Lecture: Some Statistical Inference Issues (2 of 3)

Lecture: Some Statistical Inference Issues (2 of 3) Stat260/CS294: Spectral Graph Methods Lecture 23-04/16/2015 Lecture: Some Statistical Inference Issues (2 of 3) Lecturer: Michael Mahoney Scribe: Michael Mahoney Warning: these notes are still very rough.

More information

The University of Texas at Austin Department of Electrical and Computer Engineering. EE381V: Large Scale Learning Spring 2013.

The University of Texas at Austin Department of Electrical and Computer Engineering. EE381V: Large Scale Learning Spring 2013. The University of Texas at Austin Department of Electrical and Computer Engineering EE381V: Large Scale Learning Spring 2013 Assignment 1 Caramanis/Sanghavi Due: Thursday, Feb. 7, 2013. (Problems 1 and

More information

Tutorial: Gaussian conditional independence and graphical models. Thomas Kahle Otto-von-Guericke Universität Magdeburg

Tutorial: Gaussian conditional independence and graphical models. Thomas Kahle Otto-von-Guericke Universität Magdeburg Tutorial: Gaussian conditional independence and graphical models Thomas Kahle Otto-von-Guericke Universität Magdeburg The central dogma of algebraic statistics Statistical models are varieties The central

More information

The Simplest Construction of Expanders

The Simplest Construction of Expanders Spectral Graph Theory Lecture 14 The Simplest Construction of Expanders Daniel A. Spielman October 16, 2009 14.1 Overview I am going to present the simplest construction of expanders that I have been able

More information

Damped random walks and the characteristic polynomial of the weighted Laplacian on a graph

Damped random walks and the characteristic polynomial of the weighted Laplacian on a graph Damped random walks and the characteristic polynomial of the weighted Laplacian on a graph arxiv:mathpr/0506460 v1 Jun 005 MADHAV P DESAI and HARIHARAN NARAYANAN September 11, 006 Abstract For λ > 0, we

More information

MATH 304 Linear Algebra Lecture 33: Bases of eigenvectors. Diagonalization.

MATH 304 Linear Algebra Lecture 33: Bases of eigenvectors. Diagonalization. MATH 304 Linear Algebra Lecture 33: Bases of eigenvectors. Diagonalization. Eigenvalues and eigenvectors of an operator Definition. Let V be a vector space and L : V V be a linear operator. A number λ

More information

Dimensionality Reduc1on

Dimensionality Reduc1on Dimensionality Reduc1on contd Aarti Singh Machine Learning 10-601 Nov 10, 2011 Slides Courtesy: Tom Mitchell, Eric Xing, Lawrence Saul 1 Principal Component Analysis (PCA) Principal Components are the

More information

Lecture 5: Random Walks and Markov Chain

Lecture 5: Random Walks and Markov Chain Spectral Graph Theory and Applications WS 20/202 Lecture 5: Random Walks and Markov Chain Lecturer: Thomas Sauerwald & He Sun Introduction to Markov Chains Definition 5.. A sequence of random variables

More information

AN EXAMPLE OF USING STAR COMPLEMENTS IN CLASSIFYING STRONGLY REGULAR GRAPHS

AN EXAMPLE OF USING STAR COMPLEMENTS IN CLASSIFYING STRONGLY REGULAR GRAPHS Faculty of Sciences and Mathematics, University of Niš, Serbia Available at: http://www.pmf.ni.ac.yu/filomat Filomat 22:2 (2008), 53 57 AN EXAMPLE OF USING STAR COMPLEMENTS IN CLASSIFYING STRONGLY REGULAR

More information

Lecture 1 and 2: Introduction and Graph theory basics. Spring EE 194, Networked estimation and control (Prof. Khan) January 23, 2012

Lecture 1 and 2: Introduction and Graph theory basics. Spring EE 194, Networked estimation and control (Prof. Khan) January 23, 2012 Lecture 1 and 2: Introduction and Graph theory basics Spring 2012 - EE 194, Networked estimation and control (Prof. Khan) January 23, 2012 Spring 2012: EE-194-02 Networked estimation and control Schedule

More information

Math 18, Linear Algebra, Lecture C00, Spring 2017 Review and Practice Problems for Final Exam

Math 18, Linear Algebra, Lecture C00, Spring 2017 Review and Practice Problems for Final Exam Math 8, Linear Algebra, Lecture C, Spring 7 Review and Practice Problems for Final Exam. The augmentedmatrix of a linear system has been transformed by row operations into 5 4 8. Determine if the system

More information

Kernels of Directed Graph Laplacians. J. S. Caughman and J.J.P. Veerman

Kernels of Directed Graph Laplacians. J. S. Caughman and J.J.P. Veerman Kernels of Directed Graph Laplacians J. S. Caughman and J.J.P. Veerman Department of Mathematics and Statistics Portland State University PO Box 751, Portland, OR 97207. caughman@pdx.edu, veerman@pdx.edu

More information

MATH 304 Linear Algebra Lecture 34: Review for Test 2.

MATH 304 Linear Algebra Lecture 34: Review for Test 2. MATH 304 Linear Algebra Lecture 34: Review for Test 2. Topics for Test 2 Linear transformations (Leon 4.1 4.3) Matrix transformations Matrix of a linear mapping Similar matrices Orthogonality (Leon 5.1

More information