Finding normalized and modularity cuts by spectral clustering. Ljubjana 2010, October

Size: px
Start display at page:

Download "Finding normalized and modularity cuts by spectral clustering. Ljubjana 2010, October"

Transcription

1 Finding normalized and modularity cuts by spectral clustering Marianna Bolla Institute of Mathematics Budapest University of Technology and Economics Ljubjana 2010, October

2 Outline Find community structure in networks. Antiexpanders, but the clusters are good expanders. How many clusters? Penalize clusters of extremely different sizes/volumes. Minimum multiway cut problems: ratio cut and normalized cut based on Laplacian and normalized Laplacian spectra. Communities with sparse between-cluster (and dense within-cluster) connections. Modularity cuts: spectral decomposition of the modularity and normalized modularity matrix. Communities with more within-cluster (and less between-cluster) connections than expected under independence.

3 Outline Find community structure in networks. Antiexpanders, but the clusters are good expanders. How many clusters? Penalize clusters of extremely different sizes/volumes. Minimum multiway cut problems: ratio cut and normalized cut based on Laplacian and normalized Laplacian spectra. Communities with sparse between-cluster (and dense within-cluster) connections. Modularity cuts: spectral decomposition of the modularity and normalized modularity matrix. Communities with more within-cluster (and less between-cluster) connections than expected under independence.

4 Outline Find community structure in networks. Antiexpanders, but the clusters are good expanders. How many clusters? Penalize clusters of extremely different sizes/volumes. Minimum multiway cut problems: ratio cut and normalized cut based on Laplacian and normalized Laplacian spectra. Communities with sparse between-cluster (and dense within-cluster) connections. Modularity cuts: spectral decomposition of the modularity and normalized modularity matrix. Communities with more within-cluster (and less between-cluster) connections than expected under independence.

5 Outline Find community structure in networks. Antiexpanders, but the clusters are good expanders. How many clusters? Penalize clusters of extremely different sizes/volumes. Minimum multiway cut problems: ratio cut and normalized cut based on Laplacian and normalized Laplacian spectra. Communities with sparse between-cluster (and dense within-cluster) connections. Modularity cuts: spectral decomposition of the modularity and normalized modularity matrix. Communities with more within-cluster (and less between-cluster) connections than expected under independence.

6 Notation G = (V, W): edge-weighted graph on n vertices, W: n n symmetric matrix, w ij 0, w ii = 0. (w ij : similarity between vertices i and j). Simple graph: 0/1 weights W.l.o.g., n n i=1 j=1 w ij = 1, joint distribution with marginal entries: n d i = w ij, i = 1,..., n j=1 (generalized vertex degrees) D = diag (d 1,..., d n ) Laplacian: L = D W Normalized Laplacian L D = I D 1/2 WD 1/2

7 1 k n P k = (V 1,..., V k ): k-partition of the vertices V 1,..., V k : disjoint, non-empty vertex subsets, clusters P k : the set of all k-partitions e(v a, V b ) = i V a j V b w ij : weighted cut between V a and V b Vol (V a ) = i V a d i : volume of V a

8 Ratio cut of P k = (V 1,..., V k ) given W: g(p k, W) = k 1 b=a+1 ( 1 V a + 1 ) e(v a, V b ) = V b Normalized cut of P k = (V 1,..., V k ) given W: f (P k, W) = = k 1 b=a+1 ( e(v a, V a ) V a ) 1 Vol (V a ) + 1 e(v a, V b ) Vol (V b ) e(v a, V a ) Vol (V a ) = k e(v a, V a ) Vol (V a ) Minimum k-way ratio cut and normalized cut of G = (V, W): g k (G) = min P k P k g(p k, W) and f k (G) = min P k P k f (P k, W)

9 The k-means algorithm The problem: given the points x 1,..., x n R d and an integer 1 k n, find the k-partition of the index set {1,..., n} (or equivalently, the clustering of the points into k disjoint non-empty subsets) which minimizes the following k-variance: S 2 k (x 1,..., x n ) = min P k P k S 2 k (P k, x 1,..., x n ) Usually, d k n. = min P k =(V 1,...,V k ) c a = 1 x j. V a j V a j V a x j c a 2,

10 To find the global minimum is NP-complete, but the iteration of the k-means algorithm, first described in MacQueen (1963) is capable to find a local minimum in polynomial time. If there exists a well-separated k-clustering of the points (even the largest within-cluster distance is smaller than the smallest between-cluster one) the convergence of the algorithm to the global minimum is proved by Dunn ( ), with a convenient starting. Under relaxed conditions, the speed of the algorithm is increased by a filtration in Kanungo et al. (2002). The algorithm runs faster if the separation between the clusters increases and an overall running time of O(kn) can be guaranteed.

11 Sometimes the points x 1,..., x n are endowed with the positive weights d 1,..., d n, w.l.o.g., n i=1 d i = 1. Weighted k-variance of the points: S k 2 (x 1,..., x n ) = min S k 2 (P k, x 1,..., x n ) P k P k = min d j x j c a 2, P k =(V 1,...,V k ) j V a 1 c a = d j x j. j V a d j j V a E.g., d 1,..., d n is a discrete probability distribution and a random vector takes on values x 1,..., x n with these probabilities; e.g., in a MANOVA (Multivariate Analysis of Variance) setup. The above algorithms can be easily adapted to this situation.

12 Partition matrices P k : n k balanced partition matrix Z k = (z 1,..., z k ) k-partition vector: z a = (z 1a,..., z na ) T, where z ia = 1 V a, if i V a and 0, otherwise. Z k is suborthogonal: Z T k Z k = I k The ratio cut of the k-partition P k given W: g(p k, W) = tr Z T k LZ k = z T a Lz a. (1) We want to minimize it over balanced k-partition matrices Z k Z B k.

13 Estimation by Laplacian eigenvalues G is connected, the spectrum of L: 0 = λ 1 < λ 2 λ n unit-norm, pairwise orthogonal eigenvectors: u 1, u 2,..., u n, u 1 = 1 The discrete problem is relaxed to a continuous one: r 1,..., r n R k : representatives of the vertices X = (r 1,..., r n ) T = (x 1,..., x k ) n 1 min X T X=I k n i=1 j=i+1 w ij r i r j 2 = min tr X T LX = X T X=I k and equality is attained with X = (u 1,..., u k ). i=1 λ i

14 g k (G) = min Z k Z B k tr Z T k LZ k λ i (2) and equality can be attained only in the k = 1 trivial case, otherwise the eigenvectors u i (i = 2,..., k) cannot be partition vectors, since their coordinates sum to 0 because of the orthogonality to the u 1 = 1 vector. Optimum choice of k? tr Z T k LZ k = n λ i i=1 i=1 (u T i z a ) 2. (3) This sum is the smallest possible if the largest (u T i z a ) 2 terms correspond to eigenvectors belonging to the smallest eigenvalues. Thus, the above sum is the most decreased by keeping only the k smallest eigenvalues in the inner summation and the corresponding eigenvectors are close to the subspace F k = Span {z 1,..., z k }.

15 g k,k (Z k, L) := λ i (u T i z a ) 2, i=1 we maximize g k,k (Z k, L) over Zk B for given L. The vectors λ i u i are projected onto the subspace F k : λi u i = λi (u T i z a )z a + ort Fk ( λ i u i ), i = 1,..., k. As λ 1 u 1 = 0, there is no use of projecting it. By the Pythagorean equality: λ i = λ i u i 2 = λ i (u T i z a ) 2 +dist 2 ( λ i u i, F k ), i = 1,..., k.

16 X k = ( λ 1 u 1,..., λ k u k ) = (r 1,..., r n) T. Sk 2 (X k) = dist 2 ( λ i u i, F k ). i=1 λ i = g k,k (Z k, L) + Sk 2 (P k, X k), (4) i=1 where the partition matrix Z k corresponds to the k-partition P k.

17 We are looking for the k-partition maximizing the first term. In view of (4), increasing g k,k (Z k, L) can be achieved by decreasing Sk 2(X k ); latter one is obtained by appying the k-means algorithm with k clusters for the k-dimensional representatives r 1,..., r n. As the first column of X k is 0, it is equivalent to apply the k-means algorithm with k clusters for the (k 1)-dimensional representatives that are the row vectors of the n (k 1) matrix obtained from X k by deleting its first column. λ k λ k+1 : we gain the most by omitting the n k largest eigenvalues.

18 Perturbation results W = W w + W b : within- and between-cluster edges w.r.t. P k D = D w + D b and L = L w + L b ρ: the smallest positive eigenvalue of L w ε: the largest eigenvalue of L b Suppose ε(p k ) < ρ(p k ) ρ = min ρi is large if the clusters are good expanders ε 2 max i {1,...,n} j: c(j) c(i) w ij, where c(i) is the cluster membership of vertex i. It is small, if from each vertex there are few, small-weight edges emaneting to clusters different of the vertex s own cluster

19 We proved (B, Tusnády, Discrete Math. 1994) that S 2 k (X k ) = min P k P k S 2 k (P k, X k ) k min P k P k ε(p k ) ρ(p k ) By the Weyl s perturbation theorem: 0 < λ k ε < ρ λ k+1 ρ + ε (5) for the ε and ρ of any k-partition with ε < ρ. Consequently, λ k ε(p k ) min λ k+1 P k P k ρ(p k ), (6)

20 Theorem If for a given k-partition ε(p k) ρ(p k ) < 1, then where X k = (u 1,..., u k ). Theorem Sk 2 [ε(p k )] 2 (P k, X k ) k [ρ(p k ) ε(p k )] 2, S 2 k (P k, X k) [ε(p k )] 2 [ρ(p k ) ε(p k )] 2 λ i, i=1 where X k = ( λ 1 u 1,..., λ k u k ). Consequently, S 2 k (X k) i=1 λ i min P k P k ε(p k ) 2 (ρ(p k ) ε(p k )) 2

21 Minimizing the normalized cut n k normalized partition matrix: Z k = (z 1,..., z k ) z a = (z 1a,..., z na ) T, where z ia = 1 Vol (Va), if i V a and 0, otherwise. The normalized cut of the k-partition P k given W: f (P k, W) = tr Z T k LZ k = tr (D 1/2 Z k ) T L D (D 1/2 Z k ) (7) or equivalently, f (P k, W) = k tr (D 1/2 Z k ) T (D 1/2 WD 1/2 )(D 1/2 Z k ). We want to minimize f (P k, W) over the k-partitions. It is equivalent to maximizing tr Z T k WZ k over normalized k-partition matrices Z k Z N k.

22 Normalized Laplacian eigenvalues G is connected, 0 = λ 1 < λ 2 λ p < 1 λ p+1 λ n 2 eigenvalues of L D with corresponding unit-norm, pairwise orthogonal eigenvectors u 1,..., u n, u 1 = ( d 1,..., d n ) T Continuous relaxation: X = (r 1,..., r n ) T = (x 1,..., x k ) n 1 min X T DX=I k n i=1 j=i+1 w ij r i r j 2 = min tr X T LX = X T DX=I k and the minimum is attained with x i = D 1/2 u i (i = 1,..., k). Especially, x 1 = 1. i=1 λ i

23 f k (G) = min Z k Z N k tr Z T k LZ k and equality can be attained only in the k = 1 trivial case, otherwise the transformed eigenvectors D 1/2 u i (i = 2,..., k) cannot be partition vectors, since their coordinates sum to 0 due to the orthogonality of the 1 vector. Equivalently, i=1 λ i max Z k Z N k tr Z T k WZ k k λ i = i=1 (1 λ i ). i=1

24 Optimum choice of k tr Z T k WZ k ) = n (1 λ i ) i=1 [(u i ) T (D 1/2 z a )] 2 increased if we neglect the terms belonging to the eigenvalues at least 1, hence, the outer summation stops at p. The inner sum is the largest in the k = p case, when the unit-norm, pairwise orthogonal vectors D 1/2 z 1,..., D 1/2 z p are close to the orthonormal eigenvectors u 1,..., u p, respectively. F p := Span {D 1/2 z 1,..., D 1/2 z p } f p,p (Z p, W) := p p (1 λ i ) [(u i) T (D 1/2 z a )] 2 i=1 tr (D 1/2 Z p ) T (D 1/2 WD 1/2 )(D 1/2 Z p ) f p,p (Z p, W)

25 For given W, we maximize f p,p (Z p, W) over Z N p. The vectors 1 λ i u i are projected onto the subspace F p : 1 λi u i = p [( 1 λ i u i ) T D 1/2 z a ] D 1/2 z a +ort Fp ( 1 λ i u i ) (i = 1,..., p) As 1 λ 1 u 1 = u 1 is in F p, its orthogonal component is 0. By the Pythagorean equality: 1 λ i = p [( 1 λ i u i ) T D 1/2 z a ] 2 + dist 2 ( 1 λ i u i, F p ) (i = 1,..., p)

26 Representation X p = (r 1,..., r n ) T = ( 1 λ 1 D 1/2 u 1,..., 1 λ p D 1/2 u p ) dist 2 ( 1 λ i u i, F p ) = n d j (r ji c ji ) 2, j=1 i = 1,..., p where r ji is the ith coordinate of the vector r j and c ji is the same for vector c j R p ; further, there are at most p different ones among the centers c 1,..., c n assigned to the vertex representatives. Namely, c ji = 1 d l r li, j V a, i = 1,..., p. Vol (V a ) l V a S 2 p(p p, X p ) = p dist 2 ( 1 λ i u i, F p ) = i=1 n d j r j c j 2. j=1

27 p (1 λ i ) = f p,p (Z p, W) + S p(p 2 p, X p). i=1 We are looking for the p-partition maximizing the first term. In view of the above formula, increasing f p,p (Z p, W) can be achieved by decreasing S 2 p(x p ); latter one is obtained by appying the k-means algorithm with p clusters for the p-dimensional representatives r 1,..., r n with respective weights d 1,..., d n. As the first column of X p is 1, it is equivalent to apply the k-means algorithm with p clusters for the (p 1)-dimensional representatives that are the row vectors of the n (p 1) matrix obtained from X p by deleting its first column.

28 Similarly, for d < k p: d (1 λ i ) = i=1 d [( d 1 λ i u i ) T D 1/2 z a ] 2 + dist 2 ( 1 λ i u i, F k ) i=1 i=1 := f k,d (Z k, W) + S 2 k (P k, X d), X d = ( 1 λ 1 D 1/2 u 1,..., 1 λ d D 1/2 u d ). In the presence of a spectral gap between λ d and λ d+1 < 1 neither d i=1 (1 λ i) nor S k 2(X d ) is increased significantly by introducing one more eigenvalue-eigenvector pair (by using (d + 1)-dimensional representatives instead of d-dimensional ones). Consequently, f k,d (Z k, W) would not change much, and k = d clusters based on d-dimensional representatives will suffice. Increasing the number of clusters (k + 1,..., p) will decrease the sum of the inner variances S k+1 2 (X d ) S k 2(X d ), but not significantly.

29 In the k = p special case tr Z T p LZ p p f p,p (Z p, W) = p λ i + S p(p 2 p, X p) i=1 that is true for the minima too: f p (G) p λ i + S p(x 2 p), i=1 a sharper lower estimator for the minimum normalized p-way cut than p i=1 λ i.

30 Spectral gap and variance In B, Tusnády, Discrete Math., 1994 Theorem In the representation X 2 = (D 1/2 u 1, D 1/2 u 2 ) = (1, D 1/2 u 2 ): Similarly, Theorem S 2 k (X 2 ) λ 2 λ 3 In the representation X 2 = ( 1 λ 1 D 1/2 u 1, 1 λ 2 D 1/2 u 2 ) = (1, 1 λ 2 D 1/2 u 2 ): Can it be generalized for k > 2? S 2 k (X 2) λ 2(1 λ 2 ) λ 3

31 Isoperimetric number Definition The Cheeger constant of the weighted graph G = (V, W ) is h(g) = min U V Vol (U) 1/2 e(u, Ū) Vol (U) Theorem (B, M-Sáska, Discrete Math. 2004). Let λ 2 be the smallest positive eigenvalue of L D. Then λ 2 2 h(g) min{1, 2λ 2 }. If λ 2 1 then the upper estimate can be improved to h(g) λ 2 (2 λ 2 ).

32 f 2 (G) is the symmetric version of h(g): f 2 (G) 2h(G) = f 2 (G) λ 2 (2 λ 2 ), λ 2 1. h(g) and f 2 (G) were estimated by Mohar,J. Comb. Theory B for simple graphs (similar formula by max-degree) Theorem Suppose that G = (V, W) is connected, and λ i s are the eigenvalues of L D. Then k i=1 λ i f k (G) and in the case when the optimal k-dimensional representatives can be classified into k well-separated clusters in such a way that the maximum cluster diameter ε satisfies the relation ε min{1/ 2k, 2 min i Vol (Vi )} with k-partiton (V 1,..., V k ) induced by the clusters above, then f k (G) c 2 λ i, where c = 1 + εc /( 2 εc ) and c = 1/ min i Vol (Vi ). i=1

33 The Newman Girvan modularity for simple graphs G = (V, A): simple graph, V = n, A is 0/1 adjacency matrix number of edges: e = 1 2 i = 1 n i = 1 n a ij d i = n j=1 a ij: the degree of vertex i h ij := d i d j 2e : expected number of i j edges by random attachment P k = (V 1,..., V k ): k-partition of the vertices (modules) Newman and Girvan, Phys. Rev. E, 2004 introduced a modularity with large values for stronger (than expected) intra-module connections.

34 Definition The Newman Girvan modularity of the k-partition P k given A: Q NG (P k, A) = 1 2e i,j V a (a ij h ij ) Obviously, Q NG takes on values less than 1, and due to n i=1 n j=1 a ij = n i=1 n j=1 h ij it is 0 for k = 1 (no community structure at all); therefore, only integers k [2, N] are of interest. For given A and k, we are looking for max Pk P k Q NG (P k, A), more precisely, for the optimum k-partition giving the arg max of it; and eventually, for the optimum k too (there may be several local maxima).

35 H = (h ij ): matrix of rank 1 B = A H: modularity matrix The row sums of B are zeros, therefore it always has a 0 eigenvalue with the trivial eigenvector 1 = (1, 1,..., 1) T. If G is the complete graph, then B is negative semidefinite; but typically, it is indefinite. β 1 β 2 β n eigenvalues of B with unit norm, pairwise orthogonal eigenvectors u 1,..., u n. Newman, Phys. Rev. E, 2006 uses the eigenvectors belonging to the positive eigenvalues of B.

36 Under the null-hypothesis, vertices i and j are connected to each other independently, with probabilities proportional (actually, because of the normalizing condition, equal) to their generalized degrees. for edge-weighted graphs G = (V, W), w.l.o.g. n i=1 n j=1 w ij = 1 supposed h ij = d i d j, i, j = 1,..., n Definition the Newman-Girvan modularity of P k given W: Q(P k, W) = = (w ij h ij ) i,j V a [e(v a, V a ) Vol 2 (V a )],

37 For or given k we maximize Q(P k, W) over P k. This task is equivalent to minimizing (w ij h ij ). i V a, j V b a b We want to penalize partitions with clusters of extremely different sizes or volumes Definition Balanced Newman Girvan modularity of P k given W: Q B (P k, W) = = 1 V a (w ij h ij ) i,j V a [ e(va, V a ) V a Vol 2 ] (V a ), V a

38 Definition Normalized Newman Girvan modularity of P k given W: Q N (P k, W) = = 1 Vol (V a ) e(v a, V a ) Vol (V a ) 1, i,j V a (w ij h ij ) Maximizing the normalized Newman Girvan modularity over P k is equivalent to minimizing the normalized cut.

39 Maximizing the balanced Newman Girvan modularity B = W H: modularity matrix Spectrum: β 1 β p > 0 = β p+1 β n Unit-norm, pairwise orthogonal eigenvectors: u 1,..., u n, u p+1 = 1/ n. Q B (P k, W) = Q B (Z k, B) = z T a Bz a = tr Z T k BZ k. We want to maximize tr Z T k BZ k over balanced k-partition matrices Z k Z B k.

40 Continuous relaxation: max tr (Y T BY) = Y T Y=I k max y T a y b=δ ab ya T By a = and equality is attained when y 1,..., y k are eigenvectors of B corresponding to β 1,..., β k. Though the vectors themselves are not necessarily unique (e.g., in case of multiple eigenvalues), the subspace Span {y 1,..., y k } is unique if β k > β k+1. max Z k Z B k Q B (Z k, B) β a p+1 β a β a. (8) Both inequalities can be attained by equality only in the k = 1, p = 0 case, when our underlying graph is the complete graph: A = 11 T I, W = 1 n(n 1) A. In this case there is only one cluster with partition vector of equal coordinates (balanced eigenvector belonging to the single 0 eigenvalue).

41 The maximum with respect to k of the maximum in (8) is attained with the choice of k = p + 1. Q B (Z k, B) = tr Z T k BZ k = = n β i i=1 n z T a ( β i u i u T i )z a (u T i z a ) 2. We can increase the last sum if we neglect the terms belonging to the negative eigenvalues, hence, the outer summation stops at p, or equivalently, at p + 1. In this case the inner sum is the largerst in the k = p + 1 case, when the partition vectors z 1,..., z p+1 are close to the eigenvectors u 1,..., u p+1, respectively. As both systems consist of orthonormal sets of vectors, the two subspaces spanned by them should be close to each other. The subspace F p+1 = Span {z 1,..., z p+1 } consists of stepwise constant vectors on p + 1 steps, therefore u p+1 F p+1, and it suffices to process only the first p eigenvectors. i=1

42 Q B (Z p+1, B) Q p+1,p(z p+1, B) := p i=1 p+1 β i and in the sequel, for given B, we want to maximize Q p+1,p (Z p+1, B) over Z B p+1. Project the vectors β i u i onto the subspace F p+1 : (u T i z a ) 2, p+1 βi u i = [( β i u i ) T z a ]z a +ort Fp+1 ( β i u i ), i = 1,..., p. (9) p+1 β i = β i u i 2 = [( β i u i ) T z a ] 2 + dist 2 ( β i u i, F p+1 ), i = 1,..., p.

43 p β i = i=1 p p+1 [( β i u i ) T z a ] 2 + i=1 p dist 2 ( β i u i, F p+1 ) i=1 = Q p+1,p(z p+1, B) + S 2 p+1(x p ), where the rows of X p = ( β 1 u 1,..., β p u p ) are regarded as p-dimensional representatives of the vertices. We could as well take (p + 1)-dimensional representatives as the last coordinates are zeros, and hence, S 2 p+1 (X p) = S 2 p+1 (X p+1). Maximizing Q p+1,p is equivalent to minimizing S2 p+1 (X p) that can be obtained by applying the k-means algorithm for the p-dimensional representatives with p + 1 clusters.

44 More generally, if there is a gap between β k β k+1 > 0, then we look for k + 1 clusters based on k-dimensional representatives of the vertices. k leading eigenvectors are projected onto the k-dimensional subspace of F k+1 orthogonal to 1. Calculating eigenvectors is costy; the Lánczos method performs well if we calculate only eigenvectors belonging to some leading eigenvalues followed by a spectral gap. Some authors suggest to use as many eigenvectors as possible. In fact, using more eigenvectors (up to p) is better from the point of view of accuracy, but using less eigenvectors (up to a gap in the positive part of the spectrum) is better from the computational point of view. We have to compromise.

45 Normalized modularity matrix B D = D 1/2 BD 1/2 Q N (P k, W) = Q N (Z k, B) = z T a Bz a = tr (D 1/2 Z k ) T B D (D 1/2 Z k ) 1 = β 1 β n 1: spectrum of B D u 1,..., u n: unit-norm, pairwise orthogonal eigenvectors u 1 = ( d 1,..., d n ) T p: number of positive eigenvalues of B D (this p not necessarily coincides with that of B)

46 max Z k Z N k Q N (Z k, B) = Q N (Z k, B) n β i i=1 p+1 β a β a. [(u i) T (D 1/2 z a )] 2. We can increase this sum if we neglect the terms belonging to the negative eigenvalues, hence, the outer summation stops at p, or equivalently, at p + 1. The inner sum is the largest in the k = p + 1 case, when the unit-norm, pairwise orthogonal vectors D 1/2 z 1,..., D 1/2 z p+1 are close to the eigenvectors u 1,..., u p+1, respectively. In fact, the two subspaces spanned by them should be close to each other.

47 F p+1 = Span {D 1/2 z 1,..., D 1/2 z p+1 } not stepwise constant vectors! X p = ( β 1 D 1/2 u 1,..., β p D 1/2 u p) S 2 p+1(p k, X p) = weighted k-means algorithm gap history = p i=1 dist 2 ( β i u i, F p+1 ) n d j x j c j 2 min. j=1

48 Pure community structure G: disjoint union of k complete graphs on n 1,..., n k vertices Eigenvalues of the modularity matrix: k 1 positive β i s (>1) In the n 1 = = n k special case it is a multiple eigenvalue. 0: single eigenvalue 1: eigenvalue with multiplicity n k Eigenvalues of the normalized modularity matrix: 1: with multiplicity k 1 0: single eigenvalue n k negative eigenvalues in (-1,0) In the n 1 = = n k special case there is one negative eigenvalue with multiplicity n k Piecewise constant eigenvectors belonging to the positive eigenvalues = k clusters

49 Pure anticommunity structure G: complete k-partite graph on n 1,..., n k vertices Eigenvalues of the modularity matrix: k 1 negative eigenvalues 0: all the other eigenvalues Eigenvalues of the normalized modularity matrix: k 1 negative eigenvalues in (-1,0) 0: all the other eigenvalues MINIMIZE THE MODULARITY! Piecewise constant eigenvectors belonging to the negative eigenvalues = k clusters

50 Block matrices Definition The n n symmetric real matrix B is a blown-up matrix, if there is a k k symmetric so-called pattern matrix P with entries 0 p ij 1, and there are positive integers n 1,..., n k with k i=1 n i = n, such that after rearranging its rows and columns the matrix B can be divided into k k blocks, where block (i, j) is an n i n j matrix with entries all equal to p ij (1 i, j n). G = (V, B) is a weighted graph with possible loops Generalized random graph: G is a random graph whose edges exist independently, with probabilities p ij. Modularity matrix: k 1 structural (large absolute value) eigenvalues = k clusters by the corresponding eigenvectors Pure community structure: p ii = 1, p ij = 0 (i j) Pure anticommunity structure: p ii = 0, p ij = 1 (i j)

51 Eigenvalues of blown-up matrices and their Laplacians GENERAL CASE p ij s are arbitrary, P = B := diag (n 1,..., n k ) with n i cn (i = 1,..., k) rank (B) = k, non-zero eigenvalues: n λ i = Θ(n) with stepwise constant eigenvectors λ i s are eigenvalues of 1/2 P 1/2 Laplacian eigenvalues: 0: single eigenvalue λ 1,..., λ k 1 = Θ(n) with stepwise constant eigenvectors γ i = Θ(n) with multiplicity n i 1, eigenspace: vectors with coordinates 0 except block i, where the sum of the coordinates is 0 (i = 1,..., k)

52 Normalized Laplacian eigenvalues: There exists δ > 0 (independent of n) such that there are k eigenvalues [0, 1 δ] [1 + δ, 2] with stepwise constant eigenvectors. All the other eigenvalues are 1.

53 COMPLETE k-partite GRAPH p ii = 0, p ij = 1 (i j) non-zero eigenvalues of B: λ 1,..., λ k =Θ(n) with stepwise constant eigenvectors λ 1,..., λ k [ max i n i, min i n i ] [n max i n i, n min i n i ] Laplacian eigenvalues: 0: single eigenvalue λ 1 = = λ k 1 = n with stepwise constant eigenvectors γ i = n n i with multiplicity n i 1, eigenspace: vectors with coordinates 0 except block i, where the sum of the coordinates is 0 (i = 1,..., k)

54 Normalized Laplacian eigenvalues: There exists δ > 0 (independent of n) such that there are k 1 eigenvalues [1 + δ, 2] with stepwise constant eigenvectors. 0 is a single eigenvalue, all the other eigenvalues are 1.

55 k DISJOINT CLUSTERS A = k i=1 A i block-diagonal Diagonal/non-diagonal entries of the n i n i matrix A i : ν i /µ i Large eigenvalues of A: λ i = (n i 1)µ i + ν i = Θ(n) with stepwise constant eigenvectors. Small eigenvalues of A: ν i µ i with multiplicity n i 1 (i = 1,..., k). Laplacian eigenvalues: 0: with multiplicity k, eigenspace: stepwise constant vectors. n i µ i with with multiplicity n i 1 (i = 1,..., k) Normalized Laplacian eigenvalues: 0: with multiplicity k. n i µ i ν i (n i 1)µ i 1 with with multiplicity n i 1 (i = 1,..., k). Union of k complete graphs (ν i = 0). Non-0 e.v. s: n i n i 1 > 1.

SVD, discrepancy, and regular structure of contingency tables

SVD, discrepancy, and regular structure of contingency tables SVD, discrepancy, and regular structure of contingency tables Marianna Bolla Institute of Mathematics Technical University of Budapest (Research supported by the TÁMOP-4.2.2.B-10/1-2010-0009 project) marib@math.bme.hu

More information

Spectral properties of modularity matrices

Spectral properties of modularity matrices Spectral properties of modularity matrices Marianna Bolla a,, Brian Bullins b, Sorathan Chaturapruek c, Shiwen Chen d, Katalin Friedl e a Institute of Mathematics, Budapest University of Technology and

More information

1 Matrix notation and preliminaries from spectral graph theory

1 Matrix notation and preliminaries from spectral graph theory Graph clustering (or community detection or graph partitioning) is one of the most studied problems in network analysis. One reason for this is that there are a variety of ways to define a cluster or community.

More information

Chapter 3 Transformations

Chapter 3 Transformations Chapter 3 Transformations An Introduction to Optimization Spring, 2014 Wei-Ta Chu 1 Linear Transformations A function is called a linear transformation if 1. for every and 2. for every If we fix the bases

More information

Data Mining and Analysis: Fundamental Concepts and Algorithms

Data Mining and Analysis: Fundamental Concepts and Algorithms : Fundamental Concepts and Algorithms dataminingbook.info Mohammed J. Zaki 1 Wagner Meira Jr. 2 1 Department of Computer Science Rensselaer Polytechnic Institute, Troy, NY, USA 2 Department of Computer

More information

A lower bound for the Laplacian eigenvalues of a graph proof of a conjecture by Guo

A lower bound for the Laplacian eigenvalues of a graph proof of a conjecture by Guo A lower bound for the Laplacian eigenvalues of a graph proof of a conjecture by Guo A. E. Brouwer & W. H. Haemers 2008-02-28 Abstract We show that if µ j is the j-th largest Laplacian eigenvalue, and d

More information

1 Matrix notation and preliminaries from spectral graph theory

1 Matrix notation and preliminaries from spectral graph theory Graph clustering (or community detection or graph partitioning) is one of the most studied problems in network analysis. One reason for this is that there are a variety of ways to define a cluster or community.

More information

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra. DS-GA 1002 Lecture notes 0 Fall 2016 Linear Algebra These notes provide a review of basic concepts in linear algebra. 1 Vector spaces You are no doubt familiar with vectors in R 2 or R 3, i.e. [ ] 1.1

More information

Statistical Inference on Large Contingency Tables: Convergence, Testability, Stability. COMPSTAT 2010 Paris, August 23, 2010

Statistical Inference on Large Contingency Tables: Convergence, Testability, Stability. COMPSTAT 2010 Paris, August 23, 2010 Statistical Inference on Large Contingency Tables: Convergence, Testability, Stability Marianna Bolla Institute of Mathematics Budapest University of Technology and Economics marib@math.bme.hu COMPSTAT

More information

Lecture 12 : Graph Laplacians and Cheeger s Inequality

Lecture 12 : Graph Laplacians and Cheeger s Inequality CPS290: Algorithmic Foundations of Data Science March 7, 2017 Lecture 12 : Graph Laplacians and Cheeger s Inequality Lecturer: Kamesh Munagala Scribe: Kamesh Munagala Graph Laplacian Maybe the most beautiful

More information

Linear algebra and applications to graphs Part 1

Linear algebra and applications to graphs Part 1 Linear algebra and applications to graphs Part 1 Written up by Mikhail Belkin and Moon Duchin Instructor: Laszlo Babai June 17, 2001 1 Basic Linear Algebra Exercise 1.1 Let V and W be linear subspaces

More information

Certifying the Global Optimality of Graph Cuts via Semidefinite Programming: A Theoretic Guarantee for Spectral Clustering

Certifying the Global Optimality of Graph Cuts via Semidefinite Programming: A Theoretic Guarantee for Spectral Clustering Certifying the Global Optimality of Graph Cuts via Semidefinite Programming: A Theoretic Guarantee for Spectral Clustering Shuyang Ling Courant Institute of Mathematical Sciences, NYU Aug 13, 2018 Joint

More information

Ma/CS 6b Class 23: Eigenvalues in Regular Graphs

Ma/CS 6b Class 23: Eigenvalues in Regular Graphs Ma/CS 6b Class 3: Eigenvalues in Regular Graphs By Adam Sheffer Recall: The Spectrum of a Graph Consider a graph G = V, E and let A be the adjacency matrix of G. The eigenvalues of G are the eigenvalues

More information

Last Time. Social Network Graphs Betweenness. Graph Laplacian. Girvan-Newman Algorithm. Spectral Bisection

Last Time. Social Network Graphs Betweenness. Graph Laplacian. Girvan-Newman Algorithm. Spectral Bisection Eigenvalue Problems Last Time Social Network Graphs Betweenness Girvan-Newman Algorithm Graph Laplacian Spectral Bisection λ 2, w 2 Today Small deviation into eigenvalue problems Formulation Standard eigenvalue

More information

Lecture Notes 1: Vector spaces

Lecture Notes 1: Vector spaces Optimization-based data analysis Fall 2017 Lecture Notes 1: Vector spaces In this chapter we review certain basic concepts of linear algebra, highlighting their application to signal processing. 1 Vector

More information

COMPSCI 514: Algorithms for Data Science

COMPSCI 514: Algorithms for Data Science COMPSCI 514: Algorithms for Data Science Arya Mazumdar University of Massachusetts at Amherst Fall 2018 Lecture 8 Spectral Clustering Spectral clustering Curse of dimensionality Dimensionality Reduction

More information

Dimension. Eigenvalue and eigenvector

Dimension. Eigenvalue and eigenvector Dimension. Eigenvalue and eigenvector Math 112, week 9 Goals: Bases, dimension, rank-nullity theorem. Eigenvalue and eigenvector. Suggested Textbook Readings: Sections 4.5, 4.6, 5.1, 5.2 Week 9: Dimension,

More information

Linear Algebra Massoud Malek

Linear Algebra Massoud Malek CSUEB Linear Algebra Massoud Malek Inner Product and Normed Space In all that follows, the n n identity matrix is denoted by I n, the n n zero matrix by Z n, and the zero vector by θ n An inner product

More information

Real symmetric matrices/1. 1 Eigenvalues and eigenvectors

Real symmetric matrices/1. 1 Eigenvalues and eigenvectors Real symmetric matrices 1 Eigenvalues and eigenvectors We use the convention that vectors are row vectors and matrices act on the right. Let A be a square matrix with entries in a field F; suppose that

More information

Multivariate Statistical Analysis

Multivariate Statistical Analysis Multivariate Statistical Analysis Fall 2011 C. L. Williams, Ph.D. Lecture 4 for Applied Multivariate Analysis Outline 1 Eigen values and eigen vectors Characteristic equation Some properties of eigendecompositions

More information

Spectra of Adjacency and Laplacian Matrices

Spectra of Adjacency and Laplacian Matrices Spectra of Adjacency and Laplacian Matrices Definition: University of Alicante (Spain) Matrix Computing (subject 3168 Degree in Maths) 30 hours (theory)) + 15 hours (practical assignment) Contents 1. Spectra

More information

LINEAR ALGEBRA 1, 2012-I PARTIAL EXAM 3 SOLUTIONS TO PRACTICE PROBLEMS

LINEAR ALGEBRA 1, 2012-I PARTIAL EXAM 3 SOLUTIONS TO PRACTICE PROBLEMS LINEAR ALGEBRA, -I PARTIAL EXAM SOLUTIONS TO PRACTICE PROBLEMS Problem (a) For each of the two matrices below, (i) determine whether it is diagonalizable, (ii) determine whether it is orthogonally diagonalizable,

More information

Spectral Clustering. Spectral Clustering? Two Moons Data. Spectral Clustering Algorithm: Bipartioning. Spectral methods

Spectral Clustering. Spectral Clustering? Two Moons Data. Spectral Clustering Algorithm: Bipartioning. Spectral methods Spectral Clustering Seungjin Choi Department of Computer Science POSTECH, Korea seungjin@postech.ac.kr 1 Spectral methods Spectral Clustering? Methods using eigenvectors of some matrices Involve eigen-decomposition

More information

2. Matrix Algebra and Random Vectors

2. Matrix Algebra and Random Vectors 2. Matrix Algebra and Random Vectors 2.1 Introduction Multivariate data can be conveniently display as array of numbers. In general, a rectangular array of numbers with, for instance, n rows and p columns

More information

Spectral Clustering. Zitao Liu

Spectral Clustering. Zitao Liu Spectral Clustering Zitao Liu Agenda Brief Clustering Review Similarity Graph Graph Laplacian Spectral Clustering Algorithm Graph Cut Point of View Random Walk Point of View Perturbation Theory Point of

More information

Lecture 10: October 27, 2016

Lecture 10: October 27, 2016 Mathematical Toolkit Autumn 206 Lecturer: Madhur Tulsiani Lecture 0: October 27, 206 The conjugate gradient method In the last lecture we saw the steepest descent or gradient descent method for finding

More information

Data Analysis and Manifold Learning Lecture 7: Spectral Clustering

Data Analysis and Manifold Learning Lecture 7: Spectral Clustering Data Analysis and Manifold Learning Lecture 7: Spectral Clustering Radu Horaud INRIA Grenoble Rhone-Alpes, France Radu.Horaud@inrialpes.fr http://perception.inrialpes.fr/ Outline of Lecture 7 What is spectral

More information

MATH 20F: LINEAR ALGEBRA LECTURE B00 (T. KEMP)

MATH 20F: LINEAR ALGEBRA LECTURE B00 (T. KEMP) MATH 20F: LINEAR ALGEBRA LECTURE B00 (T KEMP) Definition 01 If T (x) = Ax is a linear transformation from R n to R m then Nul (T ) = {x R n : T (x) = 0} = Nul (A) Ran (T ) = {Ax R m : x R n } = {b R m

More information

8.1 Concentration inequality for Gaussian random matrix (cont d)

8.1 Concentration inequality for Gaussian random matrix (cont d) MGMT 69: Topics in High-dimensional Data Analysis Falll 26 Lecture 8: Spectral clustering and Laplacian matrices Lecturer: Jiaming Xu Scribe: Hyun-Ju Oh and Taotao He, October 4, 26 Outline Concentration

More information

7. Symmetric Matrices and Quadratic Forms

7. Symmetric Matrices and Quadratic Forms Linear Algebra 7. Symmetric Matrices and Quadratic Forms CSIE NCU 1 7. Symmetric Matrices and Quadratic Forms 7.1 Diagonalization of symmetric matrices 2 7.2 Quadratic forms.. 9 7.4 The singular value

More information

1. What is the determinant of the following matrix? a 1 a 2 4a 3 2a 2 b 1 b 2 4b 3 2b c 1. = 4, then det

1. What is the determinant of the following matrix? a 1 a 2 4a 3 2a 2 b 1 b 2 4b 3 2b c 1. = 4, then det What is the determinant of the following matrix? 3 4 3 4 3 4 4 3 A 0 B 8 C 55 D 0 E 60 If det a a a 3 b b b 3 c c c 3 = 4, then det a a 4a 3 a b b 4b 3 b c c c 3 c = A 8 B 6 C 4 D E 3 Let A be an n n matrix

More information

The Eigenvalue Problem: Perturbation Theory

The Eigenvalue Problem: Perturbation Theory Jim Lambers MAT 610 Summer Session 2009-10 Lecture 13 Notes These notes correspond to Sections 7.2 and 8.1 in the text. The Eigenvalue Problem: Perturbation Theory The Unsymmetric Eigenvalue Problem Just

More information

Foundations of Matrix Analysis

Foundations of Matrix Analysis 1 Foundations of Matrix Analysis In this chapter we recall the basic elements of linear algebra which will be employed in the remainder of the text For most of the proofs as well as for the details, the

More information

Semidefinite and Second Order Cone Programming Seminar Fall 2001 Lecture 5

Semidefinite and Second Order Cone Programming Seminar Fall 2001 Lecture 5 Semidefinite and Second Order Cone Programming Seminar Fall 2001 Lecture 5 Instructor: Farid Alizadeh Scribe: Anton Riabov 10/08/2001 1 Overview We continue studying the maximum eigenvalue SDP, and generalize

More information

1 Review: symmetric matrices, their eigenvalues and eigenvectors

1 Review: symmetric matrices, their eigenvalues and eigenvectors Cornell University, Fall 2012 Lecture notes on spectral methods in algorithm design CS 6820: Algorithms Studying the eigenvalues and eigenvectors of matrices has powerful consequences for at least three

More information

1 Last time: least-squares problems

1 Last time: least-squares problems MATH Linear algebra (Fall 07) Lecture Last time: least-squares problems Definition. If A is an m n matrix and b R m, then a least-squares solution to the linear system Ax = b is a vector x R n such that

More information

Throughout these notes we assume V, W are finite dimensional inner product spaces over C.

Throughout these notes we assume V, W are finite dimensional inner product spaces over C. Math 342 - Linear Algebra II Notes Throughout these notes we assume V, W are finite dimensional inner product spaces over C 1 Upper Triangular Representation Proposition: Let T L(V ) There exists an orthonormal

More information

Linear Algebra in Actuarial Science: Slides to the lecture

Linear Algebra in Actuarial Science: Slides to the lecture Linear Algebra in Actuarial Science: Slides to the lecture Fall Semester 2010/2011 Linear Algebra is a Tool-Box Linear Equation Systems Discretization of differential equations: solving linear equations

More information

Linear Algebra using Dirac Notation: Pt. 2

Linear Algebra using Dirac Notation: Pt. 2 Linear Algebra using Dirac Notation: Pt. 2 PHYS 476Q - Southern Illinois University February 6, 2018 PHYS 476Q - Southern Illinois University Linear Algebra using Dirac Notation: Pt. 2 February 6, 2018

More information

Quantum Computing Lecture 2. Review of Linear Algebra

Quantum Computing Lecture 2. Review of Linear Algebra Quantum Computing Lecture 2 Review of Linear Algebra Maris Ozols Linear algebra States of a quantum system form a vector space and their transformations are described by linear operators Vector spaces

More information

MATH 304 Linear Algebra Lecture 34: Review for Test 2.

MATH 304 Linear Algebra Lecture 34: Review for Test 2. MATH 304 Linear Algebra Lecture 34: Review for Test 2. Topics for Test 2 Linear transformations (Leon 4.1 4.3) Matrix transformations Matrix of a linear mapping Similar matrices Orthogonality (Leon 5.1

More information

Singular value decomposition (SVD) of large random matrices. India, 2010

Singular value decomposition (SVD) of large random matrices. India, 2010 Singular value decomposition (SVD) of large random matrices Marianna Bolla Budapest University of Technology and Economics marib@math.bme.hu India, 2010 Motivation New challenge of multivariate statistics:

More information

Part I: Preliminary Results. Pak K. Chan, Martine Schlag and Jason Zien. Computer Engineering Board of Studies. University of California, Santa Cruz

Part I: Preliminary Results. Pak K. Chan, Martine Schlag and Jason Zien. Computer Engineering Board of Studies. University of California, Santa Cruz Spectral K-Way Ratio-Cut Partitioning Part I: Preliminary Results Pak K. Chan, Martine Schlag and Jason Zien Computer Engineering Board of Studies University of California, Santa Cruz May, 99 Abstract

More information

Boolean Inner-Product Spaces and Boolean Matrices

Boolean Inner-Product Spaces and Boolean Matrices Boolean Inner-Product Spaces and Boolean Matrices Stan Gudder Department of Mathematics, University of Denver, Denver CO 80208 Frédéric Latrémolière Department of Mathematics, University of Denver, Denver

More information

Knowledge Discovery and Data Mining 1 (VO) ( )

Knowledge Discovery and Data Mining 1 (VO) ( ) Knowledge Discovery and Data Mining 1 (VO) (707.003) Review of Linear Algebra Denis Helic KTI, TU Graz Oct 9, 2014 Denis Helic (KTI, TU Graz) KDDM1 Oct 9, 2014 1 / 74 Big picture: KDDM Probability Theory

More information

Econ Slides from Lecture 7

Econ Slides from Lecture 7 Econ 205 Sobel Econ 205 - Slides from Lecture 7 Joel Sobel August 31, 2010 Linear Algebra: Main Theory A linear combination of a collection of vectors {x 1,..., x k } is a vector of the form k λ ix i for

More information

MATH 240 Spring, Chapter 1: Linear Equations and Matrices

MATH 240 Spring, Chapter 1: Linear Equations and Matrices MATH 240 Spring, 2006 Chapter Summaries for Kolman / Hill, Elementary Linear Algebra, 8th Ed. Sections 1.1 1.6, 2.1 2.2, 3.2 3.8, 4.3 4.5, 5.1 5.3, 5.5, 6.1 6.5, 7.1 7.2, 7.4 DEFINITIONS Chapter 1: Linear

More information

Vector spaces. DS-GA 1013 / MATH-GA 2824 Optimization-based Data Analysis.

Vector spaces. DS-GA 1013 / MATH-GA 2824 Optimization-based Data Analysis. Vector spaces DS-GA 1013 / MATH-GA 2824 Optimization-based Data Analysis http://www.cims.nyu.edu/~cfgranda/pages/obda_fall17/index.html Carlos Fernandez-Granda Vector space Consists of: A set V A scalar

More information

Math 4153 Exam 3 Review. The syllabus for Exam 3 is Chapter 6 (pages ), Chapter 7 through page 137, and Chapter 8 through page 182 in Axler.

Math 4153 Exam 3 Review. The syllabus for Exam 3 is Chapter 6 (pages ), Chapter 7 through page 137, and Chapter 8 through page 182 in Axler. Math 453 Exam 3 Review The syllabus for Exam 3 is Chapter 6 (pages -2), Chapter 7 through page 37, and Chapter 8 through page 82 in Axler.. You should be sure to know precise definition of the terms we

More information

Spectral Graph Theory and its Applications. Daniel A. Spielman Dept. of Computer Science Program in Applied Mathematics Yale Unviersity

Spectral Graph Theory and its Applications. Daniel A. Spielman Dept. of Computer Science Program in Applied Mathematics Yale Unviersity Spectral Graph Theory and its Applications Daniel A. Spielman Dept. of Computer Science Program in Applied Mathematics Yale Unviersity Outline Adjacency matrix and Laplacian Intuition, spectral graph drawing

More information

Laplacian Integral Graphs with Maximum Degree 3

Laplacian Integral Graphs with Maximum Degree 3 Laplacian Integral Graphs with Maximum Degree Steve Kirkland Department of Mathematics and Statistics University of Regina Regina, Saskatchewan, Canada S4S 0A kirkland@math.uregina.ca Submitted: Nov 5,

More information

LINEAR ALGEBRA QUESTION BANK

LINEAR ALGEBRA QUESTION BANK LINEAR ALGEBRA QUESTION BANK () ( points total) Circle True or False: TRUE / FALSE: If A is any n n matrix, and I n is the n n identity matrix, then I n A = AI n = A. TRUE / FALSE: If A, B are n n matrices,

More information

Math Matrix Algebra

Math Matrix Algebra Math 44 - Matrix Algebra Review notes - (Alberto Bressan, Spring 7) sec: Orthogonal diagonalization of symmetric matrices When we seek to diagonalize a general n n matrix A, two difficulties may arise:

More information

The University of Texas at Austin Department of Electrical and Computer Engineering. EE381V: Large Scale Learning Spring 2013.

The University of Texas at Austin Department of Electrical and Computer Engineering. EE381V: Large Scale Learning Spring 2013. The University of Texas at Austin Department of Electrical and Computer Engineering EE381V: Large Scale Learning Spring 2013 Assignment 1 Caramanis/Sanghavi Due: Thursday, Feb. 7, 2013. (Problems 1 and

More information

COMMUNITY DETECTION IN SPARSE NETWORKS VIA GROTHENDIECK S INEQUALITY

COMMUNITY DETECTION IN SPARSE NETWORKS VIA GROTHENDIECK S INEQUALITY COMMUNITY DETECTION IN SPARSE NETWORKS VIA GROTHENDIECK S INEQUALITY OLIVIER GUÉDON AND ROMAN VERSHYNIN Abstract. We present a simple and flexible method to prove consistency of semidefinite optimization

More information

Machine Learning for Data Science (CS4786) Lecture 11

Machine Learning for Data Science (CS4786) Lecture 11 Machine Learning for Data Science (CS4786) Lecture 11 Spectral clustering Course Webpage : http://www.cs.cornell.edu/courses/cs4786/2016sp/ ANNOUNCEMENT 1 Assignment P1 the Diagnostic assignment 1 will

More information

MATH 315 Linear Algebra Homework #1 Assigned: August 20, 2018

MATH 315 Linear Algebra Homework #1 Assigned: August 20, 2018 Homework #1 Assigned: August 20, 2018 Review the following subjects involving systems of equations and matrices from Calculus II. Linear systems of equations Converting systems to matrix form Pivot entry

More information

Introduction to Spectral Graph Theory and Graph Clustering

Introduction to Spectral Graph Theory and Graph Clustering Introduction to Spectral Graph Theory and Graph Clustering Chengming Jiang ECS 231 Spring 2016 University of California, Davis 1 / 40 Motivation Image partitioning in computer vision 2 / 40 Motivation

More information

Notes on the Matrix-Tree theorem and Cayley s tree enumerator

Notes on the Matrix-Tree theorem and Cayley s tree enumerator Notes on the Matrix-Tree theorem and Cayley s tree enumerator 1 Cayley s tree enumerator Recall that the degree of a vertex in a tree (or in any graph) is the number of edges emanating from it We will

More information

Data Analysis and Manifold Learning Lecture 3: Graphs, Graph Matrices, and Graph Embeddings

Data Analysis and Manifold Learning Lecture 3: Graphs, Graph Matrices, and Graph Embeddings Data Analysis and Manifold Learning Lecture 3: Graphs, Graph Matrices, and Graph Embeddings Radu Horaud INRIA Grenoble Rhone-Alpes, France Radu.Horaud@inrialpes.fr http://perception.inrialpes.fr/ Outline

More information

MATH 221, Spring Homework 10 Solutions

MATH 221, Spring Homework 10 Solutions MATH 22, Spring 28 - Homework Solutions Due Tuesday, May Section 52 Page 279, Problem 2: 4 λ A λi = and the characteristic polynomial is det(a λi) = ( 4 λ)( λ) ( )(6) = λ 6 λ 2 +λ+2 The solutions to the

More information

Bare minimum on matrix algebra. Psychology 588: Covariance structure and factor models

Bare minimum on matrix algebra. Psychology 588: Covariance structure and factor models Bare minimum on matrix algebra Psychology 588: Covariance structure and factor models Matrix multiplication 2 Consider three notations for linear combinations y11 y1 m x11 x 1p b11 b 1m y y x x b b n1

More information

Math Solutions to homework 5

Math Solutions to homework 5 Math 75 - Solutions to homework 5 Cédric De Groote November 9, 207 Problem (7. in the book): Let {e n } be a complete orthonormal sequence in a Hilbert space H and let λ n C for n N. Show that there is

More information

CSCI 1951-G Optimization Methods in Finance Part 10: Conic Optimization

CSCI 1951-G Optimization Methods in Finance Part 10: Conic Optimization CSCI 1951-G Optimization Methods in Finance Part 10: Conic Optimization April 6, 2018 1 / 34 This material is covered in the textbook, Chapters 9 and 10. Some of the materials are taken from it. Some of

More information

MAT Linear Algebra Collection of sample exams

MAT Linear Algebra Collection of sample exams MAT 342 - Linear Algebra Collection of sample exams A-x. (0 pts Give the precise definition of the row echelon form. 2. ( 0 pts After performing row reductions on the augmented matrix for a certain system

More information

Eigenvalues and Eigenvectors A =

Eigenvalues and Eigenvectors A = Eigenvalues and Eigenvectors Definition 0 Let A R n n be an n n real matrix A number λ R is a real eigenvalue of A if there exists a nonzero vector v R n such that A v = λ v The vector v is called an eigenvector

More information

CS168: The Modern Algorithmic Toolbox Lectures #11 and #12: Spectral Graph Theory

CS168: The Modern Algorithmic Toolbox Lectures #11 and #12: Spectral Graph Theory CS168: The Modern Algorithmic Toolbox Lectures #11 and #12: Spectral Graph Theory Tim Roughgarden & Gregory Valiant May 2, 2016 Spectral graph theory is the powerful and beautiful theory that arises from

More information

Ir O D = D = ( ) Section 2.6 Example 1. (Bottom of page 119) dim(v ) = dim(l(v, W )) = dim(v ) dim(f ) = dim(v )

Ir O D = D = ( ) Section 2.6 Example 1. (Bottom of page 119) dim(v ) = dim(l(v, W )) = dim(v ) dim(f ) = dim(v ) Section 3.2 Theorem 3.6. Let A be an m n matrix of rank r. Then r m, r n, and, by means of a finite number of elementary row and column operations, A can be transformed into the matrix ( ) Ir O D = 1 O

More information

Spectral Clustering. Guokun Lai 2016/10

Spectral Clustering. Guokun Lai 2016/10 Spectral Clustering Guokun Lai 2016/10 1 / 37 Organization Graph Cut Fundamental Limitations of Spectral Clustering Ng 2002 paper (if we have time) 2 / 37 Notation We define a undirected weighted graph

More information

Math Camp Lecture 4: Linear Algebra. Xiao Yu Wang. Aug 2010 MIT. Xiao Yu Wang (MIT) Math Camp /10 1 / 88

Math Camp Lecture 4: Linear Algebra. Xiao Yu Wang. Aug 2010 MIT. Xiao Yu Wang (MIT) Math Camp /10 1 / 88 Math Camp 2010 Lecture 4: Linear Algebra Xiao Yu Wang MIT Aug 2010 Xiao Yu Wang (MIT) Math Camp 2010 08/10 1 / 88 Linear Algebra Game Plan Vector Spaces Linear Transformations and Matrices Determinant

More information

Lecture 7: Positive Semidefinite Matrices

Lecture 7: Positive Semidefinite Matrices Lecture 7: Positive Semidefinite Matrices Rajat Mittal IIT Kanpur The main aim of this lecture note is to prepare your background for semidefinite programming. We have already seen some linear algebra.

More information

2.2. Show that U 0 is a vector space. For each α 0 in F, show by example that U α does not satisfy closure.

2.2. Show that U 0 is a vector space. For each α 0 in F, show by example that U α does not satisfy closure. Hints for Exercises 1.3. This diagram says that f α = β g. I will prove f injective g injective. You should show g injective f injective. Assume f is injective. Now suppose g(x) = g(y) for some x, y A.

More information

Using Laplacian Eigenvalues and Eigenvectors in the Analysis of Frequency Assignment Problems

Using Laplacian Eigenvalues and Eigenvectors in the Analysis of Frequency Assignment Problems Using Laplacian Eigenvalues and Eigenvectors in the Analysis of Frequency Assignment Problems Jan van den Heuvel and Snežana Pejić Department of Mathematics London School of Economics Houghton Street,

More information

The following definition is fundamental.

The following definition is fundamental. 1. Some Basics from Linear Algebra With these notes, I will try and clarify certain topics that I only quickly mention in class. First and foremost, I will assume that you are familiar with many basic

More information

Elementary linear algebra

Elementary linear algebra Chapter 1 Elementary linear algebra 1.1 Vector spaces Vector spaces owe their importance to the fact that so many models arising in the solutions of specific problems turn out to be vector spaces. The

More information

Lecture Note 12: The Eigenvalue Problem

Lecture Note 12: The Eigenvalue Problem MATH 5330: Computational Methods of Linear Algebra Lecture Note 12: The Eigenvalue Problem 1 Theoretical Background Xianyi Zeng Department of Mathematical Sciences, UTEP The eigenvalue problem is a classical

More information

SVD, Power method, and Planted Graph problems (+ eigenvalues of random matrices)

SVD, Power method, and Planted Graph problems (+ eigenvalues of random matrices) Chapter 14 SVD, Power method, and Planted Graph problems (+ eigenvalues of random matrices) Today we continue the topic of low-dimensional approximation to datasets and matrices. Last time we saw the singular

More information

MTH 2032 SemesterII

MTH 2032 SemesterII MTH 202 SemesterII 2010-11 Linear Algebra Worked Examples Dr. Tony Yee Department of Mathematics and Information Technology The Hong Kong Institute of Education December 28, 2011 ii Contents Table of Contents

More information

Principal Component Analysis

Principal Component Analysis Machine Learning Michaelmas 2017 James Worrell Principal Component Analysis 1 Introduction 1.1 Goals of PCA Principal components analysis (PCA) is a dimensionality reduction technique that can be used

More information

Symmetric and self-adjoint matrices

Symmetric and self-adjoint matrices Symmetric and self-adjoint matrices A matrix A in M n (F) is called symmetric if A T = A, ie A ij = A ji for each i, j; and self-adjoint if A = A, ie A ij = A ji or each i, j Note for A in M n (R) that

More information

MATH 1120 (LINEAR ALGEBRA 1), FINAL EXAM FALL 2011 SOLUTIONS TO PRACTICE VERSION

MATH 1120 (LINEAR ALGEBRA 1), FINAL EXAM FALL 2011 SOLUTIONS TO PRACTICE VERSION MATH (LINEAR ALGEBRA ) FINAL EXAM FALL SOLUTIONS TO PRACTICE VERSION Problem (a) For each matrix below (i) find a basis for its column space (ii) find a basis for its row space (iii) determine whether

More information

The University of Texas at Austin Department of Electrical and Computer Engineering. EE381V: Large Scale Learning Spring 2013.

The University of Texas at Austin Department of Electrical and Computer Engineering. EE381V: Large Scale Learning Spring 2013. The University of Texas at Austin Department of Electrical and Computer Engineering EE381V: Large Scale Learning Spring 2013 Assignment Two Caramanis/Sanghavi Due: Tuesday, Feb. 19, 2013. Computational

More information

CSC Linear Programming and Combinatorial Optimization Lecture 10: Semidefinite Programming

CSC Linear Programming and Combinatorial Optimization Lecture 10: Semidefinite Programming CSC2411 - Linear Programming and Combinatorial Optimization Lecture 10: Semidefinite Programming Notes taken by Mike Jamieson March 28, 2005 Summary: In this lecture, we introduce semidefinite programming

More information

Relations Between Adjacency And Modularity Graph Partitioning: Principal Component Analysis vs. Modularity Component Analysis

Relations Between Adjacency And Modularity Graph Partitioning: Principal Component Analysis vs. Modularity Component Analysis Relations Between Adjacency And Modularity Graph Partitioning: Principal Component Analysis vs. Modularity Component Analysis Hansi Jiang Carl Meyer North Carolina State University October 27, 2015 1 /

More information

1. General Vector Spaces

1. General Vector Spaces 1.1. Vector space axioms. 1. General Vector Spaces Definition 1.1. Let V be a nonempty set of objects on which the operations of addition and scalar multiplication are defined. By addition we mean a rule

More information

Definitions for Quizzes

Definitions for Quizzes Definitions for Quizzes Italicized text (or something close to it) will be given to you. Plain text is (an example of) what you should write as a definition. [Bracketed text will not be given, nor does

More information

Basic Concepts in Matrix Algebra

Basic Concepts in Matrix Algebra Basic Concepts in Matrix Algebra An column array of p elements is called a vector of dimension p and is written as x p 1 = x 1 x 2. x p. The transpose of the column vector x p 1 is row vector x = [x 1

More information

Review problems for MA 54, Fall 2004.

Review problems for MA 54, Fall 2004. Review problems for MA 54, Fall 2004. Below are the review problems for the final. They are mostly homework problems, or very similar. If you are comfortable doing these problems, you should be fine on

More information

ON THE QUALITY OF SPECTRAL SEPARATORS

ON THE QUALITY OF SPECTRAL SEPARATORS ON THE QUALITY OF SPECTRAL SEPARATORS STEPHEN GUATTERY AND GARY L. MILLER Abstract. Computing graph separators is an important step in many graph algorithms. A popular technique for finding separators

More information

Spectral Graph Theory and You: Matrix Tree Theorem and Centrality Metrics

Spectral Graph Theory and You: Matrix Tree Theorem and Centrality Metrics Spectral Graph Theory and You: and Centrality Metrics Jonathan Gootenberg March 11, 2013 1 / 19 Outline of Topics 1 Motivation Basics of Spectral Graph Theory Understanding the characteristic polynomial

More information

Generating p-extremal graphs

Generating p-extremal graphs Generating p-extremal graphs Derrick Stolee Department of Mathematics Department of Computer Science University of Nebraska Lincoln s-dstolee1@math.unl.edu August 2, 2011 Abstract Let f(n, p be the maximum

More information

Graph Partitioning Using Random Walks

Graph Partitioning Using Random Walks Graph Partitioning Using Random Walks A Convex Optimization Perspective Lorenzo Orecchia Computer Science Why Spectral Algorithms for Graph Problems in practice? Simple to implement Can exploit very efficient

More information

Stat 159/259: Linear Algebra Notes

Stat 159/259: Linear Algebra Notes Stat 159/259: Linear Algebra Notes Jarrod Millman November 16, 2015 Abstract These notes assume you ve taken a semester of undergraduate linear algebra. In particular, I assume you are familiar with the

More information

Chap 3. Linear Algebra

Chap 3. Linear Algebra Chap 3. Linear Algebra Outlines 1. Introduction 2. Basis, Representation, and Orthonormalization 3. Linear Algebraic Equations 4. Similarity Transformation 5. Diagonal Form and Jordan Form 6. Functions

More information

Spectral clustering. Two ideal clusters, with two points each. Spectral clustering algorithms

Spectral clustering. Two ideal clusters, with two points each. Spectral clustering algorithms A simple example Two ideal clusters, with two points each Spectral clustering Lecture 2 Spectral clustering algorithms 4 2 3 A = Ideally permuted Ideal affinities 2 Indicator vectors Each cluster has an

More information

1 Directional Derivatives and Differentiability

1 Directional Derivatives and Differentiability Wednesday, January 18, 2012 1 Directional Derivatives and Differentiability Let E R N, let f : E R and let x 0 E. Given a direction v R N, let L be the line through x 0 in the direction v, that is, L :=

More information

235 Final exam review questions

235 Final exam review questions 5 Final exam review questions Paul Hacking December 4, 0 () Let A be an n n matrix and T : R n R n, T (x) = Ax the linear transformation with matrix A. What does it mean to say that a vector v R n is an

More information

The Lanczos and conjugate gradient algorithms

The Lanczos and conjugate gradient algorithms The Lanczos and conjugate gradient algorithms Gérard MEURANT October, 2008 1 The Lanczos algorithm 2 The Lanczos algorithm in finite precision 3 The nonsymmetric Lanczos algorithm 4 The Golub Kahan bidiagonalization

More information

The cocycle lattice of binary matroids

The cocycle lattice of binary matroids Published in: Europ. J. Comb. 14 (1993), 241 250. The cocycle lattice of binary matroids László Lovász Eötvös University, Budapest, Hungary, H-1088 Princeton University, Princeton, NJ 08544 Ákos Seress*

More information

1 Linear Algebra Problems

1 Linear Algebra Problems Linear Algebra Problems. Let A be the conjugate transpose of the complex matrix A; i.e., A = A t : A is said to be Hermitian if A = A; real symmetric if A is real and A t = A; skew-hermitian if A = A and

More information