Relations Between Adjacency And Modularity Graph Partitioning: Principal Component Analysis vs. Modularity Component Analysis

Similar documents
Data Analysis and Manifold Learning Lecture 3: Graphs, Graph Matrices, and Graph Embeddings

Lecture 13. Principal Component Analysis. Brett Bernstein. April 25, CDS at NYU. Brett Bernstein (CDS at NYU) Lecture 13 April 25, / 26

1 Matrix notation and preliminaries from spectral graph theory

Principal Component Analysis (PCA) CSC411/2515 Tutorial

1 Matrix notation and preliminaries from spectral graph theory

Spectral Clustering on Handwritten Digits Database

Data Analysis and Manifold Learning Lecture 7: Spectral Clustering

EECS 275 Matrix Computation

Basic Calculus Review

Principal Component Analysis -- PCA (also called Karhunen-Loeve transformation)

MLCC Clustering. Lorenzo Rosasco UNIGE-MIT-IIT

COMPSCI 514: Algorithms for Data Science

Spectral Clustering. Zitao Liu

Problem # Max points possible Actual score Total 120

Spectra of Adjacency and Laplacian Matrices

Machine learning for pervasive systems Classification in high-dimensional spaces

Machine Learning for Data Science (CS4786) Lecture 11

Statistical Machine Learning

Lecture 14: SVD, Power method, and Planted Graph problems (+ eigenvalues of random matrices) Lecturer: Sanjeev Arora

A Dimensionality Reduction Framework for Detection of Multiscale Structure in Heterogeneous Networks

20 Unsupervised Learning and Principal Components Analysis (PCA)

Principal Component Analysis

Spectral Clustering on Handwritten Digits Database Mid-Year Pr

Dimensionality Reduction: PCA. Nicholas Ruozzi University of Texas at Dallas

Finding normalized and modularity cuts by spectral clustering. Ljubjana 2010, October

Spectral Clustering - Survey, Ideas and Propositions

Functional Analysis Review

Data Mining and Analysis: Fundamental Concepts and Algorithms

Cholesky Decomposition Rectification for Non-negative Matrix Factorization

Learning Eigenfunctions: Links with Spectral Clustering and Kernel PCA

Deep learning / Ian Goodfellow, Yoshua Bengio and Aaron Courville. - Cambridge, MA ; London, Spis treści

Connection of Local Linear Embedding, ISOMAP, and Kernel Principal Component Analysis

Machine Learning (CSE 446): Unsupervised Learning: K-means and Principal Component Analysis

PRINCIPAL COMPONENTS ANALYSIS

Machine Learning 11. week

Machine Learning - MT & 14. PCA and MDS

Spectral clustering. Two ideal clusters, with two points each. Spectral clustering algorithms

Dimension Reduction and Iterative Consensus Clustering

PCA & ICA. CE-717: Machine Learning Sharif University of Technology Spring Soleymani

Unsupervised Learning: Dimensionality Reduction

DATA MINING LECTURE 8. Dimensionality Reduction PCA -- SVD

Collaborative Filtering: A Machine Learning Perspective

Markov Chains and Spectral Clustering

MATH 829: Introduction to Data Mining and Analysis Clustering II

Inverse Power Method for Non-linear Eigenproblems

Probabilistic Latent Semantic Analysis

Chapter 7: Symmetric Matrices and Quadratic Forms

MATH 567: Mathematical Techniques in Data Science Clustering II

CS4495/6495 Introduction to Computer Vision. 8B-L2 Principle Component Analysis (and its use in Computer Vision)

Analysis of Spectral Kernel Design based Semi-supervised Learning

Eigenvalues, Eigenvectors, and an Intro to PCA

Mathematical foundations - linear algebra

Robot Image Credit: Viktoriya Sukhanova 123RF.com. Dimensionality Reduction

Introduction to Machine Learning. PCA and Spectral Clustering. Introduction to Machine Learning, Slides: Eran Halperin

Positive Definite Matrix

Eigenvalues, Eigenvectors, and an Intro to PCA

The Mathematics of Facial Recognition

SVD, Power method, and Planted Graph problems (+ eigenvalues of random matrices)

Math 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination

Numerical Methods I Singular Value Decomposition

Natural Gradients via the Variational Predictive Distribution

Lecture 12 : Graph Laplacians and Cheeger s Inequality

STA141C: Big Data & High Performance Statistical Computing

Principal Component Analysis

Family Feud Review. Linear Algebra. October 22, 2013

Covariance and Correlation Matrix

Unsupervised Learning: K- Means & PCA

Focus was on solving matrix inversion problems Now we look at other properties of matrices Useful when A represents a transformations.

Eigenvalues, Eigenvectors, and an Intro to PCA

STA141C: Big Data & High Performance Statistical Computing

Nonparametric Inference for Auto-Encoding Variational Bayes

What is Principal Component Analysis?

On the Adjacency Spectra of Hypertrees

Lecture 13: Spectral Graph Theory

LECTURE NOTE #11 PROF. ALAN YUILLE

December 20, MAA704, Multivariate analysis. Christopher Engström. Multivariate. analysis. Principal component analysis

Bipartite spectral graph partitioning to co-cluster varieties and sound correspondences

Ma/CS 6b Class 23: Eigenvalues in Regular Graphs

Singular Value Decomposition

Principal Component Analysis (PCA)

Dimensionality Reduction Using PCA/LDA. Hongyu Li School of Software Engineering TongJi University Fall, 2014

LEC 2: Principal Component Analysis (PCA) A First Dimensionality Reduction Approach

Spectral Clustering. Spectral Clustering? Two Moons Data. Spectral Clustering Algorithm: Bipartioning. Spectral methods

7. Symmetric Matrices and Quadratic Forms

EE 381V: Large Scale Learning Spring Lecture 16 March 7

Lecture: Face Recognition and Feature Reduction

An Inverse Power Method for Nonlinear Eigenproblems with Applications in 1-Spectral Clustering and Sparse PCA

EVERYTHING YOU NEED TO KNOW TO BUILD YOUR FIRST CONVOLUTIONAL NEURAL NETWORK (CNN)

Statistical Learning Notes III-Section 2.4.3

Principal Component Analysis

ANSWERS (5 points) Let A be a 2 2 matrix such that A =. Compute A. 2

Mathematical foundations - linear algebra

A few applications of the SVD

Robust Principal Component Analysis

Data Mining and Analysis: Fundamental Concepts and Algorithms

MATH 567: Mathematical Techniques in Data Science Clustering II

Principal component analysis (PCA) for clustering gene expression data

Concentration Ellipsoids

Multivariate Statistical Analysis

Lecture: Face Recognition and Feature Reduction

Transcription:

Relations Between Adjacency And Modularity Graph Partitioning: Principal Component Analysis vs. Modularity Component Analysis Hansi Jiang Carl Meyer North Carolina State University October 27, 2015 1 / 15

Application and Methodology Application Graph partitioning or data clustering Methodology Spectral methods based on modularity components or principal components Relation between dominant eigenvectors of unnormalized modularity and similarity matrices Comparing modularity components and principal components 2 / 15

Definitions Given a graph or data, we have: Modularity matrix: B = A dd T /(2m) A is an adjacency matrix (graph) or similarity matrix (data) d = Ae is the degree vector m = d T e is the number of edges in the graph 3 / 15

Definitions Modularity: Q(s) = s T Bs Invented by Newman and Girvan (2004) Goal: Pick s, s 2 = 1 s.t. Q is maximized Dominant eigenvector of B maximizes Q 4 / 15

Properties Modularity matrix B Symmetric, not semi-definite (0, e) is an eigenpair of B The eigenvector corresp. to the largest eigenvalue is used in partitioning Spectral clustering uses the eigenvector corresp. to the second smallest eigenvector of L = D A 5 / 15

Dominant Eigenvectors of Modularity and Similarity Matrices Modularity matrix: B = A dd T /(2m) We want to give an explicit expression of b 1 in terms of eigenvalues and eigenvectors of A. Theorem 1 The dominant eigenvector of B is b 1 = 1 n vi T d d 2 i=1 α i β 1 v i. (α i, v i ) s are eigenpairs of A The proof is based on Cauchy s Interlacing Theorem. 6 / 15

Dominant Eigenvectors of B when A = X T X Suppose X p n = UΣV T is the uncentered data matrix Consider a special case: A = X T X Suppose A has k positive eigenvalues and they are simple The first k 1 largest eigenvalues of B are simple by the interlacing theorem The result in theorem 1 can be extended 7 / 15

Dominant Eigenvectors of B when A = X T X Theorem 2 Suppose the largest k 1 eigenvalues of B are β 1 > β 2 > > β k 1 and the nonzero eigenvalues of A = X T X are α 1 > α 2 > > α k. Further suppose that for 1 i k 1 we have β i α i and β i α i+1. Then the k 1 dominant eigenvectors of B can be written by b i = k γ ij v j, j=1 where γ ij = v T j d (α j β i ) d 2. 8 / 15

Application: Modularity Components Definition Let m T i = b T i X = k j=1 γ ij σ j u T j, then the i-th modularity component is defined to be c i = m i m i 2. σ j is the j-th singular value of X, u j corresp. singular vector Can help to explain why using several eigenvectors of B to cluster data is reasonable 9 / 15

Review: Properties of Principal Components Principal Components derived from X are orthogonal to each other Require centering data Projection of the centered data onto to the span of principal components gives clusters The first principal component has maximal variance Each succeeding principal component has maximal variance with the constraint that it is orthogonal to all prior principal components 10 / 15

Properties of Modularity Components Modularity Components derived from X are orthogonal to each other Does not require centering Projection of the uncentered data onto to the span of modularity components gives clusters The first modularity component has maximal modularity Each succeeding modularity component has maximal modularity with the constraint that it is orthogonal to all prior modularity components 11 / 15

Significance of Modularity Components Analogous to the principal components Does not require centering Gives reason to use multiple eigenvectors of B to cluster data 12 / 15

Example: PenDigit Dataset (Subset of MNIST) 12,000 data points Data points: vectors converted from a grey scale image Subset used: Digits 1, 5 and 7 13 / 15

Conclusion The exact linear relation between the dominant eigenvectors of B in terms of the eigenvectors of A is given The definition of modularity components and their properties are given The comparison between modularity components and principal components is given An example comparing the results from MCA and PCA is given 14 / 15

Thanks Thank you! Questions and Answers 15 / 15

[1] Hansi Jiang and Carl Meyer. Modularity component analysis versus principal component analysis. submitted(http://arxiv.org/pdf/1510.05492v1.pdf), 2015. [2] Yann LeCun, Léon Bottou, Yoshua Bengio, and Patrick Haffner. Gradient-based learning applied to document recognition. Proceedings of the IEEE, 86(11):2278 2324, 1998. [3] Mark EJ Newman and Michelle Girvan. Finding and evaluating community structure in networks. Physical review E, 69(2):026113, 2004. [4] Ulrike Von Luxburg. A tutorial on spectral clustering. Statistics and computing, 17(4):395 416, 2007. 15 / 15