Ensuring Strong Dominance of the Leading Eigenvalues for Cluster Ensembles

Similar documents
RITZ VALUE BOUNDS THAT EXPLOIT QUASI-SPARSITY

SIGN-BASED SPECTRAL CLUSTERING

GAUSSIAN PROCESS TRANSFORMS

Relations Between Adjacency And Modularity Graph Partitioning: Principal Component Analysis vs. Modularity Component Analysis

Matrix Theory, Math6304 Lecture Notes from October 25, 2012

CLASSIFICATION OF TREES EACH OF WHOSE ASSOCIATED ACYCLIC MATRICES WITH DISTINCT DIAGONAL ENTRIES HAS DISTINCT EIGENVALUES

1 A factor can be considered to be an underlying latent variable: (a) on which people differ. (b) that is explained by unknown variables

Spectral Graph Theory and its Applications. Daniel A. Spielman Dept. of Computer Science Program in Applied Mathematics Yale Unviersity

1 Matrix notation and preliminaries from spectral graph theory

Physical Metaphors for Graphs

Eigenvectors Via Graph Theory

Computational Lower Bounds for Community Detection on Random Graphs

8.1 Concentration inequality for Gaussian random matrix (cont d)

On marker set distance Laplacian eigenvalues in graphs.

PROOF OF TWO MATRIX THEOREMS VIA TRIANGULAR FACTORIZATIONS ROY MATHIAS

A Note on Eigenvalues of Perturbed Hermitian Matrices

Interlacing Inequalities for Totally Nonnegative Matrices

SECTIONS 5.2/5.4 BASIC PROPERTIES OF EIGENVALUES AND EIGENVECTORS / SIMILARITY TRANSFORMATIONS

Semidefinite and Second Order Cone Programming Seminar Fall 2001 Lecture 5

Network Reconstruction from Intrinsic Noise: Non-Minimum-Phase Systems

General Properties for Determining Power Loss and Efficiency of Passive Multi-Port Microwave Networks

Spectral inequalities and equalities involving products of matrices

Note on deleting a vertex and weak interlacing of the Laplacian spectrum

Majorizations for the Eigenvectors of Graph-Adjacency Matrices: A Tool for Complex Network Design

Graphs and matrices with maximal energy

CANONICAL LOSSLESS STATE-SPACE SYSTEMS: STAIRCASE FORMS AND THE SCHUR ALGORITHM

Lecture 9: SVD, Low Rank Approximation

Spectral Generative Models for Graphs

ORIE 6334 Spectral Graph Theory October 13, Lecture 15

AN ASYMPTOTIC BEHAVIOR OF QR DECOMPOSITION

Inverse Eigenvalue Problems for Two Special Acyclic Matrices

Stat 159/259: Linear Algebra Notes

13-2 Text: 28-30; AB: 1.3.3, 3.2.3, 3.4.2, 3.5, 3.6.2; GvL Eigen2

Laplacian Integral Graphs with Maximum Degree 3

THE SPECTRUM OF THE LAPLACIAN MATRIX OF A BALANCED 2 p -ARY TREE

STA141C: Big Data & High Performance Statistical Computing

Fundamentals of Matrices

Introduction to Search Engine Technology Introduction to Link Structure Analysis. Ronny Lempel Yahoo Labs, Haifa

Spectral Graph Theory

Spectral Graph Theory Lecture 2. The Laplacian. Daniel A. Spielman September 4, x T M x. ψ i = arg min

Spectral Methods for Subgraph Detection

Markov Chains and Spectral Clustering

A Random Dot Product Model for Weighted Networks arxiv: v1 [stat.ap] 8 Nov 2016

Spectra of Adjacency and Laplacian Matrices

Matrix Inequalities by Means of Block Matrices 1

A New Spectral Technique Using Normalized Adjacency Matrices for Graph Matching 1

U.C. Berkeley Better-than-Worst-Case Analysis Handout 3 Luca Trevisan May 24, 2018

Secondary κ-kernel Symmetric Fuzzy Matrices

CS168: The Modern Algorithmic Toolbox Lectures #11 and #12: Spectral Graph Theory

Zero controllability in discrete-time structured systems

a Λ q 1. Introduction

Solving Homogeneous Systems with Sub-matrices

On the adjacency matrix of a block graph

Finding normalized and modularity cuts by spectral clustering. Ljubjana 2010, October

Data Mining and Analysis: Fundamental Concepts and Algorithms

Copositive matrices and periodic dynamical systems

Four new upper bounds for the stability number of a graph

Background on Linear Algebra - Lecture 2

Iterative Methods for Eigenvalues of Symmetric Matrices as Fixed Point Theorems

BIPARTITE SUBGRAPHS AND THE SIGNLESS LAPLACIAN MATRIX

Machine Learning for Data Science (CS4786) Lecture 11

Research Article A Note on Kantorovich Inequality for Hermite Matrices

Root systems and optimal block designs

Lecture Introduction. 2 Brief Recap of Lecture 10. CS-621 Theory Gems October 24, 2012

Spectral Graph Theory Lecture 3. Fundamental Graphs. Daniel A. Spielman September 5, 2018

Combining Association Schemes

Lecture 10 - Eigenvalues problem

Compressed Statistical Testing and Application to Radar

DRAGAN STEVANOVIĆ. Key words. Modularity matrix; Community structure; Largest eigenvalue; Complete multipartite

Obtaining Consensus of Multi-agent Linear Dynamic Systems

Decentralized Stabilization of Heterogeneous Linear Multi-Agent Systems

Frame Diagonalization of Matrices

USING SINGULAR VALUE DECOMPOSITION (SVD) AS A SOLUTION FOR SEARCH RESULT CLUSTERING

Clustering-based State Aggregation of Dynamical Networks

Complex Networks CSYS/MATH 303, Spring, Prof. Peter Dodds

Peter J. Dukes. 22 August, 2012

arxiv:quant-ph/ v1 22 Aug 2005

PRODUCT DISTANCE MATRIX OF A GRAPH AND SQUARED DISTANCE MATRIX OF A TREE. R. B. Bapat and S. Sivasubramanian

Multivariate Statistical Analysis

ALGEBRAIC COMBINATORICS. c1993 C. D. Godsil

642:550, Summer 2004, Supplement 6 The Perron-Frobenius Theorem. Summer 2004

Lecture 1 and 2: Introduction and Graph theory basics. Spring EE 194, Networked estimation and control (Prof. Khan) January 23, 2012

The Eigenvalue Problem: Perturbation Theory

Markov Chains, Random Walks on Graphs, and the Laplacian

NP-hardness of the stable matrix in unit interval family problem in discrete time

Proofs for Large Sample Properties of Generalized Method of Moments Estimators

Central Groupoids, Central Digraphs, and Zero-One Matrices A Satisfying A 2 = J

ON THE HÖLDER CONTINUITY OF MATRIX FUNCTIONS FOR NORMAL MATRICES

Networks and Their Spectra

Analysis of Spectral Kernel Design based Semi-supervised Learning

An Even Order Symmetric B Tensor is Positive Definite

Discrete Simulation of Power Law Noise

ELA A NOTE ON GRAPHS WHOSE LARGEST EIGENVALUE OF THE MODULARITY MATRIX EQUALS ZERO

Latent Semantic Models. Reference: Introduction to Information Retrieval by C. Manning, P. Raghavan, H. Schutze

An Algorithmist s Toolkit September 10, Lecture 1

Characterization of Diagonalizable Matrices: An Algorithm

(Kac Moody) Chevalley groups and Lie algebras with built in structure constants Lecture 1. Lisa Carbone Rutgers University

A New Space for Comparing Graphs

Reconstruction in the Generalized Stochastic Block Model

1 Matrix notation and preliminaries from spectral graph theory

Transcription:

Ensuring Strong Dominance of the Leading Eigenvalues for Cluster Ensembles H.. Kung School of Engineering Applied Sciences Harvard University Cambridge, MA 0478 Bruce W. Suter Air Force Research Laboratory/RIB 55 Brooks Road Rome, NY Abstract Spectral analysis is a popular mathematical tool in analyzing a variety of network distributed systems. For a special class of networks, called cluster ensembles, which are made of interconnected clusters, we can characterize those which exhibit strong dominance of the leading eigenvalues in terms of the cluster structure. For such systems, only these leading eigenvalues their corresponding eigenvectors will need to be examined in studying important properties of the underlying system. his paper establishes several bounds on eigenvalue separation ratios in terms of the number of clusters, their sizes cluster interconnection topologies. Keywords-eigenvalue;eigenvector; spectral analysis; network; cluster I. INRODUCION Spectral analysis is a widely used mathematical tool in analyzing network distributed systems (see, e.g., []). Given such a system, e.g., a sensor-target measurement system (see, e.g., [5]), it's often sufficient to examine only those dominating eigenvectors for which the corresponding eigenvalues are much larger than the rest of eigenvalues. hat is, these dominating eigenvectors alone can reveal important properties of the system, e.g., its clusters (see, e.g., [5]). From the resulting clustering structure one can detect, for example, malfunctioning or malicious sensor nodes in sensor network applications [5]. More precisely, for a given system, let A denote the adjacency matrix of the graph reflecting the underlying network topology. We are interested in the eigen structure of A A, in particular, the dominance of the leading eigenvalues of A A. In this paper we consider a class of network-based systems, called cluster ensembles, which are collections of clusters connected via some cluster interconnection network as depicted in Figure. We show that for some cluster ensembles with regular structures, such as alpha-beta cluster ensembles defined in Section III, we can characterize those ensembles which exhibit strong dominance of the leading eigenvalues in terms of their underlying network topologies. hese characterizations can provide guidelines in designing cluster ensembles where such strong dominance is desired. Figure. Cluster ensemble II. AN EXAMPLE OF CLUSER ENSEMBLE o provide a concrete example, we consider a cluster ensemble composed of three clusters interconnected via a clique network, as shown in Figure. We call those cluster nodes which are part of the cluster interconnection network hub nodes. hus, in this example, nodes 0, are hub nodes. Figure. A three-cluster ensemble, where the cluster interconnection network is a clique A more readable drawing of the same cluster ensemble is depicted Figure 3, where the cluster interconnection network is placed in the center. III. ALPHA-BEA CLUSER ENSEMBLE We focus on a special class of cluster ensembles, called alpha-beta cluster ensembles, which satisfy the following properties: P. he cluster ensemble is composed of α identical clusters.

P. Each cluster contains one hub node β non-hub nodes. his means that the size of each cluster is β +. P3. In the adjacency matrix C of each cluster, columns of non-hub nodes are orthogonal to each other. It is possible to relax this orthogonality condition so C C goes from a diagonal matrix to a diagonally dominant one as described in [4]. he alpha-beta clusters represent a natural model of connecting multiple clusters. Figure 3 depicts an instance of an alpha-beta cluster ensemble with α = 3 β = 3. eigenvalues the rest of non-zero eigenvalues will increase by a factor of at least β/u. his means that when β is large, there are α dominating leading eigenvalues the number of these leading eigenvalues is exactly equal to the number of clusters in the alpha-beta cluster ensemble. hus we have succeeded in relating dominance of the leading eigenvalues to cluster size. o establish the heorem other similar results of this paper, we will carry out the following steps: S. Consider a general m-node, k-cluster ensemble. S. Partition the adjacency matrix A of the cluster ensemble introduce the cluster interconnection submatrix B. S3. Derive bounds for the eigenvalues of A A in terms of those of B. S4. Consider the special case of alpha-beta cluster ensembles, derive bounds for the eigenvalues of B in terms of α β. Figure 3. Another drawing of the cluster ensemble in Figure. It is an alpha-beta cluster ensemble with α = 3 β = 3. In what follows, the eigenvalues λ of a symmetric matrix i nxn P R are always ordered such that λ λ... λn. When it is useful to indicate that they are the eigenvalues on a specific matrix P, we write λ ( P) λ ( P)... λ n ( P) instead. IV. PREVIEW OF AN EXEMPLAR RESUL o have a quick reading on the work of this paper, we preview an exemplar result here. Note that for an alpha-beta cluster ensemble, the total number of nodes is α(β+). We use m to denote α (β + ). Moreover, we use μ to denote the largest in-degree of a nonhub node. For the network topology in Figure 3, we can check that nodes, 5 8 are those non-hub nodes which have the largest in-degree value of. hus in this case μ =. Let A be the m m adjacency matrix of the entire ensemble. hen we can show the following (Corollary 4 of Section VI): Exemplar heorem. ( λ ) m α + A A β m α We describe some system implication of the heorem. Suppose that we increase β, while keeping α μ fixed. hen, by the heorem, the separation ratio of the α leading S5. Derive bounds on the separation ratio for the α leading eigenvalues of an alpha-beta cluster ensemble, including the result of the Exemplar heorem described above. In the rest of the paper, we present our work in the order of these steps. V. PARIIONING A A AND CLUSER INERCONNECION SUBMARIX For a given m-node, k-cluster ensemble, we consider its adjacency matrix A. Since the eigenvalues of A A are invariant under similarity transformations, we can permute the columns of A so that its last k columns correspond to the k hub nodes. We express A A in a partitioned form: X A A= where, X B are (m-k) (m-k), (m-k) k k k, respectively. he B submatrix is of particular interest, as it reflects the interconnections among the hub nodes those between the hub nodes other nodes in their respective clusters. We call B the cluster interconnection submatrix. In the rest of this section, we will show that the eigenvalues of B can contribute to upper lower bounds for those of A A (see Corollary ). heorem. (Cauchy s Interlacing heorem). Let U be an m m symmetric matrix V a (m-k) (m-k) principal submatrix of U, for k < m. hen, for j =,, m-k, λj( U) λj( V) λ k + j( U ) Proof : See, for example, page 89 of [3].

heorem. Let G be an m m positive semi-definite symmetric matrix partitioned as H X G = F where H is (m-k) (m-k) for some 0 < k < m. hen, for j = 0,, k, λ ( F) λ ( G) λ ( F) + λ ( H ) k j m j k j Proof: he stated lower bound on λm j( G) follows from heorem 0.. of [6]. he stated upper bound on λm j( G) can be derived from Aronszajn's inequality on page 64 of []. Corollary. Suppose that A A is an m m symmetric positive semi-definite matrix that B are (m-k) (m-k) k k sub-matrices, respectively, defined by hen for j =,, m k, for j = 0,, k, X A A= λj( A A) λj λ k + j( A A ) λ ( B) λ ( B) + λ ( ) k j m j k j Proof: he first inequality follows from heorem. he second inequality follows from heorem. Corollary. Suppose that X A A= where A A, B are m m, (m - k) (m - k) k k respectively. hen λm k+ ( A A) λ( B) λ Proof: By letting j = m k in the first inequalities expression of Corollary, we have λ ( A A) λ ( ) By letting j = k in the second inequalities expression of Corollary, we have λ( B) λm k + ( A A ) Corollary follows from the above two inequalities. Corollary 3. Suppose that D X A A= where A A, D B are m m, (m - k) (m - k) k k matrices, respectively, D is a diagonal matrix () (,..., m k D = diag d d () ( m k) ) with d... d. hen λm k+ ( A A) λ( B) d Proof: he result follows from Corollary, by noting that in this case. λ m k d = VI. ALPHA-BEA CLUSER ENSEMBLES AND HEIR CLUSER INERCONNECION NEWORKS o establish further bounds on eigenvalues of A A, we consider alpha-beta cluster ensembles as defined in Section III. Note that an alpha-beta cluster ensemble is a m-node, k-cluster ensemble with m = α(β+) k = α. Let A be the adjacency A is α(β+) α(β+), can be written as: matrix. hen A D X A A= where D, X B are αβ αβ, αβ α, α α matrices, respectively, D is a diagonal matrix () D = diag( d,..., d αβ ) with d ()... d ( αβ ). Since each diagonal element of D is the in-degree of a non-hub node, we have d αβ = where μ is the largest in-degree of a non-hub node. he cluster interconnection submatrix B reflects the topology of the interconnection network connecting hub nodes of the clusters. We consider here two specific structures, which represent two design extremes. In the first case, hub nodes are densely connected as a clique, i.e., every hub node has a directed edge to every other hub node, including itself. In the second case, hub nodes are sparsely connected as a ring, i.e., if the clusters are labeled as cluster through α then, for each i =, α, the hub node in cluster i has a directed edge to the hub node in cluster (i+) mod α, also a directed edge to itself. hese two cases are depicted in Figure 4. 3

Figure 4. wo cluster interconnection networks: (a) clique (b) ring. Note that the alpha-beta cluster ensemble of Figure 3 uses the clique interconnection network heorem 3. Suppose that the hub nodes are connected as a clique. hen, for j =,,α-, λ ( B ) = β j λ B = α + β α Proof: When the hub nodes are connected as a clique, we have: B = αj + I α β, α x where each entry of J α α α R is equal to one α x I R α is the identity matrix. α Note the vector u = (,,, ) with all of its components equal to is an eigenvector of the cluster interconnection submatrix B with the corresponding eigenvalue being α + β. Let v be any other eigenvector of B. Since v must be orthogonal to u, we have C ν = 0 wherec is a α α matrix with all its components being α. Since B-C is a diagonal matrix with all its diagonals being β, we have Bν = ( B C) ν + Cν = ( B C ) ν = βν. hus, the eigenvalues of B are λj ( B) = β for j =,..., α λ α B = α + β. heorem 4. Suppose that the hub nodes are connected as a ring. hen, for j =,,α, β λ ( B) β + 4 j Proof: When the hub nodes are connected as a ring, we have + β β + B = + β + β By the Gershgorin Disc heorem, eigenvalues of B are bounded below by ( + β ) above by( + β ) +. From heorems 3 4, we note that when the clusters increase their size β, λ ( B) will increase thus, by Corollary, also the eigenvalue separation. More precisely, we can establish the corollaries below.. Corollary 4. Suppose that the hub nodes in the system are connected as a clique or ring. hen ( λ ) m k+ A A β Proof: he result follows from Corollary 3 heorems 3 4, while noting that d αβ =. Corollary 5. Suppose that the hub nodes in the system are connected as a clique. hen λ β λm A A α + β m ( A A) + Proof: From heorem, we note that, for j = 0,..., k, λ ( B) λ ( B) + λ ( D ) k j m j k j For j = 0, we have λk( B) λm( A A ) For j =, we have λm ( A A) λk ( B) + λ( D ) By heorem 3, we have λk ( B) = α + β λ k B = β It follows that λm( A A) λk( B) = α + β λm ( A A) λk ( B) + λ( D ) = β + by noting that λ m k( D ) =. he Corollary follows from the above two inequalities. 4

Corollary 6. Suppose that the hub nodes in the system are connected as a ring. hen λm( A A) β + 4+ β m k+ Proof: From Corollary, we note that, for j = 0,, k, λ ( B) λ ( B) + λ ( D ) k j m j k j For j = 0, we have ( A A) ( B) + ( D ) λ λ λ m k m k For j = k, we have ( B) ( A A ) By heorem 4, we have It follows that λ λ + m k λ ( B) β + 4 λ ( B ) β k λ ( B) + λ ( D) β + 4+ m k λm k+ ( A A) λ( B ) β he Corollary follows from the above two inequalities. VII. SYSEM IMPLICAION We can design a cluster ensemble system which will lead to a desired eigenvalue separation ratio of the leading eigenvalues the rest of the non-zero eigenvalues. For example, we can form an alpha-beta cluster ensemble with a relatively large β, so that β/μ will be large, where μ is the largest in-degree of a non-hub node in a cluster. hen, when the hub nodes are connected as either a clique or ring, Corollary 4 implies that the resulting system will exhibit strong dominance by the α leading eigenvalues. his means that the overall system behavior will essentially be governed by the corresponding α eigenvectors. For example, the values or signs of the components in these eigenvectors will reveal clustering structures of the system [7]. he resulting clusters, for example, can help detect malfunctioning or malicious sensors in a sensor network application [5]. Furthermore, suppose that values for α μ are fixed the hub nodes are connected as a ring. Corollary 6 implies that separation ratios among the largest α eigenvalues will be bounded above by a quantity which will approach when β increases. hus, for large β, there will be large separations between the α leading eigenvalues the rest of eigenvalues (Corollary 4), but there are no significant separations among the α leading eigenvalues themselves (Corollary 6). his means that when the hub nodes are connected as ring all the corresponding α eigenvectors will be of equal significance in characterizing the system. As a contrast, consider the case when hub nodes are connected as a clique. Assume that the values for β μ are fixed. hen, by Corollary 5, when (α + β)/(β +μ) is large (e.g., when α is large relatively to β μ), the principal eigenvalue will exhibit strong dominance. hus, for example, component values of the principal eigenvector alone will basically characterize the clustering structure of the system [7]. From the above discussion, we see that by choosing proper values for α, β μ in the design of a cluster ensemble, we can tune the dominance properties of its leading eigenvalues. VIII. SUMMARY AND CONCLUDING REMARKS We have motivated the problem of designing distributed systems which exhibit strong dominance of the leading eigenvalues, in terms of underlying network topologies. We have introduced a special class of networks called cluster ensembles, established some fundamental mathematical results of ensuring such dominance for this class of networks. As future work, we plan to generalize these characterizations by allowing more general topology than the alpha-beta ensemble develop applications in system design based on spectral properties established in this paper. ACKOWLEDGMENS his material is based on research sponsored by Air Force Research Laboratory under agreement number FA8750-08-- 09. he U.S. Government is authorized to reproduce distribute reprints for Governmental purposes notwithsting any copyright notation thereon. he views conclusions contained herein are those of the authors should not be interpreted as necessarily representing the official policies or endorsements, either expressed or implied, of Air Force Research Laboratory or the U.S. Government. REFERENCES [] R. Bhatia, Matrix Analysis, Springer-Verlag, New York, 997. [] F. R. K. Chung, Spectral Graph heory, American Mathematical Society, 997. [3] R. A. Horn C. R. Johnson, Matrix Analysis, Cambridge University Press, Cambridge, UK, 985. [4] H.. Kung B. W. Suter, Hub Matrix heory Applications to Wireless Communications, EURASIP Journal on Advances in Signal Processing, Vol. 007, Issue, Article ID 3659, 007. [5] H.. Kung D. Vlah, Validating Sensors in the Field via Spectral Clustering Based on heir Measurement Data, Military Communications Conference (MILCOM 009), October 009. [6] B. N. Parlett, he Symmetric Eigenvalue Problem. Prentice-Hall, Englewood Cliffs, NJ, 980. [7] H.. Kung D. Vlah, Sign-based spectral clustering, 5th Biennial Symposium on Communications, May 00. 5