Complex Networks, Course 303A, Spring, Prof. Peter Dodds

Similar documents
Assortativity and Mixing. Outline. Definition. General mixing. Definition. Assortativity by degree. Contagion. References. Contagion.

Random Networks. Complex Networks, CSYS/MATH 303, Spring, Prof. Peter Dodds

Random Networks. Complex Networks CSYS/MATH 303, Spring, Prof. Peter Dodds

Contagion. Complex Networks CSYS/MATH 303, Spring, Prof. Peter Dodds

arxiv: v4 [cond-mat.dis-nn] 16 May 2011

Generating Functions for Random Networks

Network Observational Methods and. Quantitative Metrics: II

Mini course on Complex Networks

Quantitative Metrics: II

Design and characterization of chemical space networks

Complex Networks CSYS/MATH 303, Spring, Prof. Peter Dodds

ECS 289 F / MAE 298, Lecture 15 May 20, Diffusion, Cascades and Influence

Contagion. Basic Contagion Models. Social Contagion Models. Network version All-to-all networks Theory. References. 1 of 58.

Self Similar (Scale Free, Power Law) Networks (I)

Communities Via Laplacian Matrices. Degree, Adjacency, and Laplacian Matrices Eigenvectors of Laplacian Matrices

Complex networks: an introduction

Liu et al.: Controllability of complex networks

6.207/14.15: Networks Lecture 12: Generalized Random Graphs

Complex Networks, Course 303A, Spring, Prof. Peter Dodds

Eigenvalues by row operations

KINETICS OF SOCIAL CONTAGION. János Kertész Central European University. SNU, June


KINETICS OF COMPLEX SOCIAL CONTAGION. János Kertész Central European University. Pohang, May 27, 2016

Web Structure Mining Nodes, Links and Influence

CSI 445/660 Part 3 (Networks and their Surrounding Contexts)

Ch. 2: Lec. 1. Basics: Outline. Importance. Usages. Key problems. Three ways of looking... Colbert on Equations. References. Ch. 2: Lec. 1.

A framework for analyzing contagion in banking networks

Modeling, Analysis, and Control of Information Propagation in Multi-layer and Multiplex Networks. Osman Yağan

Spectral Graph Theory for. Dynamic Processes on Networks

Diffusion of Innovation and Influence Maximization

Networks as a tool for Complex systems

Scale-Free Networks. Outline. Redner & Krapivisky s model. Superlinear attachment kernels. References. References. Scale-Free Networks

Lecture 10. Under Attack!

6.842 Randomness and Computation March 3, Lecture 8

Flip Your Grade Book With Concept-Based Grading CMC 3 South Presentation, Pomona, California, 2016 March 05

Social Networks- Stanley Milgram (1967)

Virgili, Tarragona (Spain) Roma (Italy) Zaragoza, Zaragoza (Spain)

Bayesian Inference for Contact Networks Given Epidemic Data

TELCOM2125: Network Science and Analysis

Erzsébet Ravasz Advisor: Albert-László Barabási

arxiv:cond-mat/ v1 3 Sep 2004

Bioinformatics 2. Yeast two hybrid. Proteomics. Proteomics

Attacks on Correlated Peer-to-Peer Networks: An Analytical Study

Biological Networks. Gavin Conant 163B ASRC

Distributed Optimization over Networks Gossip-Based Algorithms

Shannon Entropy and Degree Correlations in Complex Networks

Epidemics and information spreading

Machine Learning and Modeling for Social Networks

1 What a Neural Network Computes

Algebra & Trigonometry for College Readiness Media Update, 2016

CSI 445/660 Part 6 (Centrality Measures for Networks) 6 1 / 68

LAPLACIAN MATRIX AND APPLICATIONS

SYSTEMS BIOLOGY 1: NETWORKS

arxiv: v1 [physics.soc-ph] 13 Oct 2016

DS504/CS586: Big Data Analytics Graph Mining II

networks in molecular biology Wolfgang Huber

Diffusion of Innovation

Turing Machines. Nicholas Geis. February 5, 2015

arxiv:cond-mat/ v2 [cond-mat.stat-mech] 7 Jan 2000

PageRank: The Math-y Version (Or, What To Do When You Can t Tear Up Little Pieces of Paper)

Nonparametric Bayesian Matrix Factorization for Assortative Networks

Degree Distribution: The case of Citation Networks

Online Social Networks and Media. Link Analysis and Web Search

Proteomics. Yeast two hybrid. Proteomics - PAGE techniques. Data obtained. What is it?

EE595A Submodular functions, their optimization and applications Spring 2011

Synchronization of a General Delayed Complex Dynamical Network via Adaptive Feedback

Complex contagions and hybrid phase transitions

Sparse Linear Algebra Issues Arising in the Analysis of Complex Networks

Learning latent structure in complex networks

Avalanches in Fractional Cascading

ECS 289 / MAE 298, Lecture 7 April 22, Percolation and Epidemiology on Networks, Part 2 Searching on networks

Lecture 1/25 Chapter 2

Spectral Analysis of Directed Complex Networks. Tetsuro Murai

Topic 4 Randomized algorithms

1 Complex Networks - A Brief Overview

Overlapping Community Structure & Percolation Transition of Community Network. Yong-Yeol Ahn CCNR, Northeastern University NetSci 2010

Bootstrap Percolation on Periodic Trees

Cascades on Temporal Networks

Nonlinear Dynamical Behavior in BS Evolution Model Based on Small-World Network Added with Nonlinear Preference

Graph Theory and Networks in Biology arxiv:q-bio/ v1 [q-bio.mn] 6 Apr 2006

CSCI 3210: Computational Game Theory. Cascading Behavior in Networks Ref: [AGT] Ch 24

6.207/14.15: Networks Lectures 4, 5 & 6: Linear Dynamics, Markov Chains, Centralities

arxiv: v1 [quant-ph] 11 Mar 2016

Graph Theory and Networks in Biology

Modularity and anti-modularity in networks with arbitrary degree distribution

Network Interventions Based on Inversity: Leveraging the Friendship Paradox in Unknown Network Structures

CS 246 Review of Linear Algebra 01/17/19

Catastrophic Cascade of Failures in Interdependent Networks

Lab I. 2D Motion. 1 Introduction. 2 Theory. 2.1 scalars and vectors LAB I. 2D MOTION 15

Groups of vertices and Core-periphery structure. By: Ralucca Gera, Applied math department, Naval Postgraduate School Monterey, CA, USA

Spectral Graph Theory and You: Matrix Tree Theorem and Centrality Metrics

Project 1 Modeling of Epidemics

Online Social Networks and Media. Link Analysis and Web Search

Learning the Semantic Correlation: An Alternative Way to Gain from Unlabeled Text

In biological terms, memory refers to the ability of neural systems to store activity patterns and later recall them when required.

Common Core State Standards for Mathematics - High School

Biological Networks Analysis

ECS 253 / MAE 253, Lecture 15 May 17, I. Probability generating function recap

SECTION 7: CURVE FITTING. MAE 4020/5020 Numerical Methods with MATLAB

Analytically tractable processes on networks

Transcription:

Complex Networks, Course 303A, Spring, 2009 Prof. Peter Dodds Department of Mathematics & Statistics University of Vermont Licensed under the Creative Commons Attribution-NonCommercial-ShareAlike 3.0 License. Frame 1/26

Outline Frame 2/26

Basic idea: Random networks with arbitrary distributions cover much territory but do not represent all networks. Moving away from pure random networks was a key first step. We can extend in many other directions and a natural one is to introduce correlations between different kinds of nodes. Node attributes may be anything, e.g.: 1. 2. demographics (age, gender, etc.) 3. group affiliation We speak of mixing patterns, correlations, biases... Networks are still random at base but now have more global structure. Build on work by Newman [3, 4]. Frame 3/26

between node categories Assume types of nodes are countable, and are assigned numbers 1, 2, 3,.... Consider networks with directed edges. ( an edge connects a node of type µ e µν = Pr to a node of type ν ) a µ = Pr(an edge comes from a node of type µ) b ν = Pr(an edge leads to a node of type ν) Write E = [e µν ], a = [a µ ], and b = [b ν ]. Requirements: e µν = 1, e µν = a µ, and e µν = b ν. ν µ µ ν Frame 4/26

Connection to distribution: Frame 5/26

Notes: Varying e µν allows us to move between the following: 1. Perfectly assortative networks where nodes only connect to like nodes, and the network breaks into subnetworks. Requires e µν = 0 if µ ν and µ e µµ = 1. 2. Uncorrelated networks (as we have studied so far) For these we must have independence: e µν = a µ b ν. 3. Disassortative networks where nodes connect to nodes distinct from themselves. Disassortative networks can be hard to build and may require constraints on the e µν. Basic story: level of assortativity reflects the to which nodes are connected to nodes within their group. Frame 6/26

Correlation coefficient: Quantify the level of assortativity with the following assortativity coefficient [4] : µ r = e µµ µ a µb µ 1 µ a = Tr E E 2 1 µb µ 1 E 2 1 where 1 is the 1-norm = sum of a matrix s entries. Tr E is the fraction of edges that are within groups. E 2 1 is the fraction of edges that would be within groups if connections were random. 1 E 2 1 is a normalization factor so r max = 1. When Tr e µµ = 1, we have r = 1. When e µµ = a µ b µ, we have r = 0. Frame 7/26

Correlation coefficient: Notes: r = 1 is inaccessible if three or more types are presents. Disassortative networks simply have nodes connected to unlike nodes no measure of how unlike nodes are. Minimum value of r occurs when all links between non-like nodes: Tr e µµ = 0. where 1 r min < 0. r min = E 2 1 1 E 2 1 Frame 8/26

Scalar quantities Now consider nodes defined by a scalar integer quantity. Examples: age in years, height in inches, number of friends,... e jk = Pr a randomly chosen edge connects a node with value j to a node with value k. a j and b k are defined as before. Can now measure correlations between nodes based on this scalar quantity using standard Pearson correlation coefficient ( ): r = j k j k(e jk a j b k ) σ a σ b = j 2 a j 2 a jk j a k b k 2 b k 2 b This is the observed normalized deviation from randomness in the product jk. Frame 9/26

Degree- correlations Natural correlation is between the s of connected nodes. Now define e jk with a slight twist: ( an edge connects a j + 1 node e jk = Pr to a k + 1 node ) ( an edge runs between a node of in- j = Pr and a node of out- k ) Useful for calculations (as per R k ) Important: Must separately define P 0 as the {e jk } contain no information about isolated nodes. Directed networks still fine but we will assume from here on that e jk = e kj. Frame 10/26

Degree- correlations Notation reconciliation for undirected networks: j k r = j k(e jk R j R k ) σ 2 R where, as before, R k is the probability that a randomly chosen edge leads to a node of k + 1, and σ 2 R = j j 2 R j j 2 jr j. Frame 11/26

Degree- correlations Error estimate for r: Remove edge i and recompute r to obtain r i. Repeat for all edges and compute using the jackknife method ( ) [1] σ 2 r = i (r i r) 2. Mildly sneaky as variables need to be independent for us to be truly happy and edges are correlated... Frame 12/26

age A to package B indicates that A relies on components of B for its operation. Biological networks: k protein-protein interaction network in the yeast S. Cerevisiae 53 ; l metabolic network of the bacterium E. Measurements Coli 54 ; m neural network of theof nematode - worm C. Elegans 5,55 ; tropic interactions correlations between species in the food webs of n Ythan Estuary, Scotland 56 and o Little Rock Lake, Wisconsin 57. Group Network Type Size n Assortativity r Error r a Physics coauthorship undirected 52 909 0.363 0.002 a Biology coauthorship undirected 1 520 251 0.127 0.0004 b Mathematics coauthorship undirected 253 339 0.120 0.002 Social c Film actor collaborations undirected 449 913 0.208 0.0002 d Company directors undirected 7 673 0.276 0.004 e Student relationships undirected 573 0.029 0.037 f Email address books directed 16 881 0.092 0.004 g Power grid undirected 4 941 0.003 0.013 Technological h Internet undirected 10 697 0.189 0.002 i World Wide Web directed 269 504 0.067 0.0002 j Software dependencies directed 3 162 0.016 0.020 k Protein interactions undirected 2 115 0.156 0.010 l Metabolic network undirected 765 0.240 0.007 Biological m Neural network directed 307 0.226 0.016 n Marine food web directed 134 0.263 0.037 o Freshwater food web directed 92 0.326 0.031 most never is. In this paper, therefore, we take an ve approach, making use of computer simulation. ould like to generate on a computer a random netving, for instance, a particular value of the matrix is also fixes the distribution, via Eq. 23. In we discussed disassortative one possible way of doing this using ithm similar to that of Sec. II C. One would draw om the desired distribution e jk and then join the de- 58,59. The algorithm is as follows. 1 Given the desired edge distribution e jk, we first calculate the corresponding distribution of excess s q k from Eq. 23, and then invert Eq. 22 to find the distribution: Frame 13/26 p k q k 1 /k. 27 q j 1 / j Social networks tend to be assortative (homophily) Technological and biological networks tend to be

Spreading on -correlated networks Next: Generalize our work for random networks to -correlated networks. As before, by allowing that a node of k is activated by one neighbor with probability β k1, we can handle various problems: 1. find the giant component size. 2. find the probability and extent of spread for simple disease models. 3. find the probability of spreading for simple threshold models. Frame 14/26

Spreading on -correlated networks Goal: Find f n,j = Pr an edge emanating from a j + 1 node leads to a finite active subcomponent of size n. Repeat: a node of k is in the game with probability β k1. Define β 1 = [β k1 ]. Plan: Find the generating function F j (x; β 1 ) = n=0 f n,jx n. Frame 15/26

Spreading on -correlated networks Recursive relationship: F j (x; β 1 ) = x 0 + x k=0 k=0 e jk R j (1 β k+1,1 ) e [ jk β k+1,1 F k (x; R k β 1 )]. j First term = Pr that the first node we reach is not in the game. Second term involves Pr we hit an active node which has k outgoing edges. Next: find average size of active components reached by following a link from a j + 1 node = F j (1; β 1 ). Frame 16/26

Spreading on -correlated networks Differentiate F j (x; β 1 ), set x = 1, and rearrange. We use F k (1; β 1 ) = 1 which is true when no giant component exists. We find: R j F j (1; β 1 ) = e jk β k+1,1 + ke jk β k+1,1 F k (1; β 1 ). k=0 k=0 Rearranging and introducing a sneaky δ jk : ( ) δjk R k kβ k+1,1 e jk F k (1; β 1 ) = e jk β k+1,1. k=0 k=0 Frame 17/26

Spreading on -correlated networks In matrix form, we have where [A E, β1 ] A E, β1 F (1; β 1 ) = E β 1 = δ jkr k kβ k+1,1 e jk, j+1,k+1 [ F (1; β 1 )] = F k (1; β 1 ), k+1 [ ] [E] j+1,k+1 = e jk, and β1 = β k+1,1. k+1 Frame 18/26

Spreading on -correlated networks So, in principle at least: F (1; β 1 ) = A 1 E, β 1 E β 1. Now: as F (1; β 1 ), the average size of an active component reached along an edge, increases, we move towards a transition to a giant component. Right at the transition, the average component size explodes. Exploding inverses of matrices occur when their determinants are 0. The condition is therefore: det A E, β1 = 0. Frame 19/26

Spreading on -correlated networks General condition details: det A E, β1 = det [ δ jk R k 1 (k 1)β k,1 e j 1,k 1 ] = 0. The above collapses to our standard contagion condition when e jk = R j R k. When β 1 = β 1, we have the condition for a simple disease model s successful spread det [ δ jk R k 1 β(k 1)e j 1,k 1 ] = 0. When β 1 = 1, we have the condition for the existence of a giant component: det [ δ jk R k 1 (k 1)e j 1,k 1 ] = 0. Bonusville: We ll find another (possibly better) version of this set of conditions later... Frame 20/26

Spreading on -correlated networks We ll next find two more pieces: 1. P trig, the probability of starting a cascade 2. S, the expected extent of activation given a small seed. Triggering probability: Generating function: H(x; β 1 ) = x P k [F k 1 (x; β k 1 )]. k=0 Generating function for vulnerable component size is more complicated. Frame 21/26

Spreading on -correlated networks Want probability of not reaching a finite component. P trig = S trig =1 H(1; β 1 ) =1 P k [F k 1 (1; β k 1 )]. k=0 Last piece: we have to compute F k 1 (1; β 1 ). Nastier (nonlinear) we have to solve the recursive expression we started with when x = 1: F j (1; β 1 ) = e jk k=0 R j (1 β k+1,1 )+ k=0 Iterative methods should work here. e jk R j β k+1,1 [F k (1; β 1 )] k. Frame 22/26

Spreading on -correlated networks Truly final piece: Find final size using approach of Gleeson [2], a generalization of that used for uncorrelated random networks. Need to compute θ j,t, the probability that an edge leading to a j node is infected at time t. Evolution of edge activity probability: θ j,t+1 = G j ( θ t ) = φ 0 + (1 φ 0 ) k=1 e j 1,k 1 R j 1 k 1 ( ) k 1 θk,t i i (1 θ k,t) k 1 i β ki. i=0 Overall active fraction s evolution: φ t+1 = φ 0 +(1 φ 0 ) k P k k=0 i=0 ( ) k θk,t i i (1 θ k,t) k i β ki. Frame 23/26

Spreading on -correlated networks As before, these equations give the actual evolution of φ t for synchronous updates. condition follows from θ t+1 = G( θ t ). Linearize G around θ 0 = 0. θ j,t+1 = G j ( 0) + k=1 G j ( 0) θ k,t θ k,t + 2 G j ( 0) θk,t 2 +... k=1 θ 2 k,t If G j ( 0) 0 for at least one j, always have some infection. [ ] If G j ( G 0) = 0 j, largest eigenvalue of j ( 0) θ k,t must exceed 1. Condition for spreading is therefore dependent on eigenvalues of this matrix: G j ( 0) θ k,t = e j 1,k 1 (k 1)β k1 R j 1 Insert question from assignment 5 ( ) Frame 24/26

tively mixed network. These findings are intuitively reasonable. If the net- How the giant component changes with assortativity work mixes assortatively, then the high- vertices will tend to stick together in a subnetwork or core group of higher mean than the network as a whole. It is reasonable to suppose that percolation would occur earlier within such a subnetwork. Conversely, since percolation will be restricted to this subnetwork, it is not surprising giant component S 1.0 0.8 0.6 0.4 0.2 0.0 1 10 100 exponential parameter κ assortative neutral disassortative FIG. 1. Size of the giant component as a fraction of graph from Newman, 2002 [3] size for graphs with the edge distribution given in Eq. (9). The points are simulation results for graphs of N 100 000 vertices, while the solid lines are the numerical solution of Eq. (8). Each point is an average over ten graphs; the resulting statistical errors are smaller than the symbols. The values of p are 0.5 (circles), p 0 0:146... (squares), and 0:05 (triangles). More assortative networks percolate for lower average s But disassortative networks end up with higher extents of spreading. Frame 25/26

I [1] B. Efron and C. Stein. The jackknife estimate of variance. The Annals of Statistics, 9:586 596, 1981. pdf ( ) [2] J. P. Gleeson. Cascades on correlated and modular random networks. Phys. Rev. E, 77:046117, 2008. pdf ( ) [3] M. Newman. Assortative mixing in networks. Phys. Rev. Lett., 89:208701, 2002. pdf ( ) [4] M. E. J. Newman. patterns in networks. Phys. Rev. E, 67:026126, 2003. pdf ( ) Frame 26/26