Multi-commodity allocation for dynamic demands using PageRank vectors

Size: px
Start display at page:

Download "Multi-commodity allocation for dynamic demands using PageRank vectors"

Transcription

1 Multi-commodity allocation for dynamic demands using PageRank vectors Fan Chung, Paul Horn, and Jacob Hughes University of California, San Diego Harvard University Abstract. We consider a variant of the contact process concerning multi-commodity allocation on networks. In this process, the demands for several types of commodities are initially given at some specified vertices and then the demands spread interactively on a contact graph. To allocate supplies in such a dynamic setting, we use a modified version of PageRank vectors, called Kronecker PageRank, to identify vertices for shipping supplies. We analyze both the situation that the demand distribution evolves mostly in clusters around the initial vertices and the case that the demands spread to the whole network. We establish sharp upper bounds for the probability that the demands are satisfied as a function of PageRank vectors. Introduction Efficient allocation of resources to meet changing demands is a task arising in numerous applications. For example, institutions such as governments or corporations respond to the needs of a populace, and wish to meet the demands within allowed expenditure of resources. In some cases where demand spreads, one has to be able to act before demand becomes unmanageable. In the case of an epidemic, for instance, one desires to find a way to distribute medicine so that the disease will be contained. Such problems have been studied in several contexts using the contact process model [8], [6], [], [0], [4]. In [4], it was demonstrated how to use PageRank vectors to both restrict the number of nodes inoculated and to provide certain containment guarantees. In this paper, we study a variant of the classical contact process, a continuous time Markov process on a contact graph. In our scenario, vertices in the graph each have varying levels of demand for multiple commodities. Demand at a vertex propagates to its neighbors at a rate depending on the current demand. Our model allows for interactions between different commodities; demand for one commodity may influence demand for another. This fits many scenarios that arise, for instance demand for iphones may accelerate the demand for ipads. As another example, demand at a node can might be viewed as a measure of discontent with the current supply of a resource. It is natural for an unhappy node to create unrest in its neighbors. As the modified contact process continues, demands at a vertex are increased at a rate based on the demands at neighboring vertices. There are also decreased at a satisfaction rate, which can be thought of of as a frequency of shipments. Demand spreads at rates which are a linear combination of demands from neighboring vertices. These rates are encapsulated in a spread matrix, B, roughly analogous to the infectivity parameter in the classical contact process. The goal of this paper is to find satisfaction rates, dependent on the spread matrix B and the geometry of the contact graph which ensure that eventually all vertices have no demand and the process dies out. Our process will be defined, in detail, in Section.

2 To satisfy the demands which evolve according to our model as defined in Section, the goal is to ship commodities and supply vertices with unsatisfied demands in an efficient way. The model here differs somewhat from typical resource allocation problems in the sense that we do not specify the location of the warehouses for the supply. We will not be concerned with either the sources of the supply or the detailed incremental costs of shipping supply. Instead, our goal is to identify how often to ship each commodity to a particular vertex, in order to contain and satisfy demands, given an initial seed set. The reader is referred to [3] for the usual resource allocation problem. Contact graphs of interest take many forms: Cities and countries exert trade pressure on neighboring cities and countries. Communication on the internet can also spread demand for products, or discontent leading to a revolution. Instead of studying this problem on particular models for these contact graphs, we study the problem on arbitrary finite graphs. Two schemes of making shipments are considered. First is a global solution which involves scheduling shipments to all vertices in the graph, and ensures that all demand is satisfied in O(log n) time, with high probability and regardless of the initial demand. This is made precise in Theorem once the model is formally defined. The next scheme is a local solution, in the sense that shipments are scheduled to only a subset of vertices which contain the initial demand. In particular, when the contact graph has some clustering structure we are interested in subsets so that the demand within the subset is satisfied quickly (in O(log n) time) and demands reach a vertex not receiving shipments with low probability. Precise results to this end are given in Section 5. This latter scheme relies on understanding the geometry of the particular contact graph being studied. Our scheme uses PageRank to identify important vertices and to bound the probability that demand in our process leaves a set. We also introduce a variant of PageRank, which we call Kronecker PageRank and is introduced in Section 3, which provides sharper bounds by better utilizing the structure of the spread matrix B as well as the geometry of the graph in its estimates. Our analysis provides a tradeoff in the following sense: we may use PageRank estimates to identify a set of vertices containing the initial demand which are important to ship to or we may use our PageRank estimates to give a guarantee on the escape probability of leaving a particular set of our choosing. Precise results to this end are given in Theorems 3 and 4, using standard PageRank and Kronecker PageRank respectively. Preliminaries and The Demand Model We model demand spreading within an undirected simple graph, G = (V, E). We write this v w when v and w are adjacent. For each vertex v V, let d v be the degree of v, which is the number of neighbors of v. While not strictly necessary, we assume that there is a self loop at each vertex. In this case, d v includes v in the count of neighbors and hence the loop counts as towards the degree. We let n = V, the number of nodes of G. An exponential random variable with parameter λ has probability density function given by f(x) = λe λx for x 0, and 0 for x < 0. This distribution will be denoted Exp(λ). One important property of exponential random variables is the memoryless property: if X is an exponential random variable then for any constants a, b > 0, P(X > a + b X > a) = P(X > b).

3 If X and Y are independent and X Exp(λ ), Y Exp(λ ) then min{x, Y } Exp(λ + λ ). A Poisson point process at rate λ is a sequence of random variables {X i } i= so that X and X i X i, for i, has distribution Exp(λ). Before we describe our model, let us briefly recall the contact process on a graph G, which we denote CP (T, β, σ, G). In the contact process (see for example [] or [4]), a disease initially infects a set T V (G). The disease has an infectivity parameter, β, and each vertex has a certain amount of medicine σ v. An infected vertex v infects its neighbor u at times given by a Poisson point process {X uv } at rate β, and each infected vertex is cured at times given by a Poisson point process at rate σ v. In the most frequently studied case, σ is constant and the host graph is an infinite graph. The process ends when all vertices are cured, and the basic problem is to determine under which conditions on σ, and β the process ends almost surely. In the case of finite graphs, if σ v > 0 for every vertex, it is easy to observe that the process ends a.s., so the problem becomes determining how fast the process ends. The k commodity dynamic demand model on a graph G is a variant of the contact process, DD(τ (0), B, σ, G, N). In this situation, B is a real valued k k-matrix (not assumed to be symmetric, or even non-negative), which we call the spread matrix. The supply function is σ : V R k, and τ (0) : V N k is the initial demand. The state of the process at time t is given by τ (t) : V R k, which gives the demands for each vertex at time t. We use τ v (t) N k to denote the demand at vertex v at time t, and τ j v (t) N to denote the demand for commodity j at time t. A node v is said to be satisfied at time t if τ v (t) = 0, and unsatisfied otherwise. N N serves as a uniform bound for the maximum demand for any resource at any point (for instance, it could be the population of a city, if a vertex is a city.) The existence of N is simply to ensure integrability of some random variables. In the case of the contact process, N =. The spread matrix B = [β ij ] describes how the demand for one commodity influences demands for other commodities. The i, j entry of B, β ij, determines the spread rate of the demand for commodity j that is caused by demand for commodity i. In particular, we can describe the rate of spread events as follows. If v is a node that is unsatisfied at time t, and w an adjacent vertex, then there are spread events from v to w with rates max{τ v (t)b, 0}. That is the rate at which τ j w increases due to the demand at v is given by max{ i τ i v(t)β ij, 0}. Here, when we say an event occurs with rate λ, we mean that the elapsed time until that event takes place is distributed as Exp(λ). Because the minimum of exponential random variables is itself an exponential random variable, we can capture the total spreading rates in a condensed form. We define the rate function at time t, ρ(t) : V R k, by ρ v = w v τ w (t)b = (τ (t)(a B)) v, where τ (t) is viewed as a vector with indices indexed by V k. ρ i v(t) is the rate at which τ i v is increasing at time t. Any spread events that would raise τ i v above N are ignored. Supply events occur with rates given by τ (t)diag(σ), independently of any neighboring supply events. That is, the time until τ i v is decreased by is distributed as Exp(σ i vτ i v). We briefly give a construction of the process, to show it is well-defined. Let E denote the set of ordered edges; that is, ordered pairs that are edges in the graph, so that uv and vu are

4 distinct. We run independent Poisson point processes {Xe j,ρ } e E (G),j [k],ρ [N] k so that Xe j,ρ is at rate max{0, [τ v B] j } and independent Poisson point processes {Xv i,n } v V (G),i [k],n [N] so that Xv i,n is at rate nσv. i Then these finitely many point processes can easily be seen to define the entire process; a spread event of type j from a vertex v to a vertex u which is currently in state ρ is controlled by the point process Xvu j,ρ with satisfaction events handled similarly. Such a formulation is that it gives an easy coupling between processes that shows that if B B pointwise, the stochastic process DD(τ (0), B, σ, G, N) stochastically dominates DD(τ (0), B, σ, G, N). That is, in the coupling the demands in the B process are always at least those in the B process. This is accomplished by noting that the rates ρb ρb pointwise for all ρ N k. We thus take point processes Ye j,ρ at rate [ρb ρb ] i. If the point processes {Xe j,ρ } and {Xv i,n } are used to determine DD(τ (0), B, σ, G, N), then the point processes {Xe j,ρ Ye j,ρ } and {Xv i,n } are used to determine DD(τ (0), B, σ, G, N). In particular, this allows us to replace B with B, where B ij = max{b ij, 0}, and conclusions about the extinction of the B process still hold for B. Furthermore, this turns out not to be entirely unreasonable. One hopes that the negative entries in B would afford better bounds on the extinction time, but in many cases with negative entries in B extinctions of some demand types mean that the process is eventually run in a non-negative case. In light of this, we will assume for the rest of this paper that B is non-negative for convenience. Given an initial demand τ (0) and spread matrix B, our goal is to find a supply function σ such that demand is satisfied. Ideally we would like to do this with small supply rates. Furthermore, the supply rates should only depend on the contact graph G, the spread matrix B, and the initial demand τ (0), but not on t or τ (t). Because we take B to be an arbitrary positive k k matrix, we will at times need to use various matrix norms in order to understand the process. For a square matrix B, there are many different matrix norms that can be used (see [7]). We will use the following notation for the following norms:. B = i,j a ij is the l norm.. B = max j i a ij is the maximum column sum norm. 3. B = max i j a ij is the maximum row sum norm. 4. B = max{ λ λ is an eigenvalue of A A} is the spectral norm. 3 PageRank and Kronecker PageRank The notion of PageRank was first introduced by Brin and Page [] in 998 for Google s search algorithms. Although PageRank was originally used for the Web graph, it is well defined on any finite graph G. The basis of PageRank is random walks on graphs. A walk is a sequence of vertices (v 0, v,..., v k ) where v i v i+. A simple random walk of length k is a sequence of random variables (x 0,..., x k ) where the starting vertex x 0 is chosen according to some distribution, and P(x i+ = v x i ) = { d xi if x i v 0 if x i v.

5 Let D be the diagonal degree matrix with entries D vv = d v, and let A be the adjacency matrix with entries { if v w A vw = 0 if otherwise. Then the transition probability matrix for a random walk on G is given by W = D A. We will mainly use a modified version of the PageRank, called personalized PageRank. Personalized PageRank has two parameters, a jumping constant α [0, ] and a seed s which is some probability distribution on the vertex set V of G. The personalized PageRank vector pr(α, s) for jumping constant α and the seed distribution s on V is given by pr(α, s) = α ( α) l sw l. l=0 Note that here we view s as a row vector, which is our convention for all vectors throughout this paper. We note that the PageRank vector is also the solution to the recurrence relation pr(α, s) = αs + ( α)pr(α, s)w. The original definition of PageRank [] is the special case where s is the uniform distribution over all the vertices. For a subset of vertices H V, the volume of H is the sum of degrees of the vertices of H. The Cheeger Ratio of H, h(h), measures the cut between H and H via the relationship h(h) = e(h, H) min{vol(h), vol( H)}. The α core of a subset H is the set of vertices { C α = v H pr(α, v ) H h }. α Personalized PageRank naturally gives bounds on the probability that demands leave a given set in the k-commodity dynamic demand model. These bounds lose some of the structure of the process given by B, however. In order to better understand the process and get tighter bounds, we will use generalization of the personalized PageRank vector, the Kronecker PageRank vector. If B is an k k matrix, and A an n n matrix, then the Kronecker product A B is the nk nk block matrix a B a n B A B =..... a n B a nn B With this, we can define Kronecker PageRank.

6 Definition (Kronecker PageRank) Let B be a square k k matrix with spectral radius strictly less than, and W be the transition matrix for a random walk on a graph G. Let s be a non-negative vector in R k V. The Kronecker PageRank vector with parameters B and s is defined as Kpr(B, s) = s(w B) l = l=0 s(w l B l ) l=0 Requiring the spectral radius of B less than is necessary to ensure convergence of the infinite sum, as the spectrum of W B is the product of the spectra of W and B. Since the eigenvalues of W have absolute value at most, the sum will converge. We note that in the case where B is a matrix B = β < and s is a probability distribution, then we have the relationship Kpr(B, s) = s(w β) l = l=0 l=0 sβ l W l = pr( β, s), β so the Kronecker PageRank is a natural extension of personalized PageRank. We will see in Theorem 4 that the Kronecker PageRank will arise naturally in our analysis in Section 3, and give better bounds than those that will be afforded by standard PageRank by incorporating the spread matrix. 4 Global analysis: Supplying every vertex Here we show that if supply rates are above a certain threshold, then with probability approaching demands will be satisfied. Recall that B is a non-negative real valued k k matrix, and σ is the vector of supply rates for the process DD(τ (0), B, σ, G, N). Theorem Consider the k commodity demand model on a graph G with n vertices parameterized by spread matrix B = [β ij ]. Let X(t) = τ (t) ), the total amount of demand at time t. If the supply + δ for δ > 0 then with probability ɛ all vertices ( rates to each vertex v are σv i > d v j are satisfied at time t for all t > δ β ij +β ji ( log(nk) + log(x(0)) + log ( )). ɛ Proof. We begin by considering the quantity te[τ (t)]. From the discussion in Section, we know that demand is increasing with rates given by ρ(t) = τ (t) (A B), but also demand decreases proportionally to the supply rates and current demand. Indeed, let S = diag(σ), the diagonal nk nk matrix with entries given by the supply vector. Then demand decreases at each vertex according to rates given by the supply rate vector τ (t)s. It is not difficult to encapsulate all of the above information in the simple expression E[τ (t)] = E[ρ(t) τ (t)s] = E[τ (t)](a B S). () t

7 A detailed proof of () is left to the appendix for space reasons. Solving the matrix differential equation with initial condition E[τ (0)] = τ (0) yields E[τ (t)] = τ (0)e t(a B S). () Let Q = A B S. Then by [5], e tq e tν, where ν is the largest eigenvalue of Q+Q. We note that Q+Q = A ( B+B ) S, which has diagonal terms β ii σv, i ranging over all values of v and i. By the Gershgorin Circle Theorem, the eigenvalues of Q+Q Since σ i v > d v ( j (d v )β ii d v j i β ij + β ji are contained in the intervals σv, i d v β ij + β ji σ i v. j ) β ij +β ji + δ all the eigenvalues of Q+Q are less than δ. Therefore E[X(t)] = τ(0)e t(a B S) nk τ(0)e t(a B S) nk τ(0) e t(a B S) nk τ(0) e tν nkx(0)e tδ Thus Markov s inequality gives that P(X(t) > 0) < ɛ if t > δ ( log(nk) + log(x(0)) + log ( ɛ )). We note that this approach works for all initial distributions τ (0), and on any graph G. In particular, it is agnostic to the shape of the graph. This indicates that in many situations this approach may be overkill and that we could have used smaller supply rates. In the next section, we analyze the process more carefully and give conditions that depend on the initial distribution of demand and the geometry of the underlying contact graph. 5 Local analysis: Supplying a small subset For the remainder of the discussion, it is convenient to introduce reformulation of the model that takes advantage of the fact that demands take on integer values. Rather than view demands as a function τ : V N k, we view demands as discrete objects sitting on each node. Borrowing language from chip-firing games on graphs (see, for example, [9]) we view units of the demand as chips located on vertices of the graph. For example, if k = 7 then for a vertex v with τ v (t) = (0,,, 0,, 0, 3) then we would say that at time t there was -chip, 3-chips, 5-chips, and 3 7-chips at vertex v, corresponding to unit of demand for commodity, etc. Unlike in classical chip-firing games, the number of chips is not static, and the process is continuous time. We restate the possible transitions in terms of demand chips. For an i chip at vertex v, there are two types of transition events:

8 For each vertex w v and each j =,..., l, a j chip is added at w at rate β ij. When this occurs we say that the new j chip is created by the i chip. The i chip itself is removed with rate σ i v. Due to the properties of exponential random variables, the rates add linearly, and the model is equivalent to the original description discussed in Section. The main advantage of this reformulation is the ability to trace back the history of a chip. If there is a chip c at vertex v at time t, then either c existed at time t = 0, or there is a sequence of l chips (c 0,..., c l = c) located at vertices along a walk π = (v 0, v,..., v l = v) with the following properties:. c 0 existed at t = 0. c r is created by c r for r =,..., l. We allow π to have repeated vertices to allow for the case where demand created more demand at the same vertex. If a chip c exists at time 0, we refer to it as an initial chip. For a path π = (v 0, v,..., v l ) and a chip c 0 located at v 0, we define the event S π,c0 to be the event there is a sequence of m chips (c 0,..., c l ) located respectively at (v 0, v,..., v l ) and c r is created by c r for r =,..., l. It is important to note that S π,c0 occurring does not imply that there is any demand at v l at time t because it could be satisfied sometime before t. However, if there is a demand at v l at time t, then S π,c must have occurred for some initial chip c at vertex v 0 and some walk π from v 0 to v l. We begin by relating P(S π,c0 ) to the length of the walk π. Inspired by Theorem we make the assumption that supply rates are proportional to the degree of the vertices. That is, we assume that σv i > µ i (d v ) for all v for constants µ i > 0. Lemma Let M = diag(µ,..., µ k ), ˆB = M B and ζ = min{ ˆB, ˆB } Then for any chip c 0 located at v 0 and any walk π = (v 0,..., v l ) of length l, P(S π,c0 ) kζ l d j=0 vj Proof. Let S r denote the event that a chip c r at v r creates a chip at v r+. If c r is an i chip, then for it to create any chip at v r+ a spread event must occur before c r is removed. The time until c r creates a j chip at v r+ is an exponential random variable with rate β ij. Since the time until c r is removed is given by Exp(σ i v), the probability of c r creating a j chip is β ij β ij +σ i v β ij σ i v < β ij µ i d vr. Thus P(S r ) < β ij i,j µ i d vr = d vr ˆB. For a walk π of length l, we want to consider the intermediate steps more carefully. Since there are l transitions that occur, we can use the same reasoning as above to obtain the bound P(S π,c ) < r=0 d vr ˆB l = r=0 d vr ˆB l k r=0 d vr ˆB l k r=0 d vr ˆB l.

9 The factor of k that appears in the final lines above is just a consequence of switching from the vector -norm ˆB l to maximum column sum norm ˆB l (see [7]). We could have just as easily switched to the maximum row sum norm and obtained the term k ˆB l, and so it follows that P(S π,c ) < min{k r=0 d vr ˆB l, k r=0 d vr ˆB l } = kζ l We note that the use of ζ = min{ ˆB, ˆB } in Lemma reflects the difficulty in working with arbitrary spread matrices B. For certain classes of spread matrices (e.g. if B is symmetric or diagonalizable) it is possible to obtain tighter bounds. Lemma will be allow us to obtain a bound using PageRank, but note that our use of matrix norms ignores some of the structure of the spread matrix B. Following the proof of Theorem 3 below, a more careful analysis fully using the structure of the spread matrix B will lead naturally to use of Kronecker PageRank, which we explore in Theorem 4. Theorem 3 Suppose that initial demand is contained in S H V with and each vertex v H has supply rates σv i > µ i d v, and σw i = 0 for w H. Let M = diag(µ,..., µ k ), ˆB = M B and ζ = min{ ˆB, ˆB }. Let x(t) be defined by x v (t) = i τ v(t), i and X(t) = τ (t). Let E H denote the event that demands spread outside the set H. Then. P(E H ) X(0) ζ pr ( ζ, τ(0) X(0) ) H. If S in the ( ζ) core of H, then P(E H ) X(0)h(H) ζ( ζ), where h(h) is the Cheeger ratio of H. Proof. Let P l denote the set of all paths of length l from an initial chip in S to H such that the first k steps are in H. Let P = l= P l. The key observation is that if w H ever has demand, then S π,c must have occurred for some initial chip c and path π from the location of c to w. Thus we can use the union bound to get that P(S π,c ) P(S π,c ) π P l (π,c) P l P(S π,c ) l v 0 S c at v 0 v l H π=(v 0,...,v l ) P l d l v 0 S r=0 vr ζ l c at v 0 v l H π=(v 0,...,v l ) P l j=0 d vj. = l = l x(0)ζ l (D A) l H x(0)ζ l W l H = X(0) pr ( x ) ζ, ζ X(0) H.

10 proving the first statement. The second statement follows the same proof as Theorem 3. of [4]. Finally we show how Kronecker PageRank arises in a natural way as the bound of the escape probability of this process. Theorem 4 Suppose that the initial demand is contained in S H V with and each vertex v H has supply rates σv i µ i d v. Let M = diag(µ,..., µ k ), ˆB = M B and ζ = ˆB. Let X(t) = τ (t), the total amount of demands at time t. Let E H denote the event that demands spread outside the set H. Then E H can be bounded above using the Kronecker PageRank vector via the relationship: ( P(E H ) X(0)Kpr ˆB, τ(0) ) H X(0) Proof. Let f be a vector indicator function of commodity type on chips, that is f(c) = e i if c is an i chip, where e i denotes the ith standard basis vector for R k. Let C 0 denote the set of initial chips. By the same methods that were used in the proof of Lemma, we can bound the probability that demand originating from c ever spreads along a path π = (v 0, v,..., v l ) by the sum P(S π,c ) f(c) ˆB l d r=0 vr Therefore using the same technique as in the proof of Theorem 3 we obtain the bound P(S π,c ) P(S π,u ) P(S π,u ) π P l u S π B l l c C 0 v l H π=(v 0,...,v l ) P l f(c) ˆB l = τ(0)(d A d ˆB) l H l π=(v 0,...,v l ) P l r=0 vr l = l c C 0 v l H τ(0)(w ˆB) l H = X(0)Kpr( τ(0) ˆB, ) H X(0) On the event of non-escape, we would like to guarantee that all demand is satisfied quickly. To make this precise, let S t denote the event that all of the vertices are satisfied at time t. In order to complete the analysis of the local case, we would like to bound P(S t E H ), where E H is as in Theorems 3 and 4. Such a bound is not immediately given by Theorem. To derive a bound on P(S t E H ), consider running a modified Dirichlet process which is identical to the standard process with the same supply rates, except demand leaving H is ignored. Let S t denote the event that in Dirichlet process, all of the events are satisfied at time t then P(S t) can be bounded directly by Theorem as this Dirichlet process restricted to vertices in H is the standard process on H. Furthermore P(S t E H ) P(S t). Therefore P(S t E H ) = P(S t E H ) P(E H ) P(S t) P(E H ).

11 Combining this observation along with Theorems 3 and 4, yields that the probability of escape from H is bounded and if the process does not escape from H it dies quickly. Theorems 3 and 4 can be used in two different ways. As stated, they provide a way to bound the probability demands escape from a given subset. However, they can be also used to construct such a bounding subset. For example, given initial demand τ (0) contained in an initial set of vertices S V, we can algorithmically construct H such that demand stays in H with probability ɛ by iteratively selecting vertices with the highest (Kroneceker) PageRank. 6 An Example An immediate question is whether anything is actually gained by the introduction of Kronecker PageRank. Suppose we have a spread matrix B, and some initial demand on a graph G. We wish to identify a subset of vertices H G to make shipments to so that the escape probability is at most ɛ. We may use either Theorem 3 or Theorem 4 to identify such a set. The bound afforded by Theorem 4 is clearly sharper than the bound in Theorem 3 as the structure of the spread matrix B is taken into account, but it is not guaranteed that the identified set is actually smaller. In many cases it actually is, though depending on B it may not be significantly smaller. To illustruate, we give a simple example calculation on synthetic data. Our graph G is an instance of the following random process, which is designed to create a graph which contains tighter clusters that are slightly more sparsely connected to neighboring clusters: We begin with a cycle on 0 vertices. Each vertex is then replaced by an instance of the random graph G 0,.3, that is a graph 0 vertices and each edge existing independently with probability.3. Inter-cluster edges are then created between vertices in neighboring clusters with probability.05. We consider the case where k = 4, and B = The initial demand is given by τ (0) = {,,, 0, 0,, 0, 0,..., 0}. In addition we assume µ i =, and ζ =.85. We demonstrate the difference between Theorems 3 and 4 in the following way. The figure below shows the graph G. The demands start in the large outlined vertices and spread outward from there. Theorem 4 states that with 95% probability, demands stay in the circular vertices. Theorem 3 states that with 95% probability, demands stay in the diamond and circular vertices. This small example illustrates how the Kronecker PageRank can be used to obtain improved results.

12 References. C. Borgs, J. Chayes, A. Ganesh, and A. Saberi, How to distribute antidote to control epidemics, Random Structures & Algorithms, 37 (00), pp S. Brin and L. Page, The anatomy of a large-scale hypertextual web search engine, Computer Networks and ISDN Systems, 30 (998), pp Y. Chevaleyre, P. Dunne, U. Endriss, J. Lang, M. Lemaitre, N. Maudet, J. Padget, S. Phelps, J. Rodriguez-Aguilar, and P. Sousa, Issues in multiagent resource allocation, Informatica, (006). 4. F. Chung, P. Horn, and A. Tsiatas, Distributing antidote using PageRank vectors, Internet Mathematics, 6 (009), pp G. Dahlquist, Stability and error bounds in the numerical integration of ordinary differential equations, Kungl. Tekn. Hgsk. Handl. Stockholm. No., 30 (959), p A. Ganesh, L. Massoulie, and D. Towsley, The effect of network topology on the spread of epidemics, vol., IEEE, 005, pp R. A. Horn and C. R. Johnson, Matrix Analysis, Cambridge University Press, I. Z. Kiss, D. M. Green, and R. R. Kao, Infectious disease control using contact tracing in random and scale-free networks, Journal of the Royal Society, Interface / the Royal Society, 3 (006), pp PMID: C. Merino, The chip-firing game, Discrete Mathematics, 30 (005), pp M. Newman, Spread of epidemic disease on networks, Physical Review E, 66 (00). 7 Appendix In this section we establish the differential equation used in the proof of Theorem, namely: Lemma 5 If τ (t) denotes the demand vector at time t, A is the adjacency matrix of G, B denotes the spread matrix and S = diag(σ) denotes the supply matrix, then E[τ (t)] = E[τ (t)](a B S). t

13 The key of Lemma 5 are the following two well known and simple facts concerning exponentially distributed random variables. We use the notation f(h) = O h 0 (g(h)) to indicate that f(h) Cġ(h) for h sufficiently small. Lemma 6 Suppose X is an exponentially distributed waiting time with rate λ, then An immediate corollary is P(X < h) = λh + O h 0 (h ). Lemma 7 Suppose X, Y are independent exponentially distributed waiting times with rates λ, λ. Then P(X, Y < h) = O h 0 (h ). Proof of Lemma 5. Fix a vertex v and commodity i. We will show that t E[ τ i v(t)] = [E[τ (t)](a B S) ] i v, Since this holds for all v, and i the result will follow. To do this, we compute the derivative by the definition, that is we compute lim h 0 E[τ i v(t) τ i v(t + h)]. h To do this, consider the conditional expectation, E[τ i v(t) τ i v(t + h) τ (t)]. Note that by Lemma 7, then probability that two independent events (either two spread events, or two satisfy events or a spread and a satisfy event) occur is O h 0 (h ). On the other hand, given a neighbor u of v, and a commodity j, the probability of a spread event originating from this neighbor and commodity in time (t, t + h) is exactly B ji τ j u(t)h + O h 0 (h ). Likewise, the probability of a satisfaction event in this time is τ i v(t)σ i vh + O h 0 (h ). Linearity of expectation yields E[τ i v(t) τ i v(t + h) τ (t)] = E[τ (t)(a B S)h + O h 0 (h ) τ (t)] = τ (t)(a B S)h + O h 0 (h ) τ (t). In particular, note that the O h 0 (h ) term means that there is a (large) constant C = C(τ(t), A, B), so that O h 0 (h ) C h for h. Note that due to our conditioning this constant depends on τ(t), but critically not on h. Note that this constant is σ(τ(t))-measurable. By the tower property of conditional expectation, E[τ i lim v(t) τ i v(t + h)] E[E[τ i = lim v(t) τ i v(t + h) τ (0)]]. h 0 h h 0 h E[τ (t)(a B S)h] + E[E[O h 0 (h ) τ (t)]] = lim h 0 h = τ (t)(a B S) + lim E[O h 0 (h)]. h 0

14 It suffices to show that lim E[O h 0 (h)] lim E[ O h 0 (h) ] = 0. h 0 h 0 But recall that the O h 0 (h) term is bounded by C(τ (t), A, B) h for h. Thus it is enough to show that lim E[ C(τ (t), A, B)h ] = 0. h 0 This follows from the monotone convergence theorem, so long as lim E[ C(τ (t), A, B) ] <. h 0 To complete the proof we note that we can give an upper bound on C(τ (t), A, B) in terms of τ (t), n and max{b ij }. Indeed, the rates of the active point processes are at most τ (t) max{b i,j }; and thus the probability that any pair of point processes both have events in the period (t, t + h) is bounded by C(τ (t), A, B)h C τ (t) max{b i,j } n h. But τ (t) << n N k, where n = G, k is the number of demands and N is our uniform upper bound for the demand at a point (indeed, this is the precisely the motivation for such a bound.)

Four graph partitioning algorithms. Fan Chung University of California, San Diego

Four graph partitioning algorithms. Fan Chung University of California, San Diego Four graph partitioning algorithms Fan Chung University of California, San Diego History of graph partitioning NP-hard approximation algorithms Spectral method, Fiedler 73, Folklore Multicommunity flow,

More information

Personal PageRank and Spilling Paint

Personal PageRank and Spilling Paint Graphs and Networks Lecture 11 Personal PageRank and Spilling Paint Daniel A. Spielman October 7, 2010 11.1 Overview These lecture notes are not complete. The paint spilling metaphor is due to Berkhin

More information

Lecture: Local Spectral Methods (2 of 4) 19 Computing spectral ranking with the push procedure

Lecture: Local Spectral Methods (2 of 4) 19 Computing spectral ranking with the push procedure Stat260/CS294: Spectral Graph Methods Lecture 19-04/02/2015 Lecture: Local Spectral Methods (2 of 4) Lecturer: Michael Mahoney Scribe: Michael Mahoney Warning: these notes are still very rough. They provide

More information

0.1 Naive formulation of PageRank

0.1 Naive formulation of PageRank PageRank is a ranking system designed to find the best pages on the web. A webpage is considered good if it is endorsed (i.e. linked to) by other good webpages. The more webpages link to it, and the more

More information

Selecting Efficient Correlated Equilibria Through Distributed Learning. Jason R. Marden

Selecting Efficient Correlated Equilibria Through Distributed Learning. Jason R. Marden 1 Selecting Efficient Correlated Equilibria Through Distributed Learning Jason R. Marden Abstract A learning rule is completely uncoupled if each player s behavior is conditioned only on his own realized

More information

Stochastic modelling of epidemic spread

Stochastic modelling of epidemic spread Stochastic modelling of epidemic spread Julien Arino Centre for Research on Inner City Health St Michael s Hospital Toronto On leave from Department of Mathematics University of Manitoba Julien Arino@umanitoba.ca

More information

25.1 Markov Chain Monte Carlo (MCMC)

25.1 Markov Chain Monte Carlo (MCMC) CS880: Approximations Algorithms Scribe: Dave Andrzejewski Lecturer: Shuchi Chawla Topic: Approx counting/sampling, MCMC methods Date: 4/4/07 The previous lecture showed that, for self-reducible problems,

More information

Affine iterations on nonnegative vectors

Affine iterations on nonnegative vectors Affine iterations on nonnegative vectors V. Blondel L. Ninove P. Van Dooren CESAME Université catholique de Louvain Av. G. Lemaître 4 B-348 Louvain-la-Neuve Belgium Introduction In this paper we consider

More information

Analytically tractable processes on networks

Analytically tractable processes on networks University of California San Diego CERTH, 25 May 2011 Outline Motivation 1 Motivation Networks Random walk and Consensus Epidemic models Spreading processes on networks 2 Networks Motivation Networks Random

More information

1 Adjacency matrix and eigenvalues

1 Adjacency matrix and eigenvalues CSC 5170: Theory of Computational Complexity Lecture 7 The Chinese University of Hong Kong 1 March 2010 Our objective of study today is the random walk algorithm for deciding if two vertices in an undirected

More information

Cutting Graphs, Personal PageRank and Spilling Paint

Cutting Graphs, Personal PageRank and Spilling Paint Graphs and Networks Lecture 11 Cutting Graphs, Personal PageRank and Spilling Paint Daniel A. Spielman October 3, 2013 11.1 Disclaimer These notes are not necessarily an accurate representation of what

More information

Detecting Sharp Drops in PageRank and a Simplified Local Partitioning Algorithm

Detecting Sharp Drops in PageRank and a Simplified Local Partitioning Algorithm Detecting Sharp Drops in PageRank and a Simplified Local Partitioning Algorithm Reid Andersen and Fan Chung University of California at San Diego, La Jolla CA 92093, USA, fan@ucsd.edu, http://www.math.ucsd.edu/~fan/,

More information

Random Walks on Graphs. One Concrete Example of a random walk Motivation applications

Random Walks on Graphs. One Concrete Example of a random walk Motivation applications Random Walks on Graphs Outline One Concrete Example of a random walk Motivation applications shuffling cards universal traverse sequence self stabilizing token management scheme random sampling enumeration

More information

Math 443/543 Graph Theory Notes 5: Graphs as matrices, spectral graph theory, and PageRank

Math 443/543 Graph Theory Notes 5: Graphs as matrices, spectral graph theory, and PageRank Math 443/543 Graph Theory Notes 5: Graphs as matrices, spectral graph theory, and PageRank David Glickenstein November 3, 4 Representing graphs as matrices It will sometimes be useful to represent graphs

More information

Lecture 10. Lecturer: Aleksander Mądry Scribes: Mani Bastani Parizi and Christos Kalaitzis

Lecture 10. Lecturer: Aleksander Mądry Scribes: Mani Bastani Parizi and Christos Kalaitzis CS-621 Theory Gems October 18, 2012 Lecture 10 Lecturer: Aleksander Mądry Scribes: Mani Bastani Parizi and Christos Kalaitzis 1 Introduction In this lecture, we will see how one can use random walks to

More information

Statistics 992 Continuous-time Markov Chains Spring 2004

Statistics 992 Continuous-time Markov Chains Spring 2004 Summary Continuous-time finite-state-space Markov chains are stochastic processes that are widely used to model the process of nucleotide substitution. This chapter aims to present much of the mathematics

More information

Decentralized Stabilization of Heterogeneous Linear Multi-Agent Systems

Decentralized Stabilization of Heterogeneous Linear Multi-Agent Systems 1 Decentralized Stabilization of Heterogeneous Linear Multi-Agent Systems Mauro Franceschelli, Andrea Gasparri, Alessandro Giua, and Giovanni Ulivi Abstract In this paper the formation stabilization problem

More information

Lab 8: Measuring Graph Centrality - PageRank. Monday, November 5 CompSci 531, Fall 2018

Lab 8: Measuring Graph Centrality - PageRank. Monday, November 5 CompSci 531, Fall 2018 Lab 8: Measuring Graph Centrality - PageRank Monday, November 5 CompSci 531, Fall 2018 Outline Measuring Graph Centrality: Motivation Random Walks, Markov Chains, and Stationarity Distributions Google

More information

Lecture: Local Spectral Methods (1 of 4)

Lecture: Local Spectral Methods (1 of 4) Stat260/CS294: Spectral Graph Methods Lecture 18-03/31/2015 Lecture: Local Spectral Methods (1 of 4) Lecturer: Michael Mahoney Scribe: Michael Mahoney Warning: these notes are still very rough. They provide

More information

6.842 Randomness and Computation March 3, Lecture 8

6.842 Randomness and Computation March 3, Lecture 8 6.84 Randomness and Computation March 3, 04 Lecture 8 Lecturer: Ronitt Rubinfeld Scribe: Daniel Grier Useful Linear Algebra Let v = (v, v,..., v n ) be a non-zero n-dimensional row vector and P an n n

More information

As it is not necessarily possible to satisfy this equation, we just ask for a solution to the more general equation

As it is not necessarily possible to satisfy this equation, we just ask for a solution to the more general equation Graphs and Networks Page 1 Lecture 2, Ranking 1 Tuesday, September 12, 2006 1:14 PM I. II. I. How search engines work: a. Crawl the web, creating a database b. Answer query somehow, e.g. grep. (ex. Funk

More information

ECEN 689 Special Topics in Data Science for Communications Networks

ECEN 689 Special Topics in Data Science for Communications Networks ECEN 689 Special Topics in Data Science for Communications Networks Nick Duffield Department of Electrical & Computer Engineering Texas A&M University Lecture 8 Random Walks, Matrices and PageRank Graphs

More information

Lecture 14: Random Walks, Local Graph Clustering, Linear Programming

Lecture 14: Random Walks, Local Graph Clustering, Linear Programming CSE 521: Design and Analysis of Algorithms I Winter 2017 Lecture 14: Random Walks, Local Graph Clustering, Linear Programming Lecturer: Shayan Oveis Gharan 3/01/17 Scribe: Laura Vonessen Disclaimer: These

More information

Lecture notes for Analysis of Algorithms : Markov decision processes

Lecture notes for Analysis of Algorithms : Markov decision processes Lecture notes for Analysis of Algorithms : Markov decision processes Lecturer: Thomas Dueholm Hansen June 6, 013 Abstract We give an introduction to infinite-horizon Markov decision processes (MDPs) with

More information

Spectral Clustering on Handwritten Digits Database

Spectral Clustering on Handwritten Digits Database University of Maryland-College Park Advance Scientific Computing I,II Spectral Clustering on Handwritten Digits Database Author: Danielle Middlebrooks Dmiddle1@math.umd.edu Second year AMSC Student Advisor:

More information

Link Analysis Ranking

Link Analysis Ranking Link Analysis Ranking How do search engines decide how to rank your query results? Guess why Google ranks the query results the way it does How would you do it? Naïve ranking of query results Given query

More information

MAE 298, Lecture 8 Feb 4, Web search and decentralized search on small-worlds

MAE 298, Lecture 8 Feb 4, Web search and decentralized search on small-worlds MAE 298, Lecture 8 Feb 4, 2008 Web search and decentralized search on small-worlds Search for information Assume some resource of interest is stored at the vertices of a network: Web pages Files in a file-sharing

More information

Online Social Networks and Media. Link Analysis and Web Search

Online Social Networks and Media. Link Analysis and Web Search Online Social Networks and Media Link Analysis and Web Search How to Organize the Web First try: Human curated Web directories Yahoo, DMOZ, LookSmart How to organize the web Second try: Web Search Information

More information

Stochastic processes. MAS275 Probability Modelling. Introduction and Markov chains. Continuous time. Markov property

Stochastic processes. MAS275 Probability Modelling. Introduction and Markov chains. Continuous time. Markov property Chapter 1: and Markov chains Stochastic processes We study stochastic processes, which are families of random variables describing the evolution of a quantity with time. In some situations, we can treat

More information

Finding normalized and modularity cuts by spectral clustering. Ljubjana 2010, October

Finding normalized and modularity cuts by spectral clustering. Ljubjana 2010, October Finding normalized and modularity cuts by spectral clustering Marianna Bolla Institute of Mathematics Budapest University of Technology and Economics marib@math.bme.hu Ljubjana 2010, October Outline Find

More information

Calculating Web Page Authority Using the PageRank Algorithm

Calculating Web Page Authority Using the PageRank Algorithm Jacob Miles Prystowsky and Levi Gill Math 45, Fall 2005 1 Introduction 1.1 Abstract In this document, we examine how the Google Internet search engine uses the PageRank algorithm to assign quantitatively

More information

Eigenvalues, random walks and Ramanujan graphs

Eigenvalues, random walks and Ramanujan graphs Eigenvalues, random walks and Ramanujan graphs David Ellis 1 The Expander Mixing lemma We have seen that a bounded-degree graph is a good edge-expander if and only if if has large spectral gap If G = (V,

More information

INTRODUCTION TO MARKOV CHAINS AND MARKOV CHAIN MIXING

INTRODUCTION TO MARKOV CHAINS AND MARKOV CHAIN MIXING INTRODUCTION TO MARKOV CHAINS AND MARKOV CHAIN MIXING ERIC SHANG Abstract. This paper provides an introduction to Markov chains and their basic classifications and interesting properties. After establishing

More information

Reinforcement Learning

Reinforcement Learning Reinforcement Learning March May, 2013 Schedule Update Introduction 03/13/2015 (10:15-12:15) Sala conferenze MDPs 03/18/2015 (10:15-12:15) Sala conferenze Solving MDPs 03/20/2015 (10:15-12:15) Aula Alpha

More information

Stochastic modelling of epidemic spread

Stochastic modelling of epidemic spread Stochastic modelling of epidemic spread Julien Arino Department of Mathematics University of Manitoba Winnipeg Julien Arino@umanitoba.ca 19 May 2012 1 Introduction 2 Stochastic processes 3 The SIS model

More information

CSI 445/660 Part 6 (Centrality Measures for Networks) 6 1 / 68

CSI 445/660 Part 6 (Centrality Measures for Networks) 6 1 / 68 CSI 445/660 Part 6 (Centrality Measures for Networks) 6 1 / 68 References 1 L. Freeman, Centrality in Social Networks: Conceptual Clarification, Social Networks, Vol. 1, 1978/1979, pp. 215 239. 2 S. Wasserman

More information

Markov Chain Monte Carlo The Metropolis-Hastings Algorithm

Markov Chain Monte Carlo The Metropolis-Hastings Algorithm Markov Chain Monte Carlo The Metropolis-Hastings Algorithm Anthony Trubiano April 11th, 2018 1 Introduction Markov Chain Monte Carlo (MCMC) methods are a class of algorithms for sampling from a probability

More information

6 Evolution of Networks

6 Evolution of Networks last revised: March 2008 WARNING for Soc 376 students: This draft adopts the demography convention for transition matrices (i.e., transitions from column to row). 6 Evolution of Networks 6. Strategic network

More information

Spectral Graph Theory Lecture 2. The Laplacian. Daniel A. Spielman September 4, x T M x. ψ i = arg min

Spectral Graph Theory Lecture 2. The Laplacian. Daniel A. Spielman September 4, x T M x. ψ i = arg min Spectral Graph Theory Lecture 2 The Laplacian Daniel A. Spielman September 4, 2015 Disclaimer These notes are not necessarily an accurate representation of what happened in class. The notes written before

More information

DATA MINING LECTURE 13. Link Analysis Ranking PageRank -- Random walks HITS

DATA MINING LECTURE 13. Link Analysis Ranking PageRank -- Random walks HITS DATA MINING LECTURE 3 Link Analysis Ranking PageRank -- Random walks HITS How to organize the web First try: Manually curated Web Directories How to organize the web Second try: Web Search Information

More information

Markov Chains, Random Walks on Graphs, and the Laplacian

Markov Chains, Random Walks on Graphs, and the Laplacian Markov Chains, Random Walks on Graphs, and the Laplacian CMPSCI 791BB: Advanced ML Sridhar Mahadevan Random Walks! There is significant interest in the problem of random walks! Markov chain analysis! Computer

More information

TOPOLOGICAL EQUIVALENCE OF LINEAR ORDINARY DIFFERENTIAL EQUATIONS

TOPOLOGICAL EQUIVALENCE OF LINEAR ORDINARY DIFFERENTIAL EQUATIONS TOPOLOGICAL EQUIVALENCE OF LINEAR ORDINARY DIFFERENTIAL EQUATIONS ALEX HUMMELS Abstract. This paper proves a theorem that gives conditions for the topological equivalence of linear ordinary differential

More information

1 Searching the World Wide Web

1 Searching the World Wide Web Hubs and Authorities in a Hyperlinked Environment 1 Searching the World Wide Web Because diverse users each modify the link structure of the WWW within a relatively small scope by creating web-pages on

More information

AN ELEMENTARY PROOF OF THE SPECTRAL RADIUS FORMULA FOR MATRICES

AN ELEMENTARY PROOF OF THE SPECTRAL RADIUS FORMULA FOR MATRICES AN ELEMENTARY PROOF OF THE SPECTRAL RADIUS FORMULA FOR MATRICES JOEL A. TROPP Abstract. We present an elementary proof that the spectral radius of a matrix A may be obtained using the formula ρ(a) lim

More information

Outlines. Discrete Time Markov Chain (DTMC) Continuous Time Markov Chain (CTMC)

Outlines. Discrete Time Markov Chain (DTMC) Continuous Time Markov Chain (CTMC) Markov Chains (2) Outlines Discrete Time Markov Chain (DTMC) Continuous Time Markov Chain (CTMC) 2 pj ( n) denotes the pmf of the random variable p ( n) P( X j) j We will only be concerned with homogenous

More information

SVD, Power method, and Planted Graph problems (+ eigenvalues of random matrices)

SVD, Power method, and Planted Graph problems (+ eigenvalues of random matrices) Chapter 14 SVD, Power method, and Planted Graph problems (+ eigenvalues of random matrices) Today we continue the topic of low-dimensional approximation to datasets and matrices. Last time we saw the singular

More information

Markov chains. 1 Discrete time Markov chains. c A. J. Ganesh, University of Bristol, 2015

Markov chains. 1 Discrete time Markov chains. c A. J. Ganesh, University of Bristol, 2015 Markov chains c A. J. Ganesh, University of Bristol, 2015 1 Discrete time Markov chains Example: A drunkard is walking home from the pub. There are n lampposts between the pub and his home, at each of

More information

Majorizations for the Eigenvectors of Graph-Adjacency Matrices: A Tool for Complex Network Design

Majorizations for the Eigenvectors of Graph-Adjacency Matrices: A Tool for Complex Network Design Majorizations for the Eigenvectors of Graph-Adjacency Matrices: A Tool for Complex Network Design Rahul Dhal Electrical Engineering and Computer Science Washington State University Pullman, WA rdhal@eecs.wsu.edu

More information

Deterministic Decentralized Search in Random Graphs

Deterministic Decentralized Search in Random Graphs Deterministic Decentralized Search in Random Graphs Esteban Arcaute 1,, Ning Chen 2,, Ravi Kumar 3, David Liben-Nowell 4,, Mohammad Mahdian 3, Hamid Nazerzadeh 1,, and Ying Xu 1, 1 Stanford University.

More information

6.207/14.15: Networks Lectures 4, 5 & 6: Linear Dynamics, Markov Chains, Centralities

6.207/14.15: Networks Lectures 4, 5 & 6: Linear Dynamics, Markov Chains, Centralities 6.207/14.15: Networks Lectures 4, 5 & 6: Linear Dynamics, Markov Chains, Centralities 1 Outline Outline Dynamical systems. Linear and Non-linear. Convergence. Linear algebra and Lyapunov functions. Markov

More information

Applications to network analysis: Eigenvector centrality indices Lecture notes

Applications to network analysis: Eigenvector centrality indices Lecture notes Applications to network analysis: Eigenvector centrality indices Lecture notes Dario Fasino, University of Udine (Italy) Lecture notes for the second part of the course Nonnegative and spectral matrix

More information

Lecture Introduction. 2 Brief Recap of Lecture 10. CS-621 Theory Gems October 24, 2012

Lecture Introduction. 2 Brief Recap of Lecture 10. CS-621 Theory Gems October 24, 2012 CS-62 Theory Gems October 24, 202 Lecture Lecturer: Aleksander Mądry Scribes: Carsten Moldenhauer and Robin Scheibler Introduction In Lecture 0, we introduced a fundamental object of spectral graph theory:

More information

Modeling and Stability Analysis of a Communication Network System

Modeling and Stability Analysis of a Communication Network System Modeling and Stability Analysis of a Communication Network System Zvi Retchkiman Königsberg Instituto Politecnico Nacional e-mail: mzvi@cic.ipn.mx Abstract In this work, the modeling and stability problem

More information

Introduction to Search Engine Technology Introduction to Link Structure Analysis. Ronny Lempel Yahoo Labs, Haifa

Introduction to Search Engine Technology Introduction to Link Structure Analysis. Ronny Lempel Yahoo Labs, Haifa Introduction to Search Engine Technology Introduction to Link Structure Analysis Ronny Lempel Yahoo Labs, Haifa Outline Anchor-text indexing Mathematical Background Motivation for link structure analysis

More information

MATH36001 Perron Frobenius Theory 2015

MATH36001 Perron Frobenius Theory 2015 MATH361 Perron Frobenius Theory 215 In addition to saying something useful, the Perron Frobenius theory is elegant. It is a testament to the fact that beautiful mathematics eventually tends to be useful,

More information

Chapter 11. Min Cut Min Cut Problem Definition Some Definitions. By Sariel Har-Peled, December 10, Version: 1.

Chapter 11. Min Cut Min Cut Problem Definition Some Definitions. By Sariel Har-Peled, December 10, Version: 1. Chapter 11 Min Cut By Sariel Har-Peled, December 10, 013 1 Version: 1.0 I built on the sand And it tumbled down, I built on a rock And it tumbled down. Now when I build, I shall begin With the smoke from

More information

Online Social Networks and Media. Link Analysis and Web Search

Online Social Networks and Media. Link Analysis and Web Search Online Social Networks and Media Link Analysis and Web Search How to Organize the Web First try: Human curated Web directories Yahoo, DMOZ, LookSmart How to organize the web Second try: Web Search Information

More information

Consensus of Information Under Dynamically Changing Interaction Topologies

Consensus of Information Under Dynamically Changing Interaction Topologies Consensus of Information Under Dynamically Changing Interaction Topologies Wei Ren and Randal W. Beard Abstract This paper considers the problem of information consensus among multiple agents in the presence

More information

Kernels of Directed Graph Laplacians. J. S. Caughman and J.J.P. Veerman

Kernels of Directed Graph Laplacians. J. S. Caughman and J.J.P. Veerman Kernels of Directed Graph Laplacians J. S. Caughman and J.J.P. Veerman Department of Mathematics and Statistics Portland State University PO Box 751, Portland, OR 97207. caughman@pdx.edu, veerman@pdx.edu

More information

MS&E 321 Spring Stochastic Systems June 1, 2013 Prof. Peter W. Glynn Page 1 of 10

MS&E 321 Spring Stochastic Systems June 1, 2013 Prof. Peter W. Glynn Page 1 of 10 MS&E 321 Spring 12-13 Stochastic Systems June 1, 2013 Prof. Peter W. Glynn Page 1 of 10 Section 3: Regenerative Processes Contents 3.1 Regeneration: The Basic Idea............................... 1 3.2

More information

The Google Markov Chain: convergence speed and eigenvalues

The Google Markov Chain: convergence speed and eigenvalues U.U.D.M. Project Report 2012:14 The Google Markov Chain: convergence speed and eigenvalues Fredrik Backåker Examensarbete i matematik, 15 hp Handledare och examinator: Jakob Björnberg Juni 2012 Department

More information

Distributed Randomized Algorithms for the PageRank Computation Hideaki Ishii, Member, IEEE, and Roberto Tempo, Fellow, IEEE

Distributed Randomized Algorithms for the PageRank Computation Hideaki Ishii, Member, IEEE, and Roberto Tempo, Fellow, IEEE IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 55, NO. 9, SEPTEMBER 2010 1987 Distributed Randomized Algorithms for the PageRank Computation Hideaki Ishii, Member, IEEE, and Roberto Tempo, Fellow, IEEE Abstract

More information

CS168: The Modern Algorithmic Toolbox Lectures #11 and #12: Spectral Graph Theory

CS168: The Modern Algorithmic Toolbox Lectures #11 and #12: Spectral Graph Theory CS168: The Modern Algorithmic Toolbox Lectures #11 and #12: Spectral Graph Theory Tim Roughgarden & Gregory Valiant May 2, 2016 Spectral graph theory is the powerful and beautiful theory that arises from

More information

CS 6820 Fall 2014 Lectures, October 3-20, 2014

CS 6820 Fall 2014 Lectures, October 3-20, 2014 Analysis of Algorithms Linear Programming Notes CS 6820 Fall 2014 Lectures, October 3-20, 2014 1 Linear programming The linear programming (LP) problem is the following optimization problem. We are given

More information

Adaptive Cut Generation for Improved Linear Programming Decoding of Binary Linear Codes

Adaptive Cut Generation for Improved Linear Programming Decoding of Binary Linear Codes Adaptive Cut Generation for Improved Linear Programming Decoding of Binary Linear Codes Xiaojie Zhang and Paul H. Siegel University of California, San Diego, La Jolla, CA 9093, U Email:{ericzhang, psiegel}@ucsd.edu

More information

Dynamic Coalitional TU Games: Distributed Bargaining among Players Neighbors

Dynamic Coalitional TU Games: Distributed Bargaining among Players Neighbors Dynamic Coalitional TU Games: Distributed Bargaining among Players Neighbors Dario Bauso and Angelia Nedić January 20, 2011 Abstract We consider a sequence of transferable utility (TU) games where, at

More information

Lecture 04: Balls and Bins: Birthday Paradox. Birthday Paradox

Lecture 04: Balls and Bins: Birthday Paradox. Birthday Paradox Lecture 04: Balls and Bins: Overview In today s lecture we will start our study of balls-and-bins problems We shall consider a fundamental problem known as the Recall: Inequalities I Lemma Before we begin,

More information

Conditioning of the Entries in the Stationary Vector of a Google-Type Matrix. Steve Kirkland University of Regina

Conditioning of the Entries in the Stationary Vector of a Google-Type Matrix. Steve Kirkland University of Regina Conditioning of the Entries in the Stationary Vector of a Google-Type Matrix Steve Kirkland University of Regina June 5, 2006 Motivation: Google s PageRank algorithm finds the stationary vector of a stochastic

More information

Probability & Computing

Probability & Computing Probability & Computing Stochastic Process time t {X t t 2 T } state space Ω X t 2 state x 2 discrete time: T is countable T = {0,, 2,...} discrete space: Ω is finite or countably infinite X 0,X,X 2,...

More information

ELEMENTS OF PROBABILITY THEORY

ELEMENTS OF PROBABILITY THEORY ELEMENTS OF PROBABILITY THEORY Elements of Probability Theory A collection of subsets of a set Ω is called a σ algebra if it contains Ω and is closed under the operations of taking complements and countable

More information

Google Page Rank Project Linear Algebra Summer 2012

Google Page Rank Project Linear Algebra Summer 2012 Google Page Rank Project Linear Algebra Summer 2012 How does an internet search engine, like Google, work? In this project you will discover how the Page Rank algorithm works to give the most relevant

More information

Eugene Wigner [4]: in the natural sciences, Communications in Pure and Applied Mathematics, XIII, (1960), 1 14.

Eugene Wigner [4]: in the natural sciences, Communications in Pure and Applied Mathematics, XIII, (1960), 1 14. Introduction Eugene Wigner [4]: The miracle of the appropriateness of the language of mathematics for the formulation of the laws of physics is a wonderful gift which we neither understand nor deserve.

More information

The Growth of Functions. A Practical Introduction with as Little Theory as possible

The Growth of Functions. A Practical Introduction with as Little Theory as possible The Growth of Functions A Practical Introduction with as Little Theory as possible Complexity of Algorithms (1) Before we talk about the growth of functions and the concept of order, let s discuss why

More information

PSRGs via Random Walks on Graphs

PSRGs via Random Walks on Graphs Spectral Graph Theory Lecture 11 PSRGs via Random Walks on Graphs Daniel A. Spielman October 3, 2012 11.1 Overview There has been a lot of work on the design of Pseudo-Random Number Generators (PSRGs)

More information

On the Spectra of General Random Graphs

On the Spectra of General Random Graphs On the Spectra of General Random Graphs Fan Chung Mary Radcliffe University of California, San Diego La Jolla, CA 92093 Abstract We consider random graphs such that each edge is determined by an independent

More information

6.207/14.15: Networks Lecture 7: Search on Networks: Navigation and Web Search

6.207/14.15: Networks Lecture 7: Search on Networks: Navigation and Web Search 6.207/14.15: Networks Lecture 7: Search on Networks: Navigation and Web Search Daron Acemoglu and Asu Ozdaglar MIT September 30, 2009 1 Networks: Lecture 7 Outline Navigation (or decentralized search)

More information

On Distributed Coordination of Mobile Agents with Changing Nearest Neighbors

On Distributed Coordination of Mobile Agents with Changing Nearest Neighbors On Distributed Coordination of Mobile Agents with Changing Nearest Neighbors Ali Jadbabaie Department of Electrical and Systems Engineering University of Pennsylvania Philadelphia, PA 19104 jadbabai@seas.upenn.edu

More information

Data Mining and Analysis: Fundamental Concepts and Algorithms

Data Mining and Analysis: Fundamental Concepts and Algorithms Data Mining and Analysis: Fundamental Concepts and Algorithms dataminingbook.info Mohammed J. Zaki 1 Wagner Meira Jr. 2 1 Department of Computer Science Rensselaer Polytechnic Institute, Troy, NY, USA

More information

SMSTC (2007/08) Probability.

SMSTC (2007/08) Probability. SMSTC (27/8) Probability www.smstc.ac.uk Contents 12 Markov chains in continuous time 12 1 12.1 Markov property and the Kolmogorov equations.................... 12 2 12.1.1 Finite state space.................................

More information

BOUNDARY VALUE PROBLEMS ON A HALF SIERPINSKI GASKET

BOUNDARY VALUE PROBLEMS ON A HALF SIERPINSKI GASKET BOUNDARY VALUE PROBLEMS ON A HALF SIERPINSKI GASKET WEILIN LI AND ROBERT S. STRICHARTZ Abstract. We study boundary value problems for the Laplacian on a domain Ω consisting of the left half of the Sierpinski

More information

Spectral Graph Theory and You: Matrix Tree Theorem and Centrality Metrics

Spectral Graph Theory and You: Matrix Tree Theorem and Centrality Metrics Spectral Graph Theory and You: and Centrality Metrics Jonathan Gootenberg March 11, 2013 1 / 19 Outline of Topics 1 Motivation Basics of Spectral Graph Theory Understanding the characteristic polynomial

More information

8.1 Concentration inequality for Gaussian random matrix (cont d)

8.1 Concentration inequality for Gaussian random matrix (cont d) MGMT 69: Topics in High-dimensional Data Analysis Falll 26 Lecture 8: Spectral clustering and Laplacian matrices Lecturer: Jiaming Xu Scribe: Hyun-Ju Oh and Taotao He, October 4, 26 Outline Concentration

More information

COMMUNITY DETECTION IN SPARSE NETWORKS VIA GROTHENDIECK S INEQUALITY

COMMUNITY DETECTION IN SPARSE NETWORKS VIA GROTHENDIECK S INEQUALITY COMMUNITY DETECTION IN SPARSE NETWORKS VIA GROTHENDIECK S INEQUALITY OLIVIER GUÉDON AND ROMAN VERSHYNIN Abstract. We present a simple and flexible method to prove consistency of semidefinite optimization

More information

Z-Pencils. November 20, Abstract

Z-Pencils. November 20, Abstract Z-Pencils J. J. McDonald D. D. Olesky H. Schneider M. J. Tsatsomeros P. van den Driessche November 20, 2006 Abstract The matrix pencil (A, B) = {tb A t C} is considered under the assumptions that A is

More information

Data Mining Recitation Notes Week 3

Data Mining Recitation Notes Week 3 Data Mining Recitation Notes Week 3 Jack Rae January 28, 2013 1 Information Retrieval Given a set of documents, pull the (k) most similar document(s) to a given query. 1.1 Setup Say we have D documents

More information

Matching Polynomials of Graphs

Matching Polynomials of Graphs Spectral Graph Theory Lecture 25 Matching Polynomials of Graphs Daniel A Spielman December 7, 2015 Disclaimer These notes are not necessarily an accurate representation of what happened in class The notes

More information

Properties of Ramanujan Graphs

Properties of Ramanujan Graphs Properties of Ramanujan Graphs Andrew Droll 1, 1 Department of Mathematics and Statistics, Jeffery Hall, Queen s University Kingston, Ontario, Canada Student Number: 5638101 Defense: 27 August, 2008, Wed.

More information

University of Chicago Autumn 2003 CS Markov Chain Monte Carlo Methods

University of Chicago Autumn 2003 CS Markov Chain Monte Carlo Methods University of Chicago Autumn 2003 CS37101-1 Markov Chain Monte Carlo Methods Lecture 4: October 21, 2003 Bounding the mixing time via coupling Eric Vigoda 4.1 Introduction In this lecture we ll use the

More information

LOCAL CHEEGER INEQUALITIES AND SPARSE CUTS

LOCAL CHEEGER INEQUALITIES AND SPARSE CUTS LOCAL CHEEGER INEQUALITIES AND SPARSE CUTS CAELAN GARRETT 18.434 FINAL PROJECT CAELAN@MIT.EDU Abstract. When computing on massive graphs, it is not tractable to iterate over the full graph. Local graph

More information

Lecture 5: Random Walks and Markov Chain

Lecture 5: Random Walks and Markov Chain Spectral Graph Theory and Applications WS 20/202 Lecture 5: Random Walks and Markov Chain Lecturer: Thomas Sauerwald & He Sun Introduction to Markov Chains Definition 5.. A sequence of random variables

More information

CS6220: DATA MINING TECHNIQUES

CS6220: DATA MINING TECHNIQUES CS6220: DATA MINING TECHNIQUES Mining Graph/Network Data Instructor: Yizhou Sun yzsun@ccs.neu.edu March 16, 2016 Methods to Learn Classification Clustering Frequent Pattern Mining Matrix Data Decision

More information

Spectral radius, symmetric and positive matrices

Spectral radius, symmetric and positive matrices Spectral radius, symmetric and positive matrices Zdeněk Dvořák April 28, 2016 1 Spectral radius Definition 1. The spectral radius of a square matrix A is ρ(a) = max{ λ : λ is an eigenvalue of A}. For an

More information

Networks as vectors of their motif frequencies and 2-norm distance as a measure of similarity

Networks as vectors of their motif frequencies and 2-norm distance as a measure of similarity Networks as vectors of their motif frequencies and 2-norm distance as a measure of similarity CS322 Project Writeup Semih Salihoglu Stanford University 353 Serra Street Stanford, CA semih@stanford.edu

More information

Information Propagation Analysis of Social Network Using the Universality of Random Matrix

Information Propagation Analysis of Social Network Using the Universality of Random Matrix Information Propagation Analysis of Social Network Using the Universality of Random Matrix Yusuke Sakumoto, Tsukasa Kameyama, Chisa Takano and Masaki Aida Tokyo Metropolitan University, 6-6 Asahigaoka,

More information

CS6220: DATA MINING TECHNIQUES

CS6220: DATA MINING TECHNIQUES CS6220: DATA MINING TECHNIQUES Mining Graph/Network Data Instructor: Yizhou Sun yzsun@ccs.neu.edu November 16, 2015 Methods to Learn Classification Clustering Frequent Pattern Mining Matrix Data Decision

More information

1 Complex Networks - A Brief Overview

1 Complex Networks - A Brief Overview Power-law Degree Distributions 1 Complex Networks - A Brief Overview Complex networks occur in many social, technological and scientific settings. Examples of complex networks include World Wide Web, Internet,

More information

An Optimal Control Problem Over Infected Networks

An Optimal Control Problem Over Infected Networks Proceedings of the International Conference of Control, Dynamic Systems, and Robotics Ottawa, Ontario, Canada, May 15-16 214 Paper No. 125 An Optimal Control Problem Over Infected Networks Ali Khanafer,

More information

Link Analysis. Leonid E. Zhukov

Link Analysis. Leonid E. Zhukov Link Analysis Leonid E. Zhukov School of Data Analysis and Artificial Intelligence Department of Computer Science National Research University Higher School of Economics Structural Analysis and Visualization

More information

Consensus on networks

Consensus on networks Consensus on networks c A. J. Ganesh, University of Bristol The spread of a rumour is one example of an absorbing Markov process on networks. It was a purely increasing process and so it reached the absorbing

More information

The Strong Largeur d Arborescence

The Strong Largeur d Arborescence The Strong Largeur d Arborescence Rik Steenkamp (5887321) November 12, 2013 Master Thesis Supervisor: prof.dr. Monique Laurent Local Supervisor: prof.dr. Alexander Schrijver KdV Institute for Mathematics

More information