ALMOST SURE CONVERGENCE OF RANDOM GOSSIP ALGORITHMS

Similar documents
Randomized Gossip Algorithms

STOCHASTIC PROCESSES Basic notions

Markov Chains, Stochastic Processes, and Matrix Decompositions

Markov Chains, Random Walks on Graphs, and the Laplacian

MET Workshop: Exercises

Distributed Optimization over Networks Gossip-Based Algorithms

Consensus and Distributed Inference Rates Using Network Divergence

Convex Optimization CMU-10725

Definition A finite Markov chain is a memoryless homogeneous discrete stochastic process with a finite number of states.

18.440: Lecture 33 Markov Chains

18.600: Lecture 32 Markov Chains

T.8. Perron-Frobenius theory of positive matrices From: H.R. Thieme, Mathematics in Population Biology, Princeton University Press, Princeton 2003

Distributed Optimization over Random Networks

Distributed Randomized Algorithms for the PageRank Computation Hideaki Ishii, Member, IEEE, and Roberto Tempo, Fellow, IEEE

Average-Consensus of Multi-Agent Systems with Direct Topology Based on Event-Triggered Control

Conceptual Questions for Review

Invertibility and stability. Irreducibly diagonally dominant. Invertibility and stability, stronger result. Reducible matrices

Asymmetric Randomized Gossip Algorithms for Consensus

Social network analysis: social learning

Consensus and Products of Random Stochastic Matrices: Exact Rate for Convergence in Probability

Necessary and sufficient conditions for strong R-positivity

Almost Sure Convergence to Consensus in Markovian Random Graphs

On the convergence of weighted-average consensus

LECTURE 3. Last time:

Markov chains (week 6) Solutions

The goal of this chapter is to study linear systems of ordinary differential equations: dt,..., dx ) T

MARKOV PROCESSES. Valerio Di Valerio

MATH 56A SPRING 2008 STOCHASTIC PROCESSES 65

Random Walks on Graphs. One Concrete Example of a random walk Motivation applications

TMA Calculus 3. Lecture 21, April 3. Toke Meier Carlsen Norwegian University of Science and Technology Spring 2013

Multi Stage Queuing Model in Level Dependent Quasi Birth Death Process

18.175: Lecture 30 Markov chains

Markov chains. Randomness and Computation. Markov chains. Markov processes

NOTES ON THE PERRON-FROBENIUS THEORY OF NONNEGATIVE MATRICES

Large Deviations Rates for Distributed Inference. Doctor of Philosophy. Electrical and Computer Engineering. Dragana Bajović

Learning distributions and hypothesis testing via social learning

Asymptotics, asynchrony, and asymmetry in distributed consensus

Weighted Gossip: Distributed Averaging Using Non-Doubly Stochastic Matrices

Perron Frobenius Theory

spring, math 204 (mitchell) list of theorems 1 Linear Systems Linear Transformations Matrix Algebra

Introduction to Machine Learning CMU-10701

Generating random spanning trees. Andrei Broder. DEC - Systems Research Center 130 Lytton Ave., Palo Alto, CA

Statistics 992 Continuous-time Markov Chains Spring 2004

Consensus Seeking in Multi-agent Systems Under Dynamically Changing Interaction Topologies

Math 1553, Introduction to Linear Algebra

Bounds for (generalised) Lyapunov exponents for random products of shears

Markov Chains and Stochastic Sampling

Agreement algorithms for synchronization of clocks in nodes of stochastic networks

Distributed Estimation and Detection for Smart Grid

Boolean Inner-Product Spaces and Boolean Matrices

Combinatorial semigroups and induced/deduced operators

Quantum walk algorithms

P i [B k ] = lim. n=1 p(n) ii <. n=1. V i :=

S-adic sequences A bridge between dynamics, arithmetic, and geometry

Markov Chains CK eqns Classes Hitting times Rec./trans. Strong Markov Stat. distr. Reversibility * Markov Chains


CHILE LECTURES: MULTIPLICATIVE ERGODIC THEOREM, FRIENDS AND APPLICATIONS

On Distributed Coordination of Mobile Agents with Changing Nearest Neighbors

6 Markov Chain Monte Carlo (MCMC)

Consensus Seeking in Multi-agent Systems Under Dynamically Changing Interaction Topologies

Alternate Characterizations of Markov Processes

Understanding MCMC. Marcel Lüthi, University of Basel. Slides based on presentation by Sandro Schönborn

Introduction to Search Engine Technology Introduction to Link Structure Analysis. Ronny Lempel Yahoo Labs, Haifa

Consensus of Information Under Dynamically Changing Interaction Topologies

Geometric Constraints II

LIMITING PROBABILITY TRANSITION MATRIX OF A CONDENSED FIBONACCI TREE

Linear Stochastic Systems: A Geometric Approach to Modeling, Estimation and Identification

Uncertainty and Randomization

6.842 Randomness and Computation February 24, Lecture 6

Section 1.7: Properties of the Leslie Matrix

Quantized average consensus via dynamic coding/decoding schemes

Convergence speed in distributed consensus and averaging

Convergence Rate for Consensus with Delays

Consensus Problems on Small World Graphs: A Structural Study

On the mathematical background of Google PageRank algorithm

Lyapunov exponents of Teichmüller flows

Lecture 5: Random Walks and Markov Chain

Copositive matrices and periodic dynamical systems

Language Acquisition and Parameters: Part II

Part I Stochastic variables and Markov chains

E-Companion to The Evolution of Beliefs over Signed Social Networks

MATH 56A: STOCHASTIC PROCESSES CHAPTER 1

Math Homework 5 Solutions

Convex Optimization of Graph Laplacian Eigenvalues

Differential Topology Final Exam With Solutions

Ergodic Theorems. Samy Tindel. Purdue University. Probability Theory 2 - MA 539. Taken from Probability: Theory and examples by R.

Average laws in analysis

Powerful tool for sampling from complicated distributions. Many use Markov chains to model events that arise in nature.

Markov processes and queueing networks

On Backward Product of Stochastic Matrices

MATH 56A: STOCHASTIC PROCESSES CHAPTER 2

No class on Thursday, October 1. No office hours on Tuesday, September 29 and Thursday, October 1.

Non-Essential Uses of Probability in Analysis Part IV Efficient Markovian Couplings. Krzysztof Burdzy University of Washington

The Theory behind PageRank

Mathematical foundations - linear algebra

Parametric Signal Modeling and Linear Prediction Theory 1. Discrete-time Stochastic Processes

A Relationship Between Minimum Bending Energy and Degree Elevation for Bézier Curves

Reaching a Consensus in a Dynamically Changing Environment A Graphical Approach

Propp-Wilson Algorithm (and sampling the Ising model)

Advanced Computer Networks Lecture 2. Markov Processes

Transcription:

ALMOST SURE CONVERGENCE OF RANDOM GOSSIP ALGORITHMS Giorgio Picci with T. Taylor, ASU Tempe AZ. Wofgang Runggaldier s Birthday, Brixen July 2007 1

CONSENSUS FOR RANDOM GOSSIP ALGORITHMS Consider a finite set of nodes representing say wireless sensors or distributed computing units, can they achieve a common goal by exchanging information only locally? exchanging information locally for the purpose of forming a common estimate of some physical variable x; Each node k forms his own estimate x k (t), t Z + and updates according to exchange of information with a neighbor. Neighboring pairs are chosen randomly Q: will all local estimates {x k (t), k = 1,...,n} converge to the same value as t?. 2

DYNAMICS OF RANDOM GOSSIP ALGORITHMS While two nodes v i and v j are in communication, they exchange information to refine their own estimate using the neighbor s estimate. Model this adjustament in discrete time by a simple symmetric linear relation x i (t + 1) = x i (t) + p(x j (t) x i (t)) x j (t + 1) = x j (t) + p(x i (t) x j (t)) where p is some positive gain parameter modeling the speed of adjustment. For stability need to impose that 1 2p 1 and hence 0 p 1. For p = 1 2 you take the average of the two measurements so that x i(t +1) = x j (t + 1). 3

DYNAMICS OF RANDOM GOSSIP ALGORITHMS The whole coordinate vector x(t) R n evolves according to x(t +1) = A(e)x(t), the matrix A(e) R n n depending on the edge e = v i v j selected at that particular time instant; A(e) = 1 0 0 0.......... 1 p p. 1....... 1. p 1 p... ) T = I n p (1 vi 1 v j )(1 vi 1 v j 1 4

EIGENSPACES The vector 1 vi has the i th entry equal to 1 and zero otherwise. A(e) is a symmetric doubly stochastic matrix. The value 1 2p is a simple eigenvalue associated to the eigenvector (1 vi 1 v j ), A(e)(1 vi 1 v j ) = (1 vi 1 v j ) p(1 vi 1 v j )2 = (1 2p)(1 vi 1 v j ) ( ) the orthogonal (codimension one) subspace 1 vi 1 v j is the eigenspace of the eigenvalue 1. Let 1 := [1,...,1]. Want x(t) to converge to the subspace {1} := {α1; α R}. This would be automatically true for a fixed irreducible d-stochastic matrix. 5

A CONTROLLABILITY LEMMA Lemma 1 Let G = (V,E) be a graph. Then iff G is connected. span{1 vi 1 v j : (v i v j ) E} = 1 ( 1 vi 1 v j ) span {1}; i.e. Corollary 1 Let G = (V,E ) with E E be a subgraph of G. Let {e i : 1 i m } be an ordering of E, and let π denote a permutation of {1,2,,m }. Let B(E,π) = m i=1 A(e πi), where the product is ordered from right to left. Then if and only if G is connected. B(E,π) 1 < 1 6

THE EDGE PROCESS Let Ω = E N, be the space of all semi-infinite sequences taking values in E, and let σ : Ω Ω denote the shift map: σ(e 0,e 1,e 2,,e n, ) = (e 1,e 2,,e n, ). Let ev k : Ω E denote the evaluation on the k th term. Let µ denote an ergodic shift invariant probability measure on Ω, so that the edge process e(k) : ω ev k (ω) is ergodic. Special cases: e(k) is iid, or an ergodic Markov chain. However, what we shall do works for general ergodic processes. Consider the function C : Ω Z R n n, C(ω,t) := A(ev i (ω)) = A(ev 0 (σ i ω)) t 1 i=0 which by stationarity of e obeys the composition rule C(ω,t +s) =C(σ t ω,s)c(ω,t) with C(ω,0) = I. Such a function is called a matrix cocycle. t 1 i=0 7

MULTIPLICATIVE ERGODIC THEOREM Theorem 1 [Oseledet s Multiplicative Ergodic Theorem] Let µ be a shift invariant probability measure on Ω and suppose that the shift map σ : Ω Ω is ergodic and that log + C(ω,t) is in L 1. Then the limit Λ = lim t ( C(ω,t) T C(ω,t)) 1 2t (1) exists with probability one, is symmetric and nonnegative definite, and is µ a.s. independent of ω. Let λ 1 < λ 2 < λ k for k n be the distinct eigenvalues of Λ, let U i denote the eigenspace of λ i, and let V i = i j=1 U j. Then for u V i V i 1, 1 lim t t log C(ω,t)u = log(λ i). (2) The numbers λ i are called the Lyapunov exponents of C. 8

MULTIPLICATIVE ERGODIC THEOREM The Lyapunov exponents control the exponential rate of convergence (or non-convergence) to consensus. The matrices A(e) are doubly stochastic as are any matrix products of them, C(ω,t). If follows that the constant functions on V, {1}, as well as the mean zero functions in {1} are invariant under the action of this cocycle and of its transpose. Thus these subspaces are also invariant under the limiting matrix Λ of the Oseledet s theorem. There is a Lyapunov exponent associated with the subspace {1} which, it is not difficult to see, is one. There are n 1 Lyapunov exponents associated with the subspace 1, so the key point is to characterize them. 9

CONVERGENCE TO CONSENSUS For x R n use the symbol x := 1 n ni=1 x i. The main convergence result follows. Theorem 2 Let G = (V,E) be a connected graph and let e(t) be an ergodic stochastic process taking values on E. Suppose that the support of the probability distribution induced by e(t) is all of E. Let the gossip algorithm be initialized at x(0) = x 0. Then there is a (deterministic) constant λ < 1 and a (random) constant K λ such that µ-almost surely. x(t) x 0 1 < K λ λ t x 0 x 0 1 10

OPEN QUESTIONS Rate of convergence (for L 2...) Multiple gossiping : more than one pair of communicating edges per time slot, Convergence is merely associated to the time T it takes the algorithm to visit a spanning tree with positive probability. Indeed, the actual rate of convergence of the algorithm is just determined by T. Much remains to be done!!! 11

REFERENCES W. Runggaldier (circa 1970): STILLE WASSER GRUNDEN TIEF, unpublished (although well known among specialists). 12