Entropy for Sparse Random Graphs With Vertex-Names

Similar documents
And now for something completely different

Entropy and Compression for Sparse Graphs with Vertex Names

arxiv: v1 [math.pr] 2 Jan 2013

Lecture 4 : Adaptive source coding algorithms

Random weak limits of self-similar Schreier graphs. Tatiana Nagnibeda

Basic Principles of Lossless Coding. Universal Lossless coding. Lempel-Ziv Coding. 2. Exploit dependences between successive symbols.

Studying large networks via local weak limit theory

Chapter 2: Source coding

Theory of Computation

Hard-Core Model on Random Graphs

Szemerédi s regularity lemma revisited. Lewis Memorial Lecture March 14, Terence Tao (UCLA)

NOTES ON MATCHINGS IN CONVERGENT GRAPH SEQUENCES

Source Coding. Master Universitario en Ingeniería de Telecomunicación. I. Santamaría Universidad de Cantabria

On the efficient approximability of constraint satisfaction problems

On bounded redundancy of universal codes

We are going to discuss what it means for a sequence to converge in three stages: First, we define what it means for a sequence to converge to zero

Multimedia Communications. Mathematical Preliminaries for Lossless Compression

Mixing Times and Hitting Times

On Universal Types. Gadiel Seroussi Hewlett-Packard Laboratories Palo Alto, California, USA. University of Minnesota, September 14, 2004

EECS 229A Spring 2007 * * (a) By stationarity and the chain rule for entropy, we have

,

Multiple Choice Tries and Distributed Hash Tables

Discrete random structures whose limits are described by a PDE: 3 open problems

Lab 8: Measuring Graph Centrality - PageRank. Monday, November 5 CompSci 531, Fall 2018

Chapter 9 Fundamental Limits in Information Theory

Lossy Compression Coding Theorems for Arbitrary Sources

Lecture 16 Symbolic dynamics.

3F1 Information Theory, Lecture 3

Introduction to Techniques for Counting

1 Mechanistic and generative models of network structure

CS 420/594: Complex Systems & Self-Organization Project 1: Edge of Chaos in 1D Cellular Automata Due: Sept. 20

Networks: Lectures 9 & 10 Random graphs

Almost giant clusters for percolation on large trees

Digital communication system. Shannon s separation principle

Motivation for Arithmetic Coding

Midterm Exam Information Theory Fall Midterm Exam. Time: 09:10 12:10 11/23, 2016

Local limits of random graphs

ECE Advanced Communication Theory, Spring 2009 Homework #1 (INCOMPLETE)

/633 Introduction to Algorithms Lecturer: Michael Dinitz Topic: Dynamic Programming II Date: 10/12/17

CS1800: Mathematical Induction. Professor Kevin Gold

Module 9: Mathematical Induction

Behaviour of Lipschitz functions on negligible sets. Non-differentiability in R. Outline

VC-DENSITY FOR TREES

Statistical Physics on Sparse Random Graphs: Mathematical Perspective

Time-bounded computations

Ramsey Theory. May 24, 2015

Information & Correlation

The Complexity of the Comparator Circuit Value Problem

arxiv: v2 [math.ds] 13 Sep 2017

The greedy independent set in a random graph with given degr

First Order Convergence and Roots

Independence and chromatic number (and random k-sat): Sparse Case. Dimitris Achlioptas Microsoft

1 Ex. 1 Verify that the function H(p 1,..., p n ) = k p k log 2 p k satisfies all 8 axioms on H.

Theory of Computation

Information Theory of Random Trees Induced by Stochastic Grammars

Lecture 6: September 22

Finite Markov Information-Exchange processes

Multimedia. Multimedia Data Compression (Lossless Compression Algorithms)

P, NP, NP-Complete, and NPhard

arxiv: v1 [math.pr] 11 Dec 2017

CSC304: Algorithmic Game Theory and Mechanism Design Fall 2016

Deterministic Decentralized Search in Random Graphs

DR.RUPNATHJI( DR.RUPAK NATH )

Introduction to information theory and coding

Lecture 11: Polar codes construction

Contents. Counting Methods and Induction

ICS141: Discrete Mathematics for Computer Science I

Analytic Information Theory: From Shannon to Knuth and Back. Knuth80: Piteaa, Sweden, 2018 Dedicated to Don E. Knuth

The Kolakoski Sequence and Some Fast Algorithms, Part 2

Nonamenable Products are not Treeable

An Intuitive Introduction to Motivic Homotopy Theory Vladimir Voevodsky

d(x n, x) d(x n, x nk ) + d(x nk, x) where we chose any fixed k > N

Erdős-Renyi random graphs basics

Solving a linear equation in a set of integers II

Tutorial on Mathematical Induction

Chaos, Complexity, and Inference (36-462)

Intro to Information Theory

Tree sets. Reinhard Diestel

Lecture Notes on Inductive Definitions

CS632 Notes on Relational Query Languages I

CS1800: Strong Induction. Professor Kevin Gold

Algorithms: COMP3121/3821/9101/9801

3F1 Information Theory, Lecture 3

Notes on cycle generating functions

Chaos, Complexity, and Inference (36-462)

Complementary Ramsey numbers, graph factorizations and Ramsey graphs

Cheng Soon Ong & Christian Walder. Canberra February June 2018

Research Statement. Janine E. Janoski

Midterm 2 for CS 170

Multiscale Approach for the Network Compression-friendly Ordering

Connectivity of addable graph classes

Context tree models for source coding

Connectivity of addable graph classes

3.2 Configuration model

CS 173: Induction. Madhusudan Parthasarathy University of Illinois at Urbana-Champaign. February 7, 2016

A Mathematical Theory of Communication

Memory in Classical Information Theory: A Brief History

Section 20: Arrow Diagrams on the Integers

Advanced Counting Techniques. Chapter 8

Phase Transitions of Random Codes and GV-Bounds

Transcription:

Entropy for Sparse Random Graphs With Vertex-Names David Aldous 11 February 2013

if a problem seems... Research strategy (for old guys like me): do-able = give to Ph.D. student maybe do-able = give to post-doc clearly not do-able = think about it myself. I m thinking about a topic in a very classical way (Shannon): lossless data compression ignoring computational complexity of coding

What is a network? A graph is a well-defined mathematical object vertices and edges etc A network is a graph with context-dependent extra structure. I want to consider a certain simple-but-interesting notion of extra structure, which is the notion vertices have names.

Consider a graph with N vertices O(1) average degree vertices have distinct names, strings of length O(log N) from a fixed finite alphabet. Envisage some association between vertex names and graph structure, as in phylogenetic trees on species road networks. I claim [key point of this set-up] this is the interesting as theory setting for data compression of sparse graphs, because the entropy of both the graph structure and of the names has the same order, N log N. And we want to exploit the association.

Formal setup. Define a finite set S = S(N, A, β, α); an element of S is a network with N vertices ave degree α (at most αn/2 edges) finite alphabet A of size A vertices have distinct names strings from A of length β log A N. Here A 2, 0 < α <, 1 < β < are fixed and we study as N. Easy to see how big S is: log S(N, A, β, α) (β 1 + α 2 )N log N. For some given model of random network G N we expect ent(g N ) cn log N for some model-dependent entropy rate 0 c (β 1 + α 2 ). [cute observation: value of c doesn t depend on base of log.]

In this setting there are two extreme lines of research you might try. Clearly do-able: Invent probability models and calculate their entropy rate. With Nathan Ross we have a little paper doing this (search on arxiv aldous entropy ). Will show 3 slides about this, soon. Clearly not do-able: Design an algorithm which, given a realization from a probability model, compresses optimally (according to the entropy of the probability model) without knowing what the model is ( universal, like Lempel-Ziv for sequences). So my goal is to find a project laying between these extremes. [Pedagogical aside: this N log N world could provide projects for a first course in IT.]

What is in the existing literature? A: More mathematical. Topic called graph entropy studies the number of automorphisms of a N-vertex unlabelled graph. See the 2012 survey by Szpankowski - Choi Compression of graphical structures: Fundamental limits, algorithms, and experiments. B: More applied. Seeking to exploit the specific structure of particular types of network. Boldi - Vigna (2003). The webgraph framework I: Compression techniques. Chierichetti - Kumar - Lattanzi - Mitzenmacher - Panconesi - Raghavan (2009). On compressing social networks.

Clearly do-able project: Invent probability models and calculate their entropy rate. Many hundreds of papers in complex networks study probability models for random N-vertex graphs, implicitly with vertices labeled 1,..., N. Two simplest ways to adapt to our setting: (i) write integer labels in binary. (ii) assign completely random distinct names. For the best-known models Erdős-Rényi, small worlds, preferential attachment, random subject to prescribed degree distribution,... it is almost trivial to calculate the entropy rate. But none of these models has a very interesting association between graph structure and names. Here is our best attempt at a model that does, and that requires a non-trivial calculation for entropy rate.

Model: Grow sparse Erdős-Rényi G(N, α/n) sequentially; an arriving vertex is linked to a previous vertex with chance α/n. Vertex 1 is given a uniform random length-l N A-ary name, where (as in other models) L N β log A N. A subsequent vertex arrives with a tentative name a 0 ; gets linked to Q 0 previous vertices with names a 1,..., a Qn ; assign the name obtained by picking the letter in each coordinate 1 u L N uniformly from the 1 + Q letters au, 0 au, 1..., au Qn. This gives a family (G N ) parametrized by (A, β, α). One can calculate its Entropy rate: 1 + α 2 + β α k J k (α)h A (k) k! log A k 0 where J k (α) := 1 0 x k e αx dx and the constants h A (k) are defined as follows:.

h A (k) := A k (a 1,...,a k ) A k ent(p [a1,...,ak ] ), (1) and where p [a1,...,a k ] is the probability distribution p on A defined by p [a1,...,a k ] (a) = 1 + A {i : a i = a}. (1 + k)a Question: Can you be more creative than us in inventing such models?

[repeat earlier slide] In this setting there are two extreme lines of research you might try. Clearly do-able: Invent probability models and calculate their entropy rate....... Clearly not do-able: Design an algorithm which, given a realization from a probability model, compresses optimally (according to the entropy of the probability model) without knowing what the model is ( universal, like Lempel-Ziv for sequences). In the rest of the talk I will outline a maybe do-able project laying between these extremes.

Background: Shannon without stationarity Finite alphabet A. For each N we will be given an A-valued sequence X (N) = (X (N) i, 1 i N). No further assumptions. Question: How well can a universal data compression algorithm do? Answer: Can guarantee that lim sup N 1 (compressed length of X (N) ) c N where c is the local entropy rate defined as follows.

First suppose we have local weak convergence, meaning: take U (N) uniform on 1,..., N; there is a process X which is the limit in the sense: for each k, ( ) (X (N) U (N) +i, 1 i k) d (X i, 1 i k). Then X is stationary and has some entropy rate c. Define c = c. In general (*) may not hold but we have compactness; different stationary processes may arise as different subsequential limits; define c as [essentially] the sup of their entropy rates. Not a particularly useful idea for sequences, because stationarity is a plausible assumption and gives stronger results. But I claim it is a natural way to start thinking about data compression in our setting of sparse graphs with vertex-names.

Given rooted unlabeled graphs g N and g where g is locally finite, there is a natural notion of local convergence g N local g meaning convergence of restrictions to within fixed distance from root. This induces a notion of convergence in distribution for random such graphs G N local G in distribution. (2) Now given unrooted unlabeled random graphs G N and a rooted unlabeled random graph G, we have a notion called local weak convergence or Benjamini-Schramm convergence G N LWC G defined by: first assign a uniform random root to G N then require (2).

Some background for LWC Random N-vertex 3-regular graph converges to the infinite 3-regular tree. The rooted binary tree of height h converges to a different tree. The limit random rooted graph G always has a property (unimodular) analogous to stationarity for sequences. In contrast (to stationarity) it is not obvious that every unimodular G is a local weak limit of some sequence of finite random graphs (Aldous-Lyons conjecture). Presumably one could develop a theory of entropy/coding for sparse graphs where vertices have marks from a fixed finite alphabet, in terms of an entropy rate for the limit infinite marked graph.

Program: I want to make correct version of following conjecture about sparse random graphs-with-vertex-names G N. Write G N,k for the restriction of G N to a k-vertex neighborhod of a uniform random vertex. For fixed k we expect and then we expect a limit ent(g N,k ) c k log N as N k 1 c k c as k. Conjecture. As with sequences, this local entropy rate is an upper bound for global entropy: lim N ent(g N ) N log N c with equality under a suitable no long range dependence condition analogous to ergodic for sequences. Such a theory result would give a target for somewhat universal compression algorithms.