Lecture 13 : Kesten-Stigum bound

Similar documents
The Tightness of the Kesten-Stigum Reconstruction Bound for a Symmetric Model With Multiple Mutations

The Kesten-Stigum Reconstruction Bound Is Tight for Roughly Symmetric Binary Channels

Lecture 11 : Asymptotic Sample Complexity

Lecture 18 : Ewens sampling formula

Belief Propagation, Robust Reconstruction and Optimal Recovery of Block Models

Lecture 19 : Brownian motion: Path properties I

Sharpness of second moment criteria for branching and tree-indexed processes

Evolutionary trees and the Ising model on the Bethe lattice: a proof of Steel s conjecture

Notes 6 : First and second moment methods

Reconstruction for Models on Random Graphs

XVI International Congress on Mathematical Physics

Optimal Phylogenetic Reconstruction

Phylogenetic Reconstruction with Insertions and Deletions

From trees to seeds:

Optimal Phylogenetic Reconstruction

Notes 18 : Optional Sampling Theorem

The range of tree-indexed random walk

Lecture 9 : Identifiability of Markov Models

Lecture 19 : Chinese restaurant process

Notes 9 : Infinitely divisible and stable laws

First Passage Percolation Has Sublinear Distance Variance

Lecture 7 Introduction to Statistical Decision Theory

Glauber Dynamics on Trees and Hyperbolic Graphs

BMI/CS 776 Lecture 4. Colin Dewey

THE NOISY VETO-VOTER MODEL: A RECURSIVE DISTRIBUTIONAL EQUATION ON [0,1] April 28, 2007

arxiv: v1 [cs.ds] 2 Nov 2018

A Conceptual Proof of the Kesten-Stigum Theorem for Multi-type Branching Processes

Notes 15 : UI Martingales

Global Alignment of Molecular Sequences via Ancestral State Reconstruction

The Tightness of the Kesten Stigum Reconstruction Bound of Symmetric Model with Multiple Mutations

How robust are reconstruction thresholds for community detection?

Lecture 8: Information Theory and Statistics

Notes 1 : Measure-theoretic foundations I

Problem Set 7 Due March, 22

Modern Discrete Probability Branching processes

Streaming Algorithms for Optimal Generation of Random Bits

Theorem 2.1 (Caratheodory). A (countably additive) probability measure on a field has an extension. n=1

Branching processes. Chapter Background Basic definitions

Spectral thresholds in the bipartite stochastic block model

Notes 20 : Tests of neutrality

Reconstruction for models on random graphs

Deep Learning and Hierarchical Generative Models

1 Distribution Property Testing

Almost sure asymptotics for the random binary search tree

Broadcasting on Random Directed Acyclic Graphs

SYDE 372 Introduction to Pattern Recognition. Probability Measures for Classification: Part I

BRANCHING PROCESSES 1. GALTON-WATSON PROCESSES

Notes 13 : Conditioning

Binary Search Trees. Lecture 29 Section Robb T. Koether. Hampden-Sydney College. Fri, Apr 8, 2016

The main results about probability measures are the following two facts:

Lecture 2: Proof of Switching Lemma

or E ( U(X) ) e zx = e ux e ivx = e ux( cos(vx) + i sin(vx) ), B X := { u R : M X (u) < } (4)

6.1 Variational representation of f-divergences

The Moment Method; Convex Duality; and Large/Medium/Small Deviations

The first bound is the strongest, the other two bounds are often easier to state and compute. Proof: Applying Markov's inequality, for any >0 we have

Iowa State University. Instructor: Alex Roitershtein Summer Exam #1. Solutions. x u = 2 x v

Data Preprocessing. Cluster Similarity

Markov Models on Trees: Reconstruction and Applications. Sébastien Jocelyn Roch

Probability Models of Information Exchange on Networks Lecture 1

Resistance Growth of Branching Random Networks

arxiv: v1 [math.pr] 20 May 2018

METASTABLE BEHAVIOR FOR BOOTSTRAP PERCOLATION ON REGULAR TREES

Lecture 4 Noisy Channel Coding

Proving the central limit theorem

ECE 564/645 - Digital Communications, Spring 2018 Homework #2 Due: March 19 (In Lecture)

The University of Hong Kong Department of Statistics and Actuarial Science STAT2802 Statistical Models Tutorial Solutions Solutions to Problems 71-80

Module 3. Function of a Random Variable and its distribution

On the Submodularity of Influence in Social Networks

How Robust are Thresholds for Community Detection?

Connection to Branching Random Walk

A Martingale Central Limit Theorem

Relative growth of the partial sums of certain random Fibonacci-like sequences

Reconstruire le passé biologique modèles, méthodes, performances, limites

The Wright-Fisher Model and Genetic Drift

Perron Frobenius Theory

arxiv:math.pr/ v1 17 May 2004

Evolutionary Tree Analysis. Overview

FINITE ABELIAN GROUPS Amin Witno

From trees to seeds: on the inference of the seed from large trees in the uniform attachment model

arxiv: v2 [math.pr] 25 Feb 2017

Linear sketching for Functions over Boolean Hypercube

CLASSICAL PROBABILITY MODES OF CONVERGENCE AND INEQUALITIES

Lecture 7: The Subgraph Isomorphism Problem

Lin Song Shuo Shao Jun Chen. 11 September 2013

Bootstrap Percolation on Periodic Trees

1 Review of random functions This paper is a sequel of our earlier paper [11]. We assume that the reader is familiar with that paper, however, we repe

ENEE 621 SPRING 2016 DETECTION AND ESTIMATION THEORY THE PARAMETER ESTIMATION PROBLEM

ON COMPOUND POISSON POPULATION MODELS

Spectral Methods for Learning Multivariate Latent Tree Structure

ROBUST RECONSTRUCTION ON TREES IS DETERMINED BY THE SECOND EIGENVALUE

Lecture 3. Econ August 12

PCMI LECTURE NOTES ON PROPERTY (T ), EXPANDER GRAPHS AND APPROXIMATE GROUPS (PRELIMINARY VERSION)

3 Integration and Expectation

Phylogenetic Reconstruction with Insertions and Deletions

COPYRIGHTED MATERIAL CONTENTS. Preface Preface to the First Edition

Mutual Information, Synergy and Some Curious Phenomena for Simple Channels

CMPUT 675: Approximation Algorithms Fall 2014

Lecture 2: From Linear Regression to Kalman Filter and Beyond

The Dirichlet s P rinciple. In this lecture we discuss an alternative formulation of the Dirichlet problem for the Laplace equation:

Considering our result for the sum and product of analytic functions, this means that for (a 0, a 1,..., a N ) C N+1, the polynomial.

Transcription:

Lecture 3 : Kesten-Stigum bound MATH85K - Spring 00 Lecturer: Sebastien Roch References: [EKPS00, Mos0, MP03, BCMR06]. Previous class DEF 3. (Ancestral reconstruction solvability) Let µ + h be the distribution µ h conditioned on the root state σ 0 being +, and similarly for µ h. We say that the ancestral reconstruction problem (under the CFN model) for 0 < p < / is solvable if lim inf h µ+ h µ h > 0, otherwise the problem is unsolvable. THM 3. (Solvability) Let θ = p = /. Then when p < p the ancestral reconstruction problem is solvable. Kesten-Stigum bound The previous theorem was proved by showing that majority is a good root estimator up to p = p. Here we show that this result is best possible. Of course, majority is not the best root estimator: in general its error probability can be higher than maximum likelihood. (See Figure 3 in [EKPS00] for an insightful example where majority and maximum likelihood differ.) However, it turns out that the critical threshold for majority, called the Kesten-Stigum bound, coincides with the critical threshold of maximum likelihood in the CFN model. Note that the latter is not true for more general models [Mos0]. THM 3.3 (Tightness of Kesten-Stigum Bound) Let θ = p = /. Then when p p the ancestral reconstrution problem is not solvable. Along each path from the root, information is lost through mutation at exponential rate maeasured by θ = p. Meanwhile, the tree is growing exponentially

Lecture 3: Kesten-Stigum bound and information is duplicated measured by the branching ratio b =. These two forces balance each other out when bθ =, the critical threshold in the theorem. To prove Theorem 3.3 we analyze the maximum likelihood estimator. Let µ h (s 0 s h ) be the conditional probability of the root state s 0 given the states s h at level h. It will be more convenient to work with the following related quantity Z h = µ h (+ σ h ) µ h ( σ h ) = [µ+ h (σ h) µ h (σ h)] = µ h (+ σ h ), which, as a function of σ h, is a random variable. Note that E[Z h ] = 0 by symmetry. It is enough to prove a bound on the variance of Z h. LEM 3.4 It holds that µ + h µ h E[Zh ]. Proof: By Bayes rule and Cauchy-Schwarz µ + h (s h) µ h (s h) = µ h (s h ) µ h (+ s h ) µ h ( s h ) s h s h = E Z h E[Zh ]. Let z h = E[Zh ]. The proof of Theorem 3.3 will follow from lim h z h = 0. We apply the same type of recursive argument we used for the analysis of majority: we condition on the root to exploit conditional independence; we apply the Markov channel on the top edge. Distributional recursion We first derive a recursion for Z h. Let σ h be the states at level h below the first child of the root and let µ h be the distribution of σ h. Define Ż h = µ h (+ σ h ) µ h ( σ h ), where µ h (s 0 ṡ h ) is the conditional probability that the root is s 0 given that σ h = ṡ h. Similarly, denote with a double dot the same quantities with respect to the subtree below the second child of the root.

Lecture 3: Kesten-Stigum bound 3 LEM 3.5 It holds pointwise that Z h = Żh + Z h + Żh Z h. Proof: Using µ + h (s h) = µ + h (ṡ h) µ + h ( s h), note that Z h = = = = 4 γ µγ h (σ h) µ γ h ( σ h) µ γ h ( σ h) ( ) ( + γżh + γ γ Z ) h (Żh + Z h ), where µ γ h ( σ h) µ γ h ( σ h) = = = 4 ( + Żh Z h ). µ γ h (σ h) µ γ h ( σ h) µ γ h ( σ h) ( ) ( + γżh + γ Z h ) Define Ż h = µ h (+ σ h ) µ h ( σ h ), where µ h (s 0 σ h ) is the condition probability that the first child of the root is s 0 given that the states at level h below the first child are σ h. Similarly, LEM 3.6 It holds pointwise that Ż h = θżh. Proof: The proof is similar to that of the previous lemma and is left as an exercise.

Lecture 3: Kesten-Stigum bound 4 3 Moment recursion We now take expectations in the previous recursion for Z h. Note that we need to compute the second moment. However, an important simplification arises from the following observation: E + h [Z h] = s h µ + h (s h)z h (s h ) = s h µ h (s h ) µ+ h (s h) µ h (s h ) Z h(s h ) = s h µ h (s h )( + Z h (s h ))Z h (s h ) = E[( + Z h )Z h ] = E[Z h ], so it suffices to compute the (conditioned) first moment. Proof:(of Theorem 3.3) Using the expansion we have that + r = r + r + r, Z h = θ(żh + Z h ) θ 3 (Żh + Z h )Żh Z h + θ 4 Ż h Z h Z h θ(żh + Z h ) θ 3 (Żh + Z h )Żh Z h + θ 4 Ż h Z h, () where we used Z h. To take expectations, we need the following lemma. LEM 3.7 We have and E + h [Żh ] = θe + h [Żh ], E + h [Ż h ] = ( θ)e[ż h ] + θe+ h [Ż h ] = E[Ż h ] = E+ h [Żh ]. Proof: For the first equality, note that by symmetry E + h [Żh ] = ( p)e + h [Żh ] + pe h [Żh ] = ( p)e + h [Żh ]. The second equality is proved similarly and is left as an exercise. Taking expectations in (), using conditional independence and symmetry z h θ z h θ 4 z h + θ4 z h = θ z h θ 4 z h.

Lecture 3: Kesten-Stigum bound 5 References [BCMR06] Christian Borgs, Jennifer T. Chayes, Elchanan Mossel, and Sébastien Roch. The Kesten-Stigum reconstruction bound is tight for roughly symmetric binary channels. In FOCS, pages 58 530, 006. [EKPS00] [Mos0] [MP03] W. S. Evans, C. Kenyon, Y. Peres, and L. J. Schulman. Broadcasting on trees and the Ising model. Ann. Appl. Probab., 0():40 433, 000. E. Mossel. Reconstruction on trees: beating the second eigenvalue. Ann. Appl. Probab., ():85 300, 00. E. Mossel and Y. Peres. Information flow on trees. Ann. Appl. Probab., 3(3):87 844, 003.