Perfect simulation for image analysis

Similar documents
Perfect simulation of repulsive point processes

7.1 Coupling from the Past

Introduction to Machine Learning CMU-10701

Perfect simulation for repulsive point processes

Lect4: Exact Sampling Techniques and MCMC Convergence Analysis

Convex Optimization CMU-10725

Advanced Sampling Algorithms

A = {(x, u) : 0 u f(x)},

Computer Vision Group Prof. Daniel Cremers. 14. Sampling Methods

The Ising model and Markov chain Monte Carlo

Convergence Rate of Markov Chains

A Search and Jump Algorithm for Markov Chain Monte Carlo Sampling. Christopher Jennison. Adriana Ibrahim. Seminar at University of Kuwait

Introduction to Machine Learning CMU-10701

6 Markov Chain Monte Carlo (MCMC)

Computer Vision Group Prof. Daniel Cremers. 11. Sampling Methods: Markov Chain Monte Carlo

Adaptive Monte Carlo Methods for Numerical Integration

Advances and Applications in Perfect Sampling

Monte Carlo Methods for Computation and Optimization (048715)

Bayesian Estimation with Sparse Grids

A note on perfect simulation for exponential random graph models

Markov Chain Monte Carlo Inference. Siamak Ravanbakhsh Winter 2018

Numerical methods for lattice field theory

Coupling. 2/3/2010 and 2/5/2010

Markov Chain Monte Carlo (MCMC)

Computer Vision Group Prof. Daniel Cremers. 10a. Markov Chain Monte Carlo

Quantifying Uncertainty

Computational statistics

Stochastic Simulation

arxiv: v2 [math.pr] 29 Jul 2012

Near-linear time simulation of linear extensions of a height-2 poset with bounded interaction

The Particle Filter. PD Dr. Rudolph Triebel Computer Vision Group. Machine Learning for Computer Vision

Stat 8501 Lecture Notes Spatial Lattice Processes Charles J. Geyer November 16, Introduction

Markov Chain Monte Carlo

Computer Vision Group Prof. Daniel Cremers. 11. Sampling Methods

Bayesian Methods with Monte Carlo Markov Chains II

DAG models and Markov Chain Monte Carlo methods a short overview

Propp-Wilson Algorithm (and sampling the Ising model)

Discrete solid-on-solid models

Lecture 8: PGM Inference

MCMC Sampling for Bayesian Inference using L1-type Priors

INTRODUCTION TO MARKOV CHAIN MONTE CARLO

References. Markov-Chain Monte Carlo. Recall: Sampling Motivation. Problem. Recall: Sampling Methods. CSE586 Computer Vision II

The simple slice sampler is a specialised type of MCMC auxiliary variable method (Swendsen and Wang, 1987; Edwards and Sokal, 1988; Besag and Green, 1

Markov and Gibbs Random Fields

Markov Chain Monte Carlo Methods

An introduction to adaptive MCMC

Numerical Integration Monte Carlo style

Lecture 3: September 10

Exact Goodness-of-Fit Testing for the Ising Model

Metropolis-Hastings Algorithm

Lecture 8: The Metropolis-Hastings Algorithm


Stat 451 Lecture Notes Markov Chain Monte Carlo. Ryan Martin UIC

Markov-Chain Monte Carlo

Machine Learning for Data Science (CS4786) Lecture 24

Bayesian networks: approximate inference

Markov Chain Monte Carlo The Metropolis-Hastings Algorithm

Math 456: Mathematical Modeling. Tuesday, April 9th, 2018

1. INTRODUCTION Propp and Wilson (1996,1998) described a protocol called \coupling from the past" (CFTP) for exact sampling from a distribution using

Lecture 15: MCMC Sanjeev Arora Elad Hazan. COS 402 Machine Learning and Artificial Intelligence Fall 2016

CPSC 540: Machine Learning

Monte Carlo importance sampling and Markov chain

MSc MT15. Further Statistical Methods: MCMC. Lecture 5-6: Markov chains; Metropolis Hastings MCMC. Notes and Practicals available at

April 20th, Advanced Topics in Machine Learning California Institute of Technology. Markov Chain Monte Carlo for Machine Learning

The Monte Carlo Method

Ergodic Theorems. Samy Tindel. Purdue University. Probability Theory 2 - MA 539. Taken from Probability: Theory and examples by R.

Doing Bayesian Integrals

Bayes Nets: Sampling

Markov Processes. Stochastic process. Markov process

The coupling method - Simons Counting Complexity Bootcamp, 2016

MH I. Metropolis-Hastings (MH) algorithm is the most popular method of getting dependent samples from a probability distribution

Probabilistic Graphical Models Lecture 17: Markov chain Monte Carlo

A quick introduction to Markov chains and Markov chain Monte Carlo (revised version)

On probability... Mark Huber (Duke University) Swap moves for spatial point processes Graduate/Faculty Seminar 1 / 42

Sampling Algorithms for Probabilistic Graphical models

CS242: Probabilistic Graphical Models Lecture 7B: Markov Chain Monte Carlo & Gibbs Sampling

Session 3A: Markov chain Monte Carlo (MCMC)

Minicourse on: Markov Chain Monte Carlo: Simulation Techniques in Statistics

Probabilistic Graphical Models Lecture Notes Fall 2009

Intelligent Systems:

Practical unbiased Monte Carlo for Uncertainty Quantification

Winter 2019 Math 106 Topics in Applied Mathematics. Lecture 9: Markov Chain Monte Carlo

Lecture 16: October 29

Undirected Graphical Models: Markov Random Fields

Research Report no. 1 January 2004

Markov chain Monte Carlo

Introduction to Bayesian methods in inverse problems

Zig-Zag Monte Carlo. Delft University of Technology. Joris Bierkens February 7, 2017

A Geometric Interpretation of the Metropolis Hastings Algorithm

Randomized Quasi-Monte Carlo for MCMC

CS 343: Artificial Intelligence

Monte Carlo integration (naive Monte Carlo)

Results: MCMC Dancers, q=10, n=500

Approximate Inference

Lecture 7 and 8: Markov Chain Monte Carlo

Bayesian GLMs and Metropolis-Hastings Algorithm

LECTURE 3. Last time:

Theory of Stochastic Processes 8. Markov chain Monte Carlo

1 Probabilities. 1.1 Basics 1 PROBABILITIES

MONTE CARLO METHODS. Hedibert Freitas Lopes

Transcription:

Perfect simulation for image analysis Mark Huber Fletcher Jones Foundation Associate Professor of Mathematics and Statistics and George R. Roberts Fellow Mathematical Sciences Claremont McKenna College

A prior for pictures An image is a (rectangular) grid of pixels Neighboring pixels often alike A simple approach is the Ising model In Ising model each pixel is either 0 (black) or 1 (white). p(x) exp( β(x(i) x(j)) 2 ) edges {i,j}

Ising model Parameter β called inverse temperature When β = 0, pixels uniform and independent When β =, all pixels same color β = 0

Better pictures use grayscale Instead of {0, 1}, use [0, 1] to allow gray pixels Can generate much more detail

Autonormal model J. Besag. On the statistical analysis of dirty pictures. Journal of the Royal Statistical Society Series B, Vol. 48, No. 3, pp. 259 302, 1986. Besag introduced the autonormal model p(x) exp( β(x(i) x(j)) 2 ) edges {i,j} exp( 0.25β) exp( 0.25β) exp( 0.36β) exp( 0.16β)

Markov chain Monte Carlo easy for this model Suppose look at value of state at a single node Conditioned on neighboring pixel values......distribution of pixel value is a truncated normal

Add a density for observed image y = observed image, x = actual image l(y x) i exp( (y(i) x(i)) 2 /[2κ]) Let N A (µ, σ) be normal dist. conditioned to lie in A. Then: [y(i) x(neighbors of i)] N [0,1] (µ neighbors, κ)

Autonormal Posterior π(y) exp( β(x(i) x(j)) 2 ) edges {i,j} [ ] exp( (y(i) x(i)) 2 /[2κ]) i

The big question How do we generate samples from π?

Two approaches Markov chain theory Create a recurrent Harris chain Then limiting distribution equals stationary distribution Perfect simulation Draws exactly from desired distribution No need to know the mixing time of the chain Often employs recursion

Markov chain theory Pros Easy to build Harris chain where lim dist = stat dist Cons Difficult to show that limit reached quickly Without that, not an algorithm At best, get approximate samples from π

Perfect simulation Pros Get exact draws from π True algorithms Cons Can be difficult to build

Perfect simulation ideas All of these can be applied to this problem Acceptance Rejection [von Neumann 1951] Coupling from the Past [Propp Wilson 1996] Bounding chains [H. 1999] FMMR [Fill et al. 2000] Randomness Recycler [Fill H. 1999] Partially Recursive Acceptance Rejection [H. 2011] Only one provably fast for this problem Monotonic Coupling from the Past

Standard Harris chain rapidly mixing A. Gibbs. Convergence in the Wasserstein metric for Markov chain Monte Carlo algorithms with applications to image restoration. Stochastic Models, Vol. 20, No. 4, pp. 473 492, 2004 Showed that the standard heat bath Markov chain distribution converged quickly (O(n ln n)) using the Wasserstein metric. Wasserstein (earth mover s metric) W p (X t, π) p = inf(e[d(x t, Y ) p ]) (Y π)

Perfect simulation M. Huber. Perfect simulation for image restoration. Stochastic Models, Vol. 23, No. 3, pp. 475 487, 2007 Utilized monotonic Coupling From the Past protocol Generates perfect samples in O(n ln n) time Tricky part is dealing with continuous state space

Coupling from the past (CFTP) Start with an unknown draw from π?

Coupling from the past (CFTP) Start with an unknown draw from π? function that preserves π (like steps in a Markov chain)?

Coupling from the past (CFTP) Start with an unknown draw from π? function that preserves π (like steps in a Markov chain)? If function output independent of input, resulting output is from π

Coupling from the past (CFTP) Start with an unknown draw from π? function that preserves π (like steps in a Markov chain)? If function output independent of input, resulting output is from π Otherwise use CFTP to draw first stat state Use same function to get final stat state

CFTP with more notation Suppose φ : Ω [0, 1] Ω satisfies if X π and U Unif([0, 1]) then φ(x, U) π, Then CFTP runs as follows: CFTP Output: Y 1) Draw U Unif([0, 1]) 2) If #(φ(ω, U)) = 1 3) Y the sole element of Φ(Ω, U) 4) Else 5) Y φ(cftp, U)

Actually running CFTP Key question when implementing CFTP: How do we know when #(φ(ω, U)) = 1? One answer, use monotonic coupling from the past

Monotonic coupling from the past (MCFTP) Suppose that chain is monotonic Means if X t X t then X t+1 X t+1 big X 0 X 0 X 0 X 0 X 0 X 0 X 0 X 0 X 0 X 0 X 0 small X 0

Monotonic coupling from the past (MCFTP) Start chains at biggest and smallest state Unknown stationary state gets squished between them big X 0 X 0?? X 0 X 0 X 0??? X 0 X 0 X 0 X 0? X 0 X 0 small X 0

The other part of CFTP What if the big chain and small chain don t come together? The big idea: Try running big and small chain together If they end up in the same state (couple) that is draw Otherwise, use recursion: Use CFTP to generate first unknown stationary state Run known stat state forward using same random choices Result is our draw

Two general frameworks for creating Markov chains Gibbs sampling [heat bath] (Random scan) choose node uniform at random Draw new value for node from π conditioned on rest of state Metropolis (Random scan) choose node uniform at random Propose new value for that node Accept new value for node with easily calculated probability

Forcing continuous chains to couple Gibbs: Often monotonic, but does not couple Metropolis: Often not monotonic, but does couple Solution: Run several steps of heat bath to get upper and lower process close Take a Metropolis step to get actual coupling

Specifics In Metropolis: propose value X t (i) moves to Y Unif([X t (i) ɛ, X t (i) + ɛ]). Want ɛ small so chance of accepting close to 1 Want ɛ large so coupling easier

How to couple uniforms X t X t

How to couple uniforms Y X t X t Draw uniformly from interval containing all intervals

How to couple uniforms Draw uniformly from interval containing all intervals X t X t If Y works, use as draw from interval Y Y Y

How to couple uniforms Draw uniformly from interval containing all intervals Y X t X t Y Y If Y works, use as draw from interval (Otherwise draw independently of Y

Difficulty when near 0 or 1 Modify move somewhat: Y Unif([max{0, X t (i) ɛ}, min{1, X t (i) + ɛ]}). Same technique applies

Other approaches MCFTP not the only game in town! Other pefect simulation protocols often easier to apply to continuous state spaces Randomness Recycler [H. & Fill 2001] Partially Recursive Acceptance Rejection [H. 2009] MCFTP still only one that provably runs in polynomial time

Summary Perfect simulation for continous state space MCFTP still useful Gibbs brings states close together......metropolis finishes the job Other protocols are out there if needed