Perfect simulation of repulsive point processes

Size: px
Start display at page:

Download "Perfect simulation of repulsive point processes"

Transcription

1 Perfect simulation of repulsive point processes Mark Huber Department of Mathematical Sciences, Claremont McKenna College 29 November, 2011 Mark Huber, CMC Perfect simulation of repulsive point processes 1/56

2 Some repulsive things Spanish towns Pine trees Mark Huber, CMC Perfect simulation of repulsive point processes 2/56

3 Two data sets Locations: Spanish towns Locations: Swedish pines Mark Huber, CMC Perfect simulation of repulsive point processes 3/56

4 Points farther apart than under uniformity Locations: Spanish towns Locations: Uniform placement Mark Huber, CMC Perfect simulation of repulsive point processes 4/56

5 Modeling repulsion Mark Huber, CMC Perfect simulation of repulsive point processes 5/56

6 Modeling repulsion Two common modeling approaches 1 Give story about how process developed (time evolution) 2 Family of densities with respect to Poisson point process (density) Both based on Poisson point process # of points has Poisson distribution Given the # of points, place points uniformly independently Mean of Poisson is λ times size of region Mark Huber, CMC Perfect simulation of repulsive point processes 6/56

7 Today s talk Matérn type III process Time evolution Density Strauss process Density Time evolution Mark Huber, CMC Perfect simulation of repulsive point processes 7/56

8 Strauss process uses density Penalize PPP as follows Distance parameter R Penalty parameter γ (0, 1) Intensity parameter λ For points x, let r(x) = # of pairs less than R apart For points x, density f (x) = γ r(x) λ #x /Z (γ, λ) Parameter effects Big R spaces points farther apart Small γ means fewer points violate R distance Higher λ means more points Mark Huber, CMC Perfect simulation of repulsive point processes 8/56

9 Example of Strauss R =.02 f (x) = γ 6 λ 55 / Z Mark Huber, CMC Perfect simulation of repulsive point processes 9/56

10 Basic perfect simulation for Strauss Acceptance rejection Draw a PPP X with intensity λ Find r(x) Accept X as Strauss draw with probability γ r(x) Otherwise, reject and start over Takes too long unless λ small or γ big Mark Huber, CMC Perfect simulation of repulsive point processes 10/56

11 Matérn Type I Matérn is a storyteller Introduced three ways of explaining the repulsion Type I, II, and III Simulate a Matérn Type I as follows First simulate a Poisson point process (PPP) Remove any point within R distance of another point Mark Huber, CMC Perfect simulation of repulsive point processes 11/56

12 Matérn Type I example a b c Circles of radius R/2 Circles touch = points eliminated Points a, b, c eliminated d e Mark Huber, CMC Perfect simulation of repulsive point processes 12/56

13 Type I: After removal Call a, b, c ghost points a b c Call d, e seen points Ghost points exert invisible pressure d e Mark Huber, CMC Perfect simulation of repulsive point processes 13/56

14 Type I: Comments Problems Too many eliminations As λ, # of points 0 Need method that preserves some points Mark Huber, CMC Perfect simulation of repulsive point processes 14/56

15 Matérn Type II: Older points rule! Simulate a Matérn Type II as follows First simulate a Poisson point process (PPP) Assign each point a birthday in [0, ) Remove any point within R distance of an older point Mark Huber, CMC Perfect simulation of repulsive point processes 15/56

16 Type II: Picture a b c.84 Circles of radius R/2 Point a eliminates b Point b eliminates c d.01 e.52 Mark Huber, CMC Perfect simulation of repulsive point processes 16/56

17 Type II: After thinning Call b, c ghost points a b c Call a, d, e seen points Ghost points exert invisible pressure d e Mark Huber, CMC Perfect simulation of repulsive point processes 17/56

18 Type II: Comments Better, but not perfect First points survive Higher number of points than Type I Why should b take out c if already killed by a? Mark Huber, CMC Perfect simulation of repulsive point processes 18/56

19 Matérn Type III Simulate a Matérn Type III as follows First simulate a Poisson point process (PPP) Assign each point a birthday in [0, ) Run time forward Only allow point birth if not within R of older born point Mark Huber, CMC Perfect simulation of repulsive point processes 19/56

20 Type III: Picture a b c.84 Circles of radius R/2 Point a eliminates b d.01 e.52 Mark Huber, CMC Perfect simulation of repulsive point processes 20/56

21 Type III: After thinning Call c ghost point a b c Call a, c, d, e seen points Ghost points exert invisible pressure d e Mark Huber, CMC Perfect simulation of repulsive point processes 21/56

22 Type III: Comments The good The bad Very natural story Try to add towns/trees If too close to existing town/tree, dies off Density not in closed form nasty high dimensional integral Makes it difficult to do maximum likelihood estimate/posterior Mark Huber, CMC Perfect simulation of repulsive point processes 22/56

23 Using Matérn Type III for inference The plan First turn story of Matérn into density (as with Strauss) Build Markov chain for density Build perfect sampler around Markov chain Build product estimator to utilize samples effectively (Those last two are my area of research) Mark Huber, CMC Perfect simulation of repulsive point processes 23/56

24 Matérn Type III: Time evolution to density Mark Huber, CMC Perfect simulation of repulsive point processes 24/56

25 Casting shadows The seen points cast a shadow across time time 1 R = space Any ghost points must lie in shaded region More shadow = more space for ghost points Mark Huber, CMC Perfect simulation of repulsive point processes 25/56

26 From shadows to density Poisson process in spacetime Looks like PPP conditioned so that no points lie in shadow Parameters θ = (λ, R), region S Let A θ (x, t) be the area of the shadow in spacetime For seen points x with no shadow violations: f seen points (x, t θ) = Cλ #x exp(λa θ (x, t)) Mark Huber, CMC Perfect simulation of repulsive point processes 26/56

27 Larger shadow means more likely Alternate way of drawing Matérn Type III process Fix points in x Draw new PPP y to add to x If all of y lies in shadow of x, accept x as Matérn Type III Otherwise draw new y and repeat Mark Huber, CMC Perfect simulation of repulsive point processes 27/56

28 Acceptance/rejection leads to density What is chance of accepting a given y? Let S be size of region, A θ (x, t) size of shadow Probability no points outside of shadow exp( λ( S A θ (x, t))) = exp(λa θ (x, t)) exp( λ S ) Mark Huber, CMC Perfect simulation of repulsive point processes 28/56

29 The problem of unknown time of birth Big problem We do not see the t values! To get the density for just x integrate out t: g(x θ) = Cλ t [0,1] #x exp(λa θ (x, t)) #x Doing this integral directly extremely nasty Even calculating A θ (x, t) hard Mark Huber, CMC Perfect simulation of repulsive point processes 29/56

30 Monte Carlo methods to the rescue! Given seen points x, want to approximate g(x θ) Monte Carlo approach: treat t values as auxiliary variables Given x: randomly choose t from f (x, t θ, x) Allows use to estimate g(x θ) Mark Huber, CMC Perfect simulation of repulsive point processes 30/56

31 How to draw time stamps given locations? Markov chain Monte Carlo (MCMC) 1 Build chain so stationary distribution = target distribution 2 Under mild conditions (φ-irreducibility, aperiodicity) limiting distribution will equal stationary distribution 3 Run chain for a long time to mix, then get samples Mark Huber, CMC Perfect simulation of repulsive point processes 31/56

32 Metropolis protocol for building Markov chain One step of Metropolis 1 Propose moving from (x, t) to (x, t ) 2 Accept move with probability { f (x, t } θ) min f (x, t θ), 1 3 Otherwise stay where you are Mark Huber, CMC Perfect simulation of repulsive point processes 32/56

33 Even better... Perfect simulation techniques For some Markov chains, possible to do better Can draw exactly from stationary distribution Without worrying about mixing time Perfect sampling protocol Coupling from the past (Propp & Wilson 1994) CFTP converts Markov chains to perfect samplers A good property of Markov chain: monotonicity Monotonicity not necessary for CFTP, is sufficient Mark Huber, CMC Perfect simulation of repulsive point processes 33/56

34 Is Metropolis for Matérn type III monotonic? Propose changing time stamp for one point in x How does the shadow change? time 1 R = space Probability accept move = exp(-area of shadow change) Mark Huber, CMC Perfect simulation of repulsive point processes 34/56

35 How to flip an exp( µ) coin Fun facts about Poisson point processes For X PPP(B), P(#X = 0) = exp( µ(b)) For regions A B, X PPP(B) then X A PPP(A) To check if move in Metropolis Draw PPP(largest possible change in shadow) If PPP restricted to actual change in shadow empty, move Otherwise, stay at current time stamp Using PPP to flip exponential coins: First appears in Beskos, Papaspiliopoulos, Roberts (2006) Perfect simulation of diffusions Mark Huber, CMC Perfect simulation of repulsive point processes 35/56

36 Example time 1 R = space Mark Huber, CMC Perfect simulation of repulsive point processes 36/56

37 Example time 1 R = space Mark Huber, CMC Perfect simulation of repulsive point processes 36/56

38 Example time 1 R = space Mark Huber, CMC Perfect simulation of repulsive point processes 36/56

39 Montonicity of Markov chain step For t 1 t 2 : Run one step of Markov chain for t 1 and t 2 Use same auxiliary PPP in change of shadow Then after step, still have t 1 t 2 Immediately gives us perfect sampling! Mark Huber, CMC Perfect simulation of repulsive point processes 37/56

40 Example of Monotonicity time 1 R = 1.5 t 2 t 2 t 2 0 t 1 t 1 t 1 10 space Mark Huber, CMC Perfect simulation of repulsive point processes 38/56

41 Example of Monotonicity time 1 R = 1.5 t 2 t 2 t 2 0 t 1 t 1 t 1 10 space Mark Huber, CMC Perfect simulation of repulsive point processes 38/56

42 Example of Monotonicity time 1 R = 1.5 t 2 t 2 t 2 0 t 1 t 1 t 1 10 space t 2 accepts move Mark Huber, CMC Perfect simulation of repulsive point processes 38/56

43 Example of Monotonicity time 1 R = 1.5 t 2 t 2 t 2 0 t 1 t 1 t 1 10 space t 2 accepts move t 1 rejects move Mark Huber, CMC Perfect simulation of repulsive point processes 38/56

44 Monotonic CFTP flowchart CFTP(k) Output t 1 YES t 1 all zeros t 2 all ones save RNG seed Run Markov chain for k steps Does t 1 = t 2? RNG = Random Number Generator NO t 1 CFTP(2k) t 2 t 1 reset RNG seed Mark Huber, CMC Perfect simulation of repulsive point processes 39/56

45 Monotonic CFTP details Running CFTP(k) 1 Set t 1 to all 0 s, t 2 to all 1 s 2 Save seed to random number generator 3 Take k steps in the Markov chain 4 If t 1 = t 2, then set T k to be this common value 5 Else 1 Get T 0 by calling CFTP(2k) recursively 2 Reset seed to random number generator to what it was in step 2 3 Get T k by taking k steps in the Markov chain 6 Output T k Mark Huber, CMC Perfect simulation of repulsive point processes 40/56

46 What s the point? Why do we need samples? Ability to sample gives approximation of nasty integral Use TPA or IS+TPA to go from samples to integral Once you have that integral Gives density of data under Matérn model Basis for maximum likelihood...or posterior analysis Mark Huber, CMC Perfect simulation of repulsive point processes 41/56

47 Results: towns Mark Huber, CMC Perfect simulation of repulsive point processes 42/56

48 Results: trees Mark Huber, CMC Perfect simulation of repulsive point processes 43/56

49 Strauss: Density to time evolution Mark Huber, CMC Perfect simulation of repulsive point processes 44/56

50 Preston (1977) Birth-Death Chains Adding time dimension to Poisson point process time Gray bar = node lifespan length bar exp(1) birth birth birth 0 space birth birth birth time between births exp(λµ(s)) Mark Huber, CMC Perfect simulation of repulsive point processes 45/56

51 Preston Birth-Death Chains Points are born, and later die Rate of births is λ times area of region Each point dies at rate 1 (Rate is parameter of exponential random variable) Points alive at time 0 form PPP Mark Huber, CMC Perfect simulation of repulsive point processes 46/56

52 Preston for Strauss Recall Strauss density Penalty γ for pair of points within distance R When point is born... If point within distance R of point already born......point only born with probability γ Mark Huber, CMC Perfect simulation of repulsive point processes 47/56

53 Picture for Strauss γ = 0 time birth birth birth 0 space birth birth birth Mark Huber, CMC Perfect simulation of repulsive point processes 48/56

54 Observations Red points Strauss process subset of earlier points Kendall and J. Møller (2000): can find red points by looking backwards in time Mark Huber, CMC Perfect simulation of repulsive point processes 49/56

55 Adding a Swap move Swap move Broder (1986): swap move for perfect matchings Dyer and Greenhill (2000): swap move for independent sets of graphs Adding Swaps to Preston Huber (2011) When point not born, give chance to swap If only blocked by one point......remove blocker, allow birth Mark Huber, CMC Perfect simulation of repulsive point processes 50/56

56 Picture with swap time birth birth birth 0 space birth birth birth Mark Huber, CMC Perfect simulation of repulsive point processes 51/56

57 Swap move helps chain mix better Easier to find red points Verified experimentally on plane Points affect fewer points in future Need to be blocked by at least two points to affect future So effect of point on later points cut in half Mark Huber, CMC Perfect simulation of repulsive point processes 52/56

58 Running times with swap Strauss model on S = [0, 1] 2, γ = 0.5, R = x 106 Running time for swap and no swap chains 3 Average number of events needed per sample No swapping Always swap when possible β 1 λ Mark Huber, CMC Perfect simulation of repulsive point processes 53/56

59 Running times with swap Time for no swap divided by time for swap 5 No swap times divided by swap times 4.5 Average no swap divided by average swap β 1 λ Mark Huber, CMC Perfect simulation of repulsive point processes 54/56

60 Conclusions For Matérn type III models Built a density (not in closed form) Can approximate density using perfect MCMC methods Allows MLE or posterior analysis For Strauss process Already had a density (but not in closed form) Added new type of Markov chain move Seems to speed up chain in practice Mark Huber, CMC Perfect simulation of repulsive point processes 55/56

61 References HUBER, M.L. (2011). Spatial Birth-Death-Swap Chains. Bernoulli (forthcoming paper, available online) HUBER, M.L. AND R.L. WOLPERT (2009). Likelihood based inference for Matérn type III repulsive point processes. Advances in Applied Probability 41, MATÉRN, B. (1986). Spatial Variation, vol. 36 of Lecture Notes in Statistics. New York, NY: Springer-Verlag, 2nd ed. (first edition published 1960 by Statens Skogsforsningsinstitut, Stockholm). Mark Huber, CMC Perfect simulation of repulsive point processes 56/56

Perfect simulation for repulsive point processes

Perfect simulation for repulsive point processes Perfect simulation for repulsive point processes Why swapping at birth is a good thing Mark Huber Department of Mathematics Claremont-McKenna College 20 May, 2009 Mark Huber (Claremont-McKenna College)

More information

Perfect simulation for image analysis

Perfect simulation for image analysis Perfect simulation for image analysis Mark Huber Fletcher Jones Foundation Associate Professor of Mathematics and Statistics and George R. Roberts Fellow Mathematical Sciences Claremont McKenna College

More information

On probability... Mark Huber (Duke University) Swap moves for spatial point processes Graduate/Faculty Seminar 1 / 42

On probability... Mark Huber (Duke University) Swap moves for spatial point processes Graduate/Faculty Seminar 1 / 42 On probability... I think you re begging the question, said Haydock, and I can see looming ahead one of those terrible exercises in probability where six men have white hats and six men have black hats

More information

A = {(x, u) : 0 u f(x)},

A = {(x, u) : 0 u f(x)}, Draw x uniformly from the region {x : f(x) u }. Markov Chain Monte Carlo Lecture 5 Slice sampler: Suppose that one is interested in sampling from a density f(x), x X. Recall that sampling x f(x) is equivalent

More information

Machine Learning for Data Science (CS4786) Lecture 24

Machine Learning for Data Science (CS4786) Lecture 24 Machine Learning for Data Science (CS4786) Lecture 24 Graphical Models: Approximate Inference Course Webpage : http://www.cs.cornell.edu/courses/cs4786/2016sp/ BELIEF PROPAGATION OR MESSAGE PASSING Each

More information

A Bayesian Approach to Phylogenetics

A Bayesian Approach to Phylogenetics A Bayesian Approach to Phylogenetics Niklas Wahlberg Based largely on slides by Paul Lewis (www.eeb.uconn.edu) An Introduction to Bayesian Phylogenetics Bayesian inference in general Markov chain Monte

More information

Markov Chains and MCMC

Markov Chains and MCMC Markov Chains and MCMC CompSci 590.02 Instructor: AshwinMachanavajjhala Lecture 4 : 590.02 Spring 13 1 Recap: Monte Carlo Method If U is a universe of items, and G is a subset satisfying some property,

More information

Who was Bayes? Bayesian Phylogenetics. What is Bayes Theorem?

Who was Bayes? Bayesian Phylogenetics. What is Bayes Theorem? Who was Bayes? Bayesian Phylogenetics Bret Larget Departments of Botany and of Statistics University of Wisconsin Madison October 6, 2011 The Reverand Thomas Bayes was born in London in 1702. He was the

More information

Bayesian Phylogenetics

Bayesian Phylogenetics Bayesian Phylogenetics Bret Larget Departments of Botany and of Statistics University of Wisconsin Madison October 6, 2011 Bayesian Phylogenetics 1 / 27 Who was Bayes? The Reverand Thomas Bayes was born

More information

Introduction to Machine Learning CMU-10701

Introduction to Machine Learning CMU-10701 Introduction to Machine Learning CMU-10701 Markov Chain Monte Carlo Methods Barnabás Póczos & Aarti Singh Contents Markov Chain Monte Carlo Methods Goal & Motivation Sampling Rejection Importance Markov

More information

Introduction to Computational Biology Lecture # 14: MCMC - Markov Chain Monte Carlo

Introduction to Computational Biology Lecture # 14: MCMC - Markov Chain Monte Carlo Introduction to Computational Biology Lecture # 14: MCMC - Markov Chain Monte Carlo Assaf Weiner Tuesday, March 13, 2007 1 Introduction Today we will return to the motif finding problem, in lecture 10

More information

Likelihood-based Inference for Matérn Type III Repulsive Point Processes

Likelihood-based Inference for Matérn Type III Repulsive Point Processes Likelihood-based Inference for Matérn Type III Repulsive Point Processes Version: April 29, 2014 Mark L. Huber Departments of Mathematics and Statistical Science, Duke University autotomic@gmail.com Robert

More information

MCMC and Gibbs Sampling. Kayhan Batmanghelich

MCMC and Gibbs Sampling. Kayhan Batmanghelich MCMC and Gibbs Sampling Kayhan Batmanghelich 1 Approaches to inference l Exact inference algorithms l l l The elimination algorithm Message-passing algorithm (sum-product, belief propagation) The junction

More information

Stat 516, Homework 1

Stat 516, Homework 1 Stat 516, Homework 1 Due date: October 7 1. Consider an urn with n distinct balls numbered 1,..., n. We sample balls from the urn with replacement. Let N be the number of draws until we encounter a ball

More information

27 : Distributed Monte Carlo Markov Chain. 1 Recap of MCMC and Naive Parallel Gibbs Sampling

27 : Distributed Monte Carlo Markov Chain. 1 Recap of MCMC and Naive Parallel Gibbs Sampling 10-708: Probabilistic Graphical Models 10-708, Spring 2014 27 : Distributed Monte Carlo Markov Chain Lecturer: Eric P. Xing Scribes: Pengtao Xie, Khoa Luu In this scribe, we are going to review the Parallel

More information

Propp-Wilson Algorithm (and sampling the Ising model)

Propp-Wilson Algorithm (and sampling the Ising model) Propp-Wilson Algorithm (and sampling the Ising model) Danny Leshem, Nov 2009 References: Haggstrom, O. (2002) Finite Markov Chains and Algorithmic Applications, ch. 10-11 Propp, J. & Wilson, D. (1996)

More information

Convergence Rate of Markov Chains

Convergence Rate of Markov Chains Convergence Rate of Markov Chains Will Perkins April 16, 2013 Convergence Last class we saw that if X n is an irreducible, aperiodic, positive recurrent Markov chain, then there exists a stationary distribution

More information

Lecture 8: Bayesian Networks

Lecture 8: Bayesian Networks Lecture 8: Bayesian Networks Bayesian Networks Inference in Bayesian Networks COMP-652 and ECSE 608, Lecture 8 - January 31, 2017 1 Bayes nets P(E) E=1 E=0 0.005 0.995 E B P(B) B=1 B=0 0.01 0.99 E=0 E=1

More information

MCMC notes by Mark Holder

MCMC notes by Mark Holder MCMC notes by Mark Holder Bayesian inference Ultimately, we want to make probability statements about true values of parameters, given our data. For example P(α 0 < α 1 X). According to Bayes theorem:

More information

7.1 Coupling from the Past

7.1 Coupling from the Past Georgia Tech Fall 2006 Markov Chain Monte Carlo Methods Lecture 7: September 12, 2006 Coupling from the Past Eric Vigoda 7.1 Coupling from the Past 7.1.1 Introduction We saw in the last lecture how Markov

More information

Theory of Stochastic Processes 8. Markov chain Monte Carlo

Theory of Stochastic Processes 8. Markov chain Monte Carlo Theory of Stochastic Processes 8. Markov chain Monte Carlo Tomonari Sei sei@mist.i.u-tokyo.ac.jp Department of Mathematical Informatics, University of Tokyo June 8, 2017 http://www.stat.t.u-tokyo.ac.jp/~sei/lec.html

More information

Advances and Applications in Perfect Sampling

Advances and Applications in Perfect Sampling and Applications in Perfect Sampling Ph.D. Dissertation Defense Ulrike Schneider advisor: Jem Corcoran May 8, 2003 Department of Applied Mathematics University of Colorado Outline Introduction (1) MCMC

More information

Bayesian model selection in graphs by using BDgraph package

Bayesian model selection in graphs by using BDgraph package Bayesian model selection in graphs by using BDgraph package A. Mohammadi and E. Wit March 26, 2013 MOTIVATION Flow cytometry data with 11 proteins from Sachs et al. (2005) RESULT FOR CELL SIGNALING DATA

More information

LIKELIHOOD-BASED INFERENCE FOR MATÉRN TYPE III REPULSIVE POINT PROCESSES

LIKELIHOOD-BASED INFERENCE FOR MATÉRN TYPE III REPULSIVE POINT PROCESSES Applied Probability Trust (7 April 2009) LIKELIHOOD-BASED INFERENCE FOR MATÉRN TYPE III REPULSIVE POINT PROCESSES MARK L. HUBER, Duke University ROBERT L. WOLPERT, Duke University Abstract In a repulsive

More information

Lect4: Exact Sampling Techniques and MCMC Convergence Analysis

Lect4: Exact Sampling Techniques and MCMC Convergence Analysis Lect4: Exact Sampling Techniques and MCMC Convergence Analysis. Exact sampling. Convergence analysis of MCMC. First-hit time analysis for MCMC--ways to analyze the proposals. Outline of the Module Definitions

More information

Minicourse on: Markov Chain Monte Carlo: Simulation Techniques in Statistics

Minicourse on: Markov Chain Monte Carlo: Simulation Techniques in Statistics Minicourse on: Markov Chain Monte Carlo: Simulation Techniques in Statistics Eric Slud, Statistics Program Lecture 1: Metropolis-Hastings Algorithm, plus background in Simulation and Markov Chains. Lecture

More information

CSC 2541: Bayesian Methods for Machine Learning

CSC 2541: Bayesian Methods for Machine Learning CSC 2541: Bayesian Methods for Machine Learning Radford M. Neal, University of Toronto, 2011 Lecture 3 More Markov Chain Monte Carlo Methods The Metropolis algorithm isn t the only way to do MCMC. We ll

More information

Markov Chain Monte Carlo The Metropolis-Hastings Algorithm

Markov Chain Monte Carlo The Metropolis-Hastings Algorithm Markov Chain Monte Carlo The Metropolis-Hastings Algorithm Anthony Trubiano April 11th, 2018 1 Introduction Markov Chain Monte Carlo (MCMC) methods are a class of algorithms for sampling from a probability

More information

April 20th, Advanced Topics in Machine Learning California Institute of Technology. Markov Chain Monte Carlo for Machine Learning

April 20th, Advanced Topics in Machine Learning California Institute of Technology. Markov Chain Monte Carlo for Machine Learning for for Advanced Topics in California Institute of Technology April 20th, 2017 1 / 50 Table of Contents for 1 2 3 4 2 / 50 History of methods for Enrico Fermi used to calculate incredibly accurate predictions

More information

Adaptive Monte Carlo Methods for Numerical Integration

Adaptive Monte Carlo Methods for Numerical Integration Adaptive Monte Carlo Methods for Numerical Integration Mark Huber 1 and Sarah Schott 2 1 Department of Mathematical Sciences, Claremont McKenna College 2 Department of Mathematics, Duke University 8 March,

More information

Monte Carlo Methods. Leon Gu CSD, CMU

Monte Carlo Methods. Leon Gu CSD, CMU Monte Carlo Methods Leon Gu CSD, CMU Approximate Inference EM: y-observed variables; x-hidden variables; θ-parameters; E-step: q(x) = p(x y, θ t 1 ) M-step: θ t = arg max E q(x) [log p(y, x θ)] θ Monte

More information

Convex Optimization CMU-10725

Convex Optimization CMU-10725 Convex Optimization CMU-10725 Simulated Annealing Barnabás Póczos & Ryan Tibshirani Andrey Markov Markov Chains 2 Markov Chains Markov chain: Homogen Markov chain: 3 Markov Chains Assume that the state

More information

Computational statistics

Computational statistics Computational statistics Markov Chain Monte Carlo methods Thierry Denœux March 2017 Thierry Denœux Computational statistics March 2017 1 / 71 Contents of this chapter When a target density f can be evaluated

More information

MCMC: Markov Chain Monte Carlo

MCMC: Markov Chain Monte Carlo I529: Machine Learning in Bioinformatics (Spring 2013) MCMC: Markov Chain Monte Carlo Yuzhen Ye School of Informatics and Computing Indiana University, Bloomington Spring 2013 Contents Review of Markov

More information

Advanced Sampling Algorithms

Advanced Sampling Algorithms + Advanced Sampling Algorithms + Mobashir Mohammad Hirak Sarkar Parvathy Sudhir Yamilet Serrano Llerena Advanced Sampling Algorithms Aditya Kulkarni Tobias Bertelsen Nirandika Wanigasekara Malay Singh

More information

Data Analysis I. Dr Martin Hendry, Dept of Physics and Astronomy University of Glasgow, UK. 10 lectures, beginning October 2006

Data Analysis I. Dr Martin Hendry, Dept of Physics and Astronomy University of Glasgow, UK. 10 lectures, beginning October 2006 Astronomical p( y x, I) p( x, I) p ( x y, I) = p( y, I) Data Analysis I Dr Martin Hendry, Dept of Physics and Astronomy University of Glasgow, UK 10 lectures, beginning October 2006 4. Monte Carlo Methods

More information

Markov Networks.

Markov Networks. Markov Networks www.biostat.wisc.edu/~dpage/cs760/ Goals for the lecture you should understand the following concepts Markov network syntax Markov network semantics Potential functions Partition function

More information

CS 343: Artificial Intelligence

CS 343: Artificial Intelligence CS 343: Artificial Intelligence Bayes Nets: Sampling Prof. Scott Niekum The University of Texas at Austin [These slides based on those of Dan Klein and Pieter Abbeel for CS188 Intro to AI at UC Berkeley.

More information

17 : Markov Chain Monte Carlo

17 : Markov Chain Monte Carlo 10-708: Probabilistic Graphical Models, Spring 2015 17 : Markov Chain Monte Carlo Lecturer: Eric P. Xing Scribes: Heran Lin, Bin Deng, Yun Huang 1 Review of Monte Carlo Methods 1.1 Overview Monte Carlo

More information

Markov Chain Monte Carlo

Markov Chain Monte Carlo Markov Chain Monte Carlo Recall: To compute the expectation E ( h(y ) ) we use the approximation E(h(Y )) 1 n n h(y ) t=1 with Y (1),..., Y (n) h(y). Thus our aim is to sample Y (1),..., Y (n) from f(y).

More information

Markov chain Monte Carlo Lecture 9

Markov chain Monte Carlo Lecture 9 Markov chain Monte Carlo Lecture 9 David Sontag New York University Slides adapted from Eric Xing and Qirong Ho (CMU) Limitations of Monte Carlo Direct (unconditional) sampling Hard to get rare events

More information

The Particle Filter. PD Dr. Rudolph Triebel Computer Vision Group. Machine Learning for Computer Vision

The Particle Filter. PD Dr. Rudolph Triebel Computer Vision Group. Machine Learning for Computer Vision The Particle Filter Non-parametric implementation of Bayes filter Represents the belief (posterior) random state samples. by a set of This representation is approximate. Can represent distributions that

More information

Model Counting for Logical Theories

Model Counting for Logical Theories Model Counting for Logical Theories Wednesday Dmitry Chistikov Rayna Dimitrova Department of Computer Science University of Oxford, UK Max Planck Institute for Software Systems (MPI-SWS) Kaiserslautern

More information

Markov chain Monte Carlo

Markov chain Monte Carlo Markov chain Monte Carlo Peter Beerli October 10, 2005 [this chapter is highly influenced by chapter 1 in Markov chain Monte Carlo in Practice, eds Gilks W. R. et al. Chapman and Hall/CRC, 1996] 1 Short

More information

Markov-Chain Monte Carlo

Markov-Chain Monte Carlo Markov-Chain Monte Carlo CSE586 Computer Vision II Spring 2010, Penn State Univ. References Recall: Sampling Motivation If we can generate random samples x i from a given distribution P(x), then we can

More information

Computer Vision Group Prof. Daniel Cremers. 11. Sampling Methods

Computer Vision Group Prof. Daniel Cremers. 11. Sampling Methods Prof. Daniel Cremers 11. Sampling Methods Sampling Methods Sampling Methods are widely used in Computer Science as an approximation of a deterministic algorithm to represent uncertainty without a parametric

More information

Stochastic optimization Markov Chain Monte Carlo

Stochastic optimization Markov Chain Monte Carlo Stochastic optimization Markov Chain Monte Carlo Ethan Fetaya Weizmann Institute of Science 1 Motivation Markov chains Stationary distribution Mixing time 2 Algorithms Metropolis-Hastings Simulated Annealing

More information

Lecture 21: Counting and Sampling Problems

Lecture 21: Counting and Sampling Problems princeton univ. F 14 cos 521: Advanced Algorithm Design Lecture 21: Counting and Sampling Problems Lecturer: Sanjeev Arora Scribe: Today s topic of counting and sampling problems is motivated by computational

More information

Lecture 16: October 29

Lecture 16: October 29 CS294 Markov Chain Monte Carlo: Foundations & Applications Fall 2009 Lecture 16: October 29 Lecturer: Alistair Sinclair Scribes: Disclaimer: These notes have not been subjected to the usual scrutiny reserved

More information

Approximate inference in Energy-Based Models

Approximate inference in Energy-Based Models CSC 2535: 2013 Lecture 3b Approximate inference in Energy-Based Models Geoffrey Hinton Two types of density model Stochastic generative model using directed acyclic graph (e.g. Bayes Net) Energy-based

More information

Approximate Inference

Approximate Inference Approximate Inference Simulation has a name: sampling Sampling is a hot topic in machine learning, and it s really simple Basic idea: Draw N samples from a sampling distribution S Compute an approximate

More information

Forward Problems and their Inverse Solutions

Forward Problems and their Inverse Solutions Forward Problems and their Inverse Solutions Sarah Zedler 1,2 1 King Abdullah University of Science and Technology 2 University of Texas at Austin February, 2013 Outline 1 Forward Problem Example Weather

More information

Bayesian GLMs and Metropolis-Hastings Algorithm

Bayesian GLMs and Metropolis-Hastings Algorithm Bayesian GLMs and Metropolis-Hastings Algorithm We have seen that with conjugate or semi-conjugate prior distributions the Gibbs sampler can be used to sample from the posterior distribution. In situations,

More information

STA 4273H: Statistical Machine Learning

STA 4273H: Statistical Machine Learning STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Computer Science! Department of Statistical Sciences! rsalakhu@cs.toronto.edu! h0p://www.cs.utoronto.ca/~rsalakhu/ Lecture 7 Approximate

More information

SC7/SM6 Bayes Methods HT18 Lecturer: Geoff Nicholls Lecture 2: Monte Carlo Methods Notes and Problem sheets are available at http://www.stats.ox.ac.uk/~nicholls/bayesmethods/ and via the MSc weblearn pages.

More information

Computer Intensive Methods in Mathematical Statistics

Computer Intensive Methods in Mathematical Statistics Computer Intensive Methods in Mathematical Statistics Department of mathematics johawes@kth.se Lecture 16 Advanced topics in computational statistics 18 May 2017 Computer Intensive Methods (1) Plan of

More information

Bayesian Inference and MCMC

Bayesian Inference and MCMC Bayesian Inference and MCMC Aryan Arbabi Partly based on MCMC slides from CSC412 Fall 2018 1 / 18 Bayesian Inference - Motivation Consider we have a data set D = {x 1,..., x n }. E.g each x i can be the

More information

Review. DS GA 1002 Statistical and Mathematical Models. Carlos Fernandez-Granda

Review. DS GA 1002 Statistical and Mathematical Models.   Carlos Fernandez-Granda Review DS GA 1002 Statistical and Mathematical Models http://www.cims.nyu.edu/~cfgranda/pages/dsga1002_fall16 Carlos Fernandez-Granda Probability and statistics Probability: Framework for dealing with

More information

Markov Chain Monte Carlo (MCMC)

Markov Chain Monte Carlo (MCMC) Markov Chain Monte Carlo (MCMC Dependent Sampling Suppose we wish to sample from a density π, and we can evaluate π as a function but have no means to directly generate a sample. Rejection sampling can

More information

Artificial Intelligence

Artificial Intelligence ICS461 Fall 2010 Nancy E. Reed nreed@hawaii.edu 1 Lecture #14B Outline Inference in Bayesian Networks Exact inference by enumeration Exact inference by variable elimination Approximate inference by stochastic

More information

Time-Sensitive Dirichlet Process Mixture Models

Time-Sensitive Dirichlet Process Mixture Models Time-Sensitive Dirichlet Process Mixture Models Xiaojin Zhu Zoubin Ghahramani John Lafferty May 25 CMU-CALD-5-4 School of Computer Science Carnegie Mellon University Pittsburgh, PA 523 Abstract We introduce

More information

Announcements. Inference. Mid-term. Inference by Enumeration. Reminder: Alarm Network. Introduction to Artificial Intelligence. V22.

Announcements. Inference. Mid-term. Inference by Enumeration. Reminder: Alarm Network. Introduction to Artificial Intelligence. V22. Introduction to Artificial Intelligence V22.0472-001 Fall 2009 Lecture 15: Bayes Nets 3 Midterms graded Assignment 2 graded Announcements Rob Fergus Dept of Computer Science, Courant Institute, NYU Slides

More information

Sampling Rejection Sampling Importance Sampling Markov Chain Monte Carlo. Sampling Methods. Oliver Schulte - CMPT 419/726. Bishop PRML Ch.

Sampling Rejection Sampling Importance Sampling Markov Chain Monte Carlo. Sampling Methods. Oliver Schulte - CMPT 419/726. Bishop PRML Ch. Sampling Methods Oliver Schulte - CMP 419/726 Bishop PRML Ch. 11 Recall Inference or General Graphs Junction tree algorithm is an exact inference method for arbitrary graphs A particular tree structure

More information

Computer Vision Group Prof. Daniel Cremers. 14. Sampling Methods

Computer Vision Group Prof. Daniel Cremers. 14. Sampling Methods Prof. Daniel Cremers 14. Sampling Methods Sampling Methods Sampling Methods are widely used in Computer Science as an approximation of a deterministic algorithm to represent uncertainty without a parametric

More information

A quick introduction to Markov chains and Markov chain Monte Carlo (revised version)

A quick introduction to Markov chains and Markov chain Monte Carlo (revised version) A quick introduction to Markov chains and Markov chain Monte Carlo (revised version) Rasmus Waagepetersen Institute of Mathematical Sciences Aalborg University 1 Introduction These notes are intended to

More information

Markov Chain Monte Carlo Inference. Siamak Ravanbakhsh Winter 2018

Markov Chain Monte Carlo Inference. Siamak Ravanbakhsh Winter 2018 Graphical Models Markov Chain Monte Carlo Inference Siamak Ravanbakhsh Winter 2018 Learning objectives Markov chains the idea behind Markov Chain Monte Carlo (MCMC) two important examples: Gibbs sampling

More information

Parameter Estimation. William H. Jefferys University of Texas at Austin Parameter Estimation 7/26/05 1

Parameter Estimation. William H. Jefferys University of Texas at Austin Parameter Estimation 7/26/05 1 Parameter Estimation William H. Jefferys University of Texas at Austin bill@bayesrules.net Parameter Estimation 7/26/05 1 Elements of Inference Inference problems contain two indispensable elements: Data

More information

The Ising model and Markov chain Monte Carlo

The Ising model and Markov chain Monte Carlo The Ising model and Markov chain Monte Carlo Ramesh Sridharan These notes give a short description of the Ising model for images and an introduction to Metropolis-Hastings and Gibbs Markov Chain Monte

More information

Machine Learning. Probabilistic KNN.

Machine Learning. Probabilistic KNN. Machine Learning. Mark Girolami girolami@dcs.gla.ac.uk Department of Computing Science University of Glasgow June 21, 2007 p. 1/3 KNN is a remarkably simple algorithm with proven error-rates June 21, 2007

More information

Reminder of some Markov Chain properties:

Reminder of some Markov Chain properties: Reminder of some Markov Chain properties: 1. a transition from one state to another occurs probabilistically 2. only state that matters is where you currently are (i.e. given present, future is independent

More information

An ABC interpretation of the multiple auxiliary variable method

An ABC interpretation of the multiple auxiliary variable method School of Mathematical and Physical Sciences Department of Mathematics and Statistics Preprint MPS-2016-07 27 April 2016 An ABC interpretation of the multiple auxiliary variable method by Dennis Prangle

More information

Bayes Nets: Sampling

Bayes Nets: Sampling Bayes Nets: Sampling [These slides were created by Dan Klein and Pieter Abbeel for CS188 Intro to AI at UC Berkeley. All CS188 materials are available at http://ai.berkeley.edu.] Approximate Inference:

More information

Informatics 2D Reasoning and Agents Semester 2,

Informatics 2D Reasoning and Agents Semester 2, Informatics 2D Reasoning and Agents Semester 2, 2018 2019 Alex Lascarides alex@inf.ed.ac.uk Lecture 25 Approximate Inference in Bayesian Networks 19th March 2019 Informatics UoE Informatics 2D 1 Where

More information

Bayesian Methods for Machine Learning

Bayesian Methods for Machine Learning Bayesian Methods for Machine Learning CS 584: Big Data Analytics Material adapted from Radford Neal s tutorial (http://ftp.cs.utoronto.ca/pub/radford/bayes-tut.pdf), Zoubin Ghahramni (http://hunch.net/~coms-4771/zoubin_ghahramani_bayesian_learning.pdf),

More information

Computer Vision Group Prof. Daniel Cremers. 10a. Markov Chain Monte Carlo

Computer Vision Group Prof. Daniel Cremers. 10a. Markov Chain Monte Carlo Group Prof. Daniel Cremers 10a. Markov Chain Monte Carlo Markov Chain Monte Carlo In high-dimensional spaces, rejection sampling and importance sampling are very inefficient An alternative is Markov Chain

More information

MCMC Methods: Gibbs and Metropolis

MCMC Methods: Gibbs and Metropolis MCMC Methods: Gibbs and Metropolis Patrick Breheny February 28 Patrick Breheny BST 701: Bayesian Modeling in Biostatistics 1/30 Introduction As we have seen, the ability to sample from the posterior distribution

More information

Bayes Networks. CS540 Bryan R Gibson University of Wisconsin-Madison. Slides adapted from those used by Prof. Jerry Zhu, CS540-1

Bayes Networks. CS540 Bryan R Gibson University of Wisconsin-Madison. Slides adapted from those used by Prof. Jerry Zhu, CS540-1 Bayes Networks CS540 Bryan R Gibson University of Wisconsin-Madison Slides adapted from those used by Prof. Jerry Zhu, CS540-1 1 / 59 Outline Joint Probability: great for inference, terrible to obtain

More information

Markov Chain Monte Carlo Lecture 6

Markov Chain Monte Carlo Lecture 6 Sequential parallel tempering With the development of science and technology, we more and more need to deal with high dimensional systems. For example, we need to align a group of protein or DNA sequences

More information

Announcements. CS 188: Artificial Intelligence Fall Causality? Example: Traffic. Topology Limits Distributions. Example: Reverse Traffic

Announcements. CS 188: Artificial Intelligence Fall Causality? Example: Traffic. Topology Limits Distributions. Example: Reverse Traffic CS 188: Artificial Intelligence Fall 2008 Lecture 16: Bayes Nets III 10/23/2008 Announcements Midterms graded, up on glookup, back Tuesday W4 also graded, back in sections / box Past homeworks in return

More information

CPSC 540: Machine Learning

CPSC 540: Machine Learning CPSC 540: Machine Learning MCMC and Non-Parametric Bayes Mark Schmidt University of British Columbia Winter 2016 Admin I went through project proposals: Some of you got a message on Piazza. No news is

More information

Introduction to Machine Learning CMU-10701

Introduction to Machine Learning CMU-10701 Introduction to Machine Learning CMU-10701 Markov Chain Monte Carlo Methods Barnabás Póczos Contents Markov Chain Monte Carlo Methods Sampling Rejection Importance Hastings-Metropolis Gibbs Markov Chains

More information

28 : Approximate Inference - Distributed MCMC

28 : Approximate Inference - Distributed MCMC 10-708: Probabilistic Graphical Models, Spring 2015 28 : Approximate Inference - Distributed MCMC Lecturer: Avinava Dubey Scribes: Hakim Sidahmed, Aman Gupta 1 Introduction For many interesting problems,

More information

STA 4273H: Statistical Machine Learning

STA 4273H: Statistical Machine Learning STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.utstat.utoronto.ca/~rsalakhu/ Sidney Smith Hall, Room 6002 Lecture 3 Linear

More information

Simulation of point process using MCMC

Simulation of point process using MCMC Simulation of point process using MCMC Jakob G. Rasmussen Department of Mathematics Aalborg University Denmark March 2, 2011 1/14 Today When do we need MCMC? Case 1: Fixed number of points Metropolis-Hastings,

More information

arxiv: v2 [stat.me] 3 Apr 2015

arxiv: v2 [stat.me] 3 Apr 2015 Bayesian inference for Matérn repulsive processes arxiv:1308.1136v2 [stat.me] 3 Apr 2015 Vinayak Rao Department of Statistics, Purdue University, USA Ryan P. Adams School of Engineering and Applied Sciences,

More information

Math 456: Mathematical Modeling. Tuesday, April 9th, 2018

Math 456: Mathematical Modeling. Tuesday, April 9th, 2018 Math 456: Mathematical Modeling Tuesday, April 9th, 2018 The Ergodic theorem Tuesday, April 9th, 2018 Today 1. Asymptotic frequency (or: How to use the stationary distribution to estimate the average amount

More information

Today. Statistical Learning. Coin Flip. Coin Flip. Experiment 1: Heads. Experiment 1: Heads. Which coin will I use? Which coin will I use?

Today. Statistical Learning. Coin Flip. Coin Flip. Experiment 1: Heads. Experiment 1: Heads. Which coin will I use? Which coin will I use? Today Statistical Learning Parameter Estimation: Maximum Likelihood (ML) Maximum A Posteriori (MAP) Bayesian Continuous case Learning Parameters for a Bayesian Network Naive Bayes Maximum Likelihood estimates

More information

Bayesian Graphical Models

Bayesian Graphical Models Graphical Models and Inference, Lecture 16, Michaelmas Term 2009 December 4, 2009 Parameter θ, data X = x, likelihood L(θ x) p(x θ). Express knowledge about θ through prior distribution π on θ. Inference

More information

PMR Learning as Inference

PMR Learning as Inference Outline PMR Learning as Inference Probabilistic Modelling and Reasoning Amos Storkey Modelling 2 The Exponential Family 3 Bayesian Sets School of Informatics, University of Edinburgh Amos Storkey PMR Learning

More information

Stochastic Simulation

Stochastic Simulation Stochastic Simulation Idea: probabilities samples Get probabilities from samples: X count x 1 n 1. x k total. n k m X probability x 1. n 1 /m. x k n k /m If we could sample from a variable s (posterior)

More information

A note on perfect simulation for exponential random graph models

A note on perfect simulation for exponential random graph models A note on perfect simulation for exponential random graph models A. Cerqueira, A. Garivier and F. Leonardi October 4, 017 arxiv:1710.00873v1 [stat.co] Oct 017 Abstract In this paper we propose a perfect

More information

SAMPLING ALGORITHMS. In general. Inference in Bayesian models

SAMPLING ALGORITHMS. In general. Inference in Bayesian models SAMPLING ALGORITHMS SAMPLING ALGORITHMS In general A sampling algorithm is an algorithm that outputs samples x 1, x 2,... from a given distribution P or density p. Sampling algorithms can for example be

More information

References. Markov-Chain Monte Carlo. Recall: Sampling Motivation. Problem. Recall: Sampling Methods. CSE586 Computer Vision II

References. Markov-Chain Monte Carlo. Recall: Sampling Motivation. Problem. Recall: Sampling Methods. CSE586 Computer Vision II References Markov-Chain Monte Carlo CSE586 Computer Vision II Spring 2010, Penn State Univ. Recall: Sampling Motivation If we can generate random samples x i from a given distribution P(x), then we can

More information

Random Walks A&T and F&S 3.1.2

Random Walks A&T and F&S 3.1.2 Random Walks A&T 110-123 and F&S 3.1.2 As we explained last time, it is very difficult to sample directly a general probability distribution. - If we sample from another distribution, the overlap will

More information

Graphical Models. Lecture 15: Approximate Inference by Sampling. Andrew McCallum

Graphical Models. Lecture 15: Approximate Inference by Sampling. Andrew McCallum Graphical Models Lecture 15: Approximate Inference by Sampling Andrew McCallum mccallum@cs.umass.edu Thanks to Noah Smith and Carlos Guestrin for some slide materials. 1 General Idea Set of random variables

More information

1 Probabilities. 1.1 Basics 1 PROBABILITIES

1 Probabilities. 1.1 Basics 1 PROBABILITIES 1 PROBABILITIES 1 Probabilities Probability is a tricky word usually meaning the likelyhood of something occuring or how frequent something is. Obviously, if something happens frequently, then its probability

More information

CPSC 540: Machine Learning

CPSC 540: Machine Learning CPSC 540: Machine Learning More Approximate Inference Mark Schmidt University of British Columbia Winter 2018 Last Time: Approximate Inference We ve been discussing graphical models for density estimation,

More information

Lecture 6: Markov Chain Monte Carlo

Lecture 6: Markov Chain Monte Carlo Lecture 6: Markov Chain Monte Carlo D. Jason Koskinen koskinen@nbi.ku.dk Photo by Howard Jackman University of Copenhagen Advanced Methods in Applied Statistics Feb - Apr 2016 Niels Bohr Institute 2 Outline

More information

Lecture 15: MCMC Sanjeev Arora Elad Hazan. COS 402 Machine Learning and Artificial Intelligence Fall 2016

Lecture 15: MCMC Sanjeev Arora Elad Hazan. COS 402 Machine Learning and Artificial Intelligence Fall 2016 Lecture 15: MCMC Sanjeev Arora Elad Hazan COS 402 Machine Learning and Artificial Intelligence Fall 2016 Course progress Learning from examples Definition + fundamental theorem of statistical learning,

More information

ELEC633: Graphical Models

ELEC633: Graphical Models ELEC633: Graphical Models Tahira isa Saleem Scribe from 7 October 2008 References: Casella and George Exploring the Gibbs sampler (1992) Chib and Greenberg Understanding the Metropolis-Hastings algorithm

More information