Perfect simulation of repulsive point processes
|
|
- Martha Flynn
- 5 years ago
- Views:
Transcription
1 Perfect simulation of repulsive point processes Mark Huber Department of Mathematical Sciences, Claremont McKenna College 29 November, 2011 Mark Huber, CMC Perfect simulation of repulsive point processes 1/56
2 Some repulsive things Spanish towns Pine trees Mark Huber, CMC Perfect simulation of repulsive point processes 2/56
3 Two data sets Locations: Spanish towns Locations: Swedish pines Mark Huber, CMC Perfect simulation of repulsive point processes 3/56
4 Points farther apart than under uniformity Locations: Spanish towns Locations: Uniform placement Mark Huber, CMC Perfect simulation of repulsive point processes 4/56
5 Modeling repulsion Mark Huber, CMC Perfect simulation of repulsive point processes 5/56
6 Modeling repulsion Two common modeling approaches 1 Give story about how process developed (time evolution) 2 Family of densities with respect to Poisson point process (density) Both based on Poisson point process # of points has Poisson distribution Given the # of points, place points uniformly independently Mean of Poisson is λ times size of region Mark Huber, CMC Perfect simulation of repulsive point processes 6/56
7 Today s talk Matérn type III process Time evolution Density Strauss process Density Time evolution Mark Huber, CMC Perfect simulation of repulsive point processes 7/56
8 Strauss process uses density Penalize PPP as follows Distance parameter R Penalty parameter γ (0, 1) Intensity parameter λ For points x, let r(x) = # of pairs less than R apart For points x, density f (x) = γ r(x) λ #x /Z (γ, λ) Parameter effects Big R spaces points farther apart Small γ means fewer points violate R distance Higher λ means more points Mark Huber, CMC Perfect simulation of repulsive point processes 8/56
9 Example of Strauss R =.02 f (x) = γ 6 λ 55 / Z Mark Huber, CMC Perfect simulation of repulsive point processes 9/56
10 Basic perfect simulation for Strauss Acceptance rejection Draw a PPP X with intensity λ Find r(x) Accept X as Strauss draw with probability γ r(x) Otherwise, reject and start over Takes too long unless λ small or γ big Mark Huber, CMC Perfect simulation of repulsive point processes 10/56
11 Matérn Type I Matérn is a storyteller Introduced three ways of explaining the repulsion Type I, II, and III Simulate a Matérn Type I as follows First simulate a Poisson point process (PPP) Remove any point within R distance of another point Mark Huber, CMC Perfect simulation of repulsive point processes 11/56
12 Matérn Type I example a b c Circles of radius R/2 Circles touch = points eliminated Points a, b, c eliminated d e Mark Huber, CMC Perfect simulation of repulsive point processes 12/56
13 Type I: After removal Call a, b, c ghost points a b c Call d, e seen points Ghost points exert invisible pressure d e Mark Huber, CMC Perfect simulation of repulsive point processes 13/56
14 Type I: Comments Problems Too many eliminations As λ, # of points 0 Need method that preserves some points Mark Huber, CMC Perfect simulation of repulsive point processes 14/56
15 Matérn Type II: Older points rule! Simulate a Matérn Type II as follows First simulate a Poisson point process (PPP) Assign each point a birthday in [0, ) Remove any point within R distance of an older point Mark Huber, CMC Perfect simulation of repulsive point processes 15/56
16 Type II: Picture a b c.84 Circles of radius R/2 Point a eliminates b Point b eliminates c d.01 e.52 Mark Huber, CMC Perfect simulation of repulsive point processes 16/56
17 Type II: After thinning Call b, c ghost points a b c Call a, d, e seen points Ghost points exert invisible pressure d e Mark Huber, CMC Perfect simulation of repulsive point processes 17/56
18 Type II: Comments Better, but not perfect First points survive Higher number of points than Type I Why should b take out c if already killed by a? Mark Huber, CMC Perfect simulation of repulsive point processes 18/56
19 Matérn Type III Simulate a Matérn Type III as follows First simulate a Poisson point process (PPP) Assign each point a birthday in [0, ) Run time forward Only allow point birth if not within R of older born point Mark Huber, CMC Perfect simulation of repulsive point processes 19/56
20 Type III: Picture a b c.84 Circles of radius R/2 Point a eliminates b d.01 e.52 Mark Huber, CMC Perfect simulation of repulsive point processes 20/56
21 Type III: After thinning Call c ghost point a b c Call a, c, d, e seen points Ghost points exert invisible pressure d e Mark Huber, CMC Perfect simulation of repulsive point processes 21/56
22 Type III: Comments The good The bad Very natural story Try to add towns/trees If too close to existing town/tree, dies off Density not in closed form nasty high dimensional integral Makes it difficult to do maximum likelihood estimate/posterior Mark Huber, CMC Perfect simulation of repulsive point processes 22/56
23 Using Matérn Type III for inference The plan First turn story of Matérn into density (as with Strauss) Build Markov chain for density Build perfect sampler around Markov chain Build product estimator to utilize samples effectively (Those last two are my area of research) Mark Huber, CMC Perfect simulation of repulsive point processes 23/56
24 Matérn Type III: Time evolution to density Mark Huber, CMC Perfect simulation of repulsive point processes 24/56
25 Casting shadows The seen points cast a shadow across time time 1 R = space Any ghost points must lie in shaded region More shadow = more space for ghost points Mark Huber, CMC Perfect simulation of repulsive point processes 25/56
26 From shadows to density Poisson process in spacetime Looks like PPP conditioned so that no points lie in shadow Parameters θ = (λ, R), region S Let A θ (x, t) be the area of the shadow in spacetime For seen points x with no shadow violations: f seen points (x, t θ) = Cλ #x exp(λa θ (x, t)) Mark Huber, CMC Perfect simulation of repulsive point processes 26/56
27 Larger shadow means more likely Alternate way of drawing Matérn Type III process Fix points in x Draw new PPP y to add to x If all of y lies in shadow of x, accept x as Matérn Type III Otherwise draw new y and repeat Mark Huber, CMC Perfect simulation of repulsive point processes 27/56
28 Acceptance/rejection leads to density What is chance of accepting a given y? Let S be size of region, A θ (x, t) size of shadow Probability no points outside of shadow exp( λ( S A θ (x, t))) = exp(λa θ (x, t)) exp( λ S ) Mark Huber, CMC Perfect simulation of repulsive point processes 28/56
29 The problem of unknown time of birth Big problem We do not see the t values! To get the density for just x integrate out t: g(x θ) = Cλ t [0,1] #x exp(λa θ (x, t)) #x Doing this integral directly extremely nasty Even calculating A θ (x, t) hard Mark Huber, CMC Perfect simulation of repulsive point processes 29/56
30 Monte Carlo methods to the rescue! Given seen points x, want to approximate g(x θ) Monte Carlo approach: treat t values as auxiliary variables Given x: randomly choose t from f (x, t θ, x) Allows use to estimate g(x θ) Mark Huber, CMC Perfect simulation of repulsive point processes 30/56
31 How to draw time stamps given locations? Markov chain Monte Carlo (MCMC) 1 Build chain so stationary distribution = target distribution 2 Under mild conditions (φ-irreducibility, aperiodicity) limiting distribution will equal stationary distribution 3 Run chain for a long time to mix, then get samples Mark Huber, CMC Perfect simulation of repulsive point processes 31/56
32 Metropolis protocol for building Markov chain One step of Metropolis 1 Propose moving from (x, t) to (x, t ) 2 Accept move with probability { f (x, t } θ) min f (x, t θ), 1 3 Otherwise stay where you are Mark Huber, CMC Perfect simulation of repulsive point processes 32/56
33 Even better... Perfect simulation techniques For some Markov chains, possible to do better Can draw exactly from stationary distribution Without worrying about mixing time Perfect sampling protocol Coupling from the past (Propp & Wilson 1994) CFTP converts Markov chains to perfect samplers A good property of Markov chain: monotonicity Monotonicity not necessary for CFTP, is sufficient Mark Huber, CMC Perfect simulation of repulsive point processes 33/56
34 Is Metropolis for Matérn type III monotonic? Propose changing time stamp for one point in x How does the shadow change? time 1 R = space Probability accept move = exp(-area of shadow change) Mark Huber, CMC Perfect simulation of repulsive point processes 34/56
35 How to flip an exp( µ) coin Fun facts about Poisson point processes For X PPP(B), P(#X = 0) = exp( µ(b)) For regions A B, X PPP(B) then X A PPP(A) To check if move in Metropolis Draw PPP(largest possible change in shadow) If PPP restricted to actual change in shadow empty, move Otherwise, stay at current time stamp Using PPP to flip exponential coins: First appears in Beskos, Papaspiliopoulos, Roberts (2006) Perfect simulation of diffusions Mark Huber, CMC Perfect simulation of repulsive point processes 35/56
36 Example time 1 R = space Mark Huber, CMC Perfect simulation of repulsive point processes 36/56
37 Example time 1 R = space Mark Huber, CMC Perfect simulation of repulsive point processes 36/56
38 Example time 1 R = space Mark Huber, CMC Perfect simulation of repulsive point processes 36/56
39 Montonicity of Markov chain step For t 1 t 2 : Run one step of Markov chain for t 1 and t 2 Use same auxiliary PPP in change of shadow Then after step, still have t 1 t 2 Immediately gives us perfect sampling! Mark Huber, CMC Perfect simulation of repulsive point processes 37/56
40 Example of Monotonicity time 1 R = 1.5 t 2 t 2 t 2 0 t 1 t 1 t 1 10 space Mark Huber, CMC Perfect simulation of repulsive point processes 38/56
41 Example of Monotonicity time 1 R = 1.5 t 2 t 2 t 2 0 t 1 t 1 t 1 10 space Mark Huber, CMC Perfect simulation of repulsive point processes 38/56
42 Example of Monotonicity time 1 R = 1.5 t 2 t 2 t 2 0 t 1 t 1 t 1 10 space t 2 accepts move Mark Huber, CMC Perfect simulation of repulsive point processes 38/56
43 Example of Monotonicity time 1 R = 1.5 t 2 t 2 t 2 0 t 1 t 1 t 1 10 space t 2 accepts move t 1 rejects move Mark Huber, CMC Perfect simulation of repulsive point processes 38/56
44 Monotonic CFTP flowchart CFTP(k) Output t 1 YES t 1 all zeros t 2 all ones save RNG seed Run Markov chain for k steps Does t 1 = t 2? RNG = Random Number Generator NO t 1 CFTP(2k) t 2 t 1 reset RNG seed Mark Huber, CMC Perfect simulation of repulsive point processes 39/56
45 Monotonic CFTP details Running CFTP(k) 1 Set t 1 to all 0 s, t 2 to all 1 s 2 Save seed to random number generator 3 Take k steps in the Markov chain 4 If t 1 = t 2, then set T k to be this common value 5 Else 1 Get T 0 by calling CFTP(2k) recursively 2 Reset seed to random number generator to what it was in step 2 3 Get T k by taking k steps in the Markov chain 6 Output T k Mark Huber, CMC Perfect simulation of repulsive point processes 40/56
46 What s the point? Why do we need samples? Ability to sample gives approximation of nasty integral Use TPA or IS+TPA to go from samples to integral Once you have that integral Gives density of data under Matérn model Basis for maximum likelihood...or posterior analysis Mark Huber, CMC Perfect simulation of repulsive point processes 41/56
47 Results: towns Mark Huber, CMC Perfect simulation of repulsive point processes 42/56
48 Results: trees Mark Huber, CMC Perfect simulation of repulsive point processes 43/56
49 Strauss: Density to time evolution Mark Huber, CMC Perfect simulation of repulsive point processes 44/56
50 Preston (1977) Birth-Death Chains Adding time dimension to Poisson point process time Gray bar = node lifespan length bar exp(1) birth birth birth 0 space birth birth birth time between births exp(λµ(s)) Mark Huber, CMC Perfect simulation of repulsive point processes 45/56
51 Preston Birth-Death Chains Points are born, and later die Rate of births is λ times area of region Each point dies at rate 1 (Rate is parameter of exponential random variable) Points alive at time 0 form PPP Mark Huber, CMC Perfect simulation of repulsive point processes 46/56
52 Preston for Strauss Recall Strauss density Penalty γ for pair of points within distance R When point is born... If point within distance R of point already born......point only born with probability γ Mark Huber, CMC Perfect simulation of repulsive point processes 47/56
53 Picture for Strauss γ = 0 time birth birth birth 0 space birth birth birth Mark Huber, CMC Perfect simulation of repulsive point processes 48/56
54 Observations Red points Strauss process subset of earlier points Kendall and J. Møller (2000): can find red points by looking backwards in time Mark Huber, CMC Perfect simulation of repulsive point processes 49/56
55 Adding a Swap move Swap move Broder (1986): swap move for perfect matchings Dyer and Greenhill (2000): swap move for independent sets of graphs Adding Swaps to Preston Huber (2011) When point not born, give chance to swap If only blocked by one point......remove blocker, allow birth Mark Huber, CMC Perfect simulation of repulsive point processes 50/56
56 Picture with swap time birth birth birth 0 space birth birth birth Mark Huber, CMC Perfect simulation of repulsive point processes 51/56
57 Swap move helps chain mix better Easier to find red points Verified experimentally on plane Points affect fewer points in future Need to be blocked by at least two points to affect future So effect of point on later points cut in half Mark Huber, CMC Perfect simulation of repulsive point processes 52/56
58 Running times with swap Strauss model on S = [0, 1] 2, γ = 0.5, R = x 106 Running time for swap and no swap chains 3 Average number of events needed per sample No swapping Always swap when possible β 1 λ Mark Huber, CMC Perfect simulation of repulsive point processes 53/56
59 Running times with swap Time for no swap divided by time for swap 5 No swap times divided by swap times 4.5 Average no swap divided by average swap β 1 λ Mark Huber, CMC Perfect simulation of repulsive point processes 54/56
60 Conclusions For Matérn type III models Built a density (not in closed form) Can approximate density using perfect MCMC methods Allows MLE or posterior analysis For Strauss process Already had a density (but not in closed form) Added new type of Markov chain move Seems to speed up chain in practice Mark Huber, CMC Perfect simulation of repulsive point processes 55/56
61 References HUBER, M.L. (2011). Spatial Birth-Death-Swap Chains. Bernoulli (forthcoming paper, available online) HUBER, M.L. AND R.L. WOLPERT (2009). Likelihood based inference for Matérn type III repulsive point processes. Advances in Applied Probability 41, MATÉRN, B. (1986). Spatial Variation, vol. 36 of Lecture Notes in Statistics. New York, NY: Springer-Verlag, 2nd ed. (first edition published 1960 by Statens Skogsforsningsinstitut, Stockholm). Mark Huber, CMC Perfect simulation of repulsive point processes 56/56
Perfect simulation for repulsive point processes
Perfect simulation for repulsive point processes Why swapping at birth is a good thing Mark Huber Department of Mathematics Claremont-McKenna College 20 May, 2009 Mark Huber (Claremont-McKenna College)
More informationPerfect simulation for image analysis
Perfect simulation for image analysis Mark Huber Fletcher Jones Foundation Associate Professor of Mathematics and Statistics and George R. Roberts Fellow Mathematical Sciences Claremont McKenna College
More informationOn probability... Mark Huber (Duke University) Swap moves for spatial point processes Graduate/Faculty Seminar 1 / 42
On probability... I think you re begging the question, said Haydock, and I can see looming ahead one of those terrible exercises in probability where six men have white hats and six men have black hats
More informationA = {(x, u) : 0 u f(x)},
Draw x uniformly from the region {x : f(x) u }. Markov Chain Monte Carlo Lecture 5 Slice sampler: Suppose that one is interested in sampling from a density f(x), x X. Recall that sampling x f(x) is equivalent
More informationMachine Learning for Data Science (CS4786) Lecture 24
Machine Learning for Data Science (CS4786) Lecture 24 Graphical Models: Approximate Inference Course Webpage : http://www.cs.cornell.edu/courses/cs4786/2016sp/ BELIEF PROPAGATION OR MESSAGE PASSING Each
More informationA Bayesian Approach to Phylogenetics
A Bayesian Approach to Phylogenetics Niklas Wahlberg Based largely on slides by Paul Lewis (www.eeb.uconn.edu) An Introduction to Bayesian Phylogenetics Bayesian inference in general Markov chain Monte
More informationMarkov Chains and MCMC
Markov Chains and MCMC CompSci 590.02 Instructor: AshwinMachanavajjhala Lecture 4 : 590.02 Spring 13 1 Recap: Monte Carlo Method If U is a universe of items, and G is a subset satisfying some property,
More informationWho was Bayes? Bayesian Phylogenetics. What is Bayes Theorem?
Who was Bayes? Bayesian Phylogenetics Bret Larget Departments of Botany and of Statistics University of Wisconsin Madison October 6, 2011 The Reverand Thomas Bayes was born in London in 1702. He was the
More informationBayesian Phylogenetics
Bayesian Phylogenetics Bret Larget Departments of Botany and of Statistics University of Wisconsin Madison October 6, 2011 Bayesian Phylogenetics 1 / 27 Who was Bayes? The Reverand Thomas Bayes was born
More informationIntroduction to Machine Learning CMU-10701
Introduction to Machine Learning CMU-10701 Markov Chain Monte Carlo Methods Barnabás Póczos & Aarti Singh Contents Markov Chain Monte Carlo Methods Goal & Motivation Sampling Rejection Importance Markov
More informationIntroduction to Computational Biology Lecture # 14: MCMC - Markov Chain Monte Carlo
Introduction to Computational Biology Lecture # 14: MCMC - Markov Chain Monte Carlo Assaf Weiner Tuesday, March 13, 2007 1 Introduction Today we will return to the motif finding problem, in lecture 10
More informationLikelihood-based Inference for Matérn Type III Repulsive Point Processes
Likelihood-based Inference for Matérn Type III Repulsive Point Processes Version: April 29, 2014 Mark L. Huber Departments of Mathematics and Statistical Science, Duke University autotomic@gmail.com Robert
More informationMCMC and Gibbs Sampling. Kayhan Batmanghelich
MCMC and Gibbs Sampling Kayhan Batmanghelich 1 Approaches to inference l Exact inference algorithms l l l The elimination algorithm Message-passing algorithm (sum-product, belief propagation) The junction
More informationStat 516, Homework 1
Stat 516, Homework 1 Due date: October 7 1. Consider an urn with n distinct balls numbered 1,..., n. We sample balls from the urn with replacement. Let N be the number of draws until we encounter a ball
More information27 : Distributed Monte Carlo Markov Chain. 1 Recap of MCMC and Naive Parallel Gibbs Sampling
10-708: Probabilistic Graphical Models 10-708, Spring 2014 27 : Distributed Monte Carlo Markov Chain Lecturer: Eric P. Xing Scribes: Pengtao Xie, Khoa Luu In this scribe, we are going to review the Parallel
More informationPropp-Wilson Algorithm (and sampling the Ising model)
Propp-Wilson Algorithm (and sampling the Ising model) Danny Leshem, Nov 2009 References: Haggstrom, O. (2002) Finite Markov Chains and Algorithmic Applications, ch. 10-11 Propp, J. & Wilson, D. (1996)
More informationConvergence Rate of Markov Chains
Convergence Rate of Markov Chains Will Perkins April 16, 2013 Convergence Last class we saw that if X n is an irreducible, aperiodic, positive recurrent Markov chain, then there exists a stationary distribution
More informationLecture 8: Bayesian Networks
Lecture 8: Bayesian Networks Bayesian Networks Inference in Bayesian Networks COMP-652 and ECSE 608, Lecture 8 - January 31, 2017 1 Bayes nets P(E) E=1 E=0 0.005 0.995 E B P(B) B=1 B=0 0.01 0.99 E=0 E=1
More informationMCMC notes by Mark Holder
MCMC notes by Mark Holder Bayesian inference Ultimately, we want to make probability statements about true values of parameters, given our data. For example P(α 0 < α 1 X). According to Bayes theorem:
More information7.1 Coupling from the Past
Georgia Tech Fall 2006 Markov Chain Monte Carlo Methods Lecture 7: September 12, 2006 Coupling from the Past Eric Vigoda 7.1 Coupling from the Past 7.1.1 Introduction We saw in the last lecture how Markov
More informationTheory of Stochastic Processes 8. Markov chain Monte Carlo
Theory of Stochastic Processes 8. Markov chain Monte Carlo Tomonari Sei sei@mist.i.u-tokyo.ac.jp Department of Mathematical Informatics, University of Tokyo June 8, 2017 http://www.stat.t.u-tokyo.ac.jp/~sei/lec.html
More informationAdvances and Applications in Perfect Sampling
and Applications in Perfect Sampling Ph.D. Dissertation Defense Ulrike Schneider advisor: Jem Corcoran May 8, 2003 Department of Applied Mathematics University of Colorado Outline Introduction (1) MCMC
More informationBayesian model selection in graphs by using BDgraph package
Bayesian model selection in graphs by using BDgraph package A. Mohammadi and E. Wit March 26, 2013 MOTIVATION Flow cytometry data with 11 proteins from Sachs et al. (2005) RESULT FOR CELL SIGNALING DATA
More informationLIKELIHOOD-BASED INFERENCE FOR MATÉRN TYPE III REPULSIVE POINT PROCESSES
Applied Probability Trust (7 April 2009) LIKELIHOOD-BASED INFERENCE FOR MATÉRN TYPE III REPULSIVE POINT PROCESSES MARK L. HUBER, Duke University ROBERT L. WOLPERT, Duke University Abstract In a repulsive
More informationLect4: Exact Sampling Techniques and MCMC Convergence Analysis
Lect4: Exact Sampling Techniques and MCMC Convergence Analysis. Exact sampling. Convergence analysis of MCMC. First-hit time analysis for MCMC--ways to analyze the proposals. Outline of the Module Definitions
More informationMinicourse on: Markov Chain Monte Carlo: Simulation Techniques in Statistics
Minicourse on: Markov Chain Monte Carlo: Simulation Techniques in Statistics Eric Slud, Statistics Program Lecture 1: Metropolis-Hastings Algorithm, plus background in Simulation and Markov Chains. Lecture
More informationCSC 2541: Bayesian Methods for Machine Learning
CSC 2541: Bayesian Methods for Machine Learning Radford M. Neal, University of Toronto, 2011 Lecture 3 More Markov Chain Monte Carlo Methods The Metropolis algorithm isn t the only way to do MCMC. We ll
More informationMarkov Chain Monte Carlo The Metropolis-Hastings Algorithm
Markov Chain Monte Carlo The Metropolis-Hastings Algorithm Anthony Trubiano April 11th, 2018 1 Introduction Markov Chain Monte Carlo (MCMC) methods are a class of algorithms for sampling from a probability
More informationApril 20th, Advanced Topics in Machine Learning California Institute of Technology. Markov Chain Monte Carlo for Machine Learning
for for Advanced Topics in California Institute of Technology April 20th, 2017 1 / 50 Table of Contents for 1 2 3 4 2 / 50 History of methods for Enrico Fermi used to calculate incredibly accurate predictions
More informationAdaptive Monte Carlo Methods for Numerical Integration
Adaptive Monte Carlo Methods for Numerical Integration Mark Huber 1 and Sarah Schott 2 1 Department of Mathematical Sciences, Claremont McKenna College 2 Department of Mathematics, Duke University 8 March,
More informationMonte Carlo Methods. Leon Gu CSD, CMU
Monte Carlo Methods Leon Gu CSD, CMU Approximate Inference EM: y-observed variables; x-hidden variables; θ-parameters; E-step: q(x) = p(x y, θ t 1 ) M-step: θ t = arg max E q(x) [log p(y, x θ)] θ Monte
More informationConvex Optimization CMU-10725
Convex Optimization CMU-10725 Simulated Annealing Barnabás Póczos & Ryan Tibshirani Andrey Markov Markov Chains 2 Markov Chains Markov chain: Homogen Markov chain: 3 Markov Chains Assume that the state
More informationComputational statistics
Computational statistics Markov Chain Monte Carlo methods Thierry Denœux March 2017 Thierry Denœux Computational statistics March 2017 1 / 71 Contents of this chapter When a target density f can be evaluated
More informationMCMC: Markov Chain Monte Carlo
I529: Machine Learning in Bioinformatics (Spring 2013) MCMC: Markov Chain Monte Carlo Yuzhen Ye School of Informatics and Computing Indiana University, Bloomington Spring 2013 Contents Review of Markov
More informationAdvanced Sampling Algorithms
+ Advanced Sampling Algorithms + Mobashir Mohammad Hirak Sarkar Parvathy Sudhir Yamilet Serrano Llerena Advanced Sampling Algorithms Aditya Kulkarni Tobias Bertelsen Nirandika Wanigasekara Malay Singh
More informationData Analysis I. Dr Martin Hendry, Dept of Physics and Astronomy University of Glasgow, UK. 10 lectures, beginning October 2006
Astronomical p( y x, I) p( x, I) p ( x y, I) = p( y, I) Data Analysis I Dr Martin Hendry, Dept of Physics and Astronomy University of Glasgow, UK 10 lectures, beginning October 2006 4. Monte Carlo Methods
More informationMarkov Networks.
Markov Networks www.biostat.wisc.edu/~dpage/cs760/ Goals for the lecture you should understand the following concepts Markov network syntax Markov network semantics Potential functions Partition function
More informationCS 343: Artificial Intelligence
CS 343: Artificial Intelligence Bayes Nets: Sampling Prof. Scott Niekum The University of Texas at Austin [These slides based on those of Dan Klein and Pieter Abbeel for CS188 Intro to AI at UC Berkeley.
More information17 : Markov Chain Monte Carlo
10-708: Probabilistic Graphical Models, Spring 2015 17 : Markov Chain Monte Carlo Lecturer: Eric P. Xing Scribes: Heran Lin, Bin Deng, Yun Huang 1 Review of Monte Carlo Methods 1.1 Overview Monte Carlo
More informationMarkov Chain Monte Carlo
Markov Chain Monte Carlo Recall: To compute the expectation E ( h(y ) ) we use the approximation E(h(Y )) 1 n n h(y ) t=1 with Y (1),..., Y (n) h(y). Thus our aim is to sample Y (1),..., Y (n) from f(y).
More informationMarkov chain Monte Carlo Lecture 9
Markov chain Monte Carlo Lecture 9 David Sontag New York University Slides adapted from Eric Xing and Qirong Ho (CMU) Limitations of Monte Carlo Direct (unconditional) sampling Hard to get rare events
More informationThe Particle Filter. PD Dr. Rudolph Triebel Computer Vision Group. Machine Learning for Computer Vision
The Particle Filter Non-parametric implementation of Bayes filter Represents the belief (posterior) random state samples. by a set of This representation is approximate. Can represent distributions that
More informationModel Counting for Logical Theories
Model Counting for Logical Theories Wednesday Dmitry Chistikov Rayna Dimitrova Department of Computer Science University of Oxford, UK Max Planck Institute for Software Systems (MPI-SWS) Kaiserslautern
More informationMarkov chain Monte Carlo
Markov chain Monte Carlo Peter Beerli October 10, 2005 [this chapter is highly influenced by chapter 1 in Markov chain Monte Carlo in Practice, eds Gilks W. R. et al. Chapman and Hall/CRC, 1996] 1 Short
More informationMarkov-Chain Monte Carlo
Markov-Chain Monte Carlo CSE586 Computer Vision II Spring 2010, Penn State Univ. References Recall: Sampling Motivation If we can generate random samples x i from a given distribution P(x), then we can
More informationComputer Vision Group Prof. Daniel Cremers. 11. Sampling Methods
Prof. Daniel Cremers 11. Sampling Methods Sampling Methods Sampling Methods are widely used in Computer Science as an approximation of a deterministic algorithm to represent uncertainty without a parametric
More informationStochastic optimization Markov Chain Monte Carlo
Stochastic optimization Markov Chain Monte Carlo Ethan Fetaya Weizmann Institute of Science 1 Motivation Markov chains Stationary distribution Mixing time 2 Algorithms Metropolis-Hastings Simulated Annealing
More informationLecture 21: Counting and Sampling Problems
princeton univ. F 14 cos 521: Advanced Algorithm Design Lecture 21: Counting and Sampling Problems Lecturer: Sanjeev Arora Scribe: Today s topic of counting and sampling problems is motivated by computational
More informationLecture 16: October 29
CS294 Markov Chain Monte Carlo: Foundations & Applications Fall 2009 Lecture 16: October 29 Lecturer: Alistair Sinclair Scribes: Disclaimer: These notes have not been subjected to the usual scrutiny reserved
More informationApproximate inference in Energy-Based Models
CSC 2535: 2013 Lecture 3b Approximate inference in Energy-Based Models Geoffrey Hinton Two types of density model Stochastic generative model using directed acyclic graph (e.g. Bayes Net) Energy-based
More informationApproximate Inference
Approximate Inference Simulation has a name: sampling Sampling is a hot topic in machine learning, and it s really simple Basic idea: Draw N samples from a sampling distribution S Compute an approximate
More informationForward Problems and their Inverse Solutions
Forward Problems and their Inverse Solutions Sarah Zedler 1,2 1 King Abdullah University of Science and Technology 2 University of Texas at Austin February, 2013 Outline 1 Forward Problem Example Weather
More informationBayesian GLMs and Metropolis-Hastings Algorithm
Bayesian GLMs and Metropolis-Hastings Algorithm We have seen that with conjugate or semi-conjugate prior distributions the Gibbs sampler can be used to sample from the posterior distribution. In situations,
More informationSTA 4273H: Statistical Machine Learning
STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Computer Science! Department of Statistical Sciences! rsalakhu@cs.toronto.edu! h0p://www.cs.utoronto.ca/~rsalakhu/ Lecture 7 Approximate
More informationSC7/SM6 Bayes Methods HT18 Lecturer: Geoff Nicholls Lecture 2: Monte Carlo Methods Notes and Problem sheets are available at http://www.stats.ox.ac.uk/~nicholls/bayesmethods/ and via the MSc weblearn pages.
More informationComputer Intensive Methods in Mathematical Statistics
Computer Intensive Methods in Mathematical Statistics Department of mathematics johawes@kth.se Lecture 16 Advanced topics in computational statistics 18 May 2017 Computer Intensive Methods (1) Plan of
More informationBayesian Inference and MCMC
Bayesian Inference and MCMC Aryan Arbabi Partly based on MCMC slides from CSC412 Fall 2018 1 / 18 Bayesian Inference - Motivation Consider we have a data set D = {x 1,..., x n }. E.g each x i can be the
More informationReview. DS GA 1002 Statistical and Mathematical Models. Carlos Fernandez-Granda
Review DS GA 1002 Statistical and Mathematical Models http://www.cims.nyu.edu/~cfgranda/pages/dsga1002_fall16 Carlos Fernandez-Granda Probability and statistics Probability: Framework for dealing with
More informationMarkov Chain Monte Carlo (MCMC)
Markov Chain Monte Carlo (MCMC Dependent Sampling Suppose we wish to sample from a density π, and we can evaluate π as a function but have no means to directly generate a sample. Rejection sampling can
More informationArtificial Intelligence
ICS461 Fall 2010 Nancy E. Reed nreed@hawaii.edu 1 Lecture #14B Outline Inference in Bayesian Networks Exact inference by enumeration Exact inference by variable elimination Approximate inference by stochastic
More informationTime-Sensitive Dirichlet Process Mixture Models
Time-Sensitive Dirichlet Process Mixture Models Xiaojin Zhu Zoubin Ghahramani John Lafferty May 25 CMU-CALD-5-4 School of Computer Science Carnegie Mellon University Pittsburgh, PA 523 Abstract We introduce
More informationAnnouncements. Inference. Mid-term. Inference by Enumeration. Reminder: Alarm Network. Introduction to Artificial Intelligence. V22.
Introduction to Artificial Intelligence V22.0472-001 Fall 2009 Lecture 15: Bayes Nets 3 Midterms graded Assignment 2 graded Announcements Rob Fergus Dept of Computer Science, Courant Institute, NYU Slides
More informationSampling Rejection Sampling Importance Sampling Markov Chain Monte Carlo. Sampling Methods. Oliver Schulte - CMPT 419/726. Bishop PRML Ch.
Sampling Methods Oliver Schulte - CMP 419/726 Bishop PRML Ch. 11 Recall Inference or General Graphs Junction tree algorithm is an exact inference method for arbitrary graphs A particular tree structure
More informationComputer Vision Group Prof. Daniel Cremers. 14. Sampling Methods
Prof. Daniel Cremers 14. Sampling Methods Sampling Methods Sampling Methods are widely used in Computer Science as an approximation of a deterministic algorithm to represent uncertainty without a parametric
More informationA quick introduction to Markov chains and Markov chain Monte Carlo (revised version)
A quick introduction to Markov chains and Markov chain Monte Carlo (revised version) Rasmus Waagepetersen Institute of Mathematical Sciences Aalborg University 1 Introduction These notes are intended to
More informationMarkov Chain Monte Carlo Inference. Siamak Ravanbakhsh Winter 2018
Graphical Models Markov Chain Monte Carlo Inference Siamak Ravanbakhsh Winter 2018 Learning objectives Markov chains the idea behind Markov Chain Monte Carlo (MCMC) two important examples: Gibbs sampling
More informationParameter Estimation. William H. Jefferys University of Texas at Austin Parameter Estimation 7/26/05 1
Parameter Estimation William H. Jefferys University of Texas at Austin bill@bayesrules.net Parameter Estimation 7/26/05 1 Elements of Inference Inference problems contain two indispensable elements: Data
More informationThe Ising model and Markov chain Monte Carlo
The Ising model and Markov chain Monte Carlo Ramesh Sridharan These notes give a short description of the Ising model for images and an introduction to Metropolis-Hastings and Gibbs Markov Chain Monte
More informationMachine Learning. Probabilistic KNN.
Machine Learning. Mark Girolami girolami@dcs.gla.ac.uk Department of Computing Science University of Glasgow June 21, 2007 p. 1/3 KNN is a remarkably simple algorithm with proven error-rates June 21, 2007
More informationReminder of some Markov Chain properties:
Reminder of some Markov Chain properties: 1. a transition from one state to another occurs probabilistically 2. only state that matters is where you currently are (i.e. given present, future is independent
More informationAn ABC interpretation of the multiple auxiliary variable method
School of Mathematical and Physical Sciences Department of Mathematics and Statistics Preprint MPS-2016-07 27 April 2016 An ABC interpretation of the multiple auxiliary variable method by Dennis Prangle
More informationBayes Nets: Sampling
Bayes Nets: Sampling [These slides were created by Dan Klein and Pieter Abbeel for CS188 Intro to AI at UC Berkeley. All CS188 materials are available at http://ai.berkeley.edu.] Approximate Inference:
More informationInformatics 2D Reasoning and Agents Semester 2,
Informatics 2D Reasoning and Agents Semester 2, 2018 2019 Alex Lascarides alex@inf.ed.ac.uk Lecture 25 Approximate Inference in Bayesian Networks 19th March 2019 Informatics UoE Informatics 2D 1 Where
More informationBayesian Methods for Machine Learning
Bayesian Methods for Machine Learning CS 584: Big Data Analytics Material adapted from Radford Neal s tutorial (http://ftp.cs.utoronto.ca/pub/radford/bayes-tut.pdf), Zoubin Ghahramni (http://hunch.net/~coms-4771/zoubin_ghahramani_bayesian_learning.pdf),
More informationComputer Vision Group Prof. Daniel Cremers. 10a. Markov Chain Monte Carlo
Group Prof. Daniel Cremers 10a. Markov Chain Monte Carlo Markov Chain Monte Carlo In high-dimensional spaces, rejection sampling and importance sampling are very inefficient An alternative is Markov Chain
More informationMCMC Methods: Gibbs and Metropolis
MCMC Methods: Gibbs and Metropolis Patrick Breheny February 28 Patrick Breheny BST 701: Bayesian Modeling in Biostatistics 1/30 Introduction As we have seen, the ability to sample from the posterior distribution
More informationBayes Networks. CS540 Bryan R Gibson University of Wisconsin-Madison. Slides adapted from those used by Prof. Jerry Zhu, CS540-1
Bayes Networks CS540 Bryan R Gibson University of Wisconsin-Madison Slides adapted from those used by Prof. Jerry Zhu, CS540-1 1 / 59 Outline Joint Probability: great for inference, terrible to obtain
More informationMarkov Chain Monte Carlo Lecture 6
Sequential parallel tempering With the development of science and technology, we more and more need to deal with high dimensional systems. For example, we need to align a group of protein or DNA sequences
More informationAnnouncements. CS 188: Artificial Intelligence Fall Causality? Example: Traffic. Topology Limits Distributions. Example: Reverse Traffic
CS 188: Artificial Intelligence Fall 2008 Lecture 16: Bayes Nets III 10/23/2008 Announcements Midterms graded, up on glookup, back Tuesday W4 also graded, back in sections / box Past homeworks in return
More informationCPSC 540: Machine Learning
CPSC 540: Machine Learning MCMC and Non-Parametric Bayes Mark Schmidt University of British Columbia Winter 2016 Admin I went through project proposals: Some of you got a message on Piazza. No news is
More informationIntroduction to Machine Learning CMU-10701
Introduction to Machine Learning CMU-10701 Markov Chain Monte Carlo Methods Barnabás Póczos Contents Markov Chain Monte Carlo Methods Sampling Rejection Importance Hastings-Metropolis Gibbs Markov Chains
More information28 : Approximate Inference - Distributed MCMC
10-708: Probabilistic Graphical Models, Spring 2015 28 : Approximate Inference - Distributed MCMC Lecturer: Avinava Dubey Scribes: Hakim Sidahmed, Aman Gupta 1 Introduction For many interesting problems,
More informationSTA 4273H: Statistical Machine Learning
STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.utstat.utoronto.ca/~rsalakhu/ Sidney Smith Hall, Room 6002 Lecture 3 Linear
More informationSimulation of point process using MCMC
Simulation of point process using MCMC Jakob G. Rasmussen Department of Mathematics Aalborg University Denmark March 2, 2011 1/14 Today When do we need MCMC? Case 1: Fixed number of points Metropolis-Hastings,
More informationarxiv: v2 [stat.me] 3 Apr 2015
Bayesian inference for Matérn repulsive processes arxiv:1308.1136v2 [stat.me] 3 Apr 2015 Vinayak Rao Department of Statistics, Purdue University, USA Ryan P. Adams School of Engineering and Applied Sciences,
More informationMath 456: Mathematical Modeling. Tuesday, April 9th, 2018
Math 456: Mathematical Modeling Tuesday, April 9th, 2018 The Ergodic theorem Tuesday, April 9th, 2018 Today 1. Asymptotic frequency (or: How to use the stationary distribution to estimate the average amount
More informationToday. Statistical Learning. Coin Flip. Coin Flip. Experiment 1: Heads. Experiment 1: Heads. Which coin will I use? Which coin will I use?
Today Statistical Learning Parameter Estimation: Maximum Likelihood (ML) Maximum A Posteriori (MAP) Bayesian Continuous case Learning Parameters for a Bayesian Network Naive Bayes Maximum Likelihood estimates
More informationBayesian Graphical Models
Graphical Models and Inference, Lecture 16, Michaelmas Term 2009 December 4, 2009 Parameter θ, data X = x, likelihood L(θ x) p(x θ). Express knowledge about θ through prior distribution π on θ. Inference
More informationPMR Learning as Inference
Outline PMR Learning as Inference Probabilistic Modelling and Reasoning Amos Storkey Modelling 2 The Exponential Family 3 Bayesian Sets School of Informatics, University of Edinburgh Amos Storkey PMR Learning
More informationStochastic Simulation
Stochastic Simulation Idea: probabilities samples Get probabilities from samples: X count x 1 n 1. x k total. n k m X probability x 1. n 1 /m. x k n k /m If we could sample from a variable s (posterior)
More informationA note on perfect simulation for exponential random graph models
A note on perfect simulation for exponential random graph models A. Cerqueira, A. Garivier and F. Leonardi October 4, 017 arxiv:1710.00873v1 [stat.co] Oct 017 Abstract In this paper we propose a perfect
More informationSAMPLING ALGORITHMS. In general. Inference in Bayesian models
SAMPLING ALGORITHMS SAMPLING ALGORITHMS In general A sampling algorithm is an algorithm that outputs samples x 1, x 2,... from a given distribution P or density p. Sampling algorithms can for example be
More informationReferences. Markov-Chain Monte Carlo. Recall: Sampling Motivation. Problem. Recall: Sampling Methods. CSE586 Computer Vision II
References Markov-Chain Monte Carlo CSE586 Computer Vision II Spring 2010, Penn State Univ. Recall: Sampling Motivation If we can generate random samples x i from a given distribution P(x), then we can
More informationRandom Walks A&T and F&S 3.1.2
Random Walks A&T 110-123 and F&S 3.1.2 As we explained last time, it is very difficult to sample directly a general probability distribution. - If we sample from another distribution, the overlap will
More informationGraphical Models. Lecture 15: Approximate Inference by Sampling. Andrew McCallum
Graphical Models Lecture 15: Approximate Inference by Sampling Andrew McCallum mccallum@cs.umass.edu Thanks to Noah Smith and Carlos Guestrin for some slide materials. 1 General Idea Set of random variables
More information1 Probabilities. 1.1 Basics 1 PROBABILITIES
1 PROBABILITIES 1 Probabilities Probability is a tricky word usually meaning the likelyhood of something occuring or how frequent something is. Obviously, if something happens frequently, then its probability
More informationCPSC 540: Machine Learning
CPSC 540: Machine Learning More Approximate Inference Mark Schmidt University of British Columbia Winter 2018 Last Time: Approximate Inference We ve been discussing graphical models for density estimation,
More informationLecture 6: Markov Chain Monte Carlo
Lecture 6: Markov Chain Monte Carlo D. Jason Koskinen koskinen@nbi.ku.dk Photo by Howard Jackman University of Copenhagen Advanced Methods in Applied Statistics Feb - Apr 2016 Niels Bohr Institute 2 Outline
More informationLecture 15: MCMC Sanjeev Arora Elad Hazan. COS 402 Machine Learning and Artificial Intelligence Fall 2016
Lecture 15: MCMC Sanjeev Arora Elad Hazan COS 402 Machine Learning and Artificial Intelligence Fall 2016 Course progress Learning from examples Definition + fundamental theorem of statistical learning,
More informationELEC633: Graphical Models
ELEC633: Graphical Models Tahira isa Saleem Scribe from 7 October 2008 References: Casella and George Exploring the Gibbs sampler (1992) Chib and Greenberg Understanding the Metropolis-Hastings algorithm
More information