Bayesian Complementary Clustering, MCMC and Anglo-Saxon placenames
|
|
- Cleopatra Holmes
- 5 years ago
- Views:
Transcription
1 Bayesian Complementary Clustering, MCMC and Anglo-Saxon placenames Giacomo Zanella Department of Statistics University of Warwick, Coventry, UK 20 May 2014
2 Overview 1. Motivation: historical problem (Anglo-Saxon placenames). 2. Modeling: complementary clustering with random partition models. 3. Computation: MCMC on space of matchings (Data Association problems). 4. Real data: mild support to the historians hypothesis.
3 A classic problem: Cluster Analysis Aim: organizing objects into groups whose members are similar. Geometrical interpretation: separate points into clusters made of close points. Figure : A point pattern being divided into two clusters. Giacomo Zanella (University of Warwick) Bayesian Complementary Clustering, MCMC and Anglo-Saxon placenames 20/05/ / 26
4 Our clustering problem: original motivation Problem posed by John Blair (History Professor from Oxford). Figure : Reconstruction of an Anglo-Saxon settlement in West-Stow, Suffolk. (Image borrowed from John Blair) Giacomo Zanella (University of Warwick) Bayesian Complementary Clustering, MCMC and Anglo-Saxon placenames 20/05/ / 26
5 Empirical Observations of cluster of settlements Stretton, Newton, Burton, Carlton in the region of G t.glen
6 Empirical Observations of cluster of settlements Stretton, Newton, Burton, Carlton in the region of G t.glen Giacomo Zanella (University of Warwick) Bayesian Complementary Clustering, MCMC and Anglo-Saxon placenames 20/05/ / 26
7 Problem considered The model MCMC Real Data Recurrent Clustering Pattern Giacomo Zanella (University of Warwick) Bayesian Complementary Clustering, MCMC and Anglo-Saxon placenames 20/05/ / 26
8 Problem considered The model MCMC Real Data Recurrent Clustering Pattern Giacomo Zanella (University of Warwick) Bayesian Complementary Clustering, MCMC and Anglo-Saxon placenames 20/05/ / 26
9 Problem considered The model MCMC Real Data Recurrent Clustering Pattern Giacomo Zanella (University of Warwick) Bayesian Complementary Clustering, MCMC and Anglo-Saxon placenames 20/05/ / 26
10 Problem considered The model MCMC Real Data Recurrent Clustering Pattern Giacomo Zanella (University of Warwick) Bayesian Complementary Clustering, MCMC and Anglo-Saxon placenames 20/05/ / 26
11 Problem considered The model MCMC Real Data Recurrent Clustering Pattern Giacomo Zanella (University of Warwick) Bayesian Complementary Clustering, MCMC and Anglo-Saxon placenames 20/05/ / 26
12 Historians hypothesis Small groups of settlements (2-6 in each cluster?) Closely located (3-10 km?) Each placename corresponds to a function (e.g. Kingston= Police station ) and can appear at most once in each cluster and. Giacomo Zanella (University of Warwick) Bayesian Complementary Clustering, MCMC and Anglo-Saxon placenames 20/05/ / 26
13 Historians hypothesis Small groups of settlements (2-6 in each cluster?) Closely located (3-10 km?) Each placename corresponds to a function (e.g. Kingston= Police station ) and can appear at most once in each cluster and. Does the geographical distribution of placenames support this? Can we provide information about: Cluster sizes and average intra-cluster distance σ? Which placenames tends to cluster together? Explicit inference on the cluster partition. Giacomo Zanella (University of Warwick) Bayesian Complementary Clustering, MCMC and Anglo-Saxon placenames 20/05/ / 26
14 The data as a multi-type point process Giacomo Zanella (University of Warwick) Bayesian Complementary Clustering, MCMC and Anglo-Saxon placenames 20/05/ / 26
15 The data as a multi-type point process Giacomo Zanella (University of Warwick) Bayesian Complementary Clustering, MCMC and Anglo-Saxon placenames 20/05/ / 26
16 The data as a multi-type point process Giacomo Zanella (University of Warwick) Bayesian Complementary Clustering, MCMC and Anglo-Saxon placenames 20/05/ / 26
17 Model requirements 1: Complementary clustering Classical clustering for marked p.p.: clusters made of points close and similar marks (hence marks are an additional dimension). Giacomo Zanella (University of Warwick) Bayesian Complementary Clustering, MCMC and Anglo-Saxon placenames 20/05/ / 26
18 Model requirements 1: Complementary clustering Complementary clustering: each placename (color) can occur at most once in each cluster. Similar to Data Association Problems Giacomo Zanella (University of Warwick) Bayesian Complementary Clustering, MCMC and Anglo-Saxon placenames 20/05/ / 26
19 Model requirements 2: Inferences on the partition x = {x 1,..., x n } Observed points ρ = {C 1,..., C N } Unobserved partition Random Partition Model (RPM) 1. Exchangeable prior distribution π(ρ) 2. Conditional distribution of x ρ 3. Bayesian inferences on the partition: π(ρ x) π(ρ) π(x ρ) Figure : ρ = { {1, 3, 5}, {4, 6}, {2, 7} } Giacomo Zanella (University of Warwick) Bayesian Complementary Clustering, MCMC and Anglo-Saxon placenames 20/05/ / 26
20 Data Generation Model Cluster centers Inhomogeneous Poisson Point Process Giacomo Zanella (University of Warwick) Bayesian Complementary Clustering, MCMC and Anglo-Saxon placenames 20/05/ / 26
21 Data Generation Model Sizes of clusters: C j (p 1,..., p k ), where k is the number of colors. Locations: i.i.d. Gaussians conditioned on the cluster centers being their means Giacomo Zanella (University of Warwick) Bayesian Complementary Clustering, MCMC and Anglo-Saxon placenames 20/05/ / 26
22 Observed Data Observed point process x is the superposition of all the clusters. Giacomo Zanella (University of Warwick) Bayesian Complementary Clustering, MCMC and Anglo-Saxon placenames 20/05/ / 26
23 Posterior distribution We are interested in the posterior distribution of ρ: π(ρ x) Giacomo Zanella (University of Warwick) Bayesian Complementary Clustering, MCMC and Anglo-Saxon placenames 20/05/ / 26
24 Posterior distribution Random elements involved x ρ σ λ p i observed marked point process Partition of x intra-cluster dispersion expected number of clusters Pr[cluster size = i] π(ρ x, σ, p, λ) ( N(ρ) g(x Cj )λ p sj j=1 σ 2 x, ρ, p, λ InvGamma c sj ( exp δ2 C j ) ) 2σ 2 i,l C j, i l 1(m i m l ) ) ( α σ + n(x) N(ρ), β σ + N(ρ) p x, ρ, σ, λ Dir (α 1 + N 1 (ρ),..., α k + N k (ρ)) λ x, ρ, σ, p Gamma (k λ + N(ρ), θ λ /(θ λ + 1)) where c s = ( k s j ) sj (2π) s j 1 and δ 2 C j = i C j ( xi x Cj ) 2. j=1 δ2 C j /2 Giacomo Zanella (University of Warwick) Bayesian Complementary Clustering, MCMC and Anglo-Saxon placenames 20/05/ / 26
25 Intractability of π (ρ) π(ρ) is known up to a normalizing constant. The sample space is of order n! (normalizing π(ρ) is NP-hard) Giacomo Zanella (University of Warwick) Bayesian Complementary Clustering, MCMC and Anglo-Saxon placenames 20/05/ / 26
26 Intractability of π (ρ) π(ρ) is known up to a normalizing constant. The sample space is of order n! (normalizing π(ρ) is NP-hard) Finding ρ max = argmax ρ π(ρ) Optimal Assignation Problem 2-color: Solvable in O(n 3 ) with Hungarian Algorithm. k-color: NP-hard optimization problem. Not approximable in polynomial time with any deterministic algorithm (not in APX). Giacomo Zanella (University of Warwick) Bayesian Complementary Clustering, MCMC and Anglo-Saxon placenames 20/05/ / 26
27 Feasible approach: Monte Carlo Markov Chains Simulate an ergodic Markov chain (X n ) n 0 with stationary distribution π. Estimate I = E π [f (X )] with Î n := 1 t+n f (X k ). n k=t Metropolis-Hasting Algorithm Obtain X n+1 from X n by 1. Sample the proposed move X Q(X n, ) 2. Compute the acceptance probability α(x n, X ) = min { } f (X )Q(X,Xn) 1, f (X n)q(x n,x ). 3. With probability α(x n, X ) set X n+1 = X, otherwise set X n+1 = X n Giacomo Zanella (University of Warwick) Bayesian Complementary Clustering, MCMC and Anglo-Saxon placenames 20/05/ / 26
28 Designing the MH-algorithm: 2D case (i.e. 2 colors) Sample Space: partial matchings contained in a complete bipartite graph. partition ρ = { {1}, {2, 6}, {3}, {4, 7}, {5} } matching X (ρ) Target Measure: π (X (ρ)) e X w(e), for suitably defined edge weights. Giacomo Zanella (University of Warwick) Bayesian Complementary Clustering, MCMC and Anglo-Saxon placenames 20/05/ / 26
29 Proposal distribution Proposal distribution Q(X old, X new ) 1. Pick a red point i and a blue point j according to q(i, j). 2. Propose the corresponding move (add/remove/switch). Giacomo Zanella (University of Warwick) Bayesian Complementary Clustering, MCMC and Anglo-Saxon placenames 20/05/ / 26
30 Proposal distribution vertices edges states of the MC moves allowed Question How to choose q(i, j) in order to achieve a good mixing? Remark: Optimal choice of q(i, j) Optimal scaling of the proposal Figure : Markov Chain represented by a graph Giacomo Zanella (University of Warwick) Bayesian Complementary Clustering, MCMC and Anglo-Saxon placenames 20/05/ / 26
31 What is the best proposal distribution? 1) q(i, j) 1 {wi,j >ε} 2) q(i, j) π(x new ) 3) q(i, j) π(x new ) π(x old )+π(x new ) Giacomo Zanella (University of Warwick) Bayesian Complementary Clustering, MCMC and Anglo-Saxon placenames 20/05/ / 26
32 Truncating the posterior distribution π(x ) 1) q(i, j) 1 {wi,j >ε} 2) Cheap version of 3) 3) q(i, j) π(x new ) π(x old )+π(x new ) Giacomo Zanella (University of Warwick) Bayesian Complementary Clustering, MCMC and Anglo-Saxon placenames 20/05/ / 26
33 Some remarks on the previous MCMC Mixing problems? For p 1 0 (complete matching case), π(x ) shows an increasing multimodality (we used Simulated Tempering in those cases). Theoretical results? Jerrum and Sinclair [1996]: upper bound for the mixing time of an MCMC in a similar problem. Unfeasible in practice. Parallel computation? Developed a parallelizable multiple proposal scheme. Figure : Intensity of gray represents estimated probability of the link. Giacomo Zanella (University of Warwick) Bayesian Complementary Clustering, MCMC and Anglo-Saxon placenames 20/05/ / 26
34 General k dimensional case Sample Space: partial matchings (i.e. hypergraphs of degree at most 1) contained in a k-partite complete hypergraphs.
35 General k dimensional case Sample Space: partial matchings (i.e. hypergraphs of degree at most 1) contained in a k-partite complete hypergraphs. Giacomo Zanella (University of Warwick) Bayesian Complementary Clustering, MCMC and Anglo-Saxon placenames 20/05/ / 26
36 kd Algorithm kd-algorithm Obtain ρ new from ρ old as follows 1) Sample k 2 colors u.a.r.; 2) Evaluate (x 2D, ρ 2D old ); 3) Obtain ρ 2D new using 2D moves; 4) Obtain ρ new from ρ 2D new. Giacomo Zanella (University of Warwick) Bayesian Complementary Clustering, MCMC and Anglo-Saxon placenames 20/05/ / 26
37 Problem considered The model MCMC Real Data Real Dataset Placenames Aston/Easton Bolton Burh-Stall Burton Centres Charlton/Charlcot Chesterton Claeg Draycot/Drayton Eaton Kingston Knighton Newbold Newton Norton Stratton Sutton Tot Walton/Walcot Weston Total number of settlements Giacomo Zanella (University of Warwick) Bayesian Complementary Clustering, MCMC and Anglo-Saxon placenames 20/05/ / 26
38 Fitting our model to the data Giacomo Zanella (University of Warwick) Bayesian Complementary Clustering, MCMC and Anglo-Saxon placenames 20/05/ / 26
39 Giacomo Zanella (University of Warwick) Bayesian Complementary Clustering, MCMC and Anglo-Saxon placenames 20/05/ / 26
40 Acknowledgments Prof. Wilfrid Kendall for PhD supervision. Prof. John Blair for collaboration and arranging supply of data. CRiSM and EPSRC for funding. Giacomo Zanella (University of Warwick) Bayesian Complementary Clustering, MCMC and Anglo-Saxon placenames 20/05/ / 26
41 Thank you Giacomo Zanella (University of Warwick) Bayesian Complementary Clustering, MCMC and Anglo-Saxon placenames 20/05/ / 26
Original citation: Zanella, Giacomo. (2015) Random partition models and complementary clustering of Anglo-Saxon place-names. The Annals of Applied Statistics, 9 (4). pp. 1792-1822. Permanent WRAP url:
More informationBAYESIAN COMPLEMENTARY CLUSTERING, MCMC AND ANGLO-SAXON PLACENAMES. By Giacomo Zanella University of Warwick
Submitted to the Annals of Applied Statistics BAYESIAN COMPLEMENTARY CLUSTERING, MCMC AND ANGLO-SAXON PLACENAMES By Giacomo Zanella University of Warwick Common cluster models for multi-type point processes
More informationMarkov Chains and MCMC
Markov Chains and MCMC CompSci 590.02 Instructor: AshwinMachanavajjhala Lecture 4 : 590.02 Spring 13 1 Recap: Monte Carlo Method If U is a universe of items, and G is a subset satisfying some property,
More informationVCMC: Variational Consensus Monte Carlo
VCMC: Variational Consensus Monte Carlo Maxim Rabinovich, Elaine Angelino, Michael I. Jordan Berkeley Vision and Learning Center September 22, 2015 probabilistic models! sky fog bridge water grass object
More informationAdvances and Applications in Perfect Sampling
and Applications in Perfect Sampling Ph.D. Dissertation Defense Ulrike Schneider advisor: Jem Corcoran May 8, 2003 Department of Applied Mathematics University of Colorado Outline Introduction (1) MCMC
More informationA Dirichlet Form approach to MCMC Optimal Scaling
A Dirichlet Form approach to MCMC Optimal Scaling Giacomo Zanella, Wilfrid S. Kendall, and Mylène Bédard. g.zanella@warwick.ac.uk, w.s.kendall@warwick.ac.uk, mylene.bedard@umontreal.ca Supported by EPSRC
More informationMachine Learning. Probabilistic KNN.
Machine Learning. Mark Girolami girolami@dcs.gla.ac.uk Department of Computing Science University of Glasgow June 21, 2007 p. 1/3 KNN is a remarkably simple algorithm with proven error-rates June 21, 2007
More informationIntroduction to Bayesian methods in inverse problems
Introduction to Bayesian methods in inverse problems Ville Kolehmainen 1 1 Department of Applied Physics, University of Eastern Finland, Kuopio, Finland March 4 2013 Manchester, UK. Contents Introduction
More informationMonte Carlo in Bayesian Statistics
Monte Carlo in Bayesian Statistics Matthew Thomas SAMBa - University of Bath m.l.thomas@bath.ac.uk December 4, 2014 Matthew Thomas (SAMBa) Monte Carlo in Bayesian Statistics December 4, 2014 1 / 16 Overview
More informationPractical unbiased Monte Carlo for Uncertainty Quantification
Practical unbiased Monte Carlo for Uncertainty Quantification Sergios Agapiou Department of Statistics, University of Warwick MiR@W day: Uncertainty in Complex Computer Models, 2nd February 2015, University
More informationBayesian Methods for Machine Learning
Bayesian Methods for Machine Learning CS 584: Big Data Analytics Material adapted from Radford Neal s tutorial (http://ftp.cs.utoronto.ca/pub/radford/bayes-tut.pdf), Zoubin Ghahramni (http://hunch.net/~coms-4771/zoubin_ghahramani_bayesian_learning.pdf),
More informationComputational statistics
Computational statistics Markov Chain Monte Carlo methods Thierry Denœux March 2017 Thierry Denœux Computational statistics March 2017 1 / 71 Contents of this chapter When a target density f can be evaluated
More information(5) Multi-parameter models - Gibbs sampling. ST440/540: Applied Bayesian Analysis
Summarizing a posterior Given the data and prior the posterior is determined Summarizing the posterior gives parameter estimates, intervals, and hypothesis tests Most of these computations are integrals
More informationMarkov Chain Monte Carlo, Numerical Integration
Markov Chain Monte Carlo, Numerical Integration (See Statistics) Trevor Gallen Fall 2015 1 / 1 Agenda Numerical Integration: MCMC methods Estimating Markov Chains Estimating latent variables 2 / 1 Numerical
More informationAdaptive Monte Carlo methods
Adaptive Monte Carlo methods Jean-Michel Marin Projet Select, INRIA Futurs, Université Paris-Sud joint with Randal Douc (École Polytechnique), Arnaud Guillin (Université de Marseille) and Christian Robert
More informationMarkov Chain Monte Carlo Methods
Markov Chain Monte Carlo Methods John Geweke University of Iowa, USA 2005 Institute on Computational Economics University of Chicago - Argonne National Laboaratories July 22, 2005 The problem p (θ, ω I)
More informationComputer Vision Group Prof. Daniel Cremers. 10a. Markov Chain Monte Carlo
Group Prof. Daniel Cremers 10a. Markov Chain Monte Carlo Markov Chain Monte Carlo In high-dimensional spaces, rejection sampling and importance sampling are very inefficient An alternative is Markov Chain
More informationLecture 15: MCMC Sanjeev Arora Elad Hazan. COS 402 Machine Learning and Artificial Intelligence Fall 2016
Lecture 15: MCMC Sanjeev Arora Elad Hazan COS 402 Machine Learning and Artificial Intelligence Fall 2016 Course progress Learning from examples Definition + fundamental theorem of statistical learning,
More informationComputer Vision Group Prof. Daniel Cremers. 11. Sampling Methods: Markov Chain Monte Carlo
Group Prof. Daniel Cremers 11. Sampling Methods: Markov Chain Monte Carlo Markov Chain Monte Carlo In high-dimensional spaces, rejection sampling and importance sampling are very inefficient An alternative
More informationRecent Advances in Bayesian Inference for Inverse Problems
Recent Advances in Bayesian Inference for Inverse Problems Felix Lucka University College London, UK f.lucka@ucl.ac.uk Applied Inverse Problems Helsinki, May 25, 2015 Bayesian Inference for Inverse Problems
More informationAges of stellar populations from color-magnitude diagrams. Paul Baines. September 30, 2008
Ages of stellar populations from color-magnitude diagrams Paul Baines Department of Statistics Harvard University September 30, 2008 Context & Example Welcome! Today we will look at using hierarchical
More informationSession 5B: A worked example EGARCH model
Session 5B: A worked example EGARCH model John Geweke Bayesian Econometrics and its Applications August 7, worked example EGARCH model August 7, / 6 EGARCH Exponential generalized autoregressive conditional
More informationMarkov Chain Monte Carlo (MCMC)
Markov Chain Monte Carlo (MCMC Dependent Sampling Suppose we wish to sample from a density π, and we can evaluate π as a function but have no means to directly generate a sample. Rejection sampling can
More informationeqr094: Hierarchical MCMC for Bayesian System Reliability
eqr094: Hierarchical MCMC for Bayesian System Reliability Alyson G. Wilson Statistical Sciences Group, Los Alamos National Laboratory P.O. Box 1663, MS F600 Los Alamos, NM 87545 USA Phone: 505-667-9167
More informationOn Bayesian Computation
On Bayesian Computation Michael I. Jordan with Elaine Angelino, Maxim Rabinovich, Martin Wainwright and Yun Yang Previous Work: Information Constraints on Inference Minimize the minimax risk under constraints
More informationIntroduction. A Dirichlet Form approach to MCMC Optimal Scaling. MCMC idea
Introuction A Dirichlet Form approach to MCMC Optimal Scaling Markov chain Monte Carlo (MCMC quotes: Metropolis et al. (1953, running coe on the Los Alamos MANIAC: a feasible approach to statistical mechanics
More information28 : Approximate Inference - Distributed MCMC
10-708: Probabilistic Graphical Models, Spring 2015 28 : Approximate Inference - Distributed MCMC Lecturer: Avinava Dubey Scribes: Hakim Sidahmed, Aman Gupta 1 Introduction For many interesting problems,
More informationA Search and Jump Algorithm for Markov Chain Monte Carlo Sampling. Christopher Jennison. Adriana Ibrahim. Seminar at University of Kuwait
A Search and Jump Algorithm for Markov Chain Monte Carlo Sampling Christopher Jennison Department of Mathematical Sciences, University of Bath, UK http://people.bath.ac.uk/mascj Adriana Ibrahim Institute
More informationSTA 294: Stochastic Processes & Bayesian Nonparametrics
MARKOV CHAINS AND CONVERGENCE CONCEPTS Markov chains are among the simplest stochastic processes, just one step beyond iid sequences of random variables. Traditionally they ve been used in modelling a
More informationControl Variates for Markov Chain Monte Carlo
Control Variates for Markov Chain Monte Carlo Dellaportas, P., Kontoyiannis, I., and Tsourti, Z. Dept of Statistics, AUEB Dept of Informatics, AUEB 1st Greek Stochastics Meeting Monte Carlo: Probability
More informationLect4: Exact Sampling Techniques and MCMC Convergence Analysis
Lect4: Exact Sampling Techniques and MCMC Convergence Analysis. Exact sampling. Convergence analysis of MCMC. First-hit time analysis for MCMC--ways to analyze the proposals. Outline of the Module Definitions
More informationIntroduction to Markov Chain Monte Carlo & Gibbs Sampling
Introduction to Markov Chain Monte Carlo & Gibbs Sampling Prof. Nicholas Zabaras Sibley School of Mechanical and Aerospace Engineering 101 Frank H. T. Rhodes Hall Ithaca, NY 14853-3801 Email: zabaras@cornell.edu
More informationMarkov Chain Monte Carlo methods
Markov Chain Monte Carlo methods Tomas McKelvey and Lennart Svensson Signal Processing Group Department of Signals and Systems Chalmers University of Technology, Sweden November 26, 2012 Today s learning
More information17 : Markov Chain Monte Carlo
10-708: Probabilistic Graphical Models, Spring 2015 17 : Markov Chain Monte Carlo Lecturer: Eric P. Xing Scribes: Heran Lin, Bin Deng, Yun Huang 1 Review of Monte Carlo Methods 1.1 Overview Monte Carlo
More informationZig-Zag Monte Carlo. Delft University of Technology. Joris Bierkens February 7, 2017
Zig-Zag Monte Carlo Delft University of Technology Joris Bierkens February 7, 2017 Joris Bierkens (TU Delft) Zig-Zag Monte Carlo February 7, 2017 1 / 33 Acknowledgements Collaborators Andrew Duncan Paul
More informationSAMPLING ALGORITHMS. In general. Inference in Bayesian models
SAMPLING ALGORITHMS SAMPLING ALGORITHMS In general A sampling algorithm is an algorithm that outputs samples x 1, x 2,... from a given distribution P or density p. Sampling algorithms can for example be
More informationMSc MT15. Further Statistical Methods: MCMC. Lecture 5-6: Markov chains; Metropolis Hastings MCMC. Notes and Practicals available at
MSc MT15. Further Statistical Methods: MCMC Lecture 5-6: Markov chains; Metropolis Hastings MCMC Notes and Practicals available at www.stats.ox.ac.uk\ nicholls\mscmcmc15 Markov chain Monte Carlo Methods
More information6 Markov Chain Monte Carlo (MCMC)
6 Markov Chain Monte Carlo (MCMC) The underlying idea in MCMC is to replace the iid samples of basic MC methods, with dependent samples from an ergodic Markov chain, whose limiting (stationary) distribution
More informationAnswers and expectations
Answers and expectations For a function f(x) and distribution P(x), the expectation of f with respect to P is The expectation is the average of f, when x is drawn from the probability distribution P E
More informationIntroduction to Machine Learning CMU-10701
Introduction to Machine Learning CMU-10701 Markov Chain Monte Carlo Methods Barnabás Póczos & Aarti Singh Contents Markov Chain Monte Carlo Methods Goal & Motivation Sampling Rejection Importance Markov
More informationAn introduction to Sequential Monte Carlo
An introduction to Sequential Monte Carlo Thang Bui Jes Frellsen Department of Engineering University of Cambridge Research and Communication Club 6 February 2014 1 Sequential Monte Carlo (SMC) methods
More informationMonte Carlo methods for sampling-based Stochastic Optimization
Monte Carlo methods for sampling-based Stochastic Optimization Gersende FORT LTCI CNRS & Telecom ParisTech Paris, France Joint works with B. Jourdain, T. Lelièvre, G. Stoltz from ENPC and E. Kuhn from
More informationLecture 7 and 8: Markov Chain Monte Carlo
Lecture 7 and 8: Markov Chain Monte Carlo 4F13: Machine Learning Zoubin Ghahramani and Carl Edward Rasmussen Department of Engineering University of Cambridge http://mlg.eng.cam.ac.uk/teaching/4f13/ Ghahramani
More informationMarkov chain Monte Carlo
Markov chain Monte Carlo Peter Beerli October 10, 2005 [this chapter is highly influenced by chapter 1 in Markov chain Monte Carlo in Practice, eds Gilks W. R. et al. Chapman and Hall/CRC, 1996] 1 Short
More informationKernel adaptive Sequential Monte Carlo
Kernel adaptive Sequential Monte Carlo Ingmar Schuster (Paris Dauphine) Heiko Strathmann (University College London) Brooks Paige (Oxford) Dino Sejdinovic (Oxford) December 7, 2015 1 / 36 Section 1 Outline
More informationMarkov Chain Monte Carlo Data Association for Multi-Target Tracking
MCMCDA 1 Markov Chain Monte Carlo Data Association for Multi-Target Tracking Songhwai Oh, Stuart Russell, and Shankar Sastry Abstract This paper presents Markov chain Monte Carlo data association (MCMCDA)
More informationSequential Monte Carlo Samplers for Applications in High Dimensions
Sequential Monte Carlo Samplers for Applications in High Dimensions Alexandros Beskos National University of Singapore KAUST, 26th February 2014 Joint work with: Dan Crisan, Ajay Jasra, Nik Kantas, Alex
More informationPaul Karapanagiotidis ECO4060
Paul Karapanagiotidis ECO4060 The way forward 1) Motivate why Markov-Chain Monte Carlo (MCMC) is useful for econometric modeling 2) Introduce Markov-Chain Monte Carlo (MCMC) - Metropolis-Hastings (MH)
More informationMCMC and Gibbs Sampling. Kayhan Batmanghelich
MCMC and Gibbs Sampling Kayhan Batmanghelich 1 Approaches to inference l Exact inference algorithms l l l The elimination algorithm Message-passing algorithm (sum-product, belief propagation) The junction
More informationTEORIA BAYESIANA Ralph S. Silva
TEORIA BAYESIANA Ralph S. Silva Departamento de Métodos Estatísticos Instituto de Matemática Universidade Federal do Rio de Janeiro Sumário Numerical Integration Polynomial quadrature is intended to approximate
More informationStochastic optimization Markov Chain Monte Carlo
Stochastic optimization Markov Chain Monte Carlo Ethan Fetaya Weizmann Institute of Science 1 Motivation Markov chains Stationary distribution Mixing time 2 Algorithms Metropolis-Hastings Simulated Annealing
More informationMarkov Chain Monte Carlo
1 Motivation 1.1 Bayesian Learning Markov Chain Monte Carlo Yale Chang In Bayesian learning, given data X, we make assumptions on the generative process of X by introducing hidden variables Z: p(z): prior
More information18 : Advanced topics in MCMC. 1 Gibbs Sampling (Continued from the last lecture)
10-708: Probabilistic Graphical Models 10-708, Spring 2014 18 : Advanced topics in MCMC Lecturer: Eric P. Xing Scribes: Jessica Chemali, Seungwhan Moon 1 Gibbs Sampling (Continued from the last lecture)
More informationComputer Vision Group Prof. Daniel Cremers. 14. Sampling Methods
Prof. Daniel Cremers 14. Sampling Methods Sampling Methods Sampling Methods are widely used in Computer Science as an approximation of a deterministic algorithm to represent uncertainty without a parametric
More informationStat 516, Homework 1
Stat 516, Homework 1 Due date: October 7 1. Consider an urn with n distinct balls numbered 1,..., n. We sample balls from the urn with replacement. Let N be the number of draws until we encounter a ball
More informationPART I INTRODUCTION The meaning of probability Basic definitions for frequentist statistics and Bayesian inference Bayesian inference Combinatorics
Table of Preface page xi PART I INTRODUCTION 1 1 The meaning of probability 3 1.1 Classical definition of probability 3 1.2 Statistical definition of probability 9 1.3 Bayesian understanding of probability
More informationBayesian Estimation of Input Output Tables for Russia
Bayesian Estimation of Input Output Tables for Russia Oleg Lugovoy (EDF, RANE) Andrey Polbin (RANE) Vladimir Potashnikov (RANE) WIOD Conference April 24, 2012 Groningen Outline Motivation Objectives Bayesian
More informationMarkov chain Monte Carlo methods in atmospheric remote sensing
1 / 45 Markov chain Monte Carlo methods in atmospheric remote sensing Johanna Tamminen johanna.tamminen@fmi.fi ESA Summer School on Earth System Monitoring and Modeling July 3 Aug 11, 212, Frascati July,
More informationBayesian Nonparametric Regression for Diabetes Deaths
Bayesian Nonparametric Regression for Diabetes Deaths Brian M. Hartman PhD Student, 2010 Texas A&M University College Station, TX, USA David B. Dahl Assistant Professor Texas A&M University College Station,
More informationBayesian GLMs and Metropolis-Hastings Algorithm
Bayesian GLMs and Metropolis-Hastings Algorithm We have seen that with conjugate or semi-conjugate prior distributions the Gibbs sampler can be used to sample from the posterior distribution. In situations,
More informationMarkov Chain Monte Carlo Data Association for Multiple-Target Tracking
OH et al.: MARKOV CHAIN MONTE CARLO DATA ASSOCIATION FOR MULTIPLE-TARGET TRACKING 1 Markov Chain Monte Carlo Data Association for Multiple-Target Tracking Songhwai Oh, Stuart Russell, and Shankar Sastry
More informationSampling from complex probability distributions
Sampling from complex probability distributions Louis J. M. Aslett (louis.aslett@durham.ac.uk) Department of Mathematical Sciences Durham University UTOPIAE Training School II 4 July 2017 1/37 Motivation
More informationWrapped Gaussian processes: a short review and some new results
Wrapped Gaussian processes: a short review and some new results Giovanna Jona Lasinio 1, Gianluca Mastrantonio 2 and Alan Gelfand 3 1-Università Sapienza di Roma 2- Università RomaTRE 3- Duke University
More information16 : Approximate Inference: Markov Chain Monte Carlo
10-708: Probabilistic Graphical Models 10-708, Spring 2017 16 : Approximate Inference: Markov Chain Monte Carlo Lecturer: Eric P. Xing Scribes: Yuan Yang, Chao-Ming Yen 1 Introduction As the target distribution
More informationSC7/SM6 Bayes Methods HT18 Lecturer: Geoff Nicholls Lecture 2: Monte Carlo Methods Notes and Problem sheets are available at http://www.stats.ox.ac.uk/~nicholls/bayesmethods/ and via the MSc weblearn pages.
More informationComputer Vision Group Prof. Daniel Cremers. 11. Sampling Methods
Prof. Daniel Cremers 11. Sampling Methods Sampling Methods Sampling Methods are widely used in Computer Science as an approximation of a deterministic algorithm to represent uncertainty without a parametric
More informationLikelihood-free MCMC
Bayesian inference for stable distributions with applications in finance Department of Mathematics University of Leicester September 2, 2011 MSc project final presentation Outline 1 2 3 4 Classical Monte
More informationReview. DS GA 1002 Statistical and Mathematical Models. Carlos Fernandez-Granda
Review DS GA 1002 Statistical and Mathematical Models http://www.cims.nyu.edu/~cfgranda/pages/dsga1002_fall16 Carlos Fernandez-Granda Probability and statistics Probability: Framework for dealing with
More informationBayesian model selection in graphs by using BDgraph package
Bayesian model selection in graphs by using BDgraph package A. Mohammadi and E. Wit March 26, 2013 MOTIVATION Flow cytometry data with 11 proteins from Sachs et al. (2005) RESULT FOR CELL SIGNALING DATA
More informationMarkov Chain Monte Carlo methods
Markov Chain Monte Carlo methods By Oleg Makhnin 1 Introduction a b c M = d e f g h i 0 f(x)dx 1.1 Motivation 1.1.1 Just here Supresses numbering 1.1.2 After this 1.2 Literature 2 Method 2.1 New math As
More informationBayesian Estimation with Sparse Grids
Bayesian Estimation with Sparse Grids Kenneth L. Judd and Thomas M. Mertens Institute on Computational Economics August 7, 27 / 48 Outline Introduction 2 Sparse grids Construction Integration with sparse
More informationLecturer: David Blei Lecture #3 Scribes: Jordan Boyd-Graber and Francisco Pereira October 1, 2007
COS 597C: Bayesian Nonparametrics Lecturer: David Blei Lecture # Scribes: Jordan Boyd-Graber and Francisco Pereira October, 7 Gibbs Sampling with a DP First, let s recapitulate the model that we re using.
More informationCS281A/Stat241A Lecture 22
CS281A/Stat241A Lecture 22 p. 1/4 CS281A/Stat241A Lecture 22 Monte Carlo Methods Peter Bartlett CS281A/Stat241A Lecture 22 p. 2/4 Key ideas of this lecture Sampling in Bayesian methods: Predictive distribution
More informationBayesian inference for multivariate extreme value distributions
Bayesian inference for multivariate extreme value distributions Sebastian Engelke Clément Dombry, Marco Oesting Toronto, Fields Institute, May 4th, 2016 Main motivation For a parametric model Z F θ of
More informationTheory of Stochastic Processes 8. Markov chain Monte Carlo
Theory of Stochastic Processes 8. Markov chain Monte Carlo Tomonari Sei sei@mist.i.u-tokyo.ac.jp Department of Mathematical Informatics, University of Tokyo June 8, 2017 http://www.stat.t.u-tokyo.ac.jp/~sei/lec.html
More informationCalibrating Environmental Engineering Models and Uncertainty Analysis
Models and Cornell University Oct 14, 2008 Project Team Christine Shoemaker, co-pi, Professor of Civil and works in applied optimization, co-pi Nikolai Blizniouk, PhD student in Operations Research now
More informationStat 451 Lecture Notes Markov Chain Monte Carlo. Ryan Martin UIC
Stat 451 Lecture Notes 07 12 Markov Chain Monte Carlo Ryan Martin UIC www.math.uic.edu/~rgmartin 1 Based on Chapters 8 9 in Givens & Hoeting, Chapters 25 27 in Lange 2 Updated: April 4, 2016 1 / 42 Outline
More informationMonte Carlo (MC) Simulation Methods. Elisa Fadda
Monte Carlo (MC) Simulation Methods Elisa Fadda 1011-CH328, Molecular Modelling & Drug Design 2011 Experimental Observables A system observable is a property of the system state. The system state i is
More informationWinter 2019 Math 106 Topics in Applied Mathematics. Lecture 9: Markov Chain Monte Carlo
Winter 2019 Math 106 Topics in Applied Mathematics Data-driven Uncertainty Quantification Yoonsang Lee (yoonsang.lee@dartmouth.edu) Lecture 9: Markov Chain Monte Carlo 9.1 Markov Chain A Markov Chain Monte
More informationThe Particle Filter. PD Dr. Rudolph Triebel Computer Vision Group. Machine Learning for Computer Vision
The Particle Filter Non-parametric implementation of Bayes filter Represents the belief (posterior) random state samples. by a set of This representation is approximate. Can represent distributions that
More informationA SCALED STOCHASTIC NEWTON ALGORITHM FOR MARKOV CHAIN MONTE CARLO SIMULATIONS
A SCALED STOCHASTIC NEWTON ALGORITHM FOR MARKOV CHAIN MONTE CARLO SIMULATIONS TAN BUI-THANH AND OMAR GHATTAS Abstract. We propose a scaled stochastic Newton algorithm ssn) for local Metropolis-Hastings
More informationBayesian Inference in GLMs. Frequentists typically base inferences on MLEs, asymptotic confidence
Bayesian Inference in GLMs Frequentists typically base inferences on MLEs, asymptotic confidence limits, and log-likelihood ratio tests Bayesians base inferences on the posterior distribution of the unknowns
More informationReminder of some Markov Chain properties:
Reminder of some Markov Chain properties: 1. a transition from one state to another occurs probabilistically 2. only state that matters is where you currently are (i.e. given present, future is independent
More informationMarkov Chain Monte Carlo
Markov Chain Monte Carlo Recall: To compute the expectation E ( h(y ) ) we use the approximation E(h(Y )) 1 n n h(y ) t=1 with Y (1),..., Y (n) h(y). Thus our aim is to sample Y (1),..., Y (n) from f(y).
More informationLecture 8: The Metropolis-Hastings Algorithm
30.10.2008 What we have seen last time: Gibbs sampler Key idea: Generate a Markov chain by updating the component of (X 1,..., X p ) in turn by drawing from the full conditionals: X (t) j Two drawbacks:
More informationBayesian Inference and MCMC
Bayesian Inference and MCMC Aryan Arbabi Partly based on MCMC slides from CSC412 Fall 2018 1 / 18 Bayesian Inference - Motivation Consider we have a data set D = {x 1,..., x n }. E.g each x i can be the
More informationSampling Algorithms for Probabilistic Graphical models
Sampling Algorithms for Probabilistic Graphical models Vibhav Gogate University of Washington References: Chapter 12 of Probabilistic Graphical models: Principles and Techniques by Daphne Koller and Nir
More informationQuantifying Uncertainty
Sai Ravela M. I. T Last Updated: Spring 2013 1 Markov Chain Monte Carlo Monte Carlo sampling made for large scale problems via Markov Chains Monte Carlo Sampling Rejection Sampling Importance Sampling
More informationSTA 4273H: Statistical Machine Learning
STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Computer Science! Department of Statistical Sciences! rsalakhu@cs.toronto.edu! h0p://www.cs.utoronto.ca/~rsalakhu/ Lecture 7 Approximate
More informationBayesian Regression Linear and Logistic Regression
When we want more than point estimates Bayesian Regression Linear and Logistic Regression Nicole Beckage Ordinary Least Squares Regression and Lasso Regression return only point estimates But what if we
More informationKernel Sequential Monte Carlo
Kernel Sequential Monte Carlo Ingmar Schuster (Paris Dauphine) Heiko Strathmann (University College London) Brooks Paige (Oxford) Dino Sejdinovic (Oxford) * equal contribution April 25, 2016 1 / 37 Section
More informationBayesian Prediction of Code Output. ASA Albuquerque Chapter Short Course October 2014
Bayesian Prediction of Code Output ASA Albuquerque Chapter Short Course October 2014 Abstract This presentation summarizes Bayesian prediction methodology for the Gaussian process (GP) surrogate representation
More informationMonte Carlo Methods in Bayesian Inference: Theory, Methods and Applications
University of Arkansas, Fayetteville ScholarWorks@UARK Theses and Dissertations 1-016 Monte Carlo Methods in Bayesian Inference: Theory, Methods and Applications Huarui Zhang University of Arkansas, Fayetteville
More informationApproximate Counting and Markov Chain Monte Carlo
Approximate Counting and Markov Chain Monte Carlo A Randomized Approach Arindam Pal Department of Computer Science and Engineering Indian Institute of Technology Delhi March 18, 2011 April 8, 2011 Arindam
More informationSampling Methods (11/30/04)
CS281A/Stat241A: Statistical Learning Theory Sampling Methods (11/30/04) Lecturer: Michael I. Jordan Scribe: Jaspal S. Sandhu 1 Gibbs Sampling Figure 1: Undirected and directed graphs, respectively, with
More informationMarkov chain Monte Carlo
Markov chain Monte Carlo Markov chain Monte Carlo (MCMC) Gibbs and Metropolis Hastings Slice sampling Practical details Iain Murray http://iainmurray.net/ Reminder Need to sample large, non-standard distributions:
More informationAdaptive HMC via the Infinite Exponential Family
Adaptive HMC via the Infinite Exponential Family Arthur Gretton Gatsby Unit, CSML, University College London RegML, 2017 Arthur Gretton (Gatsby Unit, UCL) Adaptive HMC via the Infinite Exponential Family
More informationGenerative Models and Stochastic Algorithms for Population Average Estimation and Image Analysis
Generative Models and Stochastic Algorithms for Population Average Estimation and Image Analysis Stéphanie Allassonnière CIS, JHU July, 15th 28 Context : Computational Anatomy Context and motivations :
More informationA short introduction to INLA and R-INLA
A short introduction to INLA and R-INLA Integrated Nested Laplace Approximation Thomas Opitz, BioSP, INRA Avignon Workshop: Theory and practice of INLA and SPDE November 7, 2018 2/21 Plan for this talk
More informationComparison of Three Calculation Methods for a Bayesian Inference of Two Poisson Parameters
Journal of Modern Applied Statistical Methods Volume 13 Issue 1 Article 26 5-1-2014 Comparison of Three Calculation Methods for a Bayesian Inference of Two Poisson Parameters Yohei Kawasaki Tokyo University
More information