Adaptive Monte Carlo Methods for Numerical Integration

Size: px
Start display at page:

Download "Adaptive Monte Carlo Methods for Numerical Integration"

Transcription

1 Adaptive Monte Carlo Methods for Numerical Integration Mark Huber 1 and Sarah Schott 2 1 Department of Mathematical Sciences, Claremont McKenna College 2 Department of Mathematics, Duke University 8 March, 2011 Mark Huber and Sarah Schott, CMC,Duke Adaptive MC integration 1/47

2 Integration Mark Huber and Sarah Schott, CMC,Duke Adaptive MC integration 2/47

3 Numerical integration The central problem: Approximate I = f ( x) dr n, f ( x) 0 x Ω Mark Huber and Sarah Schott, CMC,Duke Adaptive MC integration 3/47

4 Applications Statistical Physics Ising model, Potts model, Gibbs distributions Dimension = number of interacting particles Statistics Frequentist p-values Bayesian posterior normalizing constant Dimension = number of data points Combinatorial optimization Counting matchings in graph Counting independent sets Dimension = number of edges or nodes in graph Mark Huber and Sarah Schott, CMC,Duke Adaptive MC integration 4/47

5 1-D: easy Fortunately, in 1-D numerical integration easy Use rectangles, trapezoidal rule, Simpson s rule, et cetera Mark Huber and Sarah Schott, CMC,Duke Adaptive MC integration 5/47

6 1-D: easy Fortunately, in 1-D numerical integration easy Use rectangles, trapezoidal rule, Simpson s rule, et cetera Mark Huber and Sarah Schott, CMC,Duke Adaptive MC integration 5/47

7 Higher dimensions, not so easy In 1-D, k intervals In 2-D, k 2 squares Mark Huber and Sarah Schott, CMC,Duke Adaptive MC integration 6/47

8 Running time exponential in number of dimensions In d dimensions, need k d evaluations Exponential growth in d Called The curse of dimensionality Monte Carlo Methods Use randomness to approximate integral Converge slowly: error = 1/square root of evaluations Error independent of # of dimensions! Mark Huber and Sarah Schott, CMC,Duke Adaptive MC integration 7/47

9 Many ways to go from samples to integrals Some Monte Carlo methods: (Sequential) Importance sampling Bridge sampling Path sampling Nested sampling Harmonic mean estimator Problem: these methods have unknown variance! SIS and HME variance can easily be infinite Estimated variance itself has unknown variance Mark Huber and Sarah Schott, CMC,Duke Adaptive MC integration 8/47

10 Our result We have a new method for approximating I = f ( x) dr n, f ( x) 0 x Ω with Î such that in time ( ) 1 P 1 + ɛ Î I 1 + ɛ > 1 δ O((log I) 2 ɛ 2 ln δ 1 ) Mark Huber and Sarah Schott, CMC,Duke Adaptive MC integration 9/47

11 Classic Monte Carlo Methods Mark Huber and Sarah Schott, CMC,Duke Adaptive MC integration 10/47

12 Abstract problem Create measure of a set A using integral: µ(a) = f ( x)d x A Mark Huber and Sarah Schott, CMC,Duke Adaptive MC integration 11/47

13 Acceptance/Rejection Finding the measure of a set Green area = B Black area = A µ(b) = µ(a) µ(b) µ(a) Mark Huber and Sarah Schott, CMC,Duke Adaptive MC integration 12/47

14 Acceptance/Rejection a.k.a Shoot at it randomly Mark Huber and Sarah Schott, CMC,Duke Adaptive MC integration 13/47

15 Acceptance/Rejection a.k.a Shoot at it randomly Mark Huber and Sarah Schott, CMC,Duke Adaptive MC integration 13/47

16 Acceptance/Rejection a.k.a Shoot at it randomly Mark Huber and Sarah Schott, CMC,Duke Adaptive MC integration 13/47

17 Acceptance/Rejection a.k.a Shoot at it randomly Mark Huber and Sarah Schott, CMC,Duke Adaptive MC integration 13/47

18 Acceptance/Rejection a.k.a Shoot at it randomly Mark Huber and Sarah Schott, CMC,Duke Adaptive MC integration 13/47

19 Acceptance/Rejection a.k.a Shoot at it randomly Mark Huber and Sarah Schott, CMC,Duke Adaptive MC integration 13/47

20 Acceptance/Rejection a.k.a Shoot at it randomly Best estimate: ˆµ(B) = 7 µ(a), A = black rectangle inside region 2 Mark Huber and Sarah Schott, CMC,Duke Adaptive MC integration 13/47

21 Analysis of Accept/Reject Algorithm Fire n times at box Say you hit H times Estimate: â = µ(a)n/h Mark Huber and Sarah Schott, CMC,Duke Adaptive MC integration 14/47

22 How well does it work? True answer: After 10 iterations 5.04 After 10 3 iterations After 10 5 iterations After 10 7 iterations About a factor of 100 per extra digit Mark Huber and Sarah Schott, CMC,Duke Adaptive MC integration 15/47

23 How well does it work? True answer: After 10 iterations 5.04 After 10 3 iterations After 10 5 iterations After 10 7 iterations About a factor of 100 per extra digit Mark Huber and Sarah Schott, CMC,Duke Adaptive MC integration 15/47

24 How well does it work? True answer: After 10 iterations 5.04 After 10 3 iterations After 10 5 iterations After 10 7 iterations About a factor of 100 per extra digit Mark Huber and Sarah Schott, CMC,Duke Adaptive MC integration 15/47

25 How well does it work? True answer: After 10 iterations 5.04 After 10 3 iterations After 10 5 iterations After 10 7 iterations About a factor of 100 per extra digit Mark Huber and Sarah Schott, CMC,Duke Adaptive MC integration 15/47

26 How well does it work? True answer: After 10 iterations 5.04 After 10 3 iterations After 10 5 iterations After 10 7 iterations About a factor of 100 per extra digit Mark Huber and Sarah Schott, CMC,Duke Adaptive MC integration 15/47

27 Bounding error Get relative error below ɛ with probability at least 1 δ: Need to analyze tails of binomial distribution to show: n 2(1 + ɛ) p 1 ɛ 2 ln 1 δ, p = µ(a) µ(b) The (1/ɛ 2 ) ln(1/δ) often called Monte Carlo error Best you can do in general So concentrate on improving 1/p part Mark Huber and Sarah Schott, CMC,Duke Adaptive MC integration 16/47

28 When p small, runtime large Usually p exponentially small in dimension of problem Mark Huber and Sarah Schott, CMC,Duke Adaptive MC integration 17/47

29 Running times Acceptance/Rejection: Product Estimator [1]: 2 1 p. [ 192 log 1 ] 2. p New method TPA: 2[log 1/p] 2 Mark Huber and Sarah Schott, CMC,Duke Adaptive MC integration 18/47

30 TPA Mark Huber and Sarah Schott, CMC,Duke Adaptive MC integration 19/47

31 TPA Idea Product estimator......plus idea from Nested sampling Result Product estimator with random cooling schedule Output can be analyzed exactly (like A/R) Mark Huber and Sarah Schott, CMC,Duke Adaptive MC integration 20/47

32 The Tootsie Pop Algorithm What is a Tootsie Pop? Hard candy lollipops with a tootsie roll (chewy chocolate) at the center In 1970, Mr. Owl was asked the question: How many licks does it take to get to the center of a Tootsie Pop? Mark Huber and Sarah Schott, CMC,Duke Adaptive MC integration 21/47

33 List of ingredients of TPA (a) A measure space (Ω, F, µ) (b) Two measurable sets: the center B and the shell B with B B (c) A family of sets {A(β)} where 1 β < β implies A(β ) A(β), 2 µ(a(β)) is continuous in β (d) Two special values β B and β B with A(β B ) = B and A(β B ) = B. Mark Huber and Sarah Schott, CMC,Duke Adaptive MC integration 22/47

34 Example of nested sets A(β) = all points within distance β of center Mark Huber and Sarah Schott, CMC,Duke Adaptive MC integration 23/47

35 Idea behind TPA Step β β β B 2 Repeat 3 Draw X µ(a(β)) 4 β inf{β : X A(β )} 5 Until β β B Mark Huber and Sarah Schott, CMC,Duke Adaptive MC integration 24/47

36 Idea behind TPA Step β β β B 2 Repeat 3 Draw X µ(a(β)) 4 β inf{β : X A(β )} 5 Until β β B Mark Huber and Sarah Schott, CMC,Duke Adaptive MC integration 24/47

37 Idea behind TPA Step β β β B 2 Repeat 3 Draw X µ(a(β)) 4 β inf{β : X A(β )} 5 Until β β B Mark Huber and Sarah Schott, CMC,Duke Adaptive MC integration 24/47

38 Idea behind TPA Step β β β B 2 Repeat 3 Draw X µ(a(β)) 4 β inf{β : X A(β )} 5 Until β β B Mark Huber and Sarah Schott, CMC,Duke Adaptive MC integration 24/47

39 Idea behind TPA Step β β β B 2 Repeat 3 Draw X µ(a(β)) 4 β inf{β : X A(β )} 5 Until β β B Mark Huber and Sarah Schott, CMC,Duke Adaptive MC integration 24/47

40 Idea behind TPA Step β β β B 2 Repeat 3 Draw X µ(a(β)) 4 β inf{β : X A(β )} 5 Until β β B Mark Huber and Sarah Schott, CMC,Duke Adaptive MC integration 24/47

41 Idea behind TPA Step β β β B 2 Repeat 3 Draw X µ(a(β)) 4 β inf{β : X A(β )} 5 Until β β B Mark Huber and Sarah Schott, CMC,Duke Adaptive MC integration 24/47

42 Idea behind TPA Step β β β B 2 Repeat 3 Draw X µ(a(β)) 4 β inf{β : X A(β )} 5 Until β β B Mark Huber and Sarah Schott, CMC,Duke Adaptive MC integration 24/47

43 Idea behind TPA Step β β β B 2 Repeat 3 Draw X µ(a(β)) 4 β inf{β : X A(β )} 5 Until β β B Mark Huber and Sarah Schott, CMC,Duke Adaptive MC integration 24/47

44 Idea behind TPA Step β β β B 2 Repeat 3 Draw X µ(a(β)) 4 β inf{β : X A(β )} 5 Until β β B Mark Huber and Sarah Schott, CMC,Duke Adaptive MC integration 24/47

45 How much is shaved off at each step? Notation: Z (β) := µ(a(β)) Lemma Say X µ(a(β)) and β = min{β : X A(β )}. Then Z (β ) Z (β) Unif([0, 1]) Proof by picture: Let b satisfy Z (b)/z (β) = 1/3 Then ( Z (β ) ) P Z (β) 1/3 = P(X A(b)) Mark Huber and Sarah Schott, CMC,Duke Adaptive MC integration 25/47

46 Each step removes on average 1/2 the measure Continue until reach center Original measure µ(b) = Z (β B ) After k steps measure Z (β k ) = Z (β B )r 1 r 2 r k, iid where r i Unif([0, 1]) Recall β B index of center Let l be number of steps until hit center l := min{k : Z (β B )r 1 r k < Z (β B )} 1 Question: what is the distribution of l? Mark Huber and Sarah Schott, CMC,Duke Adaptive MC integration 26/47

47 Logarithms Recall if U Unif([0, 1]), Since Consider the points ( ) Z (βk ) P i = ln Z (β B ) ln U Exp(1) Z (β k ) Z (β B ) r 1r 2 r k, where r i iid Unif([0, 1]), e 1 + e e k, where e i iid Exp(1) Mark Huber and Sarah Schott, CMC,Duke Adaptive MC integration 27/47

48 The Poisson point process ln Z (β B ) ln Z (β B ) Better to work in ln Z (β i ) space Recall: U Unif([0, 1]) ln U Exp(1) In log space, each step moves down Exp(1) Result: a Poisson point process from ln Z (β B ) to ln Z (β B ) Mark Huber and Sarah Schott, CMC,Duke Adaptive MC integration 28/47

49 The Poisson point process ln Z (β B ) ln Z (β 1 ) ln Z (β B ) Better to work in ln Z (β i ) space Recall: U Unif([0, 1]) ln U Exp(1) In log space, each step moves down Exp(1) Result: a Poisson point process from ln Z (β B ) to ln Z (β B ) Mark Huber and Sarah Schott, CMC,Duke Adaptive MC integration 28/47

50 The Poisson point process ln Z (β B ) ln Z (β 1 ) ln Z (β 2 ) ln Z (β B ) Better to work in ln Z (β i ) space Recall: U Unif([0, 1]) ln U Exp(1) In log space, each step moves down Exp(1) Result: a Poisson point process from ln Z (β B ) to ln Z (β B ) Mark Huber and Sarah Schott, CMC,Duke Adaptive MC integration 28/47

51 The Poisson point process ln Z (β B ) ln Z (β 1 ) ln Z (β 2 ) ln Z (β 3 ) ln Z (β B ) Better to work in ln Z (β i ) space Recall: U Unif([0, 1]) ln U Exp(1) In log space, each step moves down Exp(1) Result: a Poisson point process from ln Z (β B ) to ln Z (β B ) Mark Huber and Sarah Schott, CMC,Duke Adaptive MC integration 28/47

52 The Poisson point process ln Z (β B ) ln Z (β 1 ) ln Z (β 2 ) ln Z (β 3 ) ln Z (β B ) Better to work in ln Z (β i ) space Recall: U Unif([0, 1]) ln U Exp(1) In log space, each step moves down Exp(1) Result: a Poisson point process from ln Z (β B ) to ln Z (β B ) Mark Huber and Sarah Schott, CMC,Duke Adaptive MC integration 28/47

53 The result Output of TPA: l Pois(ln(Z (β B )/Z (β B ))) Output of A/R: H Bin(n, Z (β B )/Z (β B )) Mark Huber and Sarah Schott, CMC,Duke Adaptive MC integration 29/47

54 Repeating the Poisson point process Suppose run the Poisson point process twice Result also Poisson point process rate 2 instead of rate 1 Now run k times Result also Poisson point process rate k instead of rate 1 Final answer Pois(k ln(z (β B )/Z (β B ))) Divide by k, result close to ln[z (β B )/Z (β B )] Exponentiate, result close to Z (β B )/Z (β B ) Can use Chernoff s Bound to choose k large enough Mark Huber and Sarah Schott, CMC,Duke Adaptive MC integration 30/47

55 Bounding the tails Theorem Let p = Z (β B )/Z (β B ). For p < exp( 1) and ɛ <.3, after [ k = 2 ln 1 ] ( 3 p ɛ + 1 ) ɛ 2 ln 1 2δ runs, each of which uses on average ln(1/p) samples, the output ˆp satisfies: P((1 + ɛ) 1 ˆp/p 1 + ɛ) > 1 δ. Mark Huber and Sarah Schott, CMC,Duke Adaptive MC integration 31/47

56 Bonus: Approximate for all parameters simultaneously Can cut Poisson point process at any point: Right half still Poisson point process Yields omnithermal approximation Approximate Z (β)/z (β B ) for all β [β B, β B ] at same time Number of runs still same: [ k = 2 ln 1 ] ( 3 p ɛ + 1 ) ɛ 2 ln 1 2δ Mark Huber and Sarah Schott, CMC,Duke Adaptive MC integration 32/47

57 Proof idea: Poisson process Let N(t) be rate r Poisson process N(t) rt is a right continuous martingale Omnithermal approximation valid means did not drift too far away from 0 Mark Huber and Sarah Schott, CMC,Duke Adaptive MC integration 33/47

58 Examples Mark Huber and Sarah Schott, CMC,Duke Adaptive MC integration 34/47

59 Example 1: Mixture Gaussian Spikes Multimodal toy example Prior uniform over cube Likelihood mixture of two normals Small spike centered at (0, 0,..., 0) Large spike centered at (0.2, 0.2,..., 0.2) p θ Unif([ 1/2, 1/2] d ) d 1 L(θ) = 100 exp ( (θ i 0.2) 2 ) 2πu 2u 2 + i=1 d i=1 ( ) 1 exp θ2 i 2πv 2v 2 Mark Huber and Sarah Schott, CMC,Duke Adaptive MC integration 35/47

60 Parameter truncation Create family by limiting distance to center of small spike Mark Huber and Sarah Schott, CMC,Duke Adaptive MC integration 36/47

61 Running time results Problem: d = 20, u =.01, v =.02 True value: ln(1/p) = Algorithm (10 5 runs): ln(1/p) Running time for Example 1 Frequency Number of steps Mark Huber and Sarah Schott, CMC,Duke Adaptive MC integration 37/47

62 Example 2: Beta-binomial model Hierarchical model Data set: free throw numbers for 429 NBA players Example data point: Kobe Bryant made 483 out of 564 Model: number made by player i is Bin(n i, p i ) n i are known, p i Beta(a, b) Hyperparameters a and b, a 1 + Exp(1), b 1 + Exp(1) Density of percentage of free throws made Density N = 429 Bandwidth = Mark Huber and Sarah Schott, CMC,Duke Adaptive MC integration 38/47

63 Again use parameter trunctation Goal: find integrated likelihood Use β to limit distance from mode 2-D Unimodal problem so sampling easy True value (via numerical integration) After 10 5 runs Mark Huber and Sarah Schott, CMC,Duke Adaptive MC integration 39/47

64 Example 3: Ising model Besag[1974] modeled soil plots as good (green) or bad (red) h(x) = 13 (# adj like colored plots) π(x) = exp(2βh(x)) Z (β) parameter β is inv temp Mark Huber and Sarah Schott, CMC,Duke Adaptive MC integration 40/47

65 Integrated likelihood for Ising Parameter space one dimensional Z = 0 p β (b) exp(2bh(x)) Z (b) db, easy to do numerically if you know Z (β) over (0, ). Mark Huber and Sarah Schott, CMC,Duke Adaptive MC integration 41/47

66 Use omnithermal approximation ln(z β ) One run of TPA β Mark Huber and Sarah Schott, CMC,Duke Adaptive MC integration 42/47

67 Use omnithermal approximation ln(z β ) Sixteen runs of TPA β Mark Huber and Sarah Schott, CMC,Duke Adaptive MC integration 43/47

68 Connection to MCMC Several sampling methods use temperatures Simulated annealing Simulated tempering TPA easy for these problems Can speed up chain by giving well balanced cooling schedule Mark Huber and Sarah Schott, CMC,Duke Adaptive MC integration 44/47

69 Future directions Improvement for Gibbs distributions A Gibbs distribution: π({x}) = exp( βh(i)) Z (β) Can improve algorithm running time for Gibbs: O ((ln(1/p))ɛ 2 ln δ 1 ) Mark Huber and Sarah Schott, CMC,Duke Adaptive MC integration 45/47

70 Conclusions Randomized adaptive cooling schedules Guaranteed performance bound for MC integration (No variance estimate or unknown derivatives appear) Speed: 2[ln(1/p)] 2 much better than previous methods Speed: O ([ln(1/p)]) even better for Gibbs distributions Future directions Extend ln(1/p) method to non-gibbs distributions Remove extra ln(n) factors Mark Huber and Sarah Schott, CMC,Duke Adaptive MC integration 46/47

71 References M. Jerrum, L. Valiant, and V. Vazirani, Random generation of combinatorial structures from a uniform distribuiton Theoret. Comput. Sci., 43, , 1986 M. Huber and S. Schott, Using TPA for Bayesian inference Bayesian Statistics 9 J. Skilling, Nested sampling for general Bayesian computation, Bayesian Anal., 1(4), , 2006 D. Stefankovi c, S. Vempala and E. Vigoda, Adaptive Simulated Annealing: A Near-Optimal Connection between Sampling and Counting, J. of the ACM, 56(3), 1 36, 2009 Mark Huber and Sarah Schott, CMC,Duke Adaptive MC integration 47/47

Nests and Tootsie Pops: Bayesian Sampling with Monte Carlo

Nests and Tootsie Pops: Bayesian Sampling with Monte Carlo Monte Carlo Methods Final Project Nests and Tootsie Pops: Bayesian Sampling with Monte Carlo Authors: Michael Khanarian David Alvarez May 21, 2013 Abstract We explore two methods for the problem of computing

More information

TPA: A New Method for Approximate Counting

TPA: A New Method for Approximate Counting TPA: A New Method for Approximate Counting by Sarah Schott Department of Mathematics Duke University Date: Approved: Mark Huber, Supervisor Jonathan Mattingly Mauro Maggioni James Nolen Dissertation submitted

More information

Numerical Integration Monte Carlo style

Numerical Integration Monte Carlo style Numerical Integration Monte Carlo style Mark Huber Dept. of Mathematics and Inst. of Statistics and Decision Sciences Duke University mhuber@math.duke.edu www.math.duke.edu/~mhuber 1 Integration is hard

More information

Computational statistics

Computational statistics Computational statistics Markov Chain Monte Carlo methods Thierry Denœux March 2017 Thierry Denœux Computational statistics March 2017 1 / 71 Contents of this chapter When a target density f can be evaluated

More information

Bayesian Methods for Machine Learning

Bayesian Methods for Machine Learning Bayesian Methods for Machine Learning CS 584: Big Data Analytics Material adapted from Radford Neal s tutorial (http://ftp.cs.utoronto.ca/pub/radford/bayes-tut.pdf), Zoubin Ghahramni (http://hunch.net/~coms-4771/zoubin_ghahramani_bayesian_learning.pdf),

More information

Statistics 352: Spatial statistics. Jonathan Taylor. Department of Statistics. Models for discrete data. Stanford University.

Statistics 352: Spatial statistics. Jonathan Taylor. Department of Statistics. Models for discrete data. Stanford University. 352: 352: Models for discrete data April 28, 2009 1 / 33 Models for discrete data 352: Outline Dependent discrete data. Image data (binary). Ising models. Simulation: Gibbs sampling. Denoising. 2 / 33

More information

Stat 451 Lecture Notes Numerical Integration

Stat 451 Lecture Notes Numerical Integration Stat 451 Lecture Notes 03 12 Numerical Integration Ryan Martin UIC www.math.uic.edu/~rgmartin 1 Based on Chapter 5 in Givens & Hoeting, and Chapters 4 & 18 of Lange 2 Updated: February 11, 2016 1 / 29

More information

Why Water Freezes (A probabilistic view of phase transitions)

Why Water Freezes (A probabilistic view of phase transitions) Why Water Freezes (A probabilistic view of phase transitions) Mark Huber Dept. of Mathematics and Institute of Statistics and Decision Sciences Duke University mhuber@math.duke.edu www.math.duke.edu/~mhuber

More information

CPSC 540: Machine Learning

CPSC 540: Machine Learning CPSC 540: Machine Learning More Approximate Inference Mark Schmidt University of British Columbia Winter 2018 Last Time: Approximate Inference We ve been discussing graphical models for density estimation,

More information

Perfect simulation for image analysis

Perfect simulation for image analysis Perfect simulation for image analysis Mark Huber Fletcher Jones Foundation Associate Professor of Mathematics and Statistics and George R. Roberts Fellow Mathematical Sciences Claremont McKenna College

More information

Computer intensive statistical methods Lecture 1

Computer intensive statistical methods Lecture 1 Computer intensive statistical methods Lecture 1 Jonas Wallin Chalmers, Gothenburg. Jonas Wallin - jonwal@chalmers.se 1/27 People and literature The following people are involved in the course: Function

More information

Markov Chain Monte Carlo Inference. Siamak Ravanbakhsh Winter 2018

Markov Chain Monte Carlo Inference. Siamak Ravanbakhsh Winter 2018 Graphical Models Markov Chain Monte Carlo Inference Siamak Ravanbakhsh Winter 2018 Learning objectives Markov chains the idea behind Markov Chain Monte Carlo (MCMC) two important examples: Gibbs sampling

More information

Perfect simulation of repulsive point processes

Perfect simulation of repulsive point processes Perfect simulation of repulsive point processes Mark Huber Department of Mathematical Sciences, Claremont McKenna College 29 November, 2011 Mark Huber, CMC Perfect simulation of repulsive point processes

More information

Bayesian Computation in Color-Magnitude Diagrams

Bayesian Computation in Color-Magnitude Diagrams Bayesian Computation in Color-Magnitude Diagrams SA, AA, PT, EE, MCMC and ASIS in CMDs Paul Baines Department of Statistics Harvard University October 19, 2009 Overview Motivation and Introduction Modelling

More information

PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 2: PROBABILITY DISTRIBUTIONS

PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 2: PROBABILITY DISTRIBUTIONS PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 2: PROBABILITY DISTRIBUTIONS Parametric Distributions Basic building blocks: Need to determine given Representation: or? Recall Curve Fitting Binary Variables

More information

Machine Learning for Data Science (CS4786) Lecture 24

Machine Learning for Data Science (CS4786) Lecture 24 Machine Learning for Data Science (CS4786) Lecture 24 Graphical Models: Approximate Inference Course Webpage : http://www.cs.cornell.edu/courses/cs4786/2016sp/ BELIEF PROPAGATION OR MESSAGE PASSING Each

More information

6 Markov Chain Monte Carlo (MCMC)

6 Markov Chain Monte Carlo (MCMC) 6 Markov Chain Monte Carlo (MCMC) The underlying idea in MCMC is to replace the iid samples of basic MC methods, with dependent samples from an ergodic Markov chain, whose limiting (stationary) distribution

More information

Review. DS GA 1002 Statistical and Mathematical Models. Carlos Fernandez-Granda

Review. DS GA 1002 Statistical and Mathematical Models.   Carlos Fernandez-Granda Review DS GA 1002 Statistical and Mathematical Models http://www.cims.nyu.edu/~cfgranda/pages/dsga1002_fall16 Carlos Fernandez-Granda Probability and statistics Probability: Framework for dealing with

More information

April 20th, Advanced Topics in Machine Learning California Institute of Technology. Markov Chain Monte Carlo for Machine Learning

April 20th, Advanced Topics in Machine Learning California Institute of Technology. Markov Chain Monte Carlo for Machine Learning for for Advanced Topics in California Institute of Technology April 20th, 2017 1 / 50 Table of Contents for 1 2 3 4 2 / 50 History of methods for Enrico Fermi used to calculate incredibly accurate predictions

More information

27 : Distributed Monte Carlo Markov Chain. 1 Recap of MCMC and Naive Parallel Gibbs Sampling

27 : Distributed Monte Carlo Markov Chain. 1 Recap of MCMC and Naive Parallel Gibbs Sampling 10-708: Probabilistic Graphical Models 10-708, Spring 2014 27 : Distributed Monte Carlo Markov Chain Lecturer: Eric P. Xing Scribes: Pengtao Xie, Khoa Luu In this scribe, we are going to review the Parallel

More information

Monte Carlo methods for sampling-based Stochastic Optimization

Monte Carlo methods for sampling-based Stochastic Optimization Monte Carlo methods for sampling-based Stochastic Optimization Gersende FORT LTCI CNRS & Telecom ParisTech Paris, France Joint works with B. Jourdain, T. Lelièvre, G. Stoltz from ENPC and E. Kuhn from

More information

Ages of stellar populations from color-magnitude diagrams. Paul Baines. September 30, 2008

Ages of stellar populations from color-magnitude diagrams. Paul Baines. September 30, 2008 Ages of stellar populations from color-magnitude diagrams Paul Baines Department of Statistics Harvard University September 30, 2008 Context & Example Welcome! Today we will look at using hierarchical

More information

Markov chain Monte Carlo

Markov chain Monte Carlo 1 / 26 Markov chain Monte Carlo Timothy Hanson 1 and Alejandro Jara 2 1 Division of Biostatistics, University of Minnesota, USA 2 Department of Statistics, Universidad de Concepción, Chile IAP-Workshop

More information

Monte Carlo Methods. Leon Gu CSD, CMU

Monte Carlo Methods. Leon Gu CSD, CMU Monte Carlo Methods Leon Gu CSD, CMU Approximate Inference EM: y-observed variables; x-hidden variables; θ-parameters; E-step: q(x) = p(x y, θ t 1 ) M-step: θ t = arg max E q(x) [log p(y, x θ)] θ Monte

More information

Notes on Machine Learning for and

Notes on Machine Learning for and Notes on Machine Learning for 16.410 and 16.413 (Notes adapted from Tom Mitchell and Andrew Moore.) Choosing Hypotheses Generally want the most probable hypothesis given the training data Maximum a posteriori

More information

Introduction to Machine Learning CMU-10701

Introduction to Machine Learning CMU-10701 Introduction to Machine Learning CMU-10701 Markov Chain Monte Carlo Methods Barnabás Póczos & Aarti Singh Contents Markov Chain Monte Carlo Methods Goal & Motivation Sampling Rejection Importance Markov

More information

Bayesian Model Comparison:

Bayesian Model Comparison: Bayesian Model Comparison: Modeling Petrobrás log-returns Hedibert Freitas Lopes February 2014 Log price: y t = log p t Time span: 12/29/2000-12/31/2013 (n = 3268 days) LOG PRICE 1 2 3 4 0 500 1000 1500

More information

Stat 516, Homework 1

Stat 516, Homework 1 Stat 516, Homework 1 Due date: October 7 1. Consider an urn with n distinct balls numbered 1,..., n. We sample balls from the urn with replacement. Let N be the number of draws until we encounter a ball

More information

Outline. Binomial, Multinomial, Normal, Beta, Dirichlet. Posterior mean, MAP, credible interval, posterior distribution

Outline. Binomial, Multinomial, Normal, Beta, Dirichlet. Posterior mean, MAP, credible interval, posterior distribution Outline A short review on Bayesian analysis. Binomial, Multinomial, Normal, Beta, Dirichlet Posterior mean, MAP, credible interval, posterior distribution Gibbs sampling Revisit the Gaussian mixture model

More information

Stat 451 Lecture Notes Simulating Random Variables

Stat 451 Lecture Notes Simulating Random Variables Stat 451 Lecture Notes 05 12 Simulating Random Variables Ryan Martin UIC www.math.uic.edu/~rgmartin 1 Based on Chapter 6 in Givens & Hoeting, Chapter 22 in Lange, and Chapter 2 in Robert & Casella 2 Updated:

More information

arxiv: v1 [stat.co] 18 Feb 2012

arxiv: v1 [stat.co] 18 Feb 2012 A LEVEL-SET HIT-AND-RUN SAMPLER FOR QUASI-CONCAVE DISTRIBUTIONS Dean Foster and Shane T. Jensen arxiv:1202.4094v1 [stat.co] 18 Feb 2012 Department of Statistics The Wharton School University of Pennsylvania

More information

Curve Fitting Re-visited, Bishop1.2.5

Curve Fitting Re-visited, Bishop1.2.5 Curve Fitting Re-visited, Bishop1.2.5 Maximum Likelihood Bishop 1.2.5 Model Likelihood differentiation p(t x, w, β) = Maximum Likelihood N N ( t n y(x n, w), β 1). (1.61) n=1 As we did in the case of the

More information

Monte Carlo-based statistical methods (MASM11/FMS091)

Monte Carlo-based statistical methods (MASM11/FMS091) Monte Carlo-based statistical methods (MASM11/FMS091) Jimmy Olsson Centre for Mathematical Sciences Lund University, Sweden Lecture 12 MCMC for Bayesian computation II March 1, 2013 J. Olsson Monte Carlo-based

More information

Zig-Zag Monte Carlo. Delft University of Technology. Joris Bierkens February 7, 2017

Zig-Zag Monte Carlo. Delft University of Technology. Joris Bierkens February 7, 2017 Zig-Zag Monte Carlo Delft University of Technology Joris Bierkens February 7, 2017 Joris Bierkens (TU Delft) Zig-Zag Monte Carlo February 7, 2017 1 / 33 Acknowledgements Collaborators Andrew Duncan Paul

More information

Introduction to Markov Chain Monte Carlo & Gibbs Sampling

Introduction to Markov Chain Monte Carlo & Gibbs Sampling Introduction to Markov Chain Monte Carlo & Gibbs Sampling Prof. Nicholas Zabaras Sibley School of Mechanical and Aerospace Engineering 101 Frank H. T. Rhodes Hall Ithaca, NY 14853-3801 Email: zabaras@cornell.edu

More information

Normalising constants and maximum likelihood inference

Normalising constants and maximum likelihood inference Normalising constants and maximum likelihood inference Jakob G. Rasmussen Department of Mathematics Aalborg University Denmark March 9, 2011 1/14 Today Normalising constants Approximation of normalising

More information

A New Approximation Scheme for Monte Carlo Applications

A New Approximation Scheme for Monte Carlo Applications Claremont Colleges Scholarship @ Claremont CMC Senior Theses CMC Student Scholarship 2017 A New Approximation Scheme for Monte Carlo Applications Bo Jones Claremont McKenna College Recommended Citation

More information

Introduction to Probabilistic Machine Learning

Introduction to Probabilistic Machine Learning Introduction to Probabilistic Machine Learning Piyush Rai Dept. of CSE, IIT Kanpur (Mini-course 1) Nov 03, 2015 Piyush Rai (IIT Kanpur) Introduction to Probabilistic Machine Learning 1 Machine Learning

More information

Markov Chain Monte Carlo Lecture 6

Markov Chain Monte Carlo Lecture 6 Sequential parallel tempering With the development of science and technology, we more and more need to deal with high dimensional systems. For example, we need to align a group of protein or DNA sequences

More information

COS513 LECTURE 8 STATISTICAL CONCEPTS

COS513 LECTURE 8 STATISTICAL CONCEPTS COS513 LECTURE 8 STATISTICAL CONCEPTS NIKOLAI SLAVOV AND ANKUR PARIKH 1. MAKING MEANINGFUL STATEMENTS FROM JOINT PROBABILITY DISTRIBUTIONS. A graphical model (GM) represents a family of probability distributions

More information

Stat 451 Lecture Notes Monte Carlo Integration

Stat 451 Lecture Notes Monte Carlo Integration Stat 451 Lecture Notes 06 12 Monte Carlo Integration Ryan Martin UIC www.math.uic.edu/~rgmartin 1 Based on Chapter 6 in Givens & Hoeting, Chapter 23 in Lange, and Chapters 3 4 in Robert & Casella 2 Updated:

More information

Bayesian inference. Fredrik Ronquist and Peter Beerli. October 3, 2007

Bayesian inference. Fredrik Ronquist and Peter Beerli. October 3, 2007 Bayesian inference Fredrik Ronquist and Peter Beerli October 3, 2007 1 Introduction The last few decades has seen a growing interest in Bayesian inference, an alternative approach to statistical inference.

More information

Computer intensive statistical methods

Computer intensive statistical methods Lecture 13 MCMC, Hybrid chains October 13, 2015 Jonas Wallin jonwal@chalmers.se Chalmers, Gothenburg university MH algorithm, Chap:6.3 The metropolis hastings requires three objects, the distribution of

More information

Reminder of some Markov Chain properties:

Reminder of some Markov Chain properties: Reminder of some Markov Chain properties: 1. a transition from one state to another occurs probabilistically 2. only state that matters is where you currently are (i.e. given present, future is independent

More information

Session 5B: A worked example EGARCH model

Session 5B: A worked example EGARCH model Session 5B: A worked example EGARCH model John Geweke Bayesian Econometrics and its Applications August 7, worked example EGARCH model August 7, / 6 EGARCH Exponential generalized autoregressive conditional

More information

Bayesian model selection in graphs by using BDgraph package

Bayesian model selection in graphs by using BDgraph package Bayesian model selection in graphs by using BDgraph package A. Mohammadi and E. Wit March 26, 2013 MOTIVATION Flow cytometry data with 11 proteins from Sachs et al. (2005) RESULT FOR CELL SIGNALING DATA

More information

Midterm Examination. STA 215: Statistical Inference. Due Wednesday, 2006 Mar 8, 1:15 pm

Midterm Examination. STA 215: Statistical Inference. Due Wednesday, 2006 Mar 8, 1:15 pm Midterm Examination STA 215: Statistical Inference Due Wednesday, 2006 Mar 8, 1:15 pm This is an open-book take-home examination. You may work on it during any consecutive 24-hour period you like; please

More information

Classical and Bayesian inference

Classical and Bayesian inference Classical and Bayesian inference AMS 132 January 18, 2018 Claudia Wehrhahn (UCSC) Classical and Bayesian inference January 18, 2018 1 / 9 Sampling from a Bernoulli Distribution Theorem (Beta-Bernoulli

More information

Stat 5421 Lecture Notes Proper Conjugate Priors for Exponential Families Charles J. Geyer March 28, 2016

Stat 5421 Lecture Notes Proper Conjugate Priors for Exponential Families Charles J. Geyer March 28, 2016 Stat 5421 Lecture Notes Proper Conjugate Priors for Exponential Families Charles J. Geyer March 28, 2016 1 Theory This section explains the theory of conjugate priors for exponential families of distributions,

More information

7.1 Coupling from the Past

7.1 Coupling from the Past Georgia Tech Fall 2006 Markov Chain Monte Carlo Methods Lecture 7: September 12, 2006 Coupling from the Past Eric Vigoda 7.1 Coupling from the Past 7.1.1 Introduction We saw in the last lecture how Markov

More information

Prior Choice, Summarizing the Posterior

Prior Choice, Summarizing the Posterior Prior Choice, Summarizing the Posterior Statistics 220 Spring 2005 Copyright c 2005 by Mark E. Irwin Informative Priors Binomial Model: y π Bin(n, π) π is the success probability. Need prior p(π) Bayes

More information

Computer intensive statistical methods

Computer intensive statistical methods Lecture 11 Markov Chain Monte Carlo cont. October 6, 2015 Jonas Wallin jonwal@chalmers.se Chalmers, Gothenburg university The two stage Gibbs sampler If the conditional distributions are easy to sample

More information

Markov Chain Monte Carlo (MCMC) and Model Evaluation. August 15, 2017

Markov Chain Monte Carlo (MCMC) and Model Evaluation. August 15, 2017 Markov Chain Monte Carlo (MCMC) and Model Evaluation August 15, 2017 Frequentist Linking Frequentist and Bayesian Statistics How can we estimate model parameters and what does it imply? Want to find the

More information

Near-linear time simulation of linear extensions of a height-2 poset with bounded interaction

Near-linear time simulation of linear extensions of a height-2 poset with bounded interaction CHICAGO JOURNAL OF THEORETICAL COMPUTER SCIENCE 2014, Article 03, pages 1 16 http://cjtcs.cs.uchicago.edu/ Near-linear time simulation of linear extensions of a height-2 poset with bounded interaction

More information

Stat 451 Lecture Notes Markov Chain Monte Carlo. Ryan Martin UIC

Stat 451 Lecture Notes Markov Chain Monte Carlo. Ryan Martin UIC Stat 451 Lecture Notes 07 12 Markov Chain Monte Carlo Ryan Martin UIC www.math.uic.edu/~rgmartin 1 Based on Chapters 8 9 in Givens & Hoeting, Chapters 25 27 in Lange 2 Updated: April 4, 2016 1 / 42 Outline

More information

PARAMETER ESTIMATION: BAYESIAN APPROACH. These notes summarize the lectures on Bayesian parameter estimation.

PARAMETER ESTIMATION: BAYESIAN APPROACH. These notes summarize the lectures on Bayesian parameter estimation. PARAMETER ESTIMATION: BAYESIAN APPROACH. These notes summarize the lectures on Bayesian parameter estimation.. Beta Distribution We ll start by learning about the Beta distribution, since we end up using

More information

Stein Variational Gradient Descent: A General Purpose Bayesian Inference Algorithm

Stein Variational Gradient Descent: A General Purpose Bayesian Inference Algorithm Stein Variational Gradient Descent: A General Purpose Bayesian Inference Algorithm Qiang Liu and Dilin Wang NIPS 2016 Discussion by Yunchen Pu March 17, 2017 March 17, 2017 1 / 8 Introduction Let x R d

More information

Integrated Non-Factorized Variational Inference

Integrated Non-Factorized Variational Inference Integrated Non-Factorized Variational Inference Shaobo Han, Xuejun Liao and Lawrence Carin Duke University February 27, 2014 S. Han et al. Integrated Non-Factorized Variational Inference February 27, 2014

More information

Statistics 135 Fall 2007 Midterm Exam

Statistics 135 Fall 2007 Midterm Exam Name: Student ID Number: Statistics 135 Fall 007 Midterm Exam Ignore the finite population correction in all relevant problems. The exam is closed book, but some possibly useful facts about probability

More information

Machine Learning CSE546 Carlos Guestrin University of Washington. September 30, 2013

Machine Learning CSE546 Carlos Guestrin University of Washington. September 30, 2013 Bayesian Methods Machine Learning CSE546 Carlos Guestrin University of Washington September 30, 2013 1 What about prior n Billionaire says: Wait, I know that the thumbtack is close to 50-50. What can you

More information

Motivation Scale Mixutres of Normals Finite Gaussian Mixtures Skew-Normal Models. Mixture Models. Econ 690. Purdue University

Motivation Scale Mixutres of Normals Finite Gaussian Mixtures Skew-Normal Models. Mixture Models. Econ 690. Purdue University Econ 690 Purdue University In virtually all of the previous lectures, our models have made use of normality assumptions. From a computational point of view, the reason for this assumption is clear: combined

More information

Bayesian Inference: Posterior Intervals

Bayesian Inference: Posterior Intervals Bayesian Inference: Posterior Intervals Simple values like the posterior mean E[θ X] and posterior variance var[θ X] can be useful in learning about θ. Quantiles of π(θ X) (especially the posterior median)

More information

CSC 2541: Bayesian Methods for Machine Learning

CSC 2541: Bayesian Methods for Machine Learning CSC 2541: Bayesian Methods for Machine Learning Radford M. Neal, University of Toronto, 2011 Lecture 3 More Markov Chain Monte Carlo Methods The Metropolis algorithm isn t the only way to do MCMC. We ll

More information

Counting Linear Extensions of a Partial Order

Counting Linear Extensions of a Partial Order Counting Linear Extensions of a Partial Order Seth Harris March 6, 20 Introduction A partially ordered set (P,

More information

Probabilistic Graphical Models Lecture 17: Markov chain Monte Carlo

Probabilistic Graphical Models Lecture 17: Markov chain Monte Carlo Probabilistic Graphical Models Lecture 17: Markov chain Monte Carlo Andrew Gordon Wilson www.cs.cmu.edu/~andrewgw Carnegie Mellon University March 18, 2015 1 / 45 Resources and Attribution Image credits,

More information

Part 1: Expectation Propagation

Part 1: Expectation Propagation Chalmers Machine Learning Summer School Approximate message passing and biomedicine Part 1: Expectation Propagation Tom Heskes Machine Learning Group, Institute for Computing and Information Sciences Radboud

More information

Monte Carlo Methods for Computation and Optimization (048715)

Monte Carlo Methods for Computation and Optimization (048715) Technion Department of Electrical Engineering Monte Carlo Methods for Computation and Optimization (048715) Lecture Notes Prof. Nahum Shimkin Spring 2015 i PREFACE These lecture notes are intended for

More information

On Bayesian Computation

On Bayesian Computation On Bayesian Computation Michael I. Jordan with Elaine Angelino, Maxim Rabinovich, Martin Wainwright and Yun Yang Previous Work: Information Constraints on Inference Minimize the minimax risk under constraints

More information

Part III. A Decision-Theoretic Approach and Bayesian testing

Part III. A Decision-Theoretic Approach and Bayesian testing Part III A Decision-Theoretic Approach and Bayesian testing 1 Chapter 10 Bayesian Inference as a Decision Problem The decision-theoretic framework starts with the following situation. We would like to

More information

MCMC: Markov Chain Monte Carlo

MCMC: Markov Chain Monte Carlo I529: Machine Learning in Bioinformatics (Spring 2013) MCMC: Markov Chain Monte Carlo Yuzhen Ye School of Informatics and Computing Indiana University, Bloomington Spring 2013 Contents Review of Markov

More information

Principles of Bayesian Inference

Principles of Bayesian Inference Principles of Bayesian Inference Sudipto Banerjee University of Minnesota July 20th, 2008 1 Bayesian Principles Classical statistics: model parameters are fixed and unknown. A Bayesian thinks of parameters

More information

Theory of Stochastic Processes 8. Markov chain Monte Carlo

Theory of Stochastic Processes 8. Markov chain Monte Carlo Theory of Stochastic Processes 8. Markov chain Monte Carlo Tomonari Sei sei@mist.i.u-tokyo.ac.jp Department of Mathematical Informatics, University of Tokyo June 8, 2017 http://www.stat.t.u-tokyo.ac.jp/~sei/lec.html

More information

Lecture 8: Bayesian Estimation of Parameters in State Space Models

Lecture 8: Bayesian Estimation of Parameters in State Space Models in State Space Models March 30, 2016 Contents 1 Bayesian estimation of parameters in state space models 2 Computational methods for parameter estimation 3 Practical parameter estimation in state space

More information

Bayesian Learning. HT2015: SC4 Statistical Data Mining and Machine Learning. Maximum Likelihood Principle. The Bayesian Learning Framework

Bayesian Learning. HT2015: SC4 Statistical Data Mining and Machine Learning. Maximum Likelihood Principle. The Bayesian Learning Framework HT5: SC4 Statistical Data Mining and Machine Learning Dino Sejdinovic Department of Statistics Oxford http://www.stats.ox.ac.uk/~sejdinov/sdmml.html Maximum Likelihood Principle A generative model for

More information

Announcements. Proposals graded

Announcements. Proposals graded Announcements Proposals graded Kevin Jamieson 2018 1 Bayesian Methods Machine Learning CSE546 Kevin Jamieson University of Washington November 1, 2018 2018 Kevin Jamieson 2 MLE Recap - coin flips Data:

More information

Lecture 8: The Metropolis-Hastings Algorithm

Lecture 8: The Metropolis-Hastings Algorithm 30.10.2008 What we have seen last time: Gibbs sampler Key idea: Generate a Markov chain by updating the component of (X 1,..., X p ) in turn by drawing from the full conditionals: X (t) j Two drawbacks:

More information

Some slides from Carlos Guestrin, Luke Zettlemoyer & K Gajos 2

Some slides from Carlos Guestrin, Luke Zettlemoyer & K Gajos 2 Logistics CSE 446: Point Estimation Winter 2012 PS2 out shortly Dan Weld Some slides from Carlos Guestrin, Luke Zettlemoyer & K Gajos 2 Last Time Random variables, distributions Marginal, joint & conditional

More information

STA 4273H: Statistical Machine Learning

STA 4273H: Statistical Machine Learning STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.utstat.utoronto.ca/~rsalakhu/ Sidney Smith Hall, Room 6002 Lecture 7 Approximate

More information

Theory of Maximum Likelihood Estimation. Konstantin Kashin

Theory of Maximum Likelihood Estimation. Konstantin Kashin Gov 2001 Section 5: Theory of Maximum Likelihood Estimation Konstantin Kashin February 28, 2013 Outline Introduction Likelihood Examples of MLE Variance of MLE Asymptotic Properties What is Statistical

More information

Bayesian GLMs and Metropolis-Hastings Algorithm

Bayesian GLMs and Metropolis-Hastings Algorithm Bayesian GLMs and Metropolis-Hastings Algorithm We have seen that with conjugate or semi-conjugate prior distributions the Gibbs sampler can be used to sample from the posterior distribution. In situations,

More information

Adaptive Monte Carlo methods

Adaptive Monte Carlo methods Adaptive Monte Carlo methods Jean-Michel Marin Projet Select, INRIA Futurs, Université Paris-Sud joint with Randal Douc (École Polytechnique), Arnaud Guillin (Université de Marseille) and Christian Robert

More information

A = {(x, u) : 0 u f(x)},

A = {(x, u) : 0 u f(x)}, Draw x uniformly from the region {x : f(x) u }. Markov Chain Monte Carlo Lecture 5 Slice sampler: Suppose that one is interested in sampling from a density f(x), x X. Recall that sampling x f(x) is equivalent

More information

Fundamentals. CS 281A: Statistical Learning Theory. Yangqing Jia. August, Based on tutorial slides by Lester Mackey and Ariel Kleiner

Fundamentals. CS 281A: Statistical Learning Theory. Yangqing Jia. August, Based on tutorial slides by Lester Mackey and Ariel Kleiner Fundamentals CS 281A: Statistical Learning Theory Yangqing Jia Based on tutorial slides by Lester Mackey and Ariel Kleiner August, 2011 Outline 1 Probability 2 Statistics 3 Linear Algebra 4 Optimization

More information

Brandon C. Kelly (Harvard Smithsonian Center for Astrophysics)

Brandon C. Kelly (Harvard Smithsonian Center for Astrophysics) Brandon C. Kelly (Harvard Smithsonian Center for Astrophysics) Probability quantifies randomness and uncertainty How do I estimate the normalization and logarithmic slope of a X ray continuum, assuming

More information

Kernel adaptive Sequential Monte Carlo

Kernel adaptive Sequential Monte Carlo Kernel adaptive Sequential Monte Carlo Ingmar Schuster (Paris Dauphine) Heiko Strathmann (University College London) Brooks Paige (Oxford) Dino Sejdinovic (Oxford) December 7, 2015 1 / 36 Section 1 Outline

More information

Lecture 6: September 22

Lecture 6: September 22 CS294 Markov Chain Monte Carlo: Foundations & Applications Fall 2009 Lecture 6: September 22 Lecturer: Prof. Alistair Sinclair Scribes: Alistair Sinclair Disclaimer: These notes have not been subjected

More information

Lecture 3. G. Cowan. Lecture 3 page 1. Lectures on Statistical Data Analysis

Lecture 3. G. Cowan. Lecture 3 page 1. Lectures on Statistical Data Analysis Lecture 3 1 Probability (90 min.) Definition, Bayes theorem, probability densities and their properties, catalogue of pdfs, Monte Carlo 2 Statistical tests (90 min.) general concepts, test statistics,

More information

(5) Multi-parameter models - Gibbs sampling. ST440/540: Applied Bayesian Analysis

(5) Multi-parameter models - Gibbs sampling. ST440/540: Applied Bayesian Analysis Summarizing a posterior Given the data and prior the posterior is determined Summarizing the posterior gives parameter estimates, intervals, and hypothesis tests Most of these computations are integrals

More information

A Search and Jump Algorithm for Markov Chain Monte Carlo Sampling. Christopher Jennison. Adriana Ibrahim. Seminar at University of Kuwait

A Search and Jump Algorithm for Markov Chain Monte Carlo Sampling. Christopher Jennison. Adriana Ibrahim. Seminar at University of Kuwait A Search and Jump Algorithm for Markov Chain Monte Carlo Sampling Christopher Jennison Department of Mathematical Sciences, University of Bath, UK http://people.bath.ac.uk/mascj Adriana Ibrahim Institute

More information

Deblurring Jupiter (sampling in GLIP faster than regularized inversion) Colin Fox Richard A. Norton, J.

Deblurring Jupiter (sampling in GLIP faster than regularized inversion) Colin Fox Richard A. Norton, J. Deblurring Jupiter (sampling in GLIP faster than regularized inversion) Colin Fox fox@physics.otago.ac.nz Richard A. Norton, J. Andrés Christen Topics... Backstory (?) Sampling in linear-gaussian hierarchical

More information

Quantitative Biology II Lecture 4: Variational Methods

Quantitative Biology II Lecture 4: Variational Methods 10 th March 2015 Quantitative Biology II Lecture 4: Variational Methods Gurinder Singh Mickey Atwal Center for Quantitative Biology Cold Spring Harbor Laboratory Image credit: Mike West Summary Approximate

More information

Metropolis-Hastings Algorithm

Metropolis-Hastings Algorithm Strength of the Gibbs sampler Metropolis-Hastings Algorithm Easy algorithm to think about. Exploits the factorization properties of the joint probability distribution. No difficult choices to be made to

More information

Nonparametric Bayesian Methods - Lecture I

Nonparametric Bayesian Methods - Lecture I Nonparametric Bayesian Methods - Lecture I Harry van Zanten Korteweg-de Vries Institute for Mathematics CRiSM Masterclass, April 4-6, 2016 Overview of the lectures I Intro to nonparametric Bayesian statistics

More information

Bayesian Complementary Clustering, MCMC and Anglo-Saxon placenames

Bayesian Complementary Clustering, MCMC and Anglo-Saxon placenames Bayesian Complementary Clustering, MCMC and Anglo-Saxon placenames Giacomo Zanella g.zanella@warwick.ac.uk Department of Statistics University of Warwick, Coventry, UK 20 May 2014 Overview 1. Motivation:

More information

Multimodal Nested Sampling

Multimodal Nested Sampling Multimodal Nested Sampling Farhan Feroz Astrophysics Group, Cavendish Lab, Cambridge Inverse Problems & Cosmology Most obvious example: standard CMB data analysis pipeline But many others: object detection,

More information

17 : Markov Chain Monte Carlo

17 : Markov Chain Monte Carlo 10-708: Probabilistic Graphical Models, Spring 2015 17 : Markov Chain Monte Carlo Lecturer: Eric P. Xing Scribes: Heran Lin, Bin Deng, Yun Huang 1 Review of Monte Carlo Methods 1.1 Overview Monte Carlo

More information

A Level-Set Hit-And-Run Sampler for Quasi- Concave Distributions

A Level-Set Hit-And-Run Sampler for Quasi- Concave Distributions University of Pennsylvania ScholarlyCommons Statistics Papers Wharton Faculty Research 2014 A Level-Set Hit-And-Run Sampler for Quasi- Concave Distributions Shane T. Jensen University of Pennsylvania Dean

More information

Bayes: All uncertainty is described using probability.

Bayes: All uncertainty is described using probability. Bayes: All uncertainty is described using probability. Let w be the data and θ be any unknown quantities. Likelihood. The probability model π(w θ) has θ fixed and w varying. The likelihood L(θ; w) is π(w

More information

Graphical Models and Kernel Methods

Graphical Models and Kernel Methods Graphical Models and Kernel Methods Jerry Zhu Department of Computer Sciences University of Wisconsin Madison, USA MLSS June 17, 2014 1 / 123 Outline Graphical Models Probabilistic Inference Directed vs.

More information

Notes on MapReduce Algorithms

Notes on MapReduce Algorithms Notes on MapReduce Algorithms Barna Saha 1 Finding Minimum Spanning Tree of a Dense Graph in MapReduce We are given a graph G = (V, E) on V = N vertices and E = m N 1+c edges for some constant c > 0. Our

More information