Stat 451 Lecture Notes Markov Chain Monte Carlo. Ryan Martin UIC

Size: px
Start display at page:

Download "Stat 451 Lecture Notes Markov Chain Monte Carlo. Ryan Martin UIC"

Transcription

1 Stat 451 Lecture Notes Markov Chain Monte Carlo Ryan Martin UIC 1 Based on Chapters 8 9 in Givens & Hoeting, Chapters in Lange 2 Updated: April 4, / 42

2 Outline 1 Introduction 2 Crash course on Markov chains 3 Motivation, revisited 4 Metropolis Hastings algorithm 5 Gibbs sampler 6 Some MCMC diagnostics 7 Conclusion 2 / 42

3 Motivation We know how to sample independent random variables from the target distribution f (x), at least approximately. Monte Carlo uses these simulated random variables to approximate integrals. But the random variables don t need to be independent in order to accurately approximate integrals! Markov Chain Monte Carlo (MCMC) constructs a dependent sequence of random variables that can be used to approximate the integrals just like for ordinary Monte Carlo. The advantage of introducing this dependence is that very general black-box algorithms (and corresponding theory) are available to perform the required simulations. 3 / 42

4 Initial remarks In some sense, MCMC is basically a black-box approach that works for almost all problems. The previous remark is dangerous it s always a bad idea to use tools without knowing that they will work. Here we will discuss some basics of Markov chains and MCMC but know that there are very important unanswered questions about how and when MCMC works. 3 3 MCMC is an active area of research; despite the many developments in MCMC, according to Diaconis ( 2008), still almost nothing is known! 4 / 42

5 Outline 1 Introduction 2 Crash course on Markov chains 3 Motivation, revisited 4 Metropolis Hastings algorithm 5 Gibbs sampler 6 Some MCMC diagnostics 7 Conclusion 5 / 42

6 Markov chains A Markov chain is just a sequence of random variables {X 1, X 2,...} with a specific type of dependence structure. In particular, a Markov chain satisfies P(X n+1 B X 1,..., X n ) = P(X n+1 B X n ), ( ) i.e., the future, given the past and present, depends only on the present. Independence is a trivial Markov chain. From ( ) we can argue that the probabilistic properties of the chain are completely determined by: initial distribution for X 0, and the transition distribution, distribution of X n+1, given X n. 4 4 Assume the Markov chain is homogeneous, so that the transition distribution does not depend on n. 6 / 42

7 Example simple random walk Let U 1, U 2,... iid Unif({ 1, 1}). Set X 0 = 0 and let X n = n i=1 U i = X n 1 + U n. The initial distribution is P{X 0 = 0} = 1. The transition distribution is determined by { X n 1 1 with prob 1/2 X n = X n with prob 1/2 While very simple, the random walk is an important example in probability, having connections to advanced things like Brownian motion. 7 / 42

8 Keywords 5 A state A is recurrent if a chain starting in A will eventually return to A with probability 1. This state is nonull if the expected time to return is finite. A chain is called recurrent if each state is recurrent. A Markov chain is irreducible if there is positive probability that a chain starting in a state A can reach any other state B. A Markov chain is aperiodic if, for a starting state A, there is no constraint on the times at which the chain can return to A. An irreducible, aperiodic Markov chain with all states being nonnull recurrent is called ergodic. 5 Not mathematically precise! 8 / 42

9 Limit theory 6 f is a stationary distribution if X 0 f implies X n f for all n. An ergodic Markov chain has at most one stationary dist. Furthermore, if the chain is ergodic, then lim P(X m+n B X m A) = f (x) dx, for all A, B, m. n Even further, if ϕ(x) is integrable, then 1 n n ϕ(x t ) t=1 B ϕ(x)f (x) dx, with prob 1. This is a version of the famous ergodic theorem. There are also central limit theorems for Markov chains, but I won t say anything about this. 6 Again, not mathematically precise! 9 / 42

10 Outline 1 Introduction 2 Crash course on Markov chains 3 Motivation, revisited 4 Metropolis Hastings algorithm 5 Gibbs sampler 6 Some MCMC diagnostics 7 Conclusion 10 / 42

11 Why MCMC? In Monte Carlo applications, we want to generate random variables with distribution f. This could be difficult or impossible to do exactly. MCMC is designed to construct an ergodic Markov chain with f as its stationary distribution. Asymptotically, the chain will resemble samples from f. In particular, by the ergodic theorem, expectations with respect to f can be approximated by averages. Somewhat surprising is that it is actually quite easy to construct and simulate a suitable Markov chain, explaining why MCMC methods have become so popular. But, of course, there are practical and theoretical challenges / 42

12 Outline 1 Introduction 2 Crash course on Markov chains 3 Motivation, revisited 4 Metropolis Hastings algorithm 5 Gibbs sampler 6 Some MCMC diagnostics 7 Conclusion 12 / 42

13 Details Let f (x) denote the target distribution pdf. Let q(x y) denote a conditional pdf for X, given Y = y; this pdf should be easy to sample from. Given X 0, the Metropolis Hasting algorithm (MH) produces a sequence of random variables as follows: 1 Sample X t q(x X t 1 ). 2 Compute { R = min 1, f (Xt ) q(x t 1 Xt ) f (X t 1 ) q(xt X t 1 ) 3 Set X t = X t with probability R; otherwise, X t = X t 1. General R code to implement MH is on course website. }. 13 / 42

14 Details (cont.) The proposal distribution is not easy to choose, and the performance of the algorithm depends on this choice. Two general strategies are: Take the proposal q(x y) = q(x); i.e., at each stage of the MH algorithm, X t does not depend on X t 1. Take q(x y) = q 0 (x y), for a symmetric distribution with pdf q 0, which amounts to a random walk proposal. This is one aspect of the MCMC implementation that requires a lot of care from the user; deeper understanding is needed to really see how the proposal affects the performance. In my examples, I will just pick a proposal that seems to work reasonably well / 42

15 Details (cont.) 7 Assuming the proposal is not too bad, then a number of things can be shown about the sequence {X t : t 1}: the chain is ergodic, and the target f is the stationary distribution. Consequently, the sequence converges to the stationary distribution and, for any integrable function ϕ(x), we can approximate integrals with sample averages. So, provided that we run the simulation long enough, we should be able to get arbitrarily good approximations. This is an interesting scenario where we, as statisticians, are able to control the sample size! 7 Again, not mathematically precise! 15 / 42

16 Example: cosine model Consider the problem from old homework, where the likelihood function is L(θ) n {1 cos(x i θ)}, π θ π. i=1 Observed data (X 1,..., X n ) given in the code. Assume that θ is given a Unif( π, π) prior distribution. Use MH to sample from the posterior: Proposal: q(θ θ) = Unif(θ θ ± 0.5). Burn-in: B = Sample size: M = / 42

17 Example: cosine model (cont.) Left figure shows histogram of the MCMC sample with posterior density overlaid. Right figure shows a trace plot of the chain. Density θ θ Iteration 17 / 42

18 Example: Weibull model Data X 1,..., X n has Weibull likelihood { L(α, η) α n η n exp α n log X i η i=1 Prior: π(α, η) e α η b 1 e cη, for some (b, c). Posterior density is proportional to { α n η n+b 1 exp α ( n i=1 n i=1 ) ( log X i 1 η c + } Xi α. n i=1 Exponential is a special case of Weibull, when α = 1. Goal: informal Bayesian test of H 0 : α = 1. )} Xi α. 18 / 42

19 Example: Weibull model (cont.) Data from Problem 7.11 in Ghosh et al (2006). Use MH to sample from posterior of (α, η). Proposal: (α, η ) (α, η) Exp(α) Exp(η). b = 2 and c = 1; B = 1000 and M = Histogram shows marginal posterior of α. Is an exponential model (α = 1) reasonable? Density α 19 / 42

20 Example: logistic regression Based on Examples 1.13 and 7.11 in Robert & Casella s book. In 1986, the Challenger space shuttle exploded during take-off, the result of an o-ring failure. Failure may have been due to the cold temperature (31 F). Goal: Analyze the relationship between temperature and o-ring failure. In particular, fit a logistic regression model. 20 / 42

21 Example: logistic regression (cont.) Model: Y x Ber(p(x)), x = temperature. Failure probability, p(x), is of the form p(x) = exp(α + βx) 1 + exp(α + βx). Using available data, fit logistic regression using glm. Coefficients: Estimate Std. Error z value Pr(> z ) (Intercept) * x * --- Signif. codes: 0 *** ** 0.01 * Note that p(31) 0.999!!! 21 / 42

22 Example: logistic regression (cont.) Can also do a Bayesian analysis of this logistic model. Use MH to obtain samples from the posterior of (α, β). Samples can be used to approximate the posterior distribution of p(x 0 ) for any fixed x 0, e.g., x 0 = 65, 31; see below. Details about the prior and proposal construction are given in the R code and a short write-up posted on the course website. Density Density p(65) p(31) 22 / 42

23 Outline 1 Introduction 2 Crash course on Markov chains 3 Motivation, revisited 4 Metropolis Hastings algorithm 5 Gibbs sampler 6 Some MCMC diagnostics 7 Conclusion 23 / 42

24 Setup Suppose we have a multivariate target distribution f. MH can be applied to such a problem, but there are challenges in constructing a good proposal over multiple dimensions. Idea: sample one dimension at a time. Question: How to carry out the sampling so that it will approximate the target, at least in a limit? Gibbs sampler is the right tool for the job. 24 / 42

25 Details Suppose we have a trivariate target f (x) = f (x 1, x 2, x 3 ). Suppose we can write down the set of full conditionals f (x 1 x 2, x 3 ), f (x 2 x 1, x 3 ), f (x 3 x 1, x 2 ) and that these can be sampled from. Gibbs sampler generates a sequence {X (t) : t 0} by iteratively sampling from the conditionals: X (t) 1 f (x 1 X (t 1) 2, X (t 1) 3 ) X (t) 2 f (x 2 X (t) 1, X (t 1) 3 ) X (t) 3 f (x 3 X (t) 1, X (t) 2 ). 25 / 42

26 Details (cont.) Gibbs sequence forms a Markov chain. In fact, Gibbs sampler is a special case of MH! Connection to MH is made by considering Gibbs as a sequence that updates one component of X at a time. The acceptance probability for this form of MH update is exactly 1, which explains why Gibbs sampler has no accept/reject step. Since Gibbs is a special kind of MH, the convergence for MH applies to Gibbs as well. 26 / 42

27 Example: bivariate normal A super-simple Gibbs example: bivariate normal. Suppose X = (X 1, X 2 ) is bivariate normal, correlation ρ. Full conditionals are easy to write down here. Gibbs steps: X (t) 1 N(ρX (t 1) 2, 1 ρ 2 ) X (t) 2 N(ρX (t) 1, 1 ρ2 ). Not as efficient as direct sampling, but works fine. 27 / 42

28 Example: many-normal-means ind Model: X i N(θ i, 1), i = 1,..., n. Hierarchical prior distribution: [(θ 1,..., θ n ) ψ] iid N(0, ψ 1 ), ψ Gamma(a, b). Takes some work, but it can be shown 8 that the full conditionals are ( Xi θ i (X i, ψ) ind N 1 + ψ, ψ ( ψ 1 (θ, X ) Gamma a + n 2, b ), i = 1,..., n n i=1 So Gibbs sampler is pretty easy to implement... θ 2 i ). 8 Easiest argument is based on standard conjugate priors / 42

29 Example: many-normal-means (cont.) Suppose the goal is to estimate θ 2 = n i=1 θ2 i. In general, the MLE X 2 is lousy. However, the Bayes estimator, E( θ 2 X ), is better and can be evaluated by running the Gibbs sampler. Can use the Rao Blackwellized estimator of E(θi 2 X ) to reduce the variance. Simulation study to compare Bayes estimator with MLE: n = 10, θ = (1, 1,..., 1); 1000 reps, 5000 Monte Carlos, a = b = 1. mle.mse bayes.mse [1,] / 42

30 Example: capture recapture Example in Lange. Consider a lake that contains N fish, N is unknown. Capture recapture study: On n occasions, fish are caught, marked, and returned. At occasion i = 1,..., n, record C i = number of fish caught at time i R i = number of recaptures at time i. C i R i is the number of new fish caught at time i. Set U i = i j=1 (C i R i ). Model assumes independent binomial sampling. 30 / 42

31 Example: capture recapture (cont.) Introduce binomial success probabilities (ω 1,..., ω n ). Likelihood for (N, ω) is n ( Ui 1 ) ( L(N, ω) = ω R i i (1 ω i ) U i 1 R i N Ui 1 ) i R i=1 i C i R i n ( Ui 1 )( N Ui 1 ) = ω C i i (1 ω i ) N C i R i=1 i C i R i N! n ( Ui 1 ) = ω C i i (1 ω i ) N C i. (N U n)! R i=1 i ω C i R i (1 ω i ) N U i 1 C i +R i Priors: N Pois(m) and ω i ind Beta(a, b). 31 / 42

32 Example: capture recapture (cont.) Posterior distribution for (N, ω) proportional to N! m N (N U n )! N! n i=1 ( Ui 1 R i ) ω C i +a 1 i (1 ω i ) N C i +b 1. To run a Gibbs sampler, we need the full conditionals. Distribution of (ω 1,..., ω n ), given N and data, is clearly ω i (N, data) ind Beta(a + C i, b + N C i ), i = 1,..., n. Distribution of N, given ω and data, is ( N (ω, data) U n + Pois m Now the Gibbs sampler is easy to run... n ) (1 ω i ). i=1 32 / 42

33 Example: probit regression ind Model: Y i Ber(Φ(xi β)), i = 1,..., n. Suppose β has a normal prior. Not directly obvious how to implement Gibbs to get a sample from the posterior distribution of β. Recall, from EM notes, that this model can be simplified by introducing some missing data. The conditional distribution of the missing data, given the observed data and β, makes up one part of the full conditionals. The other part of the full conditionals is simple since the model for the complete data is, by construction, nice. 33 / 42

34 Example: probit regression (cont.) Missing data: Z 1,..., Z n where Z i N(x i β, 1) and Y i = I Zi >0, i = 1,..., n. Full conditionals: Distribution of β, given (Y, Z), only depends on Z and is easy because normal prior for β is conjugate. Distribution of Z, given (Y, β), is a truncated normal... Though I ve not given the precise details, the steps for constructing a Gibbs sampler are not too difficult; 9 see Section in Ghosh et al (2006) for details. 9 The only potential difficulty is simulating from a truncated normal when the truncation point is extreme, but remember we have talked about extreme normal tail probabilities before / 42

35 Example: Dirichlet process mixture In Bayesian nonparametrics, the Dirichlet process mixture model is probably the most widely used. Flexible model for density estimation normal mixture density with unspecified component means (and variances), but also doesn t specify the number of components. The main challenge with using mixture models is choosing how many components to use; this DPM selects the number of components automatically. Interesting that, despite being a nonparametric model, the computations are not too hard, just a Gibbs sampler. Simplest algorithm in Escobar & West (JASA 1995); a nice slice sampler is proposed in Kalli et al (Stat Comp 2011). 35 / 42

36 Example: Dirichlet process mixture (cont.) Using the slice sampler from Kalli et al to fit same normal mixture model to the galaxy data from HW. R code for this on my research page. Density Post mean Kernel Probability Y new Number of components 36 / 42

37 Outline 1 Introduction 2 Crash course on Markov chains 3 Motivation, revisited 4 Metropolis Hastings algorithm 5 Gibbs sampler 6 Some MCMC diagnostics 7 Conclusion 37 / 42

38 Diagnostic plots Sample path plot, or trace plot: Can reveal any residual dependence after burn-in. Idea is that a sample path of iid samples should show no trend, so if there is minimal trend in our sample plot, then we can be comfortable treating samples as independent. Autocorrelation plot: Plot sample correlation of {(X t, X t+r ) : t = 1, 2,...} as a function of the lag r. What to see the autocorrelation plot decay rapidly, suggesting that the dependence along the chain is not too strong. If these plots indicate that chain has not yet converged to stationarity, then you can run the chain longer or make some other modifications, e.g., Transformations thinning 38 / 42

39 Other considerations The practical/theoretical rate of convergence can depend on the parametrization; see homework. There is no agreement in the stat community about how many chains to run, how long should the burn in be, etc. Charles Geyer (Univ Minnesota) strongly supports running only one long chain check out his rants Gelman & Rubin suggest running several shorter chains with different starting points; textbook gives their diagnostic test. 39 / 42

40 Outline 1 Introduction 2 Crash course on Markov chains 3 Motivation, revisited 4 Metropolis Hastings algorithm 5 Gibbs sampler 6 Some MCMC diagnostics 7 Conclusion 40 / 42

41 Remarks MCMC methods are powerful because they give fairly general procedures able to solve a variety of important problems. There are black-box software implementations: The mcmc package in R will do random walk MH; In SAS, PROC MCMC does similar things; BUGS ( Bayes Using Gibbs Sampler ). However, it s a bad idea to blindly use these things without fully understanding what they re doing and whether or not they will actually work in your problem. Also important to look at convergence diagnostics before the simulation results be used for inference. 41 / 42

42 Remarks (cont.) Our focus here was on relatively simple MCMC methods. Don t think that MH, Gibbs, etc are all separate methods, they can be combined. For example, if one of the full conditionals is difficult to sample from, one might consider a MH step 10 within Gibbs to sample this conditional. Book by Robert & Casella has some details about more advanced MCMC methods, including various combinations of these standard techniques. 10 Could also use accept reject for this / 42

Bayesian Methods for Machine Learning

Bayesian Methods for Machine Learning Bayesian Methods for Machine Learning CS 584: Big Data Analytics Material adapted from Radford Neal s tutorial (http://ftp.cs.utoronto.ca/pub/radford/bayes-tut.pdf), Zoubin Ghahramni (http://hunch.net/~coms-4771/zoubin_ghahramani_bayesian_learning.pdf),

More information

Stat 451 Lecture Notes Monte Carlo Integration

Stat 451 Lecture Notes Monte Carlo Integration Stat 451 Lecture Notes 06 12 Monte Carlo Integration Ryan Martin UIC www.math.uic.edu/~rgmartin 1 Based on Chapter 6 in Givens & Hoeting, Chapter 23 in Lange, and Chapters 3 4 in Robert & Casella 2 Updated:

More information

Stat 535 C - Statistical Computing & Monte Carlo Methods. Lecture 15-7th March Arnaud Doucet

Stat 535 C - Statistical Computing & Monte Carlo Methods. Lecture 15-7th March Arnaud Doucet Stat 535 C - Statistical Computing & Monte Carlo Methods Lecture 15-7th March 2006 Arnaud Doucet Email: arnaud@cs.ubc.ca 1 1.1 Outline Mixture and composition of kernels. Hybrid algorithms. Examples Overview

More information

CPSC 540: Machine Learning

CPSC 540: Machine Learning CPSC 540: Machine Learning MCMC and Non-Parametric Bayes Mark Schmidt University of British Columbia Winter 2016 Admin I went through project proposals: Some of you got a message on Piazza. No news is

More information

Stat 451 Lecture Notes Simulating Random Variables

Stat 451 Lecture Notes Simulating Random Variables Stat 451 Lecture Notes 05 12 Simulating Random Variables Ryan Martin UIC www.math.uic.edu/~rgmartin 1 Based on Chapter 6 in Givens & Hoeting, Chapter 22 in Lange, and Chapter 2 in Robert & Casella 2 Updated:

More information

MCMC algorithms for fitting Bayesian models

MCMC algorithms for fitting Bayesian models MCMC algorithms for fitting Bayesian models p. 1/1 MCMC algorithms for fitting Bayesian models Sudipto Banerjee sudiptob@biostat.umn.edu University of Minnesota MCMC algorithms for fitting Bayesian models

More information

Stat 451 Lecture Notes Numerical Integration

Stat 451 Lecture Notes Numerical Integration Stat 451 Lecture Notes 03 12 Numerical Integration Ryan Martin UIC www.math.uic.edu/~rgmartin 1 Based on Chapter 5 in Givens & Hoeting, and Chapters 4 & 18 of Lange 2 Updated: February 11, 2016 1 / 29

More information

Markov chain Monte Carlo Lecture 9

Markov chain Monte Carlo Lecture 9 Markov chain Monte Carlo Lecture 9 David Sontag New York University Slides adapted from Eric Xing and Qirong Ho (CMU) Limitations of Monte Carlo Direct (unconditional) sampling Hard to get rare events

More information

Computational statistics

Computational statistics Computational statistics Markov Chain Monte Carlo methods Thierry Denœux March 2017 Thierry Denœux Computational statistics March 2017 1 / 71 Contents of this chapter When a target density f can be evaluated

More information

Markov Chain Monte Carlo

Markov Chain Monte Carlo Markov Chain Monte Carlo Recall: To compute the expectation E ( h(y ) ) we use the approximation E(h(Y )) 1 n n h(y ) t=1 with Y (1),..., Y (n) h(y). Thus our aim is to sample Y (1),..., Y (n) from f(y).

More information

Markov Chain Monte Carlo methods

Markov Chain Monte Carlo methods Markov Chain Monte Carlo methods By Oleg Makhnin 1 Introduction a b c M = d e f g h i 0 f(x)dx 1.1 Motivation 1.1.1 Just here Supresses numbering 1.1.2 After this 1.2 Literature 2 Method 2.1 New math As

More information

Introduction to Machine Learning CMU-10701

Introduction to Machine Learning CMU-10701 Introduction to Machine Learning CMU-10701 Markov Chain Monte Carlo Methods Barnabás Póczos & Aarti Singh Contents Markov Chain Monte Carlo Methods Goal & Motivation Sampling Rejection Importance Markov

More information

Markov Chain Monte Carlo (MCMC)

Markov Chain Monte Carlo (MCMC) Markov Chain Monte Carlo (MCMC Dependent Sampling Suppose we wish to sample from a density π, and we can evaluate π as a function but have no means to directly generate a sample. Rejection sampling can

More information

Markov Chain Monte Carlo methods

Markov Chain Monte Carlo methods Markov Chain Monte Carlo methods Tomas McKelvey and Lennart Svensson Signal Processing Group Department of Signals and Systems Chalmers University of Technology, Sweden November 26, 2012 Today s learning

More information

Lecture 7 and 8: Markov Chain Monte Carlo

Lecture 7 and 8: Markov Chain Monte Carlo Lecture 7 and 8: Markov Chain Monte Carlo 4F13: Machine Learning Zoubin Ghahramani and Carl Edward Rasmussen Department of Engineering University of Cambridge http://mlg.eng.cam.ac.uk/teaching/4f13/ Ghahramani

More information

Lecture 8: The Metropolis-Hastings Algorithm

Lecture 8: The Metropolis-Hastings Algorithm 30.10.2008 What we have seen last time: Gibbs sampler Key idea: Generate a Markov chain by updating the component of (X 1,..., X p ) in turn by drawing from the full conditionals: X (t) j Two drawbacks:

More information

27 : Distributed Monte Carlo Markov Chain. 1 Recap of MCMC and Naive Parallel Gibbs Sampling

27 : Distributed Monte Carlo Markov Chain. 1 Recap of MCMC and Naive Parallel Gibbs Sampling 10-708: Probabilistic Graphical Models 10-708, Spring 2014 27 : Distributed Monte Carlo Markov Chain Lecturer: Eric P. Xing Scribes: Pengtao Xie, Khoa Luu In this scribe, we are going to review the Parallel

More information

Introduction to Bayesian Statistics and Markov Chain Monte Carlo Estimation. EPSY 905: Multivariate Analysis Spring 2016 Lecture #10: April 6, 2016

Introduction to Bayesian Statistics and Markov Chain Monte Carlo Estimation. EPSY 905: Multivariate Analysis Spring 2016 Lecture #10: April 6, 2016 Introduction to Bayesian Statistics and Markov Chain Monte Carlo Estimation EPSY 905: Multivariate Analysis Spring 2016 Lecture #10: April 6, 2016 EPSY 905: Intro to Bayesian and MCMC Today s Class An

More information

Lecture 6: Markov Chain Monte Carlo

Lecture 6: Markov Chain Monte Carlo Lecture 6: Markov Chain Monte Carlo D. Jason Koskinen koskinen@nbi.ku.dk Photo by Howard Jackman University of Copenhagen Advanced Methods in Applied Statistics Feb - Apr 2016 Niels Bohr Institute 2 Outline

More information

16 : Approximate Inference: Markov Chain Monte Carlo

16 : Approximate Inference: Markov Chain Monte Carlo 10-708: Probabilistic Graphical Models 10-708, Spring 2017 16 : Approximate Inference: Markov Chain Monte Carlo Lecturer: Eric P. Xing Scribes: Yuan Yang, Chao-Ming Yen 1 Introduction As the target distribution

More information

Probabilistic Graphical Models Lecture 17: Markov chain Monte Carlo

Probabilistic Graphical Models Lecture 17: Markov chain Monte Carlo Probabilistic Graphical Models Lecture 17: Markov chain Monte Carlo Andrew Gordon Wilson www.cs.cmu.edu/~andrewgw Carnegie Mellon University March 18, 2015 1 / 45 Resources and Attribution Image credits,

More information

Bayesian GLMs and Metropolis-Hastings Algorithm

Bayesian GLMs and Metropolis-Hastings Algorithm Bayesian GLMs and Metropolis-Hastings Algorithm We have seen that with conjugate or semi-conjugate prior distributions the Gibbs sampler can be used to sample from the posterior distribution. In situations,

More information

SC7/SM6 Bayes Methods HT18 Lecturer: Geoff Nicholls Lecture 2: Monte Carlo Methods Notes and Problem sheets are available at http://www.stats.ox.ac.uk/~nicholls/bayesmethods/ and via the MSc weblearn pages.

More information

SAMPLING ALGORITHMS. In general. Inference in Bayesian models

SAMPLING ALGORITHMS. In general. Inference in Bayesian models SAMPLING ALGORITHMS SAMPLING ALGORITHMS In general A sampling algorithm is an algorithm that outputs samples x 1, x 2,... from a given distribution P or density p. Sampling algorithms can for example be

More information

The Bayesian Choice. Christian P. Robert. From Decision-Theoretic Foundations to Computational Implementation. Second Edition.

The Bayesian Choice. Christian P. Robert. From Decision-Theoretic Foundations to Computational Implementation. Second Edition. Christian P. Robert The Bayesian Choice From Decision-Theoretic Foundations to Computational Implementation Second Edition With 23 Illustrations ^Springer" Contents Preface to the Second Edition Preface

More information

17 : Markov Chain Monte Carlo

17 : Markov Chain Monte Carlo 10-708: Probabilistic Graphical Models, Spring 2015 17 : Markov Chain Monte Carlo Lecturer: Eric P. Xing Scribes: Heran Lin, Bin Deng, Yun Huang 1 Review of Monte Carlo Methods 1.1 Overview Monte Carlo

More information

13 Notes on Markov Chain Monte Carlo

13 Notes on Markov Chain Monte Carlo 13 Notes on Markov Chain Monte Carlo Markov Chain Monte Carlo is a big, and currently very rapidly developing, subject in statistical computation. Many complex and multivariate types of random data, useful

More information

Markov Chain Monte Carlo Inference. Siamak Ravanbakhsh Winter 2018

Markov Chain Monte Carlo Inference. Siamak Ravanbakhsh Winter 2018 Graphical Models Markov Chain Monte Carlo Inference Siamak Ravanbakhsh Winter 2018 Learning objectives Markov chains the idea behind Markov Chain Monte Carlo (MCMC) two important examples: Gibbs sampling

More information

MCMC: Markov Chain Monte Carlo

MCMC: Markov Chain Monte Carlo I529: Machine Learning in Bioinformatics (Spring 2013) MCMC: Markov Chain Monte Carlo Yuzhen Ye School of Informatics and Computing Indiana University, Bloomington Spring 2013 Contents Review of Markov

More information

Principles of Bayesian Inference

Principles of Bayesian Inference Principles of Bayesian Inference Sudipto Banerjee University of Minnesota July 20th, 2008 1 Bayesian Principles Classical statistics: model parameters are fixed and unknown. A Bayesian thinks of parameters

More information

STAT 425: Introduction to Bayesian Analysis

STAT 425: Introduction to Bayesian Analysis STAT 425: Introduction to Bayesian Analysis Marina Vannucci Rice University, USA Fall 2017 Marina Vannucci (Rice University, USA) Bayesian Analysis (Part 2) Fall 2017 1 / 19 Part 2: Markov chain Monte

More information

Bayesian Regression Linear and Logistic Regression

Bayesian Regression Linear and Logistic Regression When we want more than point estimates Bayesian Regression Linear and Logistic Regression Nicole Beckage Ordinary Least Squares Regression and Lasso Regression return only point estimates But what if we

More information

Bayesian Networks in Educational Assessment

Bayesian Networks in Educational Assessment Bayesian Networks in Educational Assessment Estimating Parameters with MCMC Bayesian Inference: Expanding Our Context Roy Levy Arizona State University Roy.Levy@asu.edu 2017 Roy Levy MCMC 1 MCMC 2 Posterior

More information

MCMC Methods: Gibbs and Metropolis

MCMC Methods: Gibbs and Metropolis MCMC Methods: Gibbs and Metropolis Patrick Breheny February 28 Patrick Breheny BST 701: Bayesian Modeling in Biostatistics 1/30 Introduction As we have seen, the ability to sample from the posterior distribution

More information

Adaptive Monte Carlo methods

Adaptive Monte Carlo methods Adaptive Monte Carlo methods Jean-Michel Marin Projet Select, INRIA Futurs, Université Paris-Sud joint with Randal Douc (École Polytechnique), Arnaud Guillin (Université de Marseille) and Christian Robert

More information

Markov chain Monte Carlo

Markov chain Monte Carlo Markov chain Monte Carlo Feng Li feng.li@cufe.edu.cn School of Statistics and Mathematics Central University of Finance and Economics Revised on April 24, 2017 Today we are going to learn... 1 Markov Chains

More information

Markov chain Monte Carlo methods in atmospheric remote sensing

Markov chain Monte Carlo methods in atmospheric remote sensing 1 / 45 Markov chain Monte Carlo methods in atmospheric remote sensing Johanna Tamminen johanna.tamminen@fmi.fi ESA Summer School on Earth System Monitoring and Modeling July 3 Aug 11, 212, Frascati July,

More information

Contents. Part I: Fundamentals of Bayesian Inference 1

Contents. Part I: Fundamentals of Bayesian Inference 1 Contents Preface xiii Part I: Fundamentals of Bayesian Inference 1 1 Probability and inference 3 1.1 The three steps of Bayesian data analysis 3 1.2 General notation for statistical inference 4 1.3 Bayesian

More information

Introduction to Markov Chain Monte Carlo & Gibbs Sampling

Introduction to Markov Chain Monte Carlo & Gibbs Sampling Introduction to Markov Chain Monte Carlo & Gibbs Sampling Prof. Nicholas Zabaras Sibley School of Mechanical and Aerospace Engineering 101 Frank H. T. Rhodes Hall Ithaca, NY 14853-3801 Email: zabaras@cornell.edu

More information

Computer intensive statistical methods

Computer intensive statistical methods Lecture 13 MCMC, Hybrid chains October 13, 2015 Jonas Wallin jonwal@chalmers.se Chalmers, Gothenburg university MH algorithm, Chap:6.3 The metropolis hastings requires three objects, the distribution of

More information

Hastings-within-Gibbs Algorithm: Introduction and Application on Hierarchical Model

Hastings-within-Gibbs Algorithm: Introduction and Application on Hierarchical Model UNIVERSITY OF TEXAS AT SAN ANTONIO Hastings-within-Gibbs Algorithm: Introduction and Application on Hierarchical Model Liang Jing April 2010 1 1 ABSTRACT In this paper, common MCMC algorithms are introduced

More information

Brief introduction to Markov Chain Monte Carlo

Brief introduction to Markov Chain Monte Carlo Brief introduction to Department of Probability and Mathematical Statistics seminar Stochastic modeling in economics and finance November 7, 2011 Brief introduction to Content 1 and motivation Classical

More information

Sampling from complex probability distributions

Sampling from complex probability distributions Sampling from complex probability distributions Louis J. M. Aslett (louis.aslett@durham.ac.uk) Department of Mathematical Sciences Durham University UTOPIAE Training School II 4 July 2017 1/37 Motivation

More information

STA 4273H: Statistical Machine Learning

STA 4273H: Statistical Machine Learning STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Computer Science! Department of Statistical Sciences! rsalakhu@cs.toronto.edu! h0p://www.cs.utoronto.ca/~rsalakhu/ Lecture 7 Approximate

More information

Markov chain Monte Carlo

Markov chain Monte Carlo Markov chain Monte Carlo Markov chain Monte Carlo (MCMC) Gibbs and Metropolis Hastings Slice sampling Practical details Iain Murray http://iainmurray.net/ Reminder Need to sample large, non-standard distributions:

More information

ST 740: Markov Chain Monte Carlo

ST 740: Markov Chain Monte Carlo ST 740: Markov Chain Monte Carlo Alyson Wilson Department of Statistics North Carolina State University October 14, 2012 A. Wilson (NCSU Stsatistics) MCMC October 14, 2012 1 / 20 Convergence Diagnostics:

More information

Here θ = (α, β) is the unknown parameter. The likelihood function is

Here θ = (α, β) is the unknown parameter. The likelihood function is Stat 591 Notes Logistic regression and Metropolis Hastings example Ryan Martin (rgmartin@uic.edu) November 2, 2013 Introduction (This part taken from Example 1.13 in Robert and Casella, 2004.) In 1986,

More information

A quick introduction to Markov chains and Markov chain Monte Carlo (revised version)

A quick introduction to Markov chains and Markov chain Monte Carlo (revised version) A quick introduction to Markov chains and Markov chain Monte Carlo (revised version) Rasmus Waagepetersen Institute of Mathematical Sciences Aalborg University 1 Introduction These notes are intended to

More information

MH I. Metropolis-Hastings (MH) algorithm is the most popular method of getting dependent samples from a probability distribution

MH I. Metropolis-Hastings (MH) algorithm is the most popular method of getting dependent samples from a probability distribution MH I Metropolis-Hastings (MH) algorithm is the most popular method of getting dependent samples from a probability distribution a lot of Bayesian mehods rely on the use of MH algorithm and it s famous

More information

The Particle Filter. PD Dr. Rudolph Triebel Computer Vision Group. Machine Learning for Computer Vision

The Particle Filter. PD Dr. Rudolph Triebel Computer Vision Group. Machine Learning for Computer Vision The Particle Filter Non-parametric implementation of Bayes filter Represents the belief (posterior) random state samples. by a set of This representation is approximate. Can represent distributions that

More information

Stat 535 C - Statistical Computing & Monte Carlo Methods. Lecture February Arnaud Doucet

Stat 535 C - Statistical Computing & Monte Carlo Methods. Lecture February Arnaud Doucet Stat 535 C - Statistical Computing & Monte Carlo Methods Lecture 13-28 February 2006 Arnaud Doucet Email: arnaud@cs.ubc.ca 1 1.1 Outline Limitations of Gibbs sampling. Metropolis-Hastings algorithm. Proof

More information

Metropolis-Hastings Algorithm

Metropolis-Hastings Algorithm Strength of the Gibbs sampler Metropolis-Hastings Algorithm Easy algorithm to think about. Exploits the factorization properties of the joint probability distribution. No difficult choices to be made to

More information

CSC 2541: Bayesian Methods for Machine Learning

CSC 2541: Bayesian Methods for Machine Learning CSC 2541: Bayesian Methods for Machine Learning Radford M. Neal, University of Toronto, 2011 Lecture 3 More Markov Chain Monte Carlo Methods The Metropolis algorithm isn t the only way to do MCMC. We ll

More information

Hierarchical models. Dr. Jarad Niemi. August 31, Iowa State University. Jarad Niemi (Iowa State) Hierarchical models August 31, / 31

Hierarchical models. Dr. Jarad Niemi. August 31, Iowa State University. Jarad Niemi (Iowa State) Hierarchical models August 31, / 31 Hierarchical models Dr. Jarad Niemi Iowa State University August 31, 2017 Jarad Niemi (Iowa State) Hierarchical models August 31, 2017 1 / 31 Normal hierarchical model Let Y ig N(θ g, σ 2 ) for i = 1,...,

More information

The Mixture Approach for Simulating New Families of Bivariate Distributions with Specified Correlations

The Mixture Approach for Simulating New Families of Bivariate Distributions with Specified Correlations The Mixture Approach for Simulating New Families of Bivariate Distributions with Specified Correlations John R. Michael, Significance, Inc. and William R. Schucany, Southern Methodist University The mixture

More information

Statistical Machine Learning Lecture 8: Markov Chain Monte Carlo Sampling

Statistical Machine Learning Lecture 8: Markov Chain Monte Carlo Sampling 1 / 27 Statistical Machine Learning Lecture 8: Markov Chain Monte Carlo Sampling Melih Kandemir Özyeğin University, İstanbul, Turkey 2 / 27 Monte Carlo Integration The big question : Evaluate E p(z) [f(z)]

More information

Stat 535 C - Statistical Computing & Monte Carlo Methods. Arnaud Doucet.

Stat 535 C - Statistical Computing & Monte Carlo Methods. Arnaud Doucet. Stat 535 C - Statistical Computing & Monte Carlo Methods Arnaud Doucet Email: arnaud@cs.ubc.ca 1 1.1 Outline Introduction to Markov chain Monte Carlo The Gibbs Sampler Examples Overview of the Lecture

More information

Bayesian Inference and MCMC

Bayesian Inference and MCMC Bayesian Inference and MCMC Aryan Arbabi Partly based on MCMC slides from CSC412 Fall 2018 1 / 18 Bayesian Inference - Motivation Consider we have a data set D = {x 1,..., x n }. E.g each x i can be the

More information

36-463/663Multilevel and Hierarchical Models

36-463/663Multilevel and Hierarchical Models 36-463/663Multilevel and Hierarchical Models From Bayes to MCMC to MLMs Brian Junker 132E Baker Hall brian@stat.cmu.edu 1 Outline Bayesian Statistics and MCMC Distribution of Skill Mastery in a Population

More information

Bayesian linear regression

Bayesian linear regression Bayesian linear regression Linear regression is the basis of most statistical modeling. The model is Y i = X T i β + ε i, where Y i is the continuous response X i = (X i1,..., X ip ) T is the corresponding

More information

Advances and Applications in Perfect Sampling

Advances and Applications in Perfect Sampling and Applications in Perfect Sampling Ph.D. Dissertation Defense Ulrike Schneider advisor: Jem Corcoran May 8, 2003 Department of Applied Mathematics University of Colorado Outline Introduction (1) MCMC

More information

Lecture 16: Mixtures of Generalized Linear Models

Lecture 16: Mixtures of Generalized Linear Models Lecture 16: Mixtures of Generalized Linear Models October 26, 2006 Setting Outline Often, a single GLM may be insufficiently flexible to characterize the data Setting Often, a single GLM may be insufficiently

More information

Markov chain Monte Carlo

Markov chain Monte Carlo 1 / 26 Markov chain Monte Carlo Timothy Hanson 1 and Alejandro Jara 2 1 Division of Biostatistics, University of Minnesota, USA 2 Department of Statistics, Universidad de Concepción, Chile IAP-Workshop

More information

Stat 516, Homework 1

Stat 516, Homework 1 Stat 516, Homework 1 Due date: October 7 1. Consider an urn with n distinct balls numbered 1,..., n. We sample balls from the urn with replacement. Let N be the number of draws until we encounter a ball

More information

Probabilistic Graphical Models

Probabilistic Graphical Models Probabilistic Graphical Models Lecture 11 CRFs, Exponential Family CS/CNS/EE 155 Andreas Krause Announcements Homework 2 due today Project milestones due next Monday (Nov 9) About half the work should

More information

Bayesian Inference in GLMs. Frequentists typically base inferences on MLEs, asymptotic confidence

Bayesian Inference in GLMs. Frequentists typically base inferences on MLEs, asymptotic confidence Bayesian Inference in GLMs Frequentists typically base inferences on MLEs, asymptotic confidence limits, and log-likelihood ratio tests Bayesians base inferences on the posterior distribution of the unknowns

More information

STA 216, GLM, Lecture 16. October 29, 2007

STA 216, GLM, Lecture 16. October 29, 2007 STA 216, GLM, Lecture 16 October 29, 2007 Efficient Posterior Computation in Factor Models Underlying Normal Models Generalized Latent Trait Models Formulation Genetic Epidemiology Illustration Structural

More information

Likelihood-free MCMC

Likelihood-free MCMC Bayesian inference for stable distributions with applications in finance Department of Mathematics University of Leicester September 2, 2011 MSc project final presentation Outline 1 2 3 4 Classical Monte

More information

eqr094: Hierarchical MCMC for Bayesian System Reliability

eqr094: Hierarchical MCMC for Bayesian System Reliability eqr094: Hierarchical MCMC for Bayesian System Reliability Alyson G. Wilson Statistical Sciences Group, Los Alamos National Laboratory P.O. Box 1663, MS F600 Los Alamos, NM 87545 USA Phone: 505-667-9167

More information

Nonparametric Bayesian Methods (Gaussian Processes)

Nonparametric Bayesian Methods (Gaussian Processes) [70240413 Statistical Machine Learning, Spring, 2015] Nonparametric Bayesian Methods (Gaussian Processes) Jun Zhu dcszj@mail.tsinghua.edu.cn http://bigml.cs.tsinghua.edu.cn/~jun State Key Lab of Intelligent

More information

INTRODUCTION TO BAYESIAN STATISTICS

INTRODUCTION TO BAYESIAN STATISTICS INTRODUCTION TO BAYESIAN STATISTICS Sarat C. Dass Department of Statistics & Probability Department of Computer Science & Engineering Michigan State University TOPICS The Bayesian Framework Different Types

More information

Markov Chain Monte Carlo in Practice

Markov Chain Monte Carlo in Practice Markov Chain Monte Carlo in Practice Edited by W.R. Gilks Medical Research Council Biostatistics Unit Cambridge UK S. Richardson French National Institute for Health and Medical Research Vilejuif France

More information

Exercises Tutorial at ICASSP 2016 Learning Nonlinear Dynamical Models Using Particle Filters

Exercises Tutorial at ICASSP 2016 Learning Nonlinear Dynamical Models Using Particle Filters Exercises Tutorial at ICASSP 216 Learning Nonlinear Dynamical Models Using Particle Filters Andreas Svensson, Johan Dahlin and Thomas B. Schön March 18, 216 Good luck! 1 [Bootstrap particle filter for

More information

A short diversion into the theory of Markov chains, with a view to Markov chain Monte Carlo methods

A short diversion into the theory of Markov chains, with a view to Markov chain Monte Carlo methods A short diversion into the theory of Markov chains, with a view to Markov chain Monte Carlo methods by Kasper K. Berthelsen and Jesper Møller June 2004 2004-01 DEPARTMENT OF MATHEMATICAL SCIENCES AALBORG

More information

Metropolis Hastings. Rebecca C. Steorts Bayesian Methods and Modern Statistics: STA 360/601. Module 9

Metropolis Hastings. Rebecca C. Steorts Bayesian Methods and Modern Statistics: STA 360/601. Module 9 Metropolis Hastings Rebecca C. Steorts Bayesian Methods and Modern Statistics: STA 360/601 Module 9 1 The Metropolis-Hastings algorithm is a general term for a family of Markov chain simulation methods

More information

Supplement to A Hierarchical Approach for Fitting Curves to Response Time Measurements

Supplement to A Hierarchical Approach for Fitting Curves to Response Time Measurements Supplement to A Hierarchical Approach for Fitting Curves to Response Time Measurements Jeffrey N. Rouder Francis Tuerlinckx Paul L. Speckman Jun Lu & Pablo Gomez May 4 008 1 The Weibull regression model

More information

April 20th, Advanced Topics in Machine Learning California Institute of Technology. Markov Chain Monte Carlo for Machine Learning

April 20th, Advanced Topics in Machine Learning California Institute of Technology. Markov Chain Monte Carlo for Machine Learning for for Advanced Topics in California Institute of Technology April 20th, 2017 1 / 50 Table of Contents for 1 2 3 4 2 / 50 History of methods for Enrico Fermi used to calculate incredibly accurate predictions

More information

Default Priors and Effcient Posterior Computation in Bayesian

Default Priors and Effcient Posterior Computation in Bayesian Default Priors and Effcient Posterior Computation in Bayesian Factor Analysis January 16, 2010 Presented by Eric Wang, Duke University Background and Motivation A Brief Review of Parameter Expansion Literature

More information

Principles of Bayesian Inference

Principles of Bayesian Inference Principles of Bayesian Inference Sudipto Banerjee 1 and Andrew O. Finley 2 1 Biostatistics, School of Public Health, University of Minnesota, Minneapolis, Minnesota, U.S.A. 2 Department of Forestry & Department

More information

Introduction to Probabilistic Machine Learning

Introduction to Probabilistic Machine Learning Introduction to Probabilistic Machine Learning Piyush Rai Dept. of CSE, IIT Kanpur (Mini-course 1) Nov 03, 2015 Piyush Rai (IIT Kanpur) Introduction to Probabilistic Machine Learning 1 Machine Learning

More information

Motivation Scale Mixutres of Normals Finite Gaussian Mixtures Skew-Normal Models. Mixture Models. Econ 690. Purdue University

Motivation Scale Mixutres of Normals Finite Gaussian Mixtures Skew-Normal Models. Mixture Models. Econ 690. Purdue University Econ 690 Purdue University In virtually all of the previous lectures, our models have made use of normality assumptions. From a computational point of view, the reason for this assumption is clear: combined

More information

ABC methods for phase-type distributions with applications in insurance risk problems

ABC methods for phase-type distributions with applications in insurance risk problems ABC methods for phase-type with applications problems Concepcion Ausin, Department of Statistics, Universidad Carlos III de Madrid Joint work with: Pedro Galeano, Universidad Carlos III de Madrid Simon

More information

Bagging During Markov Chain Monte Carlo for Smoother Predictions

Bagging During Markov Chain Monte Carlo for Smoother Predictions Bagging During Markov Chain Monte Carlo for Smoother Predictions Herbert K. H. Lee University of California, Santa Cruz Abstract: Making good predictions from noisy data is a challenging problem. Methods

More information

Quantifying Uncertainty

Quantifying Uncertainty Sai Ravela M. I. T Last Updated: Spring 2013 1 Markov Chain Monte Carlo Monte Carlo sampling made for large scale problems via Markov Chains Monte Carlo Sampling Rejection Sampling Importance Sampling

More information

Lecture 5. G. Cowan Lectures on Statistical Data Analysis Lecture 5 page 1

Lecture 5. G. Cowan Lectures on Statistical Data Analysis Lecture 5 page 1 Lecture 5 1 Probability (90 min.) Definition, Bayes theorem, probability densities and their properties, catalogue of pdfs, Monte Carlo 2 Statistical tests (90 min.) general concepts, test statistics,

More information

Doing Bayesian Integrals

Doing Bayesian Integrals ASTR509-13 Doing Bayesian Integrals The Reverend Thomas Bayes (c.1702 1761) Philosopher, theologian, mathematician Presbyterian (non-conformist) minister Tunbridge Wells, UK Elected FRS, perhaps due to

More information

Markov Chain Monte Carlo (MCMC) and Model Evaluation. August 15, 2017

Markov Chain Monte Carlo (MCMC) and Model Evaluation. August 15, 2017 Markov Chain Monte Carlo (MCMC) and Model Evaluation August 15, 2017 Frequentist Linking Frequentist and Bayesian Statistics How can we estimate model parameters and what does it imply? Want to find the

More information

Markov chain Monte Carlo

Markov chain Monte Carlo Markov chain Monte Carlo Peter Beerli October 10, 2005 [this chapter is highly influenced by chapter 1 in Markov chain Monte Carlo in Practice, eds Gilks W. R. et al. Chapman and Hall/CRC, 1996] 1 Short

More information

Bayesian data analysis in practice: Three simple examples

Bayesian data analysis in practice: Three simple examples Bayesian data analysis in practice: Three simple examples Martin P. Tingley Introduction These notes cover three examples I presented at Climatea on 5 October 0. Matlab code is available by request to

More information

STA 294: Stochastic Processes & Bayesian Nonparametrics

STA 294: Stochastic Processes & Bayesian Nonparametrics MARKOV CHAINS AND CONVERGENCE CONCEPTS Markov chains are among the simplest stochastic processes, just one step beyond iid sequences of random variables. Traditionally they ve been used in modelling a

More information

SAMSI Astrostatistics Tutorial. More Markov chain Monte Carlo & Demo of Mathematica software

SAMSI Astrostatistics Tutorial. More Markov chain Monte Carlo & Demo of Mathematica software SAMSI Astrostatistics Tutorial More Markov chain Monte Carlo & Demo of Mathematica software Phil Gregory University of British Columbia 26 Bayesian Logical Data Analysis for the Physical Sciences Contents:

More information

Lect4: Exact Sampling Techniques and MCMC Convergence Analysis

Lect4: Exact Sampling Techniques and MCMC Convergence Analysis Lect4: Exact Sampling Techniques and MCMC Convergence Analysis. Exact sampling. Convergence analysis of MCMC. First-hit time analysis for MCMC--ways to analyze the proposals. Outline of the Module Definitions

More information

Monte Carlo Inference Methods

Monte Carlo Inference Methods Monte Carlo Inference Methods Iain Murray University of Edinburgh http://iainmurray.net Monte Carlo and Insomnia Enrico Fermi (1901 1954) took great delight in astonishing his colleagues with his remarkably

More information

Reminder of some Markov Chain properties:

Reminder of some Markov Chain properties: Reminder of some Markov Chain properties: 1. a transition from one state to another occurs probabilistically 2. only state that matters is where you currently are (i.e. given present, future is independent

More information

Control Variates for Markov Chain Monte Carlo

Control Variates for Markov Chain Monte Carlo Control Variates for Markov Chain Monte Carlo Dellaportas, P., Kontoyiannis, I., and Tsourti, Z. Dept of Statistics, AUEB Dept of Informatics, AUEB 1st Greek Stochastics Meeting Monte Carlo: Probability

More information

Introduction to Bayesian methods in inverse problems

Introduction to Bayesian methods in inverse problems Introduction to Bayesian methods in inverse problems Ville Kolehmainen 1 1 Department of Applied Physics, University of Eastern Finland, Kuopio, Finland March 4 2013 Manchester, UK. Contents Introduction

More information

Pattern Recognition and Machine Learning. Bishop Chapter 11: Sampling Methods

Pattern Recognition and Machine Learning. Bishop Chapter 11: Sampling Methods Pattern Recognition and Machine Learning Chapter 11: Sampling Methods Elise Arnaud Jakob Verbeek May 22, 2008 Outline of the chapter 11.1 Basic Sampling Algorithms 11.2 Markov Chain Monte Carlo 11.3 Gibbs

More information

(5) Multi-parameter models - Gibbs sampling. ST440/540: Applied Bayesian Analysis

(5) Multi-parameter models - Gibbs sampling. ST440/540: Applied Bayesian Analysis Summarizing a posterior Given the data and prior the posterior is determined Summarizing the posterior gives parameter estimates, intervals, and hypothesis tests Most of these computations are integrals

More information

Bayesian model selection: methodology, computation and applications

Bayesian model selection: methodology, computation and applications Bayesian model selection: methodology, computation and applications David Nott Department of Statistics and Applied Probability National University of Singapore Statistical Genomics Summer School Program

More information

The Pennsylvania State University The Graduate School RATIO-OF-UNIFORMS MARKOV CHAIN MONTE CARLO FOR GAUSSIAN PROCESS MODELS

The Pennsylvania State University The Graduate School RATIO-OF-UNIFORMS MARKOV CHAIN MONTE CARLO FOR GAUSSIAN PROCESS MODELS The Pennsylvania State University The Graduate School RATIO-OF-UNIFORMS MARKOV CHAIN MONTE CARLO FOR GAUSSIAN PROCESS MODELS A Thesis in Statistics by Chris Groendyke c 2008 Chris Groendyke Submitted in

More information