Stat 451 Lecture Notes Markov Chain Monte Carlo. Ryan Martin UIC
|
|
- Louise Hawkins
- 5 years ago
- Views:
Transcription
1 Stat 451 Lecture Notes Markov Chain Monte Carlo Ryan Martin UIC 1 Based on Chapters 8 9 in Givens & Hoeting, Chapters in Lange 2 Updated: April 4, / 42
2 Outline 1 Introduction 2 Crash course on Markov chains 3 Motivation, revisited 4 Metropolis Hastings algorithm 5 Gibbs sampler 6 Some MCMC diagnostics 7 Conclusion 2 / 42
3 Motivation We know how to sample independent random variables from the target distribution f (x), at least approximately. Monte Carlo uses these simulated random variables to approximate integrals. But the random variables don t need to be independent in order to accurately approximate integrals! Markov Chain Monte Carlo (MCMC) constructs a dependent sequence of random variables that can be used to approximate the integrals just like for ordinary Monte Carlo. The advantage of introducing this dependence is that very general black-box algorithms (and corresponding theory) are available to perform the required simulations. 3 / 42
4 Initial remarks In some sense, MCMC is basically a black-box approach that works for almost all problems. The previous remark is dangerous it s always a bad idea to use tools without knowing that they will work. Here we will discuss some basics of Markov chains and MCMC but know that there are very important unanswered questions about how and when MCMC works. 3 3 MCMC is an active area of research; despite the many developments in MCMC, according to Diaconis ( 2008), still almost nothing is known! 4 / 42
5 Outline 1 Introduction 2 Crash course on Markov chains 3 Motivation, revisited 4 Metropolis Hastings algorithm 5 Gibbs sampler 6 Some MCMC diagnostics 7 Conclusion 5 / 42
6 Markov chains A Markov chain is just a sequence of random variables {X 1, X 2,...} with a specific type of dependence structure. In particular, a Markov chain satisfies P(X n+1 B X 1,..., X n ) = P(X n+1 B X n ), ( ) i.e., the future, given the past and present, depends only on the present. Independence is a trivial Markov chain. From ( ) we can argue that the probabilistic properties of the chain are completely determined by: initial distribution for X 0, and the transition distribution, distribution of X n+1, given X n. 4 4 Assume the Markov chain is homogeneous, so that the transition distribution does not depend on n. 6 / 42
7 Example simple random walk Let U 1, U 2,... iid Unif({ 1, 1}). Set X 0 = 0 and let X n = n i=1 U i = X n 1 + U n. The initial distribution is P{X 0 = 0} = 1. The transition distribution is determined by { X n 1 1 with prob 1/2 X n = X n with prob 1/2 While very simple, the random walk is an important example in probability, having connections to advanced things like Brownian motion. 7 / 42
8 Keywords 5 A state A is recurrent if a chain starting in A will eventually return to A with probability 1. This state is nonull if the expected time to return is finite. A chain is called recurrent if each state is recurrent. A Markov chain is irreducible if there is positive probability that a chain starting in a state A can reach any other state B. A Markov chain is aperiodic if, for a starting state A, there is no constraint on the times at which the chain can return to A. An irreducible, aperiodic Markov chain with all states being nonnull recurrent is called ergodic. 5 Not mathematically precise! 8 / 42
9 Limit theory 6 f is a stationary distribution if X 0 f implies X n f for all n. An ergodic Markov chain has at most one stationary dist. Furthermore, if the chain is ergodic, then lim P(X m+n B X m A) = f (x) dx, for all A, B, m. n Even further, if ϕ(x) is integrable, then 1 n n ϕ(x t ) t=1 B ϕ(x)f (x) dx, with prob 1. This is a version of the famous ergodic theorem. There are also central limit theorems for Markov chains, but I won t say anything about this. 6 Again, not mathematically precise! 9 / 42
10 Outline 1 Introduction 2 Crash course on Markov chains 3 Motivation, revisited 4 Metropolis Hastings algorithm 5 Gibbs sampler 6 Some MCMC diagnostics 7 Conclusion 10 / 42
11 Why MCMC? In Monte Carlo applications, we want to generate random variables with distribution f. This could be difficult or impossible to do exactly. MCMC is designed to construct an ergodic Markov chain with f as its stationary distribution. Asymptotically, the chain will resemble samples from f. In particular, by the ergodic theorem, expectations with respect to f can be approximated by averages. Somewhat surprising is that it is actually quite easy to construct and simulate a suitable Markov chain, explaining why MCMC methods have become so popular. But, of course, there are practical and theoretical challenges / 42
12 Outline 1 Introduction 2 Crash course on Markov chains 3 Motivation, revisited 4 Metropolis Hastings algorithm 5 Gibbs sampler 6 Some MCMC diagnostics 7 Conclusion 12 / 42
13 Details Let f (x) denote the target distribution pdf. Let q(x y) denote a conditional pdf for X, given Y = y; this pdf should be easy to sample from. Given X 0, the Metropolis Hasting algorithm (MH) produces a sequence of random variables as follows: 1 Sample X t q(x X t 1 ). 2 Compute { R = min 1, f (Xt ) q(x t 1 Xt ) f (X t 1 ) q(xt X t 1 ) 3 Set X t = X t with probability R; otherwise, X t = X t 1. General R code to implement MH is on course website. }. 13 / 42
14 Details (cont.) The proposal distribution is not easy to choose, and the performance of the algorithm depends on this choice. Two general strategies are: Take the proposal q(x y) = q(x); i.e., at each stage of the MH algorithm, X t does not depend on X t 1. Take q(x y) = q 0 (x y), for a symmetric distribution with pdf q 0, which amounts to a random walk proposal. This is one aspect of the MCMC implementation that requires a lot of care from the user; deeper understanding is needed to really see how the proposal affects the performance. In my examples, I will just pick a proposal that seems to work reasonably well / 42
15 Details (cont.) 7 Assuming the proposal is not too bad, then a number of things can be shown about the sequence {X t : t 1}: the chain is ergodic, and the target f is the stationary distribution. Consequently, the sequence converges to the stationary distribution and, for any integrable function ϕ(x), we can approximate integrals with sample averages. So, provided that we run the simulation long enough, we should be able to get arbitrarily good approximations. This is an interesting scenario where we, as statisticians, are able to control the sample size! 7 Again, not mathematically precise! 15 / 42
16 Example: cosine model Consider the problem from old homework, where the likelihood function is L(θ) n {1 cos(x i θ)}, π θ π. i=1 Observed data (X 1,..., X n ) given in the code. Assume that θ is given a Unif( π, π) prior distribution. Use MH to sample from the posterior: Proposal: q(θ θ) = Unif(θ θ ± 0.5). Burn-in: B = Sample size: M = / 42
17 Example: cosine model (cont.) Left figure shows histogram of the MCMC sample with posterior density overlaid. Right figure shows a trace plot of the chain. Density θ θ Iteration 17 / 42
18 Example: Weibull model Data X 1,..., X n has Weibull likelihood { L(α, η) α n η n exp α n log X i η i=1 Prior: π(α, η) e α η b 1 e cη, for some (b, c). Posterior density is proportional to { α n η n+b 1 exp α ( n i=1 n i=1 ) ( log X i 1 η c + } Xi α. n i=1 Exponential is a special case of Weibull, when α = 1. Goal: informal Bayesian test of H 0 : α = 1. )} Xi α. 18 / 42
19 Example: Weibull model (cont.) Data from Problem 7.11 in Ghosh et al (2006). Use MH to sample from posterior of (α, η). Proposal: (α, η ) (α, η) Exp(α) Exp(η). b = 2 and c = 1; B = 1000 and M = Histogram shows marginal posterior of α. Is an exponential model (α = 1) reasonable? Density α 19 / 42
20 Example: logistic regression Based on Examples 1.13 and 7.11 in Robert & Casella s book. In 1986, the Challenger space shuttle exploded during take-off, the result of an o-ring failure. Failure may have been due to the cold temperature (31 F). Goal: Analyze the relationship between temperature and o-ring failure. In particular, fit a logistic regression model. 20 / 42
21 Example: logistic regression (cont.) Model: Y x Ber(p(x)), x = temperature. Failure probability, p(x), is of the form p(x) = exp(α + βx) 1 + exp(α + βx). Using available data, fit logistic regression using glm. Coefficients: Estimate Std. Error z value Pr(> z ) (Intercept) * x * --- Signif. codes: 0 *** ** 0.01 * Note that p(31) 0.999!!! 21 / 42
22 Example: logistic regression (cont.) Can also do a Bayesian analysis of this logistic model. Use MH to obtain samples from the posterior of (α, β). Samples can be used to approximate the posterior distribution of p(x 0 ) for any fixed x 0, e.g., x 0 = 65, 31; see below. Details about the prior and proposal construction are given in the R code and a short write-up posted on the course website. Density Density p(65) p(31) 22 / 42
23 Outline 1 Introduction 2 Crash course on Markov chains 3 Motivation, revisited 4 Metropolis Hastings algorithm 5 Gibbs sampler 6 Some MCMC diagnostics 7 Conclusion 23 / 42
24 Setup Suppose we have a multivariate target distribution f. MH can be applied to such a problem, but there are challenges in constructing a good proposal over multiple dimensions. Idea: sample one dimension at a time. Question: How to carry out the sampling so that it will approximate the target, at least in a limit? Gibbs sampler is the right tool for the job. 24 / 42
25 Details Suppose we have a trivariate target f (x) = f (x 1, x 2, x 3 ). Suppose we can write down the set of full conditionals f (x 1 x 2, x 3 ), f (x 2 x 1, x 3 ), f (x 3 x 1, x 2 ) and that these can be sampled from. Gibbs sampler generates a sequence {X (t) : t 0} by iteratively sampling from the conditionals: X (t) 1 f (x 1 X (t 1) 2, X (t 1) 3 ) X (t) 2 f (x 2 X (t) 1, X (t 1) 3 ) X (t) 3 f (x 3 X (t) 1, X (t) 2 ). 25 / 42
26 Details (cont.) Gibbs sequence forms a Markov chain. In fact, Gibbs sampler is a special case of MH! Connection to MH is made by considering Gibbs as a sequence that updates one component of X at a time. The acceptance probability for this form of MH update is exactly 1, which explains why Gibbs sampler has no accept/reject step. Since Gibbs is a special kind of MH, the convergence for MH applies to Gibbs as well. 26 / 42
27 Example: bivariate normal A super-simple Gibbs example: bivariate normal. Suppose X = (X 1, X 2 ) is bivariate normal, correlation ρ. Full conditionals are easy to write down here. Gibbs steps: X (t) 1 N(ρX (t 1) 2, 1 ρ 2 ) X (t) 2 N(ρX (t) 1, 1 ρ2 ). Not as efficient as direct sampling, but works fine. 27 / 42
28 Example: many-normal-means ind Model: X i N(θ i, 1), i = 1,..., n. Hierarchical prior distribution: [(θ 1,..., θ n ) ψ] iid N(0, ψ 1 ), ψ Gamma(a, b). Takes some work, but it can be shown 8 that the full conditionals are ( Xi θ i (X i, ψ) ind N 1 + ψ, ψ ( ψ 1 (θ, X ) Gamma a + n 2, b ), i = 1,..., n n i=1 So Gibbs sampler is pretty easy to implement... θ 2 i ). 8 Easiest argument is based on standard conjugate priors / 42
29 Example: many-normal-means (cont.) Suppose the goal is to estimate θ 2 = n i=1 θ2 i. In general, the MLE X 2 is lousy. However, the Bayes estimator, E( θ 2 X ), is better and can be evaluated by running the Gibbs sampler. Can use the Rao Blackwellized estimator of E(θi 2 X ) to reduce the variance. Simulation study to compare Bayes estimator with MLE: n = 10, θ = (1, 1,..., 1); 1000 reps, 5000 Monte Carlos, a = b = 1. mle.mse bayes.mse [1,] / 42
30 Example: capture recapture Example in Lange. Consider a lake that contains N fish, N is unknown. Capture recapture study: On n occasions, fish are caught, marked, and returned. At occasion i = 1,..., n, record C i = number of fish caught at time i R i = number of recaptures at time i. C i R i is the number of new fish caught at time i. Set U i = i j=1 (C i R i ). Model assumes independent binomial sampling. 30 / 42
31 Example: capture recapture (cont.) Introduce binomial success probabilities (ω 1,..., ω n ). Likelihood for (N, ω) is n ( Ui 1 ) ( L(N, ω) = ω R i i (1 ω i ) U i 1 R i N Ui 1 ) i R i=1 i C i R i n ( Ui 1 )( N Ui 1 ) = ω C i i (1 ω i ) N C i R i=1 i C i R i N! n ( Ui 1 ) = ω C i i (1 ω i ) N C i. (N U n)! R i=1 i ω C i R i (1 ω i ) N U i 1 C i +R i Priors: N Pois(m) and ω i ind Beta(a, b). 31 / 42
32 Example: capture recapture (cont.) Posterior distribution for (N, ω) proportional to N! m N (N U n )! N! n i=1 ( Ui 1 R i ) ω C i +a 1 i (1 ω i ) N C i +b 1. To run a Gibbs sampler, we need the full conditionals. Distribution of (ω 1,..., ω n ), given N and data, is clearly ω i (N, data) ind Beta(a + C i, b + N C i ), i = 1,..., n. Distribution of N, given ω and data, is ( N (ω, data) U n + Pois m Now the Gibbs sampler is easy to run... n ) (1 ω i ). i=1 32 / 42
33 Example: probit regression ind Model: Y i Ber(Φ(xi β)), i = 1,..., n. Suppose β has a normal prior. Not directly obvious how to implement Gibbs to get a sample from the posterior distribution of β. Recall, from EM notes, that this model can be simplified by introducing some missing data. The conditional distribution of the missing data, given the observed data and β, makes up one part of the full conditionals. The other part of the full conditionals is simple since the model for the complete data is, by construction, nice. 33 / 42
34 Example: probit regression (cont.) Missing data: Z 1,..., Z n where Z i N(x i β, 1) and Y i = I Zi >0, i = 1,..., n. Full conditionals: Distribution of β, given (Y, Z), only depends on Z and is easy because normal prior for β is conjugate. Distribution of Z, given (Y, β), is a truncated normal... Though I ve not given the precise details, the steps for constructing a Gibbs sampler are not too difficult; 9 see Section in Ghosh et al (2006) for details. 9 The only potential difficulty is simulating from a truncated normal when the truncation point is extreme, but remember we have talked about extreme normal tail probabilities before / 42
35 Example: Dirichlet process mixture In Bayesian nonparametrics, the Dirichlet process mixture model is probably the most widely used. Flexible model for density estimation normal mixture density with unspecified component means (and variances), but also doesn t specify the number of components. The main challenge with using mixture models is choosing how many components to use; this DPM selects the number of components automatically. Interesting that, despite being a nonparametric model, the computations are not too hard, just a Gibbs sampler. Simplest algorithm in Escobar & West (JASA 1995); a nice slice sampler is proposed in Kalli et al (Stat Comp 2011). 35 / 42
36 Example: Dirichlet process mixture (cont.) Using the slice sampler from Kalli et al to fit same normal mixture model to the galaxy data from HW. R code for this on my research page. Density Post mean Kernel Probability Y new Number of components 36 / 42
37 Outline 1 Introduction 2 Crash course on Markov chains 3 Motivation, revisited 4 Metropolis Hastings algorithm 5 Gibbs sampler 6 Some MCMC diagnostics 7 Conclusion 37 / 42
38 Diagnostic plots Sample path plot, or trace plot: Can reveal any residual dependence after burn-in. Idea is that a sample path of iid samples should show no trend, so if there is minimal trend in our sample plot, then we can be comfortable treating samples as independent. Autocorrelation plot: Plot sample correlation of {(X t, X t+r ) : t = 1, 2,...} as a function of the lag r. What to see the autocorrelation plot decay rapidly, suggesting that the dependence along the chain is not too strong. If these plots indicate that chain has not yet converged to stationarity, then you can run the chain longer or make some other modifications, e.g., Transformations thinning 38 / 42
39 Other considerations The practical/theoretical rate of convergence can depend on the parametrization; see homework. There is no agreement in the stat community about how many chains to run, how long should the burn in be, etc. Charles Geyer (Univ Minnesota) strongly supports running only one long chain check out his rants Gelman & Rubin suggest running several shorter chains with different starting points; textbook gives their diagnostic test. 39 / 42
40 Outline 1 Introduction 2 Crash course on Markov chains 3 Motivation, revisited 4 Metropolis Hastings algorithm 5 Gibbs sampler 6 Some MCMC diagnostics 7 Conclusion 40 / 42
41 Remarks MCMC methods are powerful because they give fairly general procedures able to solve a variety of important problems. There are black-box software implementations: The mcmc package in R will do random walk MH; In SAS, PROC MCMC does similar things; BUGS ( Bayes Using Gibbs Sampler ). However, it s a bad idea to blindly use these things without fully understanding what they re doing and whether or not they will actually work in your problem. Also important to look at convergence diagnostics before the simulation results be used for inference. 41 / 42
42 Remarks (cont.) Our focus here was on relatively simple MCMC methods. Don t think that MH, Gibbs, etc are all separate methods, they can be combined. For example, if one of the full conditionals is difficult to sample from, one might consider a MH step 10 within Gibbs to sample this conditional. Book by Robert & Casella has some details about more advanced MCMC methods, including various combinations of these standard techniques. 10 Could also use accept reject for this / 42
Bayesian Methods for Machine Learning
Bayesian Methods for Machine Learning CS 584: Big Data Analytics Material adapted from Radford Neal s tutorial (http://ftp.cs.utoronto.ca/pub/radford/bayes-tut.pdf), Zoubin Ghahramni (http://hunch.net/~coms-4771/zoubin_ghahramani_bayesian_learning.pdf),
More informationStat 451 Lecture Notes Monte Carlo Integration
Stat 451 Lecture Notes 06 12 Monte Carlo Integration Ryan Martin UIC www.math.uic.edu/~rgmartin 1 Based on Chapter 6 in Givens & Hoeting, Chapter 23 in Lange, and Chapters 3 4 in Robert & Casella 2 Updated:
More informationStat 535 C - Statistical Computing & Monte Carlo Methods. Lecture 15-7th March Arnaud Doucet
Stat 535 C - Statistical Computing & Monte Carlo Methods Lecture 15-7th March 2006 Arnaud Doucet Email: arnaud@cs.ubc.ca 1 1.1 Outline Mixture and composition of kernels. Hybrid algorithms. Examples Overview
More informationCPSC 540: Machine Learning
CPSC 540: Machine Learning MCMC and Non-Parametric Bayes Mark Schmidt University of British Columbia Winter 2016 Admin I went through project proposals: Some of you got a message on Piazza. No news is
More informationStat 451 Lecture Notes Simulating Random Variables
Stat 451 Lecture Notes 05 12 Simulating Random Variables Ryan Martin UIC www.math.uic.edu/~rgmartin 1 Based on Chapter 6 in Givens & Hoeting, Chapter 22 in Lange, and Chapter 2 in Robert & Casella 2 Updated:
More informationMCMC algorithms for fitting Bayesian models
MCMC algorithms for fitting Bayesian models p. 1/1 MCMC algorithms for fitting Bayesian models Sudipto Banerjee sudiptob@biostat.umn.edu University of Minnesota MCMC algorithms for fitting Bayesian models
More informationStat 451 Lecture Notes Numerical Integration
Stat 451 Lecture Notes 03 12 Numerical Integration Ryan Martin UIC www.math.uic.edu/~rgmartin 1 Based on Chapter 5 in Givens & Hoeting, and Chapters 4 & 18 of Lange 2 Updated: February 11, 2016 1 / 29
More informationMarkov chain Monte Carlo Lecture 9
Markov chain Monte Carlo Lecture 9 David Sontag New York University Slides adapted from Eric Xing and Qirong Ho (CMU) Limitations of Monte Carlo Direct (unconditional) sampling Hard to get rare events
More informationComputational statistics
Computational statistics Markov Chain Monte Carlo methods Thierry Denœux March 2017 Thierry Denœux Computational statistics March 2017 1 / 71 Contents of this chapter When a target density f can be evaluated
More informationMarkov Chain Monte Carlo
Markov Chain Monte Carlo Recall: To compute the expectation E ( h(y ) ) we use the approximation E(h(Y )) 1 n n h(y ) t=1 with Y (1),..., Y (n) h(y). Thus our aim is to sample Y (1),..., Y (n) from f(y).
More informationMarkov Chain Monte Carlo methods
Markov Chain Monte Carlo methods By Oleg Makhnin 1 Introduction a b c M = d e f g h i 0 f(x)dx 1.1 Motivation 1.1.1 Just here Supresses numbering 1.1.2 After this 1.2 Literature 2 Method 2.1 New math As
More informationIntroduction to Machine Learning CMU-10701
Introduction to Machine Learning CMU-10701 Markov Chain Monte Carlo Methods Barnabás Póczos & Aarti Singh Contents Markov Chain Monte Carlo Methods Goal & Motivation Sampling Rejection Importance Markov
More informationMarkov Chain Monte Carlo (MCMC)
Markov Chain Monte Carlo (MCMC Dependent Sampling Suppose we wish to sample from a density π, and we can evaluate π as a function but have no means to directly generate a sample. Rejection sampling can
More informationMarkov Chain Monte Carlo methods
Markov Chain Monte Carlo methods Tomas McKelvey and Lennart Svensson Signal Processing Group Department of Signals and Systems Chalmers University of Technology, Sweden November 26, 2012 Today s learning
More informationLecture 7 and 8: Markov Chain Monte Carlo
Lecture 7 and 8: Markov Chain Monte Carlo 4F13: Machine Learning Zoubin Ghahramani and Carl Edward Rasmussen Department of Engineering University of Cambridge http://mlg.eng.cam.ac.uk/teaching/4f13/ Ghahramani
More informationLecture 8: The Metropolis-Hastings Algorithm
30.10.2008 What we have seen last time: Gibbs sampler Key idea: Generate a Markov chain by updating the component of (X 1,..., X p ) in turn by drawing from the full conditionals: X (t) j Two drawbacks:
More information27 : Distributed Monte Carlo Markov Chain. 1 Recap of MCMC and Naive Parallel Gibbs Sampling
10-708: Probabilistic Graphical Models 10-708, Spring 2014 27 : Distributed Monte Carlo Markov Chain Lecturer: Eric P. Xing Scribes: Pengtao Xie, Khoa Luu In this scribe, we are going to review the Parallel
More informationIntroduction to Bayesian Statistics and Markov Chain Monte Carlo Estimation. EPSY 905: Multivariate Analysis Spring 2016 Lecture #10: April 6, 2016
Introduction to Bayesian Statistics and Markov Chain Monte Carlo Estimation EPSY 905: Multivariate Analysis Spring 2016 Lecture #10: April 6, 2016 EPSY 905: Intro to Bayesian and MCMC Today s Class An
More informationLecture 6: Markov Chain Monte Carlo
Lecture 6: Markov Chain Monte Carlo D. Jason Koskinen koskinen@nbi.ku.dk Photo by Howard Jackman University of Copenhagen Advanced Methods in Applied Statistics Feb - Apr 2016 Niels Bohr Institute 2 Outline
More information16 : Approximate Inference: Markov Chain Monte Carlo
10-708: Probabilistic Graphical Models 10-708, Spring 2017 16 : Approximate Inference: Markov Chain Monte Carlo Lecturer: Eric P. Xing Scribes: Yuan Yang, Chao-Ming Yen 1 Introduction As the target distribution
More informationProbabilistic Graphical Models Lecture 17: Markov chain Monte Carlo
Probabilistic Graphical Models Lecture 17: Markov chain Monte Carlo Andrew Gordon Wilson www.cs.cmu.edu/~andrewgw Carnegie Mellon University March 18, 2015 1 / 45 Resources and Attribution Image credits,
More informationBayesian GLMs and Metropolis-Hastings Algorithm
Bayesian GLMs and Metropolis-Hastings Algorithm We have seen that with conjugate or semi-conjugate prior distributions the Gibbs sampler can be used to sample from the posterior distribution. In situations,
More informationSC7/SM6 Bayes Methods HT18 Lecturer: Geoff Nicholls Lecture 2: Monte Carlo Methods Notes and Problem sheets are available at http://www.stats.ox.ac.uk/~nicholls/bayesmethods/ and via the MSc weblearn pages.
More informationSAMPLING ALGORITHMS. In general. Inference in Bayesian models
SAMPLING ALGORITHMS SAMPLING ALGORITHMS In general A sampling algorithm is an algorithm that outputs samples x 1, x 2,... from a given distribution P or density p. Sampling algorithms can for example be
More informationThe Bayesian Choice. Christian P. Robert. From Decision-Theoretic Foundations to Computational Implementation. Second Edition.
Christian P. Robert The Bayesian Choice From Decision-Theoretic Foundations to Computational Implementation Second Edition With 23 Illustrations ^Springer" Contents Preface to the Second Edition Preface
More information17 : Markov Chain Monte Carlo
10-708: Probabilistic Graphical Models, Spring 2015 17 : Markov Chain Monte Carlo Lecturer: Eric P. Xing Scribes: Heran Lin, Bin Deng, Yun Huang 1 Review of Monte Carlo Methods 1.1 Overview Monte Carlo
More information13 Notes on Markov Chain Monte Carlo
13 Notes on Markov Chain Monte Carlo Markov Chain Monte Carlo is a big, and currently very rapidly developing, subject in statistical computation. Many complex and multivariate types of random data, useful
More informationMarkov Chain Monte Carlo Inference. Siamak Ravanbakhsh Winter 2018
Graphical Models Markov Chain Monte Carlo Inference Siamak Ravanbakhsh Winter 2018 Learning objectives Markov chains the idea behind Markov Chain Monte Carlo (MCMC) two important examples: Gibbs sampling
More informationMCMC: Markov Chain Monte Carlo
I529: Machine Learning in Bioinformatics (Spring 2013) MCMC: Markov Chain Monte Carlo Yuzhen Ye School of Informatics and Computing Indiana University, Bloomington Spring 2013 Contents Review of Markov
More informationPrinciples of Bayesian Inference
Principles of Bayesian Inference Sudipto Banerjee University of Minnesota July 20th, 2008 1 Bayesian Principles Classical statistics: model parameters are fixed and unknown. A Bayesian thinks of parameters
More informationSTAT 425: Introduction to Bayesian Analysis
STAT 425: Introduction to Bayesian Analysis Marina Vannucci Rice University, USA Fall 2017 Marina Vannucci (Rice University, USA) Bayesian Analysis (Part 2) Fall 2017 1 / 19 Part 2: Markov chain Monte
More informationBayesian Regression Linear and Logistic Regression
When we want more than point estimates Bayesian Regression Linear and Logistic Regression Nicole Beckage Ordinary Least Squares Regression and Lasso Regression return only point estimates But what if we
More informationBayesian Networks in Educational Assessment
Bayesian Networks in Educational Assessment Estimating Parameters with MCMC Bayesian Inference: Expanding Our Context Roy Levy Arizona State University Roy.Levy@asu.edu 2017 Roy Levy MCMC 1 MCMC 2 Posterior
More informationMCMC Methods: Gibbs and Metropolis
MCMC Methods: Gibbs and Metropolis Patrick Breheny February 28 Patrick Breheny BST 701: Bayesian Modeling in Biostatistics 1/30 Introduction As we have seen, the ability to sample from the posterior distribution
More informationAdaptive Monte Carlo methods
Adaptive Monte Carlo methods Jean-Michel Marin Projet Select, INRIA Futurs, Université Paris-Sud joint with Randal Douc (École Polytechnique), Arnaud Guillin (Université de Marseille) and Christian Robert
More informationMarkov chain Monte Carlo
Markov chain Monte Carlo Feng Li feng.li@cufe.edu.cn School of Statistics and Mathematics Central University of Finance and Economics Revised on April 24, 2017 Today we are going to learn... 1 Markov Chains
More informationMarkov chain Monte Carlo methods in atmospheric remote sensing
1 / 45 Markov chain Monte Carlo methods in atmospheric remote sensing Johanna Tamminen johanna.tamminen@fmi.fi ESA Summer School on Earth System Monitoring and Modeling July 3 Aug 11, 212, Frascati July,
More informationContents. Part I: Fundamentals of Bayesian Inference 1
Contents Preface xiii Part I: Fundamentals of Bayesian Inference 1 1 Probability and inference 3 1.1 The three steps of Bayesian data analysis 3 1.2 General notation for statistical inference 4 1.3 Bayesian
More informationIntroduction to Markov Chain Monte Carlo & Gibbs Sampling
Introduction to Markov Chain Monte Carlo & Gibbs Sampling Prof. Nicholas Zabaras Sibley School of Mechanical and Aerospace Engineering 101 Frank H. T. Rhodes Hall Ithaca, NY 14853-3801 Email: zabaras@cornell.edu
More informationComputer intensive statistical methods
Lecture 13 MCMC, Hybrid chains October 13, 2015 Jonas Wallin jonwal@chalmers.se Chalmers, Gothenburg university MH algorithm, Chap:6.3 The metropolis hastings requires three objects, the distribution of
More informationHastings-within-Gibbs Algorithm: Introduction and Application on Hierarchical Model
UNIVERSITY OF TEXAS AT SAN ANTONIO Hastings-within-Gibbs Algorithm: Introduction and Application on Hierarchical Model Liang Jing April 2010 1 1 ABSTRACT In this paper, common MCMC algorithms are introduced
More informationBrief introduction to Markov Chain Monte Carlo
Brief introduction to Department of Probability and Mathematical Statistics seminar Stochastic modeling in economics and finance November 7, 2011 Brief introduction to Content 1 and motivation Classical
More informationSampling from complex probability distributions
Sampling from complex probability distributions Louis J. M. Aslett (louis.aslett@durham.ac.uk) Department of Mathematical Sciences Durham University UTOPIAE Training School II 4 July 2017 1/37 Motivation
More informationSTA 4273H: Statistical Machine Learning
STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Computer Science! Department of Statistical Sciences! rsalakhu@cs.toronto.edu! h0p://www.cs.utoronto.ca/~rsalakhu/ Lecture 7 Approximate
More informationMarkov chain Monte Carlo
Markov chain Monte Carlo Markov chain Monte Carlo (MCMC) Gibbs and Metropolis Hastings Slice sampling Practical details Iain Murray http://iainmurray.net/ Reminder Need to sample large, non-standard distributions:
More informationST 740: Markov Chain Monte Carlo
ST 740: Markov Chain Monte Carlo Alyson Wilson Department of Statistics North Carolina State University October 14, 2012 A. Wilson (NCSU Stsatistics) MCMC October 14, 2012 1 / 20 Convergence Diagnostics:
More informationHere θ = (α, β) is the unknown parameter. The likelihood function is
Stat 591 Notes Logistic regression and Metropolis Hastings example Ryan Martin (rgmartin@uic.edu) November 2, 2013 Introduction (This part taken from Example 1.13 in Robert and Casella, 2004.) In 1986,
More informationA quick introduction to Markov chains and Markov chain Monte Carlo (revised version)
A quick introduction to Markov chains and Markov chain Monte Carlo (revised version) Rasmus Waagepetersen Institute of Mathematical Sciences Aalborg University 1 Introduction These notes are intended to
More informationMH I. Metropolis-Hastings (MH) algorithm is the most popular method of getting dependent samples from a probability distribution
MH I Metropolis-Hastings (MH) algorithm is the most popular method of getting dependent samples from a probability distribution a lot of Bayesian mehods rely on the use of MH algorithm and it s famous
More informationThe Particle Filter. PD Dr. Rudolph Triebel Computer Vision Group. Machine Learning for Computer Vision
The Particle Filter Non-parametric implementation of Bayes filter Represents the belief (posterior) random state samples. by a set of This representation is approximate. Can represent distributions that
More informationStat 535 C - Statistical Computing & Monte Carlo Methods. Lecture February Arnaud Doucet
Stat 535 C - Statistical Computing & Monte Carlo Methods Lecture 13-28 February 2006 Arnaud Doucet Email: arnaud@cs.ubc.ca 1 1.1 Outline Limitations of Gibbs sampling. Metropolis-Hastings algorithm. Proof
More informationMetropolis-Hastings Algorithm
Strength of the Gibbs sampler Metropolis-Hastings Algorithm Easy algorithm to think about. Exploits the factorization properties of the joint probability distribution. No difficult choices to be made to
More informationCSC 2541: Bayesian Methods for Machine Learning
CSC 2541: Bayesian Methods for Machine Learning Radford M. Neal, University of Toronto, 2011 Lecture 3 More Markov Chain Monte Carlo Methods The Metropolis algorithm isn t the only way to do MCMC. We ll
More informationHierarchical models. Dr. Jarad Niemi. August 31, Iowa State University. Jarad Niemi (Iowa State) Hierarchical models August 31, / 31
Hierarchical models Dr. Jarad Niemi Iowa State University August 31, 2017 Jarad Niemi (Iowa State) Hierarchical models August 31, 2017 1 / 31 Normal hierarchical model Let Y ig N(θ g, σ 2 ) for i = 1,...,
More informationThe Mixture Approach for Simulating New Families of Bivariate Distributions with Specified Correlations
The Mixture Approach for Simulating New Families of Bivariate Distributions with Specified Correlations John R. Michael, Significance, Inc. and William R. Schucany, Southern Methodist University The mixture
More informationStatistical Machine Learning Lecture 8: Markov Chain Monte Carlo Sampling
1 / 27 Statistical Machine Learning Lecture 8: Markov Chain Monte Carlo Sampling Melih Kandemir Özyeğin University, İstanbul, Turkey 2 / 27 Monte Carlo Integration The big question : Evaluate E p(z) [f(z)]
More informationStat 535 C - Statistical Computing & Monte Carlo Methods. Arnaud Doucet.
Stat 535 C - Statistical Computing & Monte Carlo Methods Arnaud Doucet Email: arnaud@cs.ubc.ca 1 1.1 Outline Introduction to Markov chain Monte Carlo The Gibbs Sampler Examples Overview of the Lecture
More informationBayesian Inference and MCMC
Bayesian Inference and MCMC Aryan Arbabi Partly based on MCMC slides from CSC412 Fall 2018 1 / 18 Bayesian Inference - Motivation Consider we have a data set D = {x 1,..., x n }. E.g each x i can be the
More information36-463/663Multilevel and Hierarchical Models
36-463/663Multilevel and Hierarchical Models From Bayes to MCMC to MLMs Brian Junker 132E Baker Hall brian@stat.cmu.edu 1 Outline Bayesian Statistics and MCMC Distribution of Skill Mastery in a Population
More informationBayesian linear regression
Bayesian linear regression Linear regression is the basis of most statistical modeling. The model is Y i = X T i β + ε i, where Y i is the continuous response X i = (X i1,..., X ip ) T is the corresponding
More informationAdvances and Applications in Perfect Sampling
and Applications in Perfect Sampling Ph.D. Dissertation Defense Ulrike Schneider advisor: Jem Corcoran May 8, 2003 Department of Applied Mathematics University of Colorado Outline Introduction (1) MCMC
More informationLecture 16: Mixtures of Generalized Linear Models
Lecture 16: Mixtures of Generalized Linear Models October 26, 2006 Setting Outline Often, a single GLM may be insufficiently flexible to characterize the data Setting Often, a single GLM may be insufficiently
More informationMarkov chain Monte Carlo
1 / 26 Markov chain Monte Carlo Timothy Hanson 1 and Alejandro Jara 2 1 Division of Biostatistics, University of Minnesota, USA 2 Department of Statistics, Universidad de Concepción, Chile IAP-Workshop
More informationStat 516, Homework 1
Stat 516, Homework 1 Due date: October 7 1. Consider an urn with n distinct balls numbered 1,..., n. We sample balls from the urn with replacement. Let N be the number of draws until we encounter a ball
More informationProbabilistic Graphical Models
Probabilistic Graphical Models Lecture 11 CRFs, Exponential Family CS/CNS/EE 155 Andreas Krause Announcements Homework 2 due today Project milestones due next Monday (Nov 9) About half the work should
More informationBayesian Inference in GLMs. Frequentists typically base inferences on MLEs, asymptotic confidence
Bayesian Inference in GLMs Frequentists typically base inferences on MLEs, asymptotic confidence limits, and log-likelihood ratio tests Bayesians base inferences on the posterior distribution of the unknowns
More informationSTA 216, GLM, Lecture 16. October 29, 2007
STA 216, GLM, Lecture 16 October 29, 2007 Efficient Posterior Computation in Factor Models Underlying Normal Models Generalized Latent Trait Models Formulation Genetic Epidemiology Illustration Structural
More informationLikelihood-free MCMC
Bayesian inference for stable distributions with applications in finance Department of Mathematics University of Leicester September 2, 2011 MSc project final presentation Outline 1 2 3 4 Classical Monte
More informationeqr094: Hierarchical MCMC for Bayesian System Reliability
eqr094: Hierarchical MCMC for Bayesian System Reliability Alyson G. Wilson Statistical Sciences Group, Los Alamos National Laboratory P.O. Box 1663, MS F600 Los Alamos, NM 87545 USA Phone: 505-667-9167
More informationNonparametric Bayesian Methods (Gaussian Processes)
[70240413 Statistical Machine Learning, Spring, 2015] Nonparametric Bayesian Methods (Gaussian Processes) Jun Zhu dcszj@mail.tsinghua.edu.cn http://bigml.cs.tsinghua.edu.cn/~jun State Key Lab of Intelligent
More informationINTRODUCTION TO BAYESIAN STATISTICS
INTRODUCTION TO BAYESIAN STATISTICS Sarat C. Dass Department of Statistics & Probability Department of Computer Science & Engineering Michigan State University TOPICS The Bayesian Framework Different Types
More informationMarkov Chain Monte Carlo in Practice
Markov Chain Monte Carlo in Practice Edited by W.R. Gilks Medical Research Council Biostatistics Unit Cambridge UK S. Richardson French National Institute for Health and Medical Research Vilejuif France
More informationExercises Tutorial at ICASSP 2016 Learning Nonlinear Dynamical Models Using Particle Filters
Exercises Tutorial at ICASSP 216 Learning Nonlinear Dynamical Models Using Particle Filters Andreas Svensson, Johan Dahlin and Thomas B. Schön March 18, 216 Good luck! 1 [Bootstrap particle filter for
More informationA short diversion into the theory of Markov chains, with a view to Markov chain Monte Carlo methods
A short diversion into the theory of Markov chains, with a view to Markov chain Monte Carlo methods by Kasper K. Berthelsen and Jesper Møller June 2004 2004-01 DEPARTMENT OF MATHEMATICAL SCIENCES AALBORG
More informationMetropolis Hastings. Rebecca C. Steorts Bayesian Methods and Modern Statistics: STA 360/601. Module 9
Metropolis Hastings Rebecca C. Steorts Bayesian Methods and Modern Statistics: STA 360/601 Module 9 1 The Metropolis-Hastings algorithm is a general term for a family of Markov chain simulation methods
More informationSupplement to A Hierarchical Approach for Fitting Curves to Response Time Measurements
Supplement to A Hierarchical Approach for Fitting Curves to Response Time Measurements Jeffrey N. Rouder Francis Tuerlinckx Paul L. Speckman Jun Lu & Pablo Gomez May 4 008 1 The Weibull regression model
More informationApril 20th, Advanced Topics in Machine Learning California Institute of Technology. Markov Chain Monte Carlo for Machine Learning
for for Advanced Topics in California Institute of Technology April 20th, 2017 1 / 50 Table of Contents for 1 2 3 4 2 / 50 History of methods for Enrico Fermi used to calculate incredibly accurate predictions
More informationDefault Priors and Effcient Posterior Computation in Bayesian
Default Priors and Effcient Posterior Computation in Bayesian Factor Analysis January 16, 2010 Presented by Eric Wang, Duke University Background and Motivation A Brief Review of Parameter Expansion Literature
More informationPrinciples of Bayesian Inference
Principles of Bayesian Inference Sudipto Banerjee 1 and Andrew O. Finley 2 1 Biostatistics, School of Public Health, University of Minnesota, Minneapolis, Minnesota, U.S.A. 2 Department of Forestry & Department
More informationIntroduction to Probabilistic Machine Learning
Introduction to Probabilistic Machine Learning Piyush Rai Dept. of CSE, IIT Kanpur (Mini-course 1) Nov 03, 2015 Piyush Rai (IIT Kanpur) Introduction to Probabilistic Machine Learning 1 Machine Learning
More informationMotivation Scale Mixutres of Normals Finite Gaussian Mixtures Skew-Normal Models. Mixture Models. Econ 690. Purdue University
Econ 690 Purdue University In virtually all of the previous lectures, our models have made use of normality assumptions. From a computational point of view, the reason for this assumption is clear: combined
More informationABC methods for phase-type distributions with applications in insurance risk problems
ABC methods for phase-type with applications problems Concepcion Ausin, Department of Statistics, Universidad Carlos III de Madrid Joint work with: Pedro Galeano, Universidad Carlos III de Madrid Simon
More informationBagging During Markov Chain Monte Carlo for Smoother Predictions
Bagging During Markov Chain Monte Carlo for Smoother Predictions Herbert K. H. Lee University of California, Santa Cruz Abstract: Making good predictions from noisy data is a challenging problem. Methods
More informationQuantifying Uncertainty
Sai Ravela M. I. T Last Updated: Spring 2013 1 Markov Chain Monte Carlo Monte Carlo sampling made for large scale problems via Markov Chains Monte Carlo Sampling Rejection Sampling Importance Sampling
More informationLecture 5. G. Cowan Lectures on Statistical Data Analysis Lecture 5 page 1
Lecture 5 1 Probability (90 min.) Definition, Bayes theorem, probability densities and their properties, catalogue of pdfs, Monte Carlo 2 Statistical tests (90 min.) general concepts, test statistics,
More informationDoing Bayesian Integrals
ASTR509-13 Doing Bayesian Integrals The Reverend Thomas Bayes (c.1702 1761) Philosopher, theologian, mathematician Presbyterian (non-conformist) minister Tunbridge Wells, UK Elected FRS, perhaps due to
More informationMarkov Chain Monte Carlo (MCMC) and Model Evaluation. August 15, 2017
Markov Chain Monte Carlo (MCMC) and Model Evaluation August 15, 2017 Frequentist Linking Frequentist and Bayesian Statistics How can we estimate model parameters and what does it imply? Want to find the
More informationMarkov chain Monte Carlo
Markov chain Monte Carlo Peter Beerli October 10, 2005 [this chapter is highly influenced by chapter 1 in Markov chain Monte Carlo in Practice, eds Gilks W. R. et al. Chapman and Hall/CRC, 1996] 1 Short
More informationBayesian data analysis in practice: Three simple examples
Bayesian data analysis in practice: Three simple examples Martin P. Tingley Introduction These notes cover three examples I presented at Climatea on 5 October 0. Matlab code is available by request to
More informationSTA 294: Stochastic Processes & Bayesian Nonparametrics
MARKOV CHAINS AND CONVERGENCE CONCEPTS Markov chains are among the simplest stochastic processes, just one step beyond iid sequences of random variables. Traditionally they ve been used in modelling a
More informationSAMSI Astrostatistics Tutorial. More Markov chain Monte Carlo & Demo of Mathematica software
SAMSI Astrostatistics Tutorial More Markov chain Monte Carlo & Demo of Mathematica software Phil Gregory University of British Columbia 26 Bayesian Logical Data Analysis for the Physical Sciences Contents:
More informationLect4: Exact Sampling Techniques and MCMC Convergence Analysis
Lect4: Exact Sampling Techniques and MCMC Convergence Analysis. Exact sampling. Convergence analysis of MCMC. First-hit time analysis for MCMC--ways to analyze the proposals. Outline of the Module Definitions
More informationMonte Carlo Inference Methods
Monte Carlo Inference Methods Iain Murray University of Edinburgh http://iainmurray.net Monte Carlo and Insomnia Enrico Fermi (1901 1954) took great delight in astonishing his colleagues with his remarkably
More informationReminder of some Markov Chain properties:
Reminder of some Markov Chain properties: 1. a transition from one state to another occurs probabilistically 2. only state that matters is where you currently are (i.e. given present, future is independent
More informationControl Variates for Markov Chain Monte Carlo
Control Variates for Markov Chain Monte Carlo Dellaportas, P., Kontoyiannis, I., and Tsourti, Z. Dept of Statistics, AUEB Dept of Informatics, AUEB 1st Greek Stochastics Meeting Monte Carlo: Probability
More informationIntroduction to Bayesian methods in inverse problems
Introduction to Bayesian methods in inverse problems Ville Kolehmainen 1 1 Department of Applied Physics, University of Eastern Finland, Kuopio, Finland March 4 2013 Manchester, UK. Contents Introduction
More informationPattern Recognition and Machine Learning. Bishop Chapter 11: Sampling Methods
Pattern Recognition and Machine Learning Chapter 11: Sampling Methods Elise Arnaud Jakob Verbeek May 22, 2008 Outline of the chapter 11.1 Basic Sampling Algorithms 11.2 Markov Chain Monte Carlo 11.3 Gibbs
More information(5) Multi-parameter models - Gibbs sampling. ST440/540: Applied Bayesian Analysis
Summarizing a posterior Given the data and prior the posterior is determined Summarizing the posterior gives parameter estimates, intervals, and hypothesis tests Most of these computations are integrals
More informationBayesian model selection: methodology, computation and applications
Bayesian model selection: methodology, computation and applications David Nott Department of Statistics and Applied Probability National University of Singapore Statistical Genomics Summer School Program
More informationThe Pennsylvania State University The Graduate School RATIO-OF-UNIFORMS MARKOV CHAIN MONTE CARLO FOR GAUSSIAN PROCESS MODELS
The Pennsylvania State University The Graduate School RATIO-OF-UNIFORMS MARKOV CHAIN MONTE CARLO FOR GAUSSIAN PROCESS MODELS A Thesis in Statistics by Chris Groendyke c 2008 Chris Groendyke Submitted in
More information