The Metropolis-Hastings Algorithm. June 8, 2012
|
|
- Kristina Stafford
- 5 years ago
- Views:
Transcription
1 The Metropolis-Hastings Algorithm June 8, 22
2 The Plan. Understand what a simulated distribution is 2. Understand why the Metropolis-Hastings algorithm works 3. Learn how to apply the Metropolis-Hastings algorithm Most of today s lecture can be found in Chib (2) and An and Schorfheide (27)
3 Bayesian statistics: From last time Variance of estimator (frequentist) vs variance of parameter (Bayesian) Subjective view of probability Probabilities are statement about our knowledge Treating model parameters as random variables thus do not mean that we think that they necessarily vary over time.
4 Bayesian statistics: From last time Bayesian procedures can be derived from 4 Bayesian Principles. The Likelihood Principle 2. The Sufficiency Principle 3. The Conditionality Principle 4. The Stopping Rule Principle
5 Main concepts and notation The main components in Bayesian inference are: Data (observables) Z T R T n A model: Parameters θ R k A prior distribution p(θ) : R k R + Likelihood function p(z θ) : R T n R k R + Posterior density p(θ Z) : R T n R k R + We need a method to construct the posterior density
6 The end product of Bayesian statistics Most of Bayesian econometrics consists of simulating distributions of parameters using numerical methods. A simulated posterior is a numerical approximation to the distribution p(z θ)p(θ) This is useful since the the distribution p(z θ)p(θ) (by Bayes rule) is proportional to p(θ Z) p(θ Z) = p(z θ)p(θ) p(z) We rely on ergodicity, i.e. that the moments of the constructed sample correspond to the moments of the distribution p(θ Z) The most popular (and general) procedure to simulate the posterior is called the Metropolis-Hastings Algorithm
7 Metropolis-Hastings Metropolis-Hastings is a way to simulate a sample from a target distribution In practice, the target distribution will in most cases be the posterior density p(θ Z) but it doesn t have to be But what does it mean to sample from a given distribution?
8 Sampling from a distribution Example: Standard Normal distribution N(, ) If distribution is ergodic, sample shares all the properties of the true distribution asymptotically Why don t we just compute the moments directly? When we can, we should (as in the Normal distribution s case) Not always possible, either because tractability reasons or for computational burden
9 Sample from Standard Normal
10 Sample from Standard Normal
11 Sample from Standard Normal x 4
12 Sample from Standard Normal
13 Sampling from a distribution Example: Standard Normal distribution N(, ) If distribution is ergodic, sample shares all the properties of the true distribution asymptotically Why don t we just compute the moments directly? When we can, we should (as in the Normal distribution s case) Not always possible, either because tractability reasons or for computational burden
14 What is a Markov Chain? A process for which the distribution of next period variables are independent of the past, once we condition on the current state A Markov chain is a stochastic process with the Markov property Markov chain is sometimes taken to mean only processes with a countably finite number of states Here, the term will be used in the broader sense Markov Chain Monte Carlo methods provide a way of generating samples that share the properties of the target density (i.e. the object of interest)
15 What is a Markov Chain? Markov property p(θ j+ θ j ) = p(θ j+ θ j, θ j, θ j 2,..., θ ) Transition density function p(θ j+ θ j ) describes the distribution of θ j+ conditional on θ j. In most applications, we know the conditional transition density and can figure out unconditional properties like E(θ) and E ( θ 2) MCMC methods can be used to do the opposite: Determine a particular conditional transition density such that the unconditional distribution converges to that of the target distribution. Let s have a first look at the Metropolis-Hastings Algorithm
16 The Random-Walk Metropolis Algorithm. Start with an arbitrary value θ 2. Update from θ j to θ j+ (j =, 2,...J) by 2. Generate θ N (θ j, Σ) 2.2 Define 2.3 Take θ j+ = 3. Repeat Step 2 J times. But why does it work? α = min ( p(θ ) ) p(θ j ), { θ with probability α θ j otherwise } ()
17 Simulating distributions using MCMC methods We can look at a simple discrete state space example: θ can take two values θ { θ L, θ H} Probabilities p(θ L ) =.4 and p(θ H ) =.6 How can we construct a sequence θ (), θ (2), θ (3),..., θ (J) such that the relative number of occurrences of θ L and θ H in the sample correspond to those of the density described by p(θ)?
18 A two-state Markov Chain for? Transition probabilities are defined as ( π i,k = p θ j+ = θ i θ j = θ k) : i, k {L, H} Unconditional probabilities solves the equation [ ] [ ] [ π πl π H = πl π LL π HL H π LH π HH ] Problem: Define transition probabilities such that unconditional probabilities equal are those of the target distribution
19 A two-state Markov Chain for? We want the chain to spend more time in the more likely state θ H but not all the time. If θ j = θ L then θ j+ = θ H If θ j = θ H then θ j+ = θ L with probability α and θ j+ = θ H with probability ( α) Finding the right α : [ πl π H ] = [ πl π H ] [ α ( α) ] We get two equations, π L = απ H, for α gives α = π L π H π H = π L + ( α)π H, solving
20 Simulating distributions using MCMC methods Let s do an example by hand.
21 Simulating distributions using MCMC methods Using the ratio of relative probability/density to decide whether to accept a candidate turns out to be extremely general: exactly the same condition ensures that Markov Chain converges to target distribution also for continuous parameter spaces Loosely speaking, it is the condition that makes sure that the Markov chain spend just the right amount of time at each point in the parameter space How the candidate is generated almost doesn t matter at all as long as chain is: Irreducible Aperiodic Partly depends on choice that determine how proposal density is generated
22 Proposal density Any proposal density with infinite support would do (e.g. normal) The choice a matter of efficiency. Some options: RW-MH Adaptive Random Walk Independent M-H
23 The Random-Walk Metropolis Algorithm. Start with an arbitrary value θ 2. Update from θ j to θ j+ (j =, 2,...J) by 2. Generate θ N (θ j, Σ) 2.2 Define 2.3 Take θ j+ = 3. Repeat Step 2 J times. α = min ( L (Z θ ) ) L (Z θ j ), { θ with probability α θ j otherwise We can use this to estimate the likelihood function of a simple DSGE model } (2)
24 A simple DSGE model x t = ρx t + ut x y t = E t (y t+ ) γ [r t E t (π t+ )] + ut y π t = E t (π t+ ) + κ [y t x t ] + ut π r t = φ π π t
25 A simple DSGE model Substitute in the interest rate in the Euler equation x t = ρx t + u x t y t = E t (y t+ ) γ [φ ππ t E t (π t+ )] + u y t π t = E t (π t+ ) + κ [y t x t ] + u π t
26 Solving model using method of undetermined coefficients Conjecture that model can be put in the form x t = ρx t + ut x y t = ax t + ut y π t = bx t + ut π Why is this a good guess?
27 Solving model using method of undetermined coefficients Substitute in conjectured form of solution (ignoring the shocks ut y and ut π for now) into structural equation ax t = aρx t γ [φ πbx t bρx t ] bx t = bρx t + κ [ax t x t ] where we used that x t = ρx t + u x t implies that E[x t+ x t ] = ρx t
28 Solving model using method of undetermined coefficients Equate coefficients on right and left hand side a = aρ γ φ πb + γ bρ b = bρ + κ [a ] or [ ( ρ) γ (φ π ρ) κ ( ρ) ] [ a b ] = [ κ ]
29 Solving model using method of undetermined coefficients Solve for a and b [ ] a = b [ ( ρ) γ (φ π ρ) κ ( ρ) ] [ κ ] or [ a b ] [ = κ φ ρ c κγ ρ c ] where c = γ κρ 2γρ + κφ + γρ 2 <
30 A simple DSGE model The solved model x t = ρx t + u x t y t = κ ρ φ π x t + ut y c π t = κγ ρ x t + ut π c where c = γ κρ 2γρ + κφ + γρ 2 < We want to estimate the distributions of θ = {ρ, γ, κ, φ, σ x, σ y, σ π, }
31 A simple DSGE model Put the solved model in state space form X t = AX t + Cu t Z t = DX t + v t where X t = x t, A = ρ, Cu t = ut x [ ] [ yt κ φπ ρ Z t =, D = c π t κγ ρ c ] [ u y, v t = t u π t ]
32 The log likelihood function of a state space system For a given state space system X t = AX t + Cu t Z t (p ) = DX t + v t we can evaluate the log likelihood by computing T [ L(Z Θ) =.5 p ln(2π) + ln Ω t + Z ] tω t Z t t= where Z t are the innovation from the Kalman filter
33 The parameter vector θ and the variance of random walk innovations in MCMC Σ Parameterize the model according to θ = {ρ, γ, κ, φ, σ x, σ y, σ π, } = {.9, 2,.,.5,,, } Generate data from true model and T = Set starting value θ () = θ (OK, this option is not available in practice.) Set covariance matrix of random walk increments in Metropolis Algorithm proportional to absolute values of true parameters Σ = ε diag(abs(θ))
34 The Random-Walk Metropolis Algorithm We now have all we need:. Start with an arbitrary value θ 2. Update from θ j to θ j+ (j =, 2,...J) by 2. Generate θ N (θ j, Σ) 2.2 Define 2.3 Take θ j+ = 3. Repeat Step 2 J times α = min ( L (Z θ ) ) L (Z θ j ), { θ with probability α θ j otherwise } (3)
35 The MCMC with J=
36 The MCMC with J= x
37 The MCMC with J=
38 The MCMC with J= x
39 The MCMC with J= x x x x x x x 4
40 The MCMC with J=
41 Identification Some posterior distribution look flat Can be a short sample issue Nothing to do if we are estimating a model on real data, but here we can check Can be a problem with mapping between parameters and likelihood
42 The MCMC with T=2 and J=
43 Identification Posterior distribution for γ and κ kappa still look flat probably a true identification issue Little can be said about identification a priori
44 Convergence How many draws do we need? Optimal J increases with number of parameters One informal check is to plot the diagonal of the recursive covariance matrix of the MCMC j j θ (i) θ (i) for j =, 2,...J i=
45 Checking for convergence 2.5 x x
46 Checking for convergence J= x
47 Combining prior and sample information Sometimes we know more about the parameters than what the data tells us, i.e. we have some prior information.
48 What do we mean by prior information? For a DSGE model, we may have information about deep parameters Range of some parameters restricted by theory, e.g. risk aversion should be positive Discount rate is inverse of average real interest rates Price stickiness can be measured by surveys We may know something about the mean of a process
49 How do we combine prior and sample information? Bayes theorem: P (θ Z) P(Z) = P (Z θ) P(θ) P (θ Z) = P (Z θ) P(θ) P(Z) Since P(Z) is a constant, we can use P (Z θ) P(θ) as the posterior likelihood (a likelihood function is any function that is proportional to the probability). We now need to choose P(θ)
50 Choosing prior distributions The beta distribution is a good choice when parameter is in [,] where P(x) = ( x)b x a B(a, b) B(a, b) = (a )!(b )! (a + b )! Easier to parameterize using expression for mean, mode and variance: µ = σ 2 = a a + b, x = a a + b 2 ab (a + b) 2 (a + b + )
51 Examples of beta distributions holding mean fixed
52 Examples of beta distributions holding s.d. fixed
53 Choosing prior distributions The inverse gamma distribution is a good choice when parameter is positive P(x) = ba Γ(a) (/x)a+ exp( b/x) where Γ(a) = (a )! Again, easier to parameterize using expression for mean, mode and variance: µ = σ 2 = b b ; a >, x = a a + b 2 (a ) 2 (a 2) ; a > 2
54 Examples of inverse gamma distributions
55 The Random-Walk Metropolis Algorithm with priors. Start with an arbitrary value θ 2. Update from θ j to θ j+ (j =, 2,...J) by 2. Generate θ N (θ j, Σ) 2.2 Define 2.3 Take α = min θ j+ = ( L (Z θ ) P(θ ) ) L (Z θ j ) P(θ j ), { θ with probability α θ j otherwise The only difference compared to before is that the priors appear in the ratio in (3). } (4)
56 Practical implementation Since we compute log-likelihood we can use the log of the likelihood ratio in the Random Walk Metropolis Algorithm ln α = min [( ln L (Z θ ) + ln P(θ ) ln L (Z θ j ) ln P(θ j )), ] If priors across parameters are independent we have that ln P(θ j ) = ln P(θ,j ) + ln P(θ 2,j ) ln P(θ q,j ) where θ j = [ θ,j θ 2,j θ q,j ]
57 Let s get deep with an old friend: x t = ρx t + u x t y t = E t (y t+ ) γ [r t E t (π t+ )] + u y t π t = E t (π t+ ) + κ [y t x t ] + u π t r t = φ π π t The parameter κ is in the benchmark 3-equation NK model given by ( δ)( δβ) κ = δ where δ is the Calvo parameter of price stickiness and β is the discount factor. We now have a new parameter vector θ = {ρ, γ, δ, β, φ, σ x, σ y, σ π, }
58 The priors The prior on relative risk aversion γ is truncated Normal with mean 2 and s.d..5. The prior on the discount factor β is Beta with mean.99 and s.d.. The prior on the Calvo parameter δ is Beta with mean.75 and s.d
59 The Random-Walk Metropolis Algorithm with priors. Start with an arbitrary value θ 2. Update from θ j to θ j+ (j =, 2,...J) by 2. Generate θ N (θ j, Σ) 2.2 Define 2.3 Take α = min θ j+ = 3. Repeat Step 2 J times ( L (Z θ ) P(θ ) ) L (Z θ j ) P(θ j ), { θ with probability α θ j otherwise } (5)
60 The log prior The log prior is given by ln P(θ j ) = ln P(θ,j ) + ln P(θ 2,j ) ln P(θ q,j ) = ln P(γ j ) + ln P(δ j ) + ln P(β j ) since we can ignore the (constant) probabilities on the uniform priors. We then have that ln [L (Z θ ) P(θ )] = ln L(Z θ) + ln P(θ j ) = α = min i.e. all we need for the RWMA ( exp [ln L(Z θ ) + ln P(θ ) )] exp [ln L(Z θ j ) + ln P(θ j )],
61 Posterior with uniform and informative priors
62 Inference about parameters Compute sample equivalent: Central tendencies: Mean, mode Variance Compute frequency of occurrence: How likely is it that θ and θ 2 are both smaller than?
63 That s it for today.
Bayesian Methods for DSGE models Lecture 1 Macro models as data generating processes
Bayesian Methods for DSGE models Lecture 1 Macro models as data generating processes Kristoffer Nimark CREI, Universitat Pompeu Fabra and Barcelona GSE June 30, 2014 Bayesian Methods for DSGE models Course
More informationML Estimation of State Space Models. March 17, 2017
ML Estimation of State Space Models March 17, 2017 Recap from last time: Unobserved Component model of Inflation π t = τ t + η t τ t = τ t 1 + ε t Decomposes inflation into permanent (τ) and transitory
More informationML Estimation of State Space Models. February 29, 2016
ML Estimation of State Space Models February 29, 2016 Recap from last time: Unobserved Component model of Inflation π t = τ t + η t τ t = τ t 1 + ε t Decomposes inflation into permanent (τ) and transitory
More informationBayesian Inference and MCMC
Bayesian Inference and MCMC Aryan Arbabi Partly based on MCMC slides from CSC412 Fall 2018 1 / 18 Bayesian Inference - Motivation Consider we have a data set D = {x 1,..., x n }. E.g each x i can be the
More informationBayesian Methods for Machine Learning
Bayesian Methods for Machine Learning CS 584: Big Data Analytics Material adapted from Radford Neal s tutorial (http://ftp.cs.utoronto.ca/pub/radford/bayes-tut.pdf), Zoubin Ghahramni (http://hunch.net/~coms-4771/zoubin_ghahramani_bayesian_learning.pdf),
More informationMarkov chain Monte Carlo
1 / 26 Markov chain Monte Carlo Timothy Hanson 1 and Alejandro Jara 2 1 Division of Biostatistics, University of Minnesota, USA 2 Department of Statistics, Universidad de Concepción, Chile IAP-Workshop
More informationDSGE Methods. Estimation of DSGE models: Maximum Likelihood & Bayesian. Willi Mutschler, M.Sc.
DSGE Methods Estimation of DSGE models: Maximum Likelihood & Bayesian Willi Mutschler, M.Sc. Institute of Econometrics and Economic Statistics University of Münster willi.mutschler@uni-muenster.de Summer
More informationLecture 2: From Linear Regression to Kalman Filter and Beyond
Lecture 2: From Linear Regression to Kalman Filter and Beyond Department of Biomedical Engineering and Computational Science Aalto University January 26, 2012 Contents 1 Batch and Recursive Estimation
More informationLecture 2: From Linear Regression to Kalman Filter and Beyond
Lecture 2: From Linear Regression to Kalman Filter and Beyond January 18, 2017 Contents 1 Batch and Recursive Estimation 2 Towards Bayesian Filtering 3 Kalman Filter and Bayesian Filtering and Smoothing
More informationPrinciples of Bayesian Inference
Principles of Bayesian Inference Sudipto Banerjee University of Minnesota July 20th, 2008 1 Bayesian Principles Classical statistics: model parameters are fixed and unknown. A Bayesian thinks of parameters
More informationDavid Giles Bayesian Econometrics
David Giles Bayesian Econometrics 1. General Background 2. Constructing Prior Distributions 3. Properties of Bayes Estimators and Tests 4. Bayesian Analysis of the Multiple Regression Model 5. Bayesian
More informationLecture 7 and 8: Markov Chain Monte Carlo
Lecture 7 and 8: Markov Chain Monte Carlo 4F13: Machine Learning Zoubin Ghahramani and Carl Edward Rasmussen Department of Engineering University of Cambridge http://mlg.eng.cam.ac.uk/teaching/4f13/ Ghahramani
More informationSC7/SM6 Bayes Methods HT18 Lecturer: Geoff Nicholls Lecture 2: Monte Carlo Methods Notes and Problem sheets are available at http://www.stats.ox.ac.uk/~nicholls/bayesmethods/ and via the MSc weblearn pages.
More informationCSC 2541: Bayesian Methods for Machine Learning
CSC 2541: Bayesian Methods for Machine Learning Radford M. Neal, University of Toronto, 2011 Lecture 3 More Markov Chain Monte Carlo Methods The Metropolis algorithm isn t the only way to do MCMC. We ll
More informationLecture 8: The Metropolis-Hastings Algorithm
30.10.2008 What we have seen last time: Gibbs sampler Key idea: Generate a Markov chain by updating the component of (X 1,..., X p ) in turn by drawing from the full conditionals: X (t) j Two drawbacks:
More informationMarkov chain Monte Carlo
Markov chain Monte Carlo Feng Li feng.li@cufe.edu.cn School of Statistics and Mathematics Central University of Finance and Economics Revised on April 24, 2017 Today we are going to learn... 1 Markov Chains
More informationMCMC algorithms for fitting Bayesian models
MCMC algorithms for fitting Bayesian models p. 1/1 MCMC algorithms for fitting Bayesian models Sudipto Banerjee sudiptob@biostat.umn.edu University of Minnesota MCMC algorithms for fitting Bayesian models
More informationMarkov Chains and MCMC
Markov Chains and MCMC CompSci 590.02 Instructor: AshwinMachanavajjhala Lecture 4 : 590.02 Spring 13 1 Recap: Monte Carlo Method If U is a universe of items, and G is a subset satisfying some property,
More informationSequential Monte Carlo Methods (for DSGE Models)
Sequential Monte Carlo Methods (for DSGE Models) Frank Schorfheide University of Pennsylvania, PIER, CEPR, and NBER October 23, 2017 Some References These lectures use material from our joint work: Tempered
More informationReminder of some Markov Chain properties:
Reminder of some Markov Chain properties: 1. a transition from one state to another occurs probabilistically 2. only state that matters is where you currently are (i.e. given present, future is independent
More informationMarkov Chain Monte Carlo methods
Markov Chain Monte Carlo methods Tomas McKelvey and Lennart Svensson Signal Processing Group Department of Signals and Systems Chalmers University of Technology, Sweden November 26, 2012 Today s learning
More informationCS281A/Stat241A Lecture 22
CS281A/Stat241A Lecture 22 p. 1/4 CS281A/Stat241A Lecture 22 Monte Carlo Methods Peter Bartlett CS281A/Stat241A Lecture 22 p. 2/4 Key ideas of this lecture Sampling in Bayesian methods: Predictive distribution
More informationAnswers and expectations
Answers and expectations For a function f(x) and distribution P(x), the expectation of f with respect to P is The expectation is the average of f, when x is drawn from the probability distribution P E
More informationHierarchical models. Dr. Jarad Niemi. August 31, Iowa State University. Jarad Niemi (Iowa State) Hierarchical models August 31, / 31
Hierarchical models Dr. Jarad Niemi Iowa State University August 31, 2017 Jarad Niemi (Iowa State) Hierarchical models August 31, 2017 1 / 31 Normal hierarchical model Let Y ig N(θ g, σ 2 ) for i = 1,...,
More informationBayesian Computations for DSGE Models
Bayesian Computations for DSGE Models Frank Schorfheide University of Pennsylvania, PIER, CEPR, and NBER October 23, 2017 This Lecture is Based on Bayesian Estimation of DSGE Models Edward P. Herbst &
More informationBayesian Estimation with Sparse Grids
Bayesian Estimation with Sparse Grids Kenneth L. Judd and Thomas M. Mertens Institute on Computational Economics August 7, 27 / 48 Outline Introduction 2 Sparse grids Construction Integration with sparse
More informationBayesian Estimation of DSGE Models
Bayesian Estimation of DSGE Models Stéphane Adjemian Université du Maine, GAINS & CEPREMAP stephane.adjemian@univ-lemans.fr http://www.dynare.org/stepan June 28, 2011 June 28, 2011 Université du Maine,
More informationSequential Monte Carlo Methods
University of Pennsylvania Bradley Visitor Lectures October 23, 2017 Introduction Unfortunately, standard MCMC can be inaccurate, especially in medium and large-scale DSGE models: disentangling importance
More informationSession 3A: Markov chain Monte Carlo (MCMC)
Session 3A: Markov chain Monte Carlo (MCMC) John Geweke Bayesian Econometrics and its Applications August 15, 2012 ohn Geweke Bayesian Econometrics and its Session Applications 3A: Markov () chain Monte
More informationOnline appendix to On the stability of the excess sensitivity of aggregate consumption growth in the US
Online appendix to On the stability of the excess sensitivity of aggregate consumption growth in the US Gerdie Everaert 1, Lorenzo Pozzi 2, and Ruben Schoonackers 3 1 Ghent University & SHERPPA 2 Erasmus
More informationSTA 4273H: Statistical Machine Learning
STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Computer Science! Department of Statistical Sciences! rsalakhu@cs.toronto.edu! h0p://www.cs.utoronto.ca/~rsalakhu/ Lecture 7 Approximate
More informationComputational statistics
Computational statistics Markov Chain Monte Carlo methods Thierry Denœux March 2017 Thierry Denœux Computational statistics March 2017 1 / 71 Contents of this chapter When a target density f can be evaluated
More informationProbabilistic Machine Learning
Probabilistic Machine Learning Bayesian Nets, MCMC, and more Marek Petrik 4/18/2017 Based on: P. Murphy, K. (2012). Machine Learning: A Probabilistic Perspective. Chapter 10. Conditional Independence Independent
More informationBayesian Inference: Concept and Practice
Inference: Concept and Practice fundamentals Johan A. Elkink School of Politics & International Relations University College Dublin 5 June 2017 1 2 3 Bayes theorem In order to estimate the parameters of
More informationMarkov Chain Monte Carlo
Chapter 5 Markov Chain Monte Carlo MCMC is a kind of improvement of the Monte Carlo method By sampling from a Markov chain whose stationary distribution is the desired sampling distributuion, it is possible
More informationLECTURE 15 Markov chain Monte Carlo
LECTURE 15 Markov chain Monte Carlo There are many settings when posterior computation is a challenge in that one does not have a closed form expression for the posterior distribution. Markov chain Monte
More informationMCMC and Gibbs Sampling. Kayhan Batmanghelich
MCMC and Gibbs Sampling Kayhan Batmanghelich 1 Approaches to inference l Exact inference algorithms l l l The elimination algorithm Message-passing algorithm (sum-product, belief propagation) The junction
More informationMarkov Chain Monte Carlo, Numerical Integration
Markov Chain Monte Carlo, Numerical Integration (See Statistics) Trevor Gallen Fall 2015 1 / 1 Agenda Numerical Integration: MCMC methods Estimating Markov Chains Estimating latent variables 2 / 1 Numerical
More informationMonetary and Exchange Rate Policy Under Remittance Fluctuations. Technical Appendix and Additional Results
Monetary and Exchange Rate Policy Under Remittance Fluctuations Technical Appendix and Additional Results Federico Mandelman February In this appendix, I provide technical details on the Bayesian estimation.
More informationMarkov Chain Monte Carlo Methods
Markov Chain Monte Carlo Methods John Geweke University of Iowa, USA 2005 Institute on Computational Economics University of Chicago - Argonne National Laboaratories July 22, 2005 The problem p (θ, ω I)
More informationCSC 2541: Bayesian Methods for Machine Learning
CSC 2541: Bayesian Methods for Machine Learning Radford M. Neal, University of Toronto, 2011 Lecture 4 Problem: Density Estimation We have observed data, y 1,..., y n, drawn independently from some unknown
More informationF denotes cumulative density. denotes probability density function; (.)
BAYESIAN ANALYSIS: FOREWORDS Notation. System means the real thing and a model is an assumed mathematical form for the system.. he probability model class M contains the set of the all admissible models
More informationBayesian Estimation of DSGE Models 1 Chapter 3: A Crash Course in Bayesian Inference
1 The views expressed in this paper are those of the authors and do not necessarily reflect the views of the Federal Reserve Board of Governors or the Federal Reserve System. Bayesian Estimation of DSGE
More informationLecture 6: Markov Chain Monte Carlo
Lecture 6: Markov Chain Monte Carlo D. Jason Koskinen koskinen@nbi.ku.dk Photo by Howard Jackman University of Copenhagen Advanced Methods in Applied Statistics Feb - Apr 2016 Niels Bohr Institute 2 Outline
More informationMarkov Chain Monte Carlo methods
Markov Chain Monte Carlo methods By Oleg Makhnin 1 Introduction a b c M = d e f g h i 0 f(x)dx 1.1 Motivation 1.1.1 Just here Supresses numbering 1.1.2 After this 1.2 Literature 2 Method 2.1 New math As
More informationIntroduction to Machine Learning
Introduction to Machine Learning Brown University CSCI 1950-F, Spring 2012 Prof. Erik Sudderth Lecture 25: Markov Chain Monte Carlo (MCMC) Course Review and Advanced Topics Many figures courtesy Kevin
More informationThe Particle Filter. PD Dr. Rudolph Triebel Computer Vision Group. Machine Learning for Computer Vision
The Particle Filter Non-parametric implementation of Bayes filter Represents the belief (posterior) random state samples. by a set of This representation is approximate. Can represent distributions that
More informationNon-Parametric Bayes
Non-Parametric Bayes Mark Schmidt UBC Machine Learning Reading Group January 2016 Current Hot Topics in Machine Learning Bayesian learning includes: Gaussian processes. Approximate inference. Bayesian
More informationBayesian Inference for DSGE Models. Lawrence J. Christiano
Bayesian Inference for DSGE Models Lawrence J. Christiano Outline State space-observer form. convenient for model estimation and many other things. Bayesian inference Bayes rule. Monte Carlo integation.
More informationESTIMATION of a DSGE MODEL
ESTIMATION of a DSGE MODEL Paris, October 17 2005 STÉPHANE ADJEMIAN stephane.adjemian@ens.fr UNIVERSITÉ DU MAINE & CEPREMAP Slides.tex ESTIMATION of a DSGE MODEL STÉPHANE ADJEMIAN 16/10/2005 21:37 p. 1/3
More informationMetropolis-Hastings Algorithm
Strength of the Gibbs sampler Metropolis-Hastings Algorithm Easy algorithm to think about. Exploits the factorization properties of the joint probability distribution. No difficult choices to be made to
More informationParameter estimation and forecasting. Cristiano Porciani AIfA, Uni-Bonn
Parameter estimation and forecasting Cristiano Porciani AIfA, Uni-Bonn Questions? C. Porciani Estimation & forecasting 2 Temperature fluctuations Variance at multipole l (angle ~180o/l) C. Porciani Estimation
More informationDensity Estimation. Seungjin Choi
Density Estimation Seungjin Choi Department of Computer Science and Engineering Pohang University of Science and Technology 77 Cheongam-ro, Nam-gu, Pohang 37673, Korea seungjin@postech.ac.kr http://mlg.postech.ac.kr/
More informationPrinciples of Bayesian Inference
Principles of Bayesian Inference Sudipto Banerjee and Andrew O. Finley 2 Biostatistics, School of Public Health, University of Minnesota, Minneapolis, Minnesota, U.S.A. 2 Department of Forestry & Department
More informationProbability and Estimation. Alan Moses
Probability and Estimation Alan Moses Random variables and probability A random variable is like a variable in algebra (e.g., y=e x ), but where at least part of the variability is taken to be stochastic.
More informationBayesian parameter estimation in predictive engineering
Bayesian parameter estimation in predictive engineering Damon McDougall Institute for Computational Engineering and Sciences, UT Austin 14th August 2014 1/27 Motivation Understand physical phenomena Observations
More informationCalibration of Stochastic Volatility Models using Particle Markov Chain Monte Carlo Methods
Calibration of Stochastic Volatility Models using Particle Markov Chain Monte Carlo Methods Jonas Hallgren 1 1 Department of Mathematics KTH Royal Institute of Technology Stockholm, Sweden BFS 2012 June
More information(5) Multi-parameter models - Gibbs sampling. ST440/540: Applied Bayesian Analysis
Summarizing a posterior Given the data and prior the posterior is determined Summarizing the posterior gives parameter estimates, intervals, and hypothesis tests Most of these computations are integrals
More informationECO 513 Fall 2009 C. Sims HIDDEN MARKOV CHAIN MODELS
ECO 513 Fall 2009 C. Sims HIDDEN MARKOV CHAIN MODELS 1. THE CLASS OF MODELS y t {y s, s < t} p(y t θ t, {y s, s < t}) θ t = θ(s t ) P[S t = i S t 1 = j] = h ij. 2. WHAT S HANDY ABOUT IT Evaluating the
More informationMARKOV CHAIN MONTE CARLO
MARKOV CHAIN MONTE CARLO RYAN WANG Abstract. This paper gives a brief introduction to Markov Chain Monte Carlo methods, which offer a general framework for calculating difficult integrals. We start with
More information1 Bayesian Linear Regression (BLR)
Statistical Techniques in Robotics (STR, S15) Lecture#10 (Wednesday, February 11) Lecturer: Byron Boots Gaussian Properties, Bayesian Linear Regression 1 Bayesian Linear Regression (BLR) In linear regression,
More informationIntroduction to Machine Learning CMU-10701
Introduction to Machine Learning CMU-10701 Markov Chain Monte Carlo Methods Barnabás Póczos & Aarti Singh Contents Markov Chain Monte Carlo Methods Goal & Motivation Sampling Rejection Importance Markov
More informationTime Series and Dynamic Models
Time Series and Dynamic Models Section 1 Intro to Bayesian Inference Carlos M. Carvalho The University of Texas at Austin 1 Outline 1 1. Foundations of Bayesian Statistics 2. Bayesian Estimation 3. The
More informationApril 20th, Advanced Topics in Machine Learning California Institute of Technology. Markov Chain Monte Carlo for Machine Learning
for for Advanced Topics in California Institute of Technology April 20th, 2017 1 / 50 Table of Contents for 1 2 3 4 2 / 50 History of methods for Enrico Fermi used to calculate incredibly accurate predictions
More informationMarkov chain Monte Carlo Lecture 9
Markov chain Monte Carlo Lecture 9 David Sontag New York University Slides adapted from Eric Xing and Qirong Ho (CMU) Limitations of Monte Carlo Direct (unconditional) sampling Hard to get rare events
More informationLabor-Supply Shifts and Economic Fluctuations. Technical Appendix
Labor-Supply Shifts and Economic Fluctuations Technical Appendix Yongsung Chang Department of Economics University of Pennsylvania Frank Schorfheide Department of Economics University of Pennsylvania January
More informationPrinciples of Bayesian Inference
Principles of Bayesian Inference Sudipto Banerjee 1 and Andrew O. Finley 2 1 Biostatistics, School of Public Health, University of Minnesota, Minneapolis, Minnesota, U.S.A. 2 Department of Forestry & Department
More informationMarkov Networks.
Markov Networks www.biostat.wisc.edu/~dpage/cs760/ Goals for the lecture you should understand the following concepts Markov network syntax Markov network semantics Potential functions Partition function
More informationIntroduction to Computational Biology Lecture # 14: MCMC - Markov Chain Monte Carlo
Introduction to Computational Biology Lecture # 14: MCMC - Markov Chain Monte Carlo Assaf Weiner Tuesday, March 13, 2007 1 Introduction Today we will return to the motif finding problem, in lecture 10
More informationBayesian Gaussian Process Regression
Bayesian Gaussian Process Regression STAT8810, Fall 2017 M.T. Pratola October 7, 2017 Today Bayesian Gaussian Process Regression Bayesian GP Regression Recall we had observations from our expensive simulator,
More informationMCMC Review. MCMC Review. Gibbs Sampling. MCMC Review
MCMC Review http://jackman.stanford.edu/mcmc/icpsr99.pdf http://students.washington.edu/fkrogsta/bayes/stat538.pdf http://www.stat.berkeley.edu/users/terry/classes/s260.1998 /Week9a/week9a/week9a.html
More informationBayesian Methods in Multilevel Regression
Bayesian Methods in Multilevel Regression Joop Hox MuLOG, 15 september 2000 mcmc What is Statistics?! Statistics is about uncertainty To err is human, to forgive divine, but to include errors in your design
More informationMarkov Chain Monte Carlo (MCMC)
Markov Chain Monte Carlo (MCMC Dependent Sampling Suppose we wish to sample from a density π, and we can evaluate π as a function but have no means to directly generate a sample. Rejection sampling can
More informationTDA231. Logistic regression
TDA231 Devdatt Dubhashi dubhashi@chalmers.se Dept. of Computer Science and Engg. Chalmers University February 19, 2016 Some data 5 x2 0 5 5 0 5 x 1 In the Bayes classifier, we built a model of each class
More informationCPSC 540: Machine Learning
CPSC 540: Machine Learning MCMC and Non-Parametric Bayes Mark Schmidt University of British Columbia Winter 2016 Admin I went through project proposals: Some of you got a message on Piazza. No news is
More informationan introduction to bayesian inference
with an application to network analysis http://jakehofman.com january 13, 2010 motivation would like models that: provide predictive and explanatory power are complex enough to describe observed phenomena
More informationDAG models and Markov Chain Monte Carlo methods a short overview
DAG models and Markov Chain Monte Carlo methods a short overview Søren Højsgaard Institute of Genetics and Biotechnology University of Aarhus August 18, 2008 Printed: August 18, 2008 File: DAGMC-Lecture.tex
More informationPattern Recognition and Machine Learning. Bishop Chapter 11: Sampling Methods
Pattern Recognition and Machine Learning Chapter 11: Sampling Methods Elise Arnaud Jakob Verbeek May 22, 2008 Outline of the chapter 11.1 Basic Sampling Algorithms 11.2 Markov Chain Monte Carlo 11.3 Gibbs
More informationLecture 4: Dynamic models
linear s Lecture 4: s Hedibert Freitas Lopes The University of Chicago Booth School of Business 5807 South Woodlawn Avenue, Chicago, IL 60637 http://faculty.chicagobooth.edu/hedibert.lopes hlopes@chicagobooth.edu
More informationLikelihood-free MCMC
Bayesian inference for stable distributions with applications in finance Department of Mathematics University of Leicester September 2, 2011 MSc project final presentation Outline 1 2 3 4 Classical Monte
More information1 Probabilities. 1.1 Basics 1 PROBABILITIES
1 PROBABILITIES 1 Probabilities Probability is a tricky word usually meaning the likelyhood of something occuring or how frequent something is. Obviously, if something happens frequently, then its probability
More informationStat 535 C - Statistical Computing & Monte Carlo Methods. Arnaud Doucet.
Stat 535 C - Statistical Computing & Monte Carlo Methods Arnaud Doucet Email: arnaud@cs.ubc.ca 1 1.1 Outline Introduction to Markov chain Monte Carlo The Gibbs Sampler Examples Overview of the Lecture
More informationBayesian Linear Models
Bayesian Linear Models Sudipto Banerjee 1 and Andrew O. Finley 2 1 Department of Forestry & Department of Geography, Michigan State University, Lansing Michigan, U.S.A. 2 Biostatistics, School of Public
More informationSimulation - Lectures - Part III Markov chain Monte Carlo
Simulation - Lectures - Part III Markov chain Monte Carlo Julien Berestycki Part A Simulation and Statistical Programming Hilary Term 2018 Part A Simulation. HT 2018. J. Berestycki. 1 / 50 Outline Markov
More informationMCMC Sampling for Bayesian Inference using L1-type Priors
MÜNSTER MCMC Sampling for Bayesian Inference using L1-type Priors (what I do whenever the ill-posedness of EEG/MEG is just not frustrating enough!) AG Imaging Seminar Felix Lucka 26.06.2012 , MÜNSTER Sampling
More informationBayesian Inference for DSGE Models. Lawrence J. Christiano
Bayesian Inference for DSGE Models Lawrence J. Christiano Outline State space-observer form. convenient for model estimation and many other things. Preliminaries. Probabilities. Maximum Likelihood. Bayesian
More informationDynamic System Identification using HDMR-Bayesian Technique
Dynamic System Identification using HDMR-Bayesian Technique *Shereena O A 1) and Dr. B N Rao 2) 1), 2) Department of Civil Engineering, IIT Madras, Chennai 600036, Tamil Nadu, India 1) ce14d020@smail.iitm.ac.in
More informationSTAT Advanced Bayesian Inference
1 / 32 STAT 625 - Advanced Bayesian Inference Meng Li Department of Statistics Jan 23, 218 The Dirichlet distribution 2 / 32 θ Dirichlet(a 1,...,a k ) with density p(θ 1,θ 2,...,θ k ) = k j=1 Γ(a j) Γ(
More informationBayesian Estimation of Input Output Tables for Russia
Bayesian Estimation of Input Output Tables for Russia Oleg Lugovoy (EDF, RANE) Andrey Polbin (RANE) Vladimir Potashnikov (RANE) WIOD Conference April 24, 2012 Groningen Outline Motivation Objectives Bayesian
More informationMetropolis Hastings. Rebecca C. Steorts Bayesian Methods and Modern Statistics: STA 360/601. Module 9
Metropolis Hastings Rebecca C. Steorts Bayesian Methods and Modern Statistics: STA 360/601 Module 9 1 The Metropolis-Hastings algorithm is a general term for a family of Markov chain simulation methods
More information16 : Approximate Inference: Markov Chain Monte Carlo
10-708: Probabilistic Graphical Models 10-708, Spring 2017 16 : Approximate Inference: Markov Chain Monte Carlo Lecturer: Eric P. Xing Scribes: Yuan Yang, Chao-Ming Yen 1 Introduction As the target distribution
More informationMarkov Chain Monte Carlo Inference. Siamak Ravanbakhsh Winter 2018
Graphical Models Markov Chain Monte Carlo Inference Siamak Ravanbakhsh Winter 2018 Learning objectives Markov chains the idea behind Markov Chain Monte Carlo (MCMC) two important examples: Gibbs sampling
More informationBayesian Semiparametric GARCH Models
Bayesian Semiparametric GARCH Models Xibin (Bill) Zhang and Maxwell L. King Department of Econometrics and Business Statistics Faculty of Business and Economics xibin.zhang@monash.edu Quantitative Methods
More informationBayes Nets: Sampling
Bayes Nets: Sampling [These slides were created by Dan Klein and Pieter Abbeel for CS188 Intro to AI at UC Berkeley. All CS188 materials are available at http://ai.berkeley.edu.] Approximate Inference:
More informationComputer Vision Group Prof. Daniel Cremers. 11. Sampling Methods
Prof. Daniel Cremers 11. Sampling Methods Sampling Methods Sampling Methods are widely used in Computer Science as an approximation of a deterministic algorithm to represent uncertainty without a parametric
More informationSUPPLEMENT TO MARKET ENTRY COSTS, PRODUCER HETEROGENEITY, AND EXPORT DYNAMICS (Econometrica, Vol. 75, No. 3, May 2007, )
Econometrica Supplementary Material SUPPLEMENT TO MARKET ENTRY COSTS, PRODUCER HETEROGENEITY, AND EXPORT DYNAMICS (Econometrica, Vol. 75, No. 3, May 2007, 653 710) BY SANGHAMITRA DAS, MARK ROBERTS, AND
More informationDS-GA 1003: Machine Learning and Computational Statistics Homework 7: Bayesian Modeling
DS-GA 1003: Machine Learning and Computational Statistics Homework 7: Bayesian Modeling Due: Tuesday, May 10, 2016, at 6pm (Submit via NYU Classes) Instructions: Your answers to the questions below, including
More informationBayesian Linear Models
Bayesian Linear Models Sudipto Banerjee 1 and Andrew O. Finley 2 1 Biostatistics, School of Public Health, University of Minnesota, Minneapolis, Minnesota, U.S.A. 2 Department of Forestry & Department
More information1 Probabilities. 1.1 Basics 1 PROBABILITIES
1 PROBABILITIES 1 Probabilities Probability is a tricky word usually meaning the likelyhood of something occuring or how frequent something is. Obviously, if something happens frequently, then its probability
More informationComputer Practical: Metropolis-Hastings-based MCMC
Computer Practical: Metropolis-Hastings-based MCMC Andrea Arnold and Franz Hamilton North Carolina State University July 30, 2016 A. Arnold / F. Hamilton (NCSU) MH-based MCMC July 30, 2016 1 / 19 Markov
More information