The Metropolis-Hastings Algorithm. June 8, 2012

Size: px
Start display at page:

Download "The Metropolis-Hastings Algorithm. June 8, 2012"

Transcription

1 The Metropolis-Hastings Algorithm June 8, 22

2 The Plan. Understand what a simulated distribution is 2. Understand why the Metropolis-Hastings algorithm works 3. Learn how to apply the Metropolis-Hastings algorithm Most of today s lecture can be found in Chib (2) and An and Schorfheide (27)

3 Bayesian statistics: From last time Variance of estimator (frequentist) vs variance of parameter (Bayesian) Subjective view of probability Probabilities are statement about our knowledge Treating model parameters as random variables thus do not mean that we think that they necessarily vary over time.

4 Bayesian statistics: From last time Bayesian procedures can be derived from 4 Bayesian Principles. The Likelihood Principle 2. The Sufficiency Principle 3. The Conditionality Principle 4. The Stopping Rule Principle

5 Main concepts and notation The main components in Bayesian inference are: Data (observables) Z T R T n A model: Parameters θ R k A prior distribution p(θ) : R k R + Likelihood function p(z θ) : R T n R k R + Posterior density p(θ Z) : R T n R k R + We need a method to construct the posterior density

6 The end product of Bayesian statistics Most of Bayesian econometrics consists of simulating distributions of parameters using numerical methods. A simulated posterior is a numerical approximation to the distribution p(z θ)p(θ) This is useful since the the distribution p(z θ)p(θ) (by Bayes rule) is proportional to p(θ Z) p(θ Z) = p(z θ)p(θ) p(z) We rely on ergodicity, i.e. that the moments of the constructed sample correspond to the moments of the distribution p(θ Z) The most popular (and general) procedure to simulate the posterior is called the Metropolis-Hastings Algorithm

7 Metropolis-Hastings Metropolis-Hastings is a way to simulate a sample from a target distribution In practice, the target distribution will in most cases be the posterior density p(θ Z) but it doesn t have to be But what does it mean to sample from a given distribution?

8 Sampling from a distribution Example: Standard Normal distribution N(, ) If distribution is ergodic, sample shares all the properties of the true distribution asymptotically Why don t we just compute the moments directly? When we can, we should (as in the Normal distribution s case) Not always possible, either because tractability reasons or for computational burden

9 Sample from Standard Normal

10 Sample from Standard Normal

11 Sample from Standard Normal x 4

12 Sample from Standard Normal

13 Sampling from a distribution Example: Standard Normal distribution N(, ) If distribution is ergodic, sample shares all the properties of the true distribution asymptotically Why don t we just compute the moments directly? When we can, we should (as in the Normal distribution s case) Not always possible, either because tractability reasons or for computational burden

14 What is a Markov Chain? A process for which the distribution of next period variables are independent of the past, once we condition on the current state A Markov chain is a stochastic process with the Markov property Markov chain is sometimes taken to mean only processes with a countably finite number of states Here, the term will be used in the broader sense Markov Chain Monte Carlo methods provide a way of generating samples that share the properties of the target density (i.e. the object of interest)

15 What is a Markov Chain? Markov property p(θ j+ θ j ) = p(θ j+ θ j, θ j, θ j 2,..., θ ) Transition density function p(θ j+ θ j ) describes the distribution of θ j+ conditional on θ j. In most applications, we know the conditional transition density and can figure out unconditional properties like E(θ) and E ( θ 2) MCMC methods can be used to do the opposite: Determine a particular conditional transition density such that the unconditional distribution converges to that of the target distribution. Let s have a first look at the Metropolis-Hastings Algorithm

16 The Random-Walk Metropolis Algorithm. Start with an arbitrary value θ 2. Update from θ j to θ j+ (j =, 2,...J) by 2. Generate θ N (θ j, Σ) 2.2 Define 2.3 Take θ j+ = 3. Repeat Step 2 J times. But why does it work? α = min ( p(θ ) ) p(θ j ), { θ with probability α θ j otherwise } ()

17 Simulating distributions using MCMC methods We can look at a simple discrete state space example: θ can take two values θ { θ L, θ H} Probabilities p(θ L ) =.4 and p(θ H ) =.6 How can we construct a sequence θ (), θ (2), θ (3),..., θ (J) such that the relative number of occurrences of θ L and θ H in the sample correspond to those of the density described by p(θ)?

18 A two-state Markov Chain for? Transition probabilities are defined as ( π i,k = p θ j+ = θ i θ j = θ k) : i, k {L, H} Unconditional probabilities solves the equation [ ] [ ] [ π πl π H = πl π LL π HL H π LH π HH ] Problem: Define transition probabilities such that unconditional probabilities equal are those of the target distribution

19 A two-state Markov Chain for? We want the chain to spend more time in the more likely state θ H but not all the time. If θ j = θ L then θ j+ = θ H If θ j = θ H then θ j+ = θ L with probability α and θ j+ = θ H with probability ( α) Finding the right α : [ πl π H ] = [ πl π H ] [ α ( α) ] We get two equations, π L = απ H, for α gives α = π L π H π H = π L + ( α)π H, solving

20 Simulating distributions using MCMC methods Let s do an example by hand.

21 Simulating distributions using MCMC methods Using the ratio of relative probability/density to decide whether to accept a candidate turns out to be extremely general: exactly the same condition ensures that Markov Chain converges to target distribution also for continuous parameter spaces Loosely speaking, it is the condition that makes sure that the Markov chain spend just the right amount of time at each point in the parameter space How the candidate is generated almost doesn t matter at all as long as chain is: Irreducible Aperiodic Partly depends on choice that determine how proposal density is generated

22 Proposal density Any proposal density with infinite support would do (e.g. normal) The choice a matter of efficiency. Some options: RW-MH Adaptive Random Walk Independent M-H

23 The Random-Walk Metropolis Algorithm. Start with an arbitrary value θ 2. Update from θ j to θ j+ (j =, 2,...J) by 2. Generate θ N (θ j, Σ) 2.2 Define 2.3 Take θ j+ = 3. Repeat Step 2 J times. α = min ( L (Z θ ) ) L (Z θ j ), { θ with probability α θ j otherwise We can use this to estimate the likelihood function of a simple DSGE model } (2)

24 A simple DSGE model x t = ρx t + ut x y t = E t (y t+ ) γ [r t E t (π t+ )] + ut y π t = E t (π t+ ) + κ [y t x t ] + ut π r t = φ π π t

25 A simple DSGE model Substitute in the interest rate in the Euler equation x t = ρx t + u x t y t = E t (y t+ ) γ [φ ππ t E t (π t+ )] + u y t π t = E t (π t+ ) + κ [y t x t ] + u π t

26 Solving model using method of undetermined coefficients Conjecture that model can be put in the form x t = ρx t + ut x y t = ax t + ut y π t = bx t + ut π Why is this a good guess?

27 Solving model using method of undetermined coefficients Substitute in conjectured form of solution (ignoring the shocks ut y and ut π for now) into structural equation ax t = aρx t γ [φ πbx t bρx t ] bx t = bρx t + κ [ax t x t ] where we used that x t = ρx t + u x t implies that E[x t+ x t ] = ρx t

28 Solving model using method of undetermined coefficients Equate coefficients on right and left hand side a = aρ γ φ πb + γ bρ b = bρ + κ [a ] or [ ( ρ) γ (φ π ρ) κ ( ρ) ] [ a b ] = [ κ ]

29 Solving model using method of undetermined coefficients Solve for a and b [ ] a = b [ ( ρ) γ (φ π ρ) κ ( ρ) ] [ κ ] or [ a b ] [ = κ φ ρ c κγ ρ c ] where c = γ κρ 2γρ + κφ + γρ 2 <

30 A simple DSGE model The solved model x t = ρx t + u x t y t = κ ρ φ π x t + ut y c π t = κγ ρ x t + ut π c where c = γ κρ 2γρ + κφ + γρ 2 < We want to estimate the distributions of θ = {ρ, γ, κ, φ, σ x, σ y, σ π, }

31 A simple DSGE model Put the solved model in state space form X t = AX t + Cu t Z t = DX t + v t where X t = x t, A = ρ, Cu t = ut x [ ] [ yt κ φπ ρ Z t =, D = c π t κγ ρ c ] [ u y, v t = t u π t ]

32 The log likelihood function of a state space system For a given state space system X t = AX t + Cu t Z t (p ) = DX t + v t we can evaluate the log likelihood by computing T [ L(Z Θ) =.5 p ln(2π) + ln Ω t + Z ] tω t Z t t= where Z t are the innovation from the Kalman filter

33 The parameter vector θ and the variance of random walk innovations in MCMC Σ Parameterize the model according to θ = {ρ, γ, κ, φ, σ x, σ y, σ π, } = {.9, 2,.,.5,,, } Generate data from true model and T = Set starting value θ () = θ (OK, this option is not available in practice.) Set covariance matrix of random walk increments in Metropolis Algorithm proportional to absolute values of true parameters Σ = ε diag(abs(θ))

34 The Random-Walk Metropolis Algorithm We now have all we need:. Start with an arbitrary value θ 2. Update from θ j to θ j+ (j =, 2,...J) by 2. Generate θ N (θ j, Σ) 2.2 Define 2.3 Take θ j+ = 3. Repeat Step 2 J times α = min ( L (Z θ ) ) L (Z θ j ), { θ with probability α θ j otherwise } (3)

35 The MCMC with J=

36 The MCMC with J= x

37 The MCMC with J=

38 The MCMC with J= x

39 The MCMC with J= x x x x x x x 4

40 The MCMC with J=

41 Identification Some posterior distribution look flat Can be a short sample issue Nothing to do if we are estimating a model on real data, but here we can check Can be a problem with mapping between parameters and likelihood

42 The MCMC with T=2 and J=

43 Identification Posterior distribution for γ and κ kappa still look flat probably a true identification issue Little can be said about identification a priori

44 Convergence How many draws do we need? Optimal J increases with number of parameters One informal check is to plot the diagonal of the recursive covariance matrix of the MCMC j j θ (i) θ (i) for j =, 2,...J i=

45 Checking for convergence 2.5 x x

46 Checking for convergence J= x

47 Combining prior and sample information Sometimes we know more about the parameters than what the data tells us, i.e. we have some prior information.

48 What do we mean by prior information? For a DSGE model, we may have information about deep parameters Range of some parameters restricted by theory, e.g. risk aversion should be positive Discount rate is inverse of average real interest rates Price stickiness can be measured by surveys We may know something about the mean of a process

49 How do we combine prior and sample information? Bayes theorem: P (θ Z) P(Z) = P (Z θ) P(θ) P (θ Z) = P (Z θ) P(θ) P(Z) Since P(Z) is a constant, we can use P (Z θ) P(θ) as the posterior likelihood (a likelihood function is any function that is proportional to the probability). We now need to choose P(θ)

50 Choosing prior distributions The beta distribution is a good choice when parameter is in [,] where P(x) = ( x)b x a B(a, b) B(a, b) = (a )!(b )! (a + b )! Easier to parameterize using expression for mean, mode and variance: µ = σ 2 = a a + b, x = a a + b 2 ab (a + b) 2 (a + b + )

51 Examples of beta distributions holding mean fixed

52 Examples of beta distributions holding s.d. fixed

53 Choosing prior distributions The inverse gamma distribution is a good choice when parameter is positive P(x) = ba Γ(a) (/x)a+ exp( b/x) where Γ(a) = (a )! Again, easier to parameterize using expression for mean, mode and variance: µ = σ 2 = b b ; a >, x = a a + b 2 (a ) 2 (a 2) ; a > 2

54 Examples of inverse gamma distributions

55 The Random-Walk Metropolis Algorithm with priors. Start with an arbitrary value θ 2. Update from θ j to θ j+ (j =, 2,...J) by 2. Generate θ N (θ j, Σ) 2.2 Define 2.3 Take α = min θ j+ = ( L (Z θ ) P(θ ) ) L (Z θ j ) P(θ j ), { θ with probability α θ j otherwise The only difference compared to before is that the priors appear in the ratio in (3). } (4)

56 Practical implementation Since we compute log-likelihood we can use the log of the likelihood ratio in the Random Walk Metropolis Algorithm ln α = min [( ln L (Z θ ) + ln P(θ ) ln L (Z θ j ) ln P(θ j )), ] If priors across parameters are independent we have that ln P(θ j ) = ln P(θ,j ) + ln P(θ 2,j ) ln P(θ q,j ) where θ j = [ θ,j θ 2,j θ q,j ]

57 Let s get deep with an old friend: x t = ρx t + u x t y t = E t (y t+ ) γ [r t E t (π t+ )] + u y t π t = E t (π t+ ) + κ [y t x t ] + u π t r t = φ π π t The parameter κ is in the benchmark 3-equation NK model given by ( δ)( δβ) κ = δ where δ is the Calvo parameter of price stickiness and β is the discount factor. We now have a new parameter vector θ = {ρ, γ, δ, β, φ, σ x, σ y, σ π, }

58 The priors The prior on relative risk aversion γ is truncated Normal with mean 2 and s.d..5. The prior on the discount factor β is Beta with mean.99 and s.d.. The prior on the Calvo parameter δ is Beta with mean.75 and s.d

59 The Random-Walk Metropolis Algorithm with priors. Start with an arbitrary value θ 2. Update from θ j to θ j+ (j =, 2,...J) by 2. Generate θ N (θ j, Σ) 2.2 Define 2.3 Take α = min θ j+ = 3. Repeat Step 2 J times ( L (Z θ ) P(θ ) ) L (Z θ j ) P(θ j ), { θ with probability α θ j otherwise } (5)

60 The log prior The log prior is given by ln P(θ j ) = ln P(θ,j ) + ln P(θ 2,j ) ln P(θ q,j ) = ln P(γ j ) + ln P(δ j ) + ln P(β j ) since we can ignore the (constant) probabilities on the uniform priors. We then have that ln [L (Z θ ) P(θ )] = ln L(Z θ) + ln P(θ j ) = α = min i.e. all we need for the RWMA ( exp [ln L(Z θ ) + ln P(θ ) )] exp [ln L(Z θ j ) + ln P(θ j )],

61 Posterior with uniform and informative priors

62 Inference about parameters Compute sample equivalent: Central tendencies: Mean, mode Variance Compute frequency of occurrence: How likely is it that θ and θ 2 are both smaller than?

63 That s it for today.

Bayesian Methods for DSGE models Lecture 1 Macro models as data generating processes

Bayesian Methods for DSGE models Lecture 1 Macro models as data generating processes Bayesian Methods for DSGE models Lecture 1 Macro models as data generating processes Kristoffer Nimark CREI, Universitat Pompeu Fabra and Barcelona GSE June 30, 2014 Bayesian Methods for DSGE models Course

More information

ML Estimation of State Space Models. March 17, 2017

ML Estimation of State Space Models. March 17, 2017 ML Estimation of State Space Models March 17, 2017 Recap from last time: Unobserved Component model of Inflation π t = τ t + η t τ t = τ t 1 + ε t Decomposes inflation into permanent (τ) and transitory

More information

ML Estimation of State Space Models. February 29, 2016

ML Estimation of State Space Models. February 29, 2016 ML Estimation of State Space Models February 29, 2016 Recap from last time: Unobserved Component model of Inflation π t = τ t + η t τ t = τ t 1 + ε t Decomposes inflation into permanent (τ) and transitory

More information

Bayesian Inference and MCMC

Bayesian Inference and MCMC Bayesian Inference and MCMC Aryan Arbabi Partly based on MCMC slides from CSC412 Fall 2018 1 / 18 Bayesian Inference - Motivation Consider we have a data set D = {x 1,..., x n }. E.g each x i can be the

More information

Bayesian Methods for Machine Learning

Bayesian Methods for Machine Learning Bayesian Methods for Machine Learning CS 584: Big Data Analytics Material adapted from Radford Neal s tutorial (http://ftp.cs.utoronto.ca/pub/radford/bayes-tut.pdf), Zoubin Ghahramni (http://hunch.net/~coms-4771/zoubin_ghahramani_bayesian_learning.pdf),

More information

Markov chain Monte Carlo

Markov chain Monte Carlo 1 / 26 Markov chain Monte Carlo Timothy Hanson 1 and Alejandro Jara 2 1 Division of Biostatistics, University of Minnesota, USA 2 Department of Statistics, Universidad de Concepción, Chile IAP-Workshop

More information

DSGE Methods. Estimation of DSGE models: Maximum Likelihood & Bayesian. Willi Mutschler, M.Sc.

DSGE Methods. Estimation of DSGE models: Maximum Likelihood & Bayesian. Willi Mutschler, M.Sc. DSGE Methods Estimation of DSGE models: Maximum Likelihood & Bayesian Willi Mutschler, M.Sc. Institute of Econometrics and Economic Statistics University of Münster willi.mutschler@uni-muenster.de Summer

More information

Lecture 2: From Linear Regression to Kalman Filter and Beyond

Lecture 2: From Linear Regression to Kalman Filter and Beyond Lecture 2: From Linear Regression to Kalman Filter and Beyond Department of Biomedical Engineering and Computational Science Aalto University January 26, 2012 Contents 1 Batch and Recursive Estimation

More information

Lecture 2: From Linear Regression to Kalman Filter and Beyond

Lecture 2: From Linear Regression to Kalman Filter and Beyond Lecture 2: From Linear Regression to Kalman Filter and Beyond January 18, 2017 Contents 1 Batch and Recursive Estimation 2 Towards Bayesian Filtering 3 Kalman Filter and Bayesian Filtering and Smoothing

More information

Principles of Bayesian Inference

Principles of Bayesian Inference Principles of Bayesian Inference Sudipto Banerjee University of Minnesota July 20th, 2008 1 Bayesian Principles Classical statistics: model parameters are fixed and unknown. A Bayesian thinks of parameters

More information

David Giles Bayesian Econometrics

David Giles Bayesian Econometrics David Giles Bayesian Econometrics 1. General Background 2. Constructing Prior Distributions 3. Properties of Bayes Estimators and Tests 4. Bayesian Analysis of the Multiple Regression Model 5. Bayesian

More information

Lecture 7 and 8: Markov Chain Monte Carlo

Lecture 7 and 8: Markov Chain Monte Carlo Lecture 7 and 8: Markov Chain Monte Carlo 4F13: Machine Learning Zoubin Ghahramani and Carl Edward Rasmussen Department of Engineering University of Cambridge http://mlg.eng.cam.ac.uk/teaching/4f13/ Ghahramani

More information

SC7/SM6 Bayes Methods HT18 Lecturer: Geoff Nicholls Lecture 2: Monte Carlo Methods Notes and Problem sheets are available at http://www.stats.ox.ac.uk/~nicholls/bayesmethods/ and via the MSc weblearn pages.

More information

CSC 2541: Bayesian Methods for Machine Learning

CSC 2541: Bayesian Methods for Machine Learning CSC 2541: Bayesian Methods for Machine Learning Radford M. Neal, University of Toronto, 2011 Lecture 3 More Markov Chain Monte Carlo Methods The Metropolis algorithm isn t the only way to do MCMC. We ll

More information

Lecture 8: The Metropolis-Hastings Algorithm

Lecture 8: The Metropolis-Hastings Algorithm 30.10.2008 What we have seen last time: Gibbs sampler Key idea: Generate a Markov chain by updating the component of (X 1,..., X p ) in turn by drawing from the full conditionals: X (t) j Two drawbacks:

More information

Markov chain Monte Carlo

Markov chain Monte Carlo Markov chain Monte Carlo Feng Li feng.li@cufe.edu.cn School of Statistics and Mathematics Central University of Finance and Economics Revised on April 24, 2017 Today we are going to learn... 1 Markov Chains

More information

MCMC algorithms for fitting Bayesian models

MCMC algorithms for fitting Bayesian models MCMC algorithms for fitting Bayesian models p. 1/1 MCMC algorithms for fitting Bayesian models Sudipto Banerjee sudiptob@biostat.umn.edu University of Minnesota MCMC algorithms for fitting Bayesian models

More information

Markov Chains and MCMC

Markov Chains and MCMC Markov Chains and MCMC CompSci 590.02 Instructor: AshwinMachanavajjhala Lecture 4 : 590.02 Spring 13 1 Recap: Monte Carlo Method If U is a universe of items, and G is a subset satisfying some property,

More information

Sequential Monte Carlo Methods (for DSGE Models)

Sequential Monte Carlo Methods (for DSGE Models) Sequential Monte Carlo Methods (for DSGE Models) Frank Schorfheide University of Pennsylvania, PIER, CEPR, and NBER October 23, 2017 Some References These lectures use material from our joint work: Tempered

More information

Reminder of some Markov Chain properties:

Reminder of some Markov Chain properties: Reminder of some Markov Chain properties: 1. a transition from one state to another occurs probabilistically 2. only state that matters is where you currently are (i.e. given present, future is independent

More information

Markov Chain Monte Carlo methods

Markov Chain Monte Carlo methods Markov Chain Monte Carlo methods Tomas McKelvey and Lennart Svensson Signal Processing Group Department of Signals and Systems Chalmers University of Technology, Sweden November 26, 2012 Today s learning

More information

CS281A/Stat241A Lecture 22

CS281A/Stat241A Lecture 22 CS281A/Stat241A Lecture 22 p. 1/4 CS281A/Stat241A Lecture 22 Monte Carlo Methods Peter Bartlett CS281A/Stat241A Lecture 22 p. 2/4 Key ideas of this lecture Sampling in Bayesian methods: Predictive distribution

More information

Answers and expectations

Answers and expectations Answers and expectations For a function f(x) and distribution P(x), the expectation of f with respect to P is The expectation is the average of f, when x is drawn from the probability distribution P E

More information

Hierarchical models. Dr. Jarad Niemi. August 31, Iowa State University. Jarad Niemi (Iowa State) Hierarchical models August 31, / 31

Hierarchical models. Dr. Jarad Niemi. August 31, Iowa State University. Jarad Niemi (Iowa State) Hierarchical models August 31, / 31 Hierarchical models Dr. Jarad Niemi Iowa State University August 31, 2017 Jarad Niemi (Iowa State) Hierarchical models August 31, 2017 1 / 31 Normal hierarchical model Let Y ig N(θ g, σ 2 ) for i = 1,...,

More information

Bayesian Computations for DSGE Models

Bayesian Computations for DSGE Models Bayesian Computations for DSGE Models Frank Schorfheide University of Pennsylvania, PIER, CEPR, and NBER October 23, 2017 This Lecture is Based on Bayesian Estimation of DSGE Models Edward P. Herbst &

More information

Bayesian Estimation with Sparse Grids

Bayesian Estimation with Sparse Grids Bayesian Estimation with Sparse Grids Kenneth L. Judd and Thomas M. Mertens Institute on Computational Economics August 7, 27 / 48 Outline Introduction 2 Sparse grids Construction Integration with sparse

More information

Bayesian Estimation of DSGE Models

Bayesian Estimation of DSGE Models Bayesian Estimation of DSGE Models Stéphane Adjemian Université du Maine, GAINS & CEPREMAP stephane.adjemian@univ-lemans.fr http://www.dynare.org/stepan June 28, 2011 June 28, 2011 Université du Maine,

More information

Sequential Monte Carlo Methods

Sequential Monte Carlo Methods University of Pennsylvania Bradley Visitor Lectures October 23, 2017 Introduction Unfortunately, standard MCMC can be inaccurate, especially in medium and large-scale DSGE models: disentangling importance

More information

Session 3A: Markov chain Monte Carlo (MCMC)

Session 3A: Markov chain Monte Carlo (MCMC) Session 3A: Markov chain Monte Carlo (MCMC) John Geweke Bayesian Econometrics and its Applications August 15, 2012 ohn Geweke Bayesian Econometrics and its Session Applications 3A: Markov () chain Monte

More information

Online appendix to On the stability of the excess sensitivity of aggregate consumption growth in the US

Online appendix to On the stability of the excess sensitivity of aggregate consumption growth in the US Online appendix to On the stability of the excess sensitivity of aggregate consumption growth in the US Gerdie Everaert 1, Lorenzo Pozzi 2, and Ruben Schoonackers 3 1 Ghent University & SHERPPA 2 Erasmus

More information

STA 4273H: Statistical Machine Learning

STA 4273H: Statistical Machine Learning STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Computer Science! Department of Statistical Sciences! rsalakhu@cs.toronto.edu! h0p://www.cs.utoronto.ca/~rsalakhu/ Lecture 7 Approximate

More information

Computational statistics

Computational statistics Computational statistics Markov Chain Monte Carlo methods Thierry Denœux March 2017 Thierry Denœux Computational statistics March 2017 1 / 71 Contents of this chapter When a target density f can be evaluated

More information

Probabilistic Machine Learning

Probabilistic Machine Learning Probabilistic Machine Learning Bayesian Nets, MCMC, and more Marek Petrik 4/18/2017 Based on: P. Murphy, K. (2012). Machine Learning: A Probabilistic Perspective. Chapter 10. Conditional Independence Independent

More information

Bayesian Inference: Concept and Practice

Bayesian Inference: Concept and Practice Inference: Concept and Practice fundamentals Johan A. Elkink School of Politics & International Relations University College Dublin 5 June 2017 1 2 3 Bayes theorem In order to estimate the parameters of

More information

Markov Chain Monte Carlo

Markov Chain Monte Carlo Chapter 5 Markov Chain Monte Carlo MCMC is a kind of improvement of the Monte Carlo method By sampling from a Markov chain whose stationary distribution is the desired sampling distributuion, it is possible

More information

LECTURE 15 Markov chain Monte Carlo

LECTURE 15 Markov chain Monte Carlo LECTURE 15 Markov chain Monte Carlo There are many settings when posterior computation is a challenge in that one does not have a closed form expression for the posterior distribution. Markov chain Monte

More information

MCMC and Gibbs Sampling. Kayhan Batmanghelich

MCMC and Gibbs Sampling. Kayhan Batmanghelich MCMC and Gibbs Sampling Kayhan Batmanghelich 1 Approaches to inference l Exact inference algorithms l l l The elimination algorithm Message-passing algorithm (sum-product, belief propagation) The junction

More information

Markov Chain Monte Carlo, Numerical Integration

Markov Chain Monte Carlo, Numerical Integration Markov Chain Monte Carlo, Numerical Integration (See Statistics) Trevor Gallen Fall 2015 1 / 1 Agenda Numerical Integration: MCMC methods Estimating Markov Chains Estimating latent variables 2 / 1 Numerical

More information

Monetary and Exchange Rate Policy Under Remittance Fluctuations. Technical Appendix and Additional Results

Monetary and Exchange Rate Policy Under Remittance Fluctuations. Technical Appendix and Additional Results Monetary and Exchange Rate Policy Under Remittance Fluctuations Technical Appendix and Additional Results Federico Mandelman February In this appendix, I provide technical details on the Bayesian estimation.

More information

Markov Chain Monte Carlo Methods

Markov Chain Monte Carlo Methods Markov Chain Monte Carlo Methods John Geweke University of Iowa, USA 2005 Institute on Computational Economics University of Chicago - Argonne National Laboaratories July 22, 2005 The problem p (θ, ω I)

More information

CSC 2541: Bayesian Methods for Machine Learning

CSC 2541: Bayesian Methods for Machine Learning CSC 2541: Bayesian Methods for Machine Learning Radford M. Neal, University of Toronto, 2011 Lecture 4 Problem: Density Estimation We have observed data, y 1,..., y n, drawn independently from some unknown

More information

F denotes cumulative density. denotes probability density function; (.)

F denotes cumulative density. denotes probability density function; (.) BAYESIAN ANALYSIS: FOREWORDS Notation. System means the real thing and a model is an assumed mathematical form for the system.. he probability model class M contains the set of the all admissible models

More information

Bayesian Estimation of DSGE Models 1 Chapter 3: A Crash Course in Bayesian Inference

Bayesian Estimation of DSGE Models 1 Chapter 3: A Crash Course in Bayesian Inference 1 The views expressed in this paper are those of the authors and do not necessarily reflect the views of the Federal Reserve Board of Governors or the Federal Reserve System. Bayesian Estimation of DSGE

More information

Lecture 6: Markov Chain Monte Carlo

Lecture 6: Markov Chain Monte Carlo Lecture 6: Markov Chain Monte Carlo D. Jason Koskinen koskinen@nbi.ku.dk Photo by Howard Jackman University of Copenhagen Advanced Methods in Applied Statistics Feb - Apr 2016 Niels Bohr Institute 2 Outline

More information

Markov Chain Monte Carlo methods

Markov Chain Monte Carlo methods Markov Chain Monte Carlo methods By Oleg Makhnin 1 Introduction a b c M = d e f g h i 0 f(x)dx 1.1 Motivation 1.1.1 Just here Supresses numbering 1.1.2 After this 1.2 Literature 2 Method 2.1 New math As

More information

Introduction to Machine Learning

Introduction to Machine Learning Introduction to Machine Learning Brown University CSCI 1950-F, Spring 2012 Prof. Erik Sudderth Lecture 25: Markov Chain Monte Carlo (MCMC) Course Review and Advanced Topics Many figures courtesy Kevin

More information

The Particle Filter. PD Dr. Rudolph Triebel Computer Vision Group. Machine Learning for Computer Vision

The Particle Filter. PD Dr. Rudolph Triebel Computer Vision Group. Machine Learning for Computer Vision The Particle Filter Non-parametric implementation of Bayes filter Represents the belief (posterior) random state samples. by a set of This representation is approximate. Can represent distributions that

More information

Non-Parametric Bayes

Non-Parametric Bayes Non-Parametric Bayes Mark Schmidt UBC Machine Learning Reading Group January 2016 Current Hot Topics in Machine Learning Bayesian learning includes: Gaussian processes. Approximate inference. Bayesian

More information

Bayesian Inference for DSGE Models. Lawrence J. Christiano

Bayesian Inference for DSGE Models. Lawrence J. Christiano Bayesian Inference for DSGE Models Lawrence J. Christiano Outline State space-observer form. convenient for model estimation and many other things. Bayesian inference Bayes rule. Monte Carlo integation.

More information

ESTIMATION of a DSGE MODEL

ESTIMATION of a DSGE MODEL ESTIMATION of a DSGE MODEL Paris, October 17 2005 STÉPHANE ADJEMIAN stephane.adjemian@ens.fr UNIVERSITÉ DU MAINE & CEPREMAP Slides.tex ESTIMATION of a DSGE MODEL STÉPHANE ADJEMIAN 16/10/2005 21:37 p. 1/3

More information

Metropolis-Hastings Algorithm

Metropolis-Hastings Algorithm Strength of the Gibbs sampler Metropolis-Hastings Algorithm Easy algorithm to think about. Exploits the factorization properties of the joint probability distribution. No difficult choices to be made to

More information

Parameter estimation and forecasting. Cristiano Porciani AIfA, Uni-Bonn

Parameter estimation and forecasting. Cristiano Porciani AIfA, Uni-Bonn Parameter estimation and forecasting Cristiano Porciani AIfA, Uni-Bonn Questions? C. Porciani Estimation & forecasting 2 Temperature fluctuations Variance at multipole l (angle ~180o/l) C. Porciani Estimation

More information

Density Estimation. Seungjin Choi

Density Estimation. Seungjin Choi Density Estimation Seungjin Choi Department of Computer Science and Engineering Pohang University of Science and Technology 77 Cheongam-ro, Nam-gu, Pohang 37673, Korea seungjin@postech.ac.kr http://mlg.postech.ac.kr/

More information

Principles of Bayesian Inference

Principles of Bayesian Inference Principles of Bayesian Inference Sudipto Banerjee and Andrew O. Finley 2 Biostatistics, School of Public Health, University of Minnesota, Minneapolis, Minnesota, U.S.A. 2 Department of Forestry & Department

More information

Probability and Estimation. Alan Moses

Probability and Estimation. Alan Moses Probability and Estimation Alan Moses Random variables and probability A random variable is like a variable in algebra (e.g., y=e x ), but where at least part of the variability is taken to be stochastic.

More information

Bayesian parameter estimation in predictive engineering

Bayesian parameter estimation in predictive engineering Bayesian parameter estimation in predictive engineering Damon McDougall Institute for Computational Engineering and Sciences, UT Austin 14th August 2014 1/27 Motivation Understand physical phenomena Observations

More information

Calibration of Stochastic Volatility Models using Particle Markov Chain Monte Carlo Methods

Calibration of Stochastic Volatility Models using Particle Markov Chain Monte Carlo Methods Calibration of Stochastic Volatility Models using Particle Markov Chain Monte Carlo Methods Jonas Hallgren 1 1 Department of Mathematics KTH Royal Institute of Technology Stockholm, Sweden BFS 2012 June

More information

(5) Multi-parameter models - Gibbs sampling. ST440/540: Applied Bayesian Analysis

(5) Multi-parameter models - Gibbs sampling. ST440/540: Applied Bayesian Analysis Summarizing a posterior Given the data and prior the posterior is determined Summarizing the posterior gives parameter estimates, intervals, and hypothesis tests Most of these computations are integrals

More information

ECO 513 Fall 2009 C. Sims HIDDEN MARKOV CHAIN MODELS

ECO 513 Fall 2009 C. Sims HIDDEN MARKOV CHAIN MODELS ECO 513 Fall 2009 C. Sims HIDDEN MARKOV CHAIN MODELS 1. THE CLASS OF MODELS y t {y s, s < t} p(y t θ t, {y s, s < t}) θ t = θ(s t ) P[S t = i S t 1 = j] = h ij. 2. WHAT S HANDY ABOUT IT Evaluating the

More information

MARKOV CHAIN MONTE CARLO

MARKOV CHAIN MONTE CARLO MARKOV CHAIN MONTE CARLO RYAN WANG Abstract. This paper gives a brief introduction to Markov Chain Monte Carlo methods, which offer a general framework for calculating difficult integrals. We start with

More information

1 Bayesian Linear Regression (BLR)

1 Bayesian Linear Regression (BLR) Statistical Techniques in Robotics (STR, S15) Lecture#10 (Wednesday, February 11) Lecturer: Byron Boots Gaussian Properties, Bayesian Linear Regression 1 Bayesian Linear Regression (BLR) In linear regression,

More information

Introduction to Machine Learning CMU-10701

Introduction to Machine Learning CMU-10701 Introduction to Machine Learning CMU-10701 Markov Chain Monte Carlo Methods Barnabás Póczos & Aarti Singh Contents Markov Chain Monte Carlo Methods Goal & Motivation Sampling Rejection Importance Markov

More information

Time Series and Dynamic Models

Time Series and Dynamic Models Time Series and Dynamic Models Section 1 Intro to Bayesian Inference Carlos M. Carvalho The University of Texas at Austin 1 Outline 1 1. Foundations of Bayesian Statistics 2. Bayesian Estimation 3. The

More information

April 20th, Advanced Topics in Machine Learning California Institute of Technology. Markov Chain Monte Carlo for Machine Learning

April 20th, Advanced Topics in Machine Learning California Institute of Technology. Markov Chain Monte Carlo for Machine Learning for for Advanced Topics in California Institute of Technology April 20th, 2017 1 / 50 Table of Contents for 1 2 3 4 2 / 50 History of methods for Enrico Fermi used to calculate incredibly accurate predictions

More information

Markov chain Monte Carlo Lecture 9

Markov chain Monte Carlo Lecture 9 Markov chain Monte Carlo Lecture 9 David Sontag New York University Slides adapted from Eric Xing and Qirong Ho (CMU) Limitations of Monte Carlo Direct (unconditional) sampling Hard to get rare events

More information

Labor-Supply Shifts and Economic Fluctuations. Technical Appendix

Labor-Supply Shifts and Economic Fluctuations. Technical Appendix Labor-Supply Shifts and Economic Fluctuations Technical Appendix Yongsung Chang Department of Economics University of Pennsylvania Frank Schorfheide Department of Economics University of Pennsylvania January

More information

Principles of Bayesian Inference

Principles of Bayesian Inference Principles of Bayesian Inference Sudipto Banerjee 1 and Andrew O. Finley 2 1 Biostatistics, School of Public Health, University of Minnesota, Minneapolis, Minnesota, U.S.A. 2 Department of Forestry & Department

More information

Markov Networks.

Markov Networks. Markov Networks www.biostat.wisc.edu/~dpage/cs760/ Goals for the lecture you should understand the following concepts Markov network syntax Markov network semantics Potential functions Partition function

More information

Introduction to Computational Biology Lecture # 14: MCMC - Markov Chain Monte Carlo

Introduction to Computational Biology Lecture # 14: MCMC - Markov Chain Monte Carlo Introduction to Computational Biology Lecture # 14: MCMC - Markov Chain Monte Carlo Assaf Weiner Tuesday, March 13, 2007 1 Introduction Today we will return to the motif finding problem, in lecture 10

More information

Bayesian Gaussian Process Regression

Bayesian Gaussian Process Regression Bayesian Gaussian Process Regression STAT8810, Fall 2017 M.T. Pratola October 7, 2017 Today Bayesian Gaussian Process Regression Bayesian GP Regression Recall we had observations from our expensive simulator,

More information

MCMC Review. MCMC Review. Gibbs Sampling. MCMC Review

MCMC Review. MCMC Review. Gibbs Sampling. MCMC Review MCMC Review http://jackman.stanford.edu/mcmc/icpsr99.pdf http://students.washington.edu/fkrogsta/bayes/stat538.pdf http://www.stat.berkeley.edu/users/terry/classes/s260.1998 /Week9a/week9a/week9a.html

More information

Bayesian Methods in Multilevel Regression

Bayesian Methods in Multilevel Regression Bayesian Methods in Multilevel Regression Joop Hox MuLOG, 15 september 2000 mcmc What is Statistics?! Statistics is about uncertainty To err is human, to forgive divine, but to include errors in your design

More information

Markov Chain Monte Carlo (MCMC)

Markov Chain Monte Carlo (MCMC) Markov Chain Monte Carlo (MCMC Dependent Sampling Suppose we wish to sample from a density π, and we can evaluate π as a function but have no means to directly generate a sample. Rejection sampling can

More information

TDA231. Logistic regression

TDA231. Logistic regression TDA231 Devdatt Dubhashi dubhashi@chalmers.se Dept. of Computer Science and Engg. Chalmers University February 19, 2016 Some data 5 x2 0 5 5 0 5 x 1 In the Bayes classifier, we built a model of each class

More information

CPSC 540: Machine Learning

CPSC 540: Machine Learning CPSC 540: Machine Learning MCMC and Non-Parametric Bayes Mark Schmidt University of British Columbia Winter 2016 Admin I went through project proposals: Some of you got a message on Piazza. No news is

More information

an introduction to bayesian inference

an introduction to bayesian inference with an application to network analysis http://jakehofman.com january 13, 2010 motivation would like models that: provide predictive and explanatory power are complex enough to describe observed phenomena

More information

DAG models and Markov Chain Monte Carlo methods a short overview

DAG models and Markov Chain Monte Carlo methods a short overview DAG models and Markov Chain Monte Carlo methods a short overview Søren Højsgaard Institute of Genetics and Biotechnology University of Aarhus August 18, 2008 Printed: August 18, 2008 File: DAGMC-Lecture.tex

More information

Pattern Recognition and Machine Learning. Bishop Chapter 11: Sampling Methods

Pattern Recognition and Machine Learning. Bishop Chapter 11: Sampling Methods Pattern Recognition and Machine Learning Chapter 11: Sampling Methods Elise Arnaud Jakob Verbeek May 22, 2008 Outline of the chapter 11.1 Basic Sampling Algorithms 11.2 Markov Chain Monte Carlo 11.3 Gibbs

More information

Lecture 4: Dynamic models

Lecture 4: Dynamic models linear s Lecture 4: s Hedibert Freitas Lopes The University of Chicago Booth School of Business 5807 South Woodlawn Avenue, Chicago, IL 60637 http://faculty.chicagobooth.edu/hedibert.lopes hlopes@chicagobooth.edu

More information

Likelihood-free MCMC

Likelihood-free MCMC Bayesian inference for stable distributions with applications in finance Department of Mathematics University of Leicester September 2, 2011 MSc project final presentation Outline 1 2 3 4 Classical Monte

More information

1 Probabilities. 1.1 Basics 1 PROBABILITIES

1 Probabilities. 1.1 Basics 1 PROBABILITIES 1 PROBABILITIES 1 Probabilities Probability is a tricky word usually meaning the likelyhood of something occuring or how frequent something is. Obviously, if something happens frequently, then its probability

More information

Stat 535 C - Statistical Computing & Monte Carlo Methods. Arnaud Doucet.

Stat 535 C - Statistical Computing & Monte Carlo Methods. Arnaud Doucet. Stat 535 C - Statistical Computing & Monte Carlo Methods Arnaud Doucet Email: arnaud@cs.ubc.ca 1 1.1 Outline Introduction to Markov chain Monte Carlo The Gibbs Sampler Examples Overview of the Lecture

More information

Bayesian Linear Models

Bayesian Linear Models Bayesian Linear Models Sudipto Banerjee 1 and Andrew O. Finley 2 1 Department of Forestry & Department of Geography, Michigan State University, Lansing Michigan, U.S.A. 2 Biostatistics, School of Public

More information

Simulation - Lectures - Part III Markov chain Monte Carlo

Simulation - Lectures - Part III Markov chain Monte Carlo Simulation - Lectures - Part III Markov chain Monte Carlo Julien Berestycki Part A Simulation and Statistical Programming Hilary Term 2018 Part A Simulation. HT 2018. J. Berestycki. 1 / 50 Outline Markov

More information

MCMC Sampling for Bayesian Inference using L1-type Priors

MCMC Sampling for Bayesian Inference using L1-type Priors MÜNSTER MCMC Sampling for Bayesian Inference using L1-type Priors (what I do whenever the ill-posedness of EEG/MEG is just not frustrating enough!) AG Imaging Seminar Felix Lucka 26.06.2012 , MÜNSTER Sampling

More information

Bayesian Inference for DSGE Models. Lawrence J. Christiano

Bayesian Inference for DSGE Models. Lawrence J. Christiano Bayesian Inference for DSGE Models Lawrence J. Christiano Outline State space-observer form. convenient for model estimation and many other things. Preliminaries. Probabilities. Maximum Likelihood. Bayesian

More information

Dynamic System Identification using HDMR-Bayesian Technique

Dynamic System Identification using HDMR-Bayesian Technique Dynamic System Identification using HDMR-Bayesian Technique *Shereena O A 1) and Dr. B N Rao 2) 1), 2) Department of Civil Engineering, IIT Madras, Chennai 600036, Tamil Nadu, India 1) ce14d020@smail.iitm.ac.in

More information

STAT Advanced Bayesian Inference

STAT Advanced Bayesian Inference 1 / 32 STAT 625 - Advanced Bayesian Inference Meng Li Department of Statistics Jan 23, 218 The Dirichlet distribution 2 / 32 θ Dirichlet(a 1,...,a k ) with density p(θ 1,θ 2,...,θ k ) = k j=1 Γ(a j) Γ(

More information

Bayesian Estimation of Input Output Tables for Russia

Bayesian Estimation of Input Output Tables for Russia Bayesian Estimation of Input Output Tables for Russia Oleg Lugovoy (EDF, RANE) Andrey Polbin (RANE) Vladimir Potashnikov (RANE) WIOD Conference April 24, 2012 Groningen Outline Motivation Objectives Bayesian

More information

Metropolis Hastings. Rebecca C. Steorts Bayesian Methods and Modern Statistics: STA 360/601. Module 9

Metropolis Hastings. Rebecca C. Steorts Bayesian Methods and Modern Statistics: STA 360/601. Module 9 Metropolis Hastings Rebecca C. Steorts Bayesian Methods and Modern Statistics: STA 360/601 Module 9 1 The Metropolis-Hastings algorithm is a general term for a family of Markov chain simulation methods

More information

16 : Approximate Inference: Markov Chain Monte Carlo

16 : Approximate Inference: Markov Chain Monte Carlo 10-708: Probabilistic Graphical Models 10-708, Spring 2017 16 : Approximate Inference: Markov Chain Monte Carlo Lecturer: Eric P. Xing Scribes: Yuan Yang, Chao-Ming Yen 1 Introduction As the target distribution

More information

Markov Chain Monte Carlo Inference. Siamak Ravanbakhsh Winter 2018

Markov Chain Monte Carlo Inference. Siamak Ravanbakhsh Winter 2018 Graphical Models Markov Chain Monte Carlo Inference Siamak Ravanbakhsh Winter 2018 Learning objectives Markov chains the idea behind Markov Chain Monte Carlo (MCMC) two important examples: Gibbs sampling

More information

Bayesian Semiparametric GARCH Models

Bayesian Semiparametric GARCH Models Bayesian Semiparametric GARCH Models Xibin (Bill) Zhang and Maxwell L. King Department of Econometrics and Business Statistics Faculty of Business and Economics xibin.zhang@monash.edu Quantitative Methods

More information

Bayes Nets: Sampling

Bayes Nets: Sampling Bayes Nets: Sampling [These slides were created by Dan Klein and Pieter Abbeel for CS188 Intro to AI at UC Berkeley. All CS188 materials are available at http://ai.berkeley.edu.] Approximate Inference:

More information

Computer Vision Group Prof. Daniel Cremers. 11. Sampling Methods

Computer Vision Group Prof. Daniel Cremers. 11. Sampling Methods Prof. Daniel Cremers 11. Sampling Methods Sampling Methods Sampling Methods are widely used in Computer Science as an approximation of a deterministic algorithm to represent uncertainty without a parametric

More information

SUPPLEMENT TO MARKET ENTRY COSTS, PRODUCER HETEROGENEITY, AND EXPORT DYNAMICS (Econometrica, Vol. 75, No. 3, May 2007, )

SUPPLEMENT TO MARKET ENTRY COSTS, PRODUCER HETEROGENEITY, AND EXPORT DYNAMICS (Econometrica, Vol. 75, No. 3, May 2007, ) Econometrica Supplementary Material SUPPLEMENT TO MARKET ENTRY COSTS, PRODUCER HETEROGENEITY, AND EXPORT DYNAMICS (Econometrica, Vol. 75, No. 3, May 2007, 653 710) BY SANGHAMITRA DAS, MARK ROBERTS, AND

More information

DS-GA 1003: Machine Learning and Computational Statistics Homework 7: Bayesian Modeling

DS-GA 1003: Machine Learning and Computational Statistics Homework 7: Bayesian Modeling DS-GA 1003: Machine Learning and Computational Statistics Homework 7: Bayesian Modeling Due: Tuesday, May 10, 2016, at 6pm (Submit via NYU Classes) Instructions: Your answers to the questions below, including

More information

Bayesian Linear Models

Bayesian Linear Models Bayesian Linear Models Sudipto Banerjee 1 and Andrew O. Finley 2 1 Biostatistics, School of Public Health, University of Minnesota, Minneapolis, Minnesota, U.S.A. 2 Department of Forestry & Department

More information

1 Probabilities. 1.1 Basics 1 PROBABILITIES

1 Probabilities. 1.1 Basics 1 PROBABILITIES 1 PROBABILITIES 1 Probabilities Probability is a tricky word usually meaning the likelyhood of something occuring or how frequent something is. Obviously, if something happens frequently, then its probability

More information

Computer Practical: Metropolis-Hastings-based MCMC

Computer Practical: Metropolis-Hastings-based MCMC Computer Practical: Metropolis-Hastings-based MCMC Andrea Arnold and Franz Hamilton North Carolina State University July 30, 2016 A. Arnold / F. Hamilton (NCSU) MH-based MCMC July 30, 2016 1 / 19 Markov

More information