Paul Karapanagiotidis ECO4060

Similar documents
Markov Chain Monte Carlo Methods

Kernel adaptive Sequential Monte Carlo

Session 3A: Markov chain Monte Carlo (MCMC)

19 : Slice Sampling and HMC

Markov Chain Monte Carlo Lecture 4

CPSC 540: Machine Learning

Large Scale Bayesian Inference

Hastings-within-Gibbs Algorithm: Introduction and Application on Hierarchical Model

MCMC and Gibbs Sampling. Kayhan Batmanghelich

Exercises Tutorial at ICASSP 2016 Learning Nonlinear Dynamical Models Using Particle Filters

Markov Chain Monte Carlo methods

STA 4273H: Statistical Machine Learning

CSC 446 Notes: Lecture 13

Monte Carlo in Bayesian Statistics

Riemann Manifold Methods in Bayesian Statistics

Markov Chain Monte Carlo, Numerical Integration

Learning Energy-Based Models of High-Dimensional Data

1 Geometry of high dimensional probability distributions

Density estimation. Computing, and avoiding, partition functions. Iain Murray

Introduction to Hamiltonian Monte Carlo Method

Markov Chain Monte Carlo methods

Introduction to Machine Learning CMU-10701

MCMC Sampling for Bayesian Inference using L1-type Priors

The Metropolis-Hastings Algorithm. June 8, 2012

The University of Auckland Applied Mathematics Bayesian Methods for Inverse Problems : why and how Colin Fox Tiangang Cui, Mike O Sullivan (Auckland),

STA 294: Stochastic Processes & Bayesian Nonparametrics

Lecture 6: Markov Chain Monte Carlo

Infinite-State Markov-switching for Dynamic. Volatility Models : Web Appendix

The Bias-Variance dilemma of the Monte Carlo. method. Technion - Israel Institute of Technology, Technion City, Haifa 32000, Israel

Notes on pseudo-marginal methods, variational Bayes and ABC

17 : Markov Chain Monte Carlo

Markov Chain Monte Carlo (MCMC)

MIT /30 Gelman, Carpenter, Hoffman, Guo, Goodrich, Lee,... Stan for Bayesian data analysis

Lecture 8: Bayesian Estimation of Parameters in State Space Models

Bayesian Inference for Discretely Sampled Diffusion Processes: A New MCMC Based Approach to Inference

Online appendix to On the stability of the excess sensitivity of aggregate consumption growth in the US

Markov Chain Monte Carlo (MCMC)

Computational statistics

LECTURE 15 Markov chain Monte Carlo

Sampling from complex probability distributions

Tutorial on Probabilistic Programming with PyMC3

Research Division Federal Reserve Bank of St. Louis Working Paper Series

Nonparametric Bayesian Methods (Gaussian Processes)

17 : Optimization and Monte Carlo Methods

Stat 535 C - Statistical Computing & Monte Carlo Methods. Lecture February Arnaud Doucet

Approximate inference in Energy-Based Models

Monetary and Exchange Rate Policy Under Remittance Fluctuations. Technical Appendix and Additional Results

A Search and Jump Algorithm for Markov Chain Monte Carlo Sampling. Christopher Jennison. Adriana Ibrahim. Seminar at University of Kuwait

Probabilistic Machine Learning

Bayesian Methods for Machine Learning

Gradient-based Monte Carlo sampling methods

16 : Approximate Inference: Markov Chain Monte Carlo

Sandwiching the marginal likelihood using bidirectional Monte Carlo. Roger Grosse

Markov Chain Monte Carlo Inference. Siamak Ravanbakhsh Winter 2018

Lecture 7 and 8: Markov Chain Monte Carlo

Pseudo-marginal MCMC methods for inference in latent variable models

Constrained Hamiltonian Monte Carlo in BEKK GARCH with Targeting

Lecture 15: MCMC Sanjeev Arora Elad Hazan. COS 402 Machine Learning and Artificial Intelligence Fall 2016

Zig-Zag Monte Carlo. Delft University of Technology. Joris Bierkens February 7, 2017

Markov Chains and MCMC

Forward Problems and their Inverse Solutions

Markov chain Monte Carlo

A Review of Pseudo-Marginal Markov Chain Monte Carlo

SAMPLING ALGORITHMS. In general. Inference in Bayesian models

Minicourse on: Markov Chain Monte Carlo: Simulation Techniques in Statistics

Random Walks A&T and F&S 3.1.2

Answers and expectations

Pattern Recognition and Machine Learning. Bishop Chapter 11: Sampling Methods

MARKOV CHAIN MONTE CARLO

CSC 2541: Bayesian Methods for Machine Learning

Markov chain Monte Carlo Lecture 9

Session 5B: A worked example EGARCH model

Bayesian Phylogenetics:

Advanced Statistical Modelling

Chapter 12 PAWL-Forced Simulated Tempering

Bayesian inference for multivariate extreme value distributions

Monte Carlo Methods for Computation and Optimization (048715)

Kernel Sequential Monte Carlo

GSHMC: An efficient Markov chain Monte Carlo sampling method. Sebastian Reich in collaboration with Elena Akhmatskaya (Fujitsu Laboratories Europe)

Quasi-Newton Methods for Markov Chain Monte Carlo

Markov chain Monte Carlo methods in atmospheric remote sensing

Gaussian kernel GARCH models

Log Gaussian Cox Processes. Chi Group Meeting February 23, 2016

Markov Chain Monte Carlo

CSC 2541: Bayesian Methods for Machine Learning


Markov Chain Monte Carlo

The zig-zag and super-efficient sampling for Bayesian analysis of big data

Bayesian Inference in GLMs. Frequentists typically base inferences on MLEs, asymptotic confidence

Markov Chain Monte Carlo Methods for Stochastic Optimization

Quantifying Uncertainty

Computer Practical: Metropolis-Hastings-based MCMC

Doing Bayesian Integrals

Hamiltonian Monte Carlo

Particle Filtering Approaches for Dynamic Stochastic Optimization

Markov chain Monte Carlo methods for visual tracking

An ABC interpretation of the multiple auxiliary variable method

Bayesian Inference for DSGE Models. Lawrence J. Christiano

Multimodal Nested Sampling

Markov chain Monte Carlo

Transcription:

Paul Karapanagiotidis ECO4060

The way forward 1) Motivate why Markov-Chain Monte Carlo (MCMC) is useful for econometric modeling 2) Introduce Markov-Chain Monte Carlo (MCMC) - Metropolis-Hastings (MH) algorithm 3) Discuss recent improvements to the MH algorithm - Hamiltonian Monte Carlo (HMC) 4) Talk about possible future research: - Apply MCMC to Stochastic Volatility modeling

Econometric Modeling What is the connection between MCMC and econometric models? (i.e. why do we even care?) Modern econometric models often involve: - A large number of parameters - Complicated non-linear relationships - Often impossible to find analytic closed form solutions - Unobserved or latent variables may require simulation MCMC is useful since: - We can now estimate the parameters of these otherwise intractable models And. - Easy to implement Bayesian approach which often simplifies problems - Turns larger estimation problems can be broken up into smaller parts

Econometric Modeling What are some examples of useful applications? 1) High-dimensional multivariate GARCH models - BEKK Specification (Engle & Kroner, 1995) 2) Stochastic Volatility (SV) models (Taylor, 1994) - Volatility is unobserved 3) Dynamic Stochastic General Equilibrium models (DSGE) - Bayesian approach facilitates estimation (Fernandez-V., 2009)

An Example What exactly is MCMC (Markov Chain Monte Carlo)? Example: ( classic MC integration) Classic MC integration of the Expectation requires being able to sample from f(x)! MCMC resolves this issue: it provides a general framework for sampling from any distribution!

Metropolis-Hastings algorithm Paper by Chib & Greenberg (1995) One of the most general MCMC methods. - Transition kernel (continuous states) is: - Invariant distribution π(x) satisfies: - Detailed balance condition: We accept proposed value to the Markov Chain if: Question: How to propose values from that are usually accepted?

Traditional proposal mechanism We typically use a Random walk! Problems: - Random walk typically suffers from correlation of the Markov Chain! - High correlation makes inference less precise - When f(x) is a very complicated distribution: - Random walk will prove even worse as it will get stuck in regions A lack of good proposal mechanisms represents the main challenge facing the M-H algorithm. Can we find a more intelligent method of proposing Markov Chain draws?

Hamiltonian Monte Carlo Paper by Radford Neal (2010) Example: Bi-Variate Normal R.V. x with parameters θ={μ,σ} The distribution f(x θ) looks like:

Hamiltonian Monte Carlo - p = x denotes current position of the object ; m, momentum; M, a mass matrix. is potential energy as a function of position is kinetic energy as a function of momentum Laws of Motion: - Boltzmann distribution of energy in a system : - Η(p,m) is the Hamiltonian energy of the system (from statistical mechanics). Target distribution we wish to sample from

Hamiltonian Monte Carlo Total Energy, should stay constant as we move through the (p,m) space! In theory HMC could intelligently traverse the entire distribution support in one step, so draws would be nearly independent! Metropolis-Hastings draws are now accepted to the Markov Chain if: Ratio of Boltzmann dist s

Hamiltonian Monte Carlo Problems: - Computers can t deal with continuous variables - need to discretize laws of motion - Often H(p,m) will grow without bound if stability conditions not met - Proposed momentum doesn t consider current position - algorithm could still be made smarter - Only really works for smooth distributions

Hamiltonian Monte Carlo Maheu & Burda (2010) Augmented HMC and high-dimension BEKK GARCH : Improvements to HMC: - Fundamental idea is to adapt M(x) according to the local curvature of f(x θ) - Mass matrix M(x) now a function of the current state of the R.V. - However, this makes H(p,m) non-separable - we now require a local symmetry condition - Therefore, authors derive a new discretization that: - ensures local symmetry condition - avoids approximation error growing without bound So we ve got some nice improvements here application to BEKK GARCH shows great potential!

Looking forward towards research proposal Possible future research area: - High-frequency data - Stochastic Volatility (SV) model is probably preferred to ARCH - Model volatility with Realized Volatility estimator - Volatility Spillover effects Benefits of MCMC in this case: - MCMC on SV automatically provides optimal non-linear smoothed state inference - No need for separate particle filter or Kalman filter! - Higher precision samples from HMC imply better inference - Bayesian approach! - lets us turn the problem into an integral - breaks up problem into constituent conditional distributions