Pseudo-marginal MCMC methods for inference in latent variable models

Size: px
Start display at page:

Download "Pseudo-marginal MCMC methods for inference in latent variable models"

Transcription

1 Pseudo-marginal MCMC methods for inference in latent variable models Arnaud Doucet Department of Statistics, Oxford University Joint work with George Deligiannidis (Oxford) & Mike Pitt (Kings) MCQMC, 19/08/2016 (MCQMC, 19/08/2016) 1 / 31

2 Organization of the talk Latent variable models (MCQMC, 19/08/2016) 2 / 31

3 Organization of the talk Latent variable models The pseudo-marginal method (MCQMC, 19/08/2016) 2 / 31

4 Organization of the talk Latent variable models The pseudo-marginal method Optimal tuning (MCQMC, 19/08/2016) 2 / 31

5 Organization of the talk Latent variable models The pseudo-marginal method Optimal tuning The correlated pseudo-marginal method (MCQMC, 19/08/2016) 2 / 31

6 Organization of the talk Latent variable models The pseudo-marginal method Optimal tuning The correlated pseudo-marginal method Illustrations (MCQMC, 19/08/2016) 2 / 31

7 Latent Variable Models Assume (Y t ) t 1 are i.i.d. random variables such that X t i.i.d. µ θ ( ), Y t (X t = x) g θ ( x) for t = 1,..., T where (X t ) t 1 are latent variables. (MCQMC, 19/08/2016) 3 / 31

8 Latent Variable Models Assume (Y t ) t 1 are i.i.d. random variables such that X t i.i.d. µ θ ( ), Y t (X t = x) g θ ( x) for t = 1,..., T where (X t ) t 1 are latent variables. The likelihood of θ Θ R d associated to observations Y 1:T = y 1:T is p θ (y 1:T ) = T t=1 p θ (y t ), where p θ (y t ) = µ θ (x t ) g θ (y t x t ) dx t. (MCQMC, 19/08/2016) 3 / 31

9 Latent Variable Models Assume (Y t ) t 1 are i.i.d. random variables such that X t i.i.d. µ θ ( ), Y t (X t = x) g θ ( x) for t = 1,..., T where (X t ) t 1 are latent variables. The likelihood of θ Θ R d associated to observations Y 1:T = y 1:T is p θ (y 1:T ) = T t=1 p θ (y t ), where p θ (y t ) = µ θ (x t ) g θ (y t x t ) dx t. In many scenarios, p θ (y 1:T ) cannot be evaluated exactly; e.g. multivariate probit model. (MCQMC, 19/08/2016) 3 / 31

10 Latent Variable Models Assume {X t } t 1 is a latent Markov process, i.e. X 1 µ θ ( ) and X t+1 (X t = x) f θ ( x), Y t (X t = x) g θ ( x). (MCQMC, 19/08/2016) 4 / 31

11 Latent Variable Models Assume {X t } t 1 is a latent Markov process, i.e. X 1 µ θ ( ) and X t+1 (X t = x) f θ ( x), Y t (X t = x) g θ ( x). The likelihood of θ R d associated to Y 1:T = y 1:T is p θ (y 1:T ) = p θ (x 1:T, y 1:T )dx 1:T where p θ (x 1:T, y 1:T ) = µ θ (x 1 )g θ (y 1 x 1 ) T t=2 f θ (x t x t 1 ) g θ (y t x t ). (MCQMC, 19/08/2016) 4 / 31

12 Latent Variable Models Assume {X t } t 1 is a latent Markov process, i.e. X 1 µ θ ( ) and X t+1 (X t = x) f θ ( x), Y t (X t = x) g θ ( x). The likelihood of θ R d associated to Y 1:T = y 1:T is p θ (y 1:T ) = p θ (x 1:T, y 1:T )dx 1:T where p θ (x 1:T, y 1:T ) = µ θ (x 1 )g θ (y 1 x 1 ) T t=2 f θ (x t x t 1 ) g θ (y t x t ). State-space models are a very popular class of time series models but inference is diffi cult as p θ (y 1:T ) is intractable for non-linear/non-gaussian models. (MCQMC, 19/08/2016) 4 / 31

13 Bayesian Inference Prior distribution of density p (θ). (MCQMC, 19/08/2016) 5 / 31

14 Bayesian Inference Prior distribution of density p (θ). Likelihood function p θ (y 1:T ). (MCQMC, 19/08/2016) 5 / 31

15 Bayesian Inference Prior distribution of density p (θ). Likelihood function p θ (y 1:T ). Bayesian inference relies on the posterior π (θ) = p ( θ y 1:T ) = p θ (y 1:T ) p (θ) Θ p θ (y 1:T ) p ( θ ) dθ. (MCQMC, 19/08/2016) 5 / 31

16 Bayesian Inference Prior distribution of density p (θ). Likelihood function p θ (y 1:T ). Bayesian inference relies on the posterior π (θ) = p ( θ y 1:T ) = p θ (y 1:T ) p (θ) Θ p θ (y 1:T ) p ( θ ) dθ. For non-trivial models, inference relies typically on MCMC. (MCQMC, 19/08/2016) 5 / 31

17 Standard MCMC Approaches Standard MCMC schemes target p ( θ, x 1:T y 1:T ) p (θ) p θ (x 1:T, y 1:T ) using Gibbs type strategy; i.e. sample alternately X 1:T p θ ( y 1:T ) and θ p ( y 1:T, X 1:T ). (MCQMC, 19/08/2016) 6 / 31

18 Standard MCMC Approaches Standard MCMC schemes target p ( θ, x 1:T y 1:T ) p (θ) p θ (x 1:T, y 1:T ) using Gibbs type strategy; i.e. sample alternately X 1:T p θ ( y 1:T ) and θ p ( y 1:T, X 1:T ). Problem 1: It can be diffi cult to sample p θ (x 1:T y 1:T ); e.g. state-space models. (MCQMC, 19/08/2016) 6 / 31

19 Standard MCMC Approaches Standard MCMC schemes target p ( θ, x 1:T y 1:T ) p (θ) p θ (x 1:T, y 1:T ) using Gibbs type strategy; i.e. sample alternately X 1:T p θ ( y 1:T ) and θ p ( y 1:T, X 1:T ). Problem 1: It can be diffi cult to sample p θ (x 1:T y 1:T ); e.g. state-space models. Problem 2: Even when it is implementable, Gibbs can converge very slowly. (MCQMC, 19/08/2016) 6 / 31

20 Standard MCMC Approaches Standard MCMC schemes target p ( θ, x 1:T y 1:T ) p (θ) p θ (x 1:T, y 1:T ) using Gibbs type strategy; i.e. sample alternately X 1:T p θ ( y 1:T ) and θ p ( y 1:T, X 1:T ). Problem 1: It can be diffi cult to sample p θ (x 1:T y 1:T ); e.g. state-space models. Problem 2: Even when it is implementable, Gibbs can converge very slowly. Pseudo-marginal methods mimic an algorithm targetting directly p ( θ y 1:T ) instead of p ( θ, x 1:T y 1:T ). (MCQMC, 19/08/2016) 6 / 31

21 Ideal Marginal Metropolis-Hastings Metropolis Hastings (MH) algorithm simulates an ergodic Markov chain {ϑ i } i 1 of limiting distribution π (θ). (MCQMC, 19/08/2016) 7 / 31

22 Ideal Marginal Metropolis-Hastings Metropolis Hastings (MH) algorithm simulates an ergodic Markov chain {ϑ i } i 1 of limiting distribution π (θ). (MCQMC, 19/08/2016) 7 / 31

23 Ideal Marginal Metropolis-Hastings Metropolis Hastings (MH) algorithm simulates an ergodic Markov chain {ϑ i } i 1 of limiting distribution π (θ). At iteration i Sample ϑ q ( ϑ i 1 ). (MCQMC, 19/08/2016) 7 / 31

24 Ideal Marginal Metropolis-Hastings Metropolis Hastings (MH) algorithm simulates an ergodic Markov chain {ϑ i } i 1 of limiting distribution π (θ). At iteration i Sample ϑ q ( ϑ i 1 ). With probability 1 π (ϑ) q ( ϑ i 1 ϑ) π (ϑ i 1 ) q ( ϑ ϑ i 1 ) = 1 p ϑ (y 1:T ) p (ϑ) q ( ϑ i 1 ϑ) p ϑi 1 (y 1:T ) p (ϑ i 1 ) q ( ϑ ϑ i 1 ), set ϑ i = ϑ, otherwise set ϑ i = ϑ i 1. (MCQMC, 19/08/2016) 7 / 31

25 Ideal Marginal Metropolis-Hastings Metropolis Hastings (MH) algorithm simulates an ergodic Markov chain {ϑ i } i 1 of limiting distribution π (θ). At iteration i Sample ϑ q ( ϑ i 1 ). With probability 1 π (ϑ) q ( ϑ i 1 ϑ) π (ϑ i 1 ) q ( ϑ ϑ i 1 ) = 1 p ϑ (y 1:T ) p (ϑ) q ( ϑ i 1 ϑ) p ϑi 1 (y 1:T ) p (ϑ i 1 ) q ( ϑ ϑ i 1 ), set ϑ i = ϑ, otherwise set ϑ i = ϑ i 1. Problem: MH cannot be implemented if p ϑ (y 1:T ) cannot be evaluated. (MCQMC, 19/08/2016) 7 / 31

26 Pseudo-Marginal Metropolis Hastings Idea: Replace p ϑ (y 1:T ) by an non-negative unbiased estimate p ϑ (y 1:T ) in MH. (MCQMC, 19/08/2016) 8 / 31

27 Pseudo-Marginal Metropolis Hastings Idea: Replace p ϑ (y 1:T ) by an non-negative unbiased estimate p ϑ (y 1:T ) in MH. (MCQMC, 19/08/2016) 8 / 31

28 Pseudo-Marginal Metropolis Hastings Idea: Replace p ϑ (y 1:T ) by an non-negative unbiased estimate p ϑ (y 1:T ) in MH. At iteration i Sample ϑ q ( ϑ i 1 ). (MCQMC, 19/08/2016) 8 / 31

29 Pseudo-Marginal Metropolis Hastings Idea: Replace p ϑ (y 1:T ) by an non-negative unbiased estimate p ϑ (y 1:T ) in MH. At iteration i Sample ϑ q ( ϑ i 1 ). Compute an estimate p ϑ (y 1:T ) of p ϑ (y 1:T ). (MCQMC, 19/08/2016) 8 / 31

30 Pseudo-Marginal Metropolis Hastings Idea: Replace p ϑ (y 1:T ) by an non-negative unbiased estimate p ϑ (y 1:T ) in MH. At iteration i Sample ϑ q ( ϑ i 1 ). Compute an estimate p ϑ (y 1:T ) of p ϑ (y 1:T ). With probability 1 = 1 p ϑ (y 1:T ) p (ϑ) q ( ϑ i 1 ϑ) p ϑi 1 (y 1:T ) p (ϑ i 1 ) q ( ϑ ϑ i 1 ) p ϑ (y 1:T ) p (ϑ) q ( ϑ i 1 ϑ) p ϑ (y 1:T ) /p ϑ (y 1:T ) p ϑi 1 (y 1:T ) p (ϑ i 1 ) q ( ϑ ϑ i 1 ) p ϑi 1 (y 1:T ) /p ϑi 1 (y 1:T ), set ϑ i = ϑ, p ϑi (y 1:T ) = p ϑ (y 1:T ) otherwise set ϑ i = ϑ i 1, p ϑi (y 1:T ) = p ϑi 1 (y 1:T ). (MCQMC, 19/08/2016) 8 / 31

31 Pseudo-Marginal Metropolis Hastings Proposition (Lin, Liu & Sloan, Phys. Rev. D, 2000): If p θ (y 1:T ) is a non-negative unbiased estimator of p θ (y 1:T ) then the pseudo-marginal MH kernel admits π (θ) as invariant density. (MCQMC, 19/08/2016) 9 / 31

32 Pseudo-Marginal Metropolis Hastings Proposition (Lin, Liu & Sloan, Phys. Rev. D, 2000): If p θ (y 1:T ) is a non-negative unbiased estimator of p θ (y 1:T ) then the pseudo-marginal MH kernel admits π (θ) as invariant density. Let U be the U-valued r.v. such that p θ (y 1:T ) = p θ (y 1:T ; U) and E [ p θ (y 1:T ; U)] = p θ (y 1:T ) when U m θ ( ). (MCQMC, 19/08/2016) 9 / 31

33 Pseudo-Marginal Metropolis Hastings Proposition (Lin, Liu & Sloan, Phys. Rev. D, 2000): If p θ (y 1:T ) is a non-negative unbiased estimator of p θ (y 1:T ) then the pseudo-marginal MH kernel admits π (θ) as invariant density. Let U be the U-valued r.v. such that p θ (y 1:T ) = p θ (y 1:T ; U) and E [ p θ (y 1:T ; U)] = p θ (y 1:T ) when U m θ ( ). Pseudo-marginal MH is a standard MH on Θ U with target π(θ, u) and proposal q ( ϑ θ) m ϑ (v) where π(θ, u) = π (θ) p θ (y 1:T ; u) p θ (y 1:T ) m θ (u) satisfies π(θ, u)du = π (θ). (MCQMC, 19/08/2016) 9 / 31

34 Likelihood Estimators For latent variable models, one has p θ (y 1:T ) = T t=1 p θ (y t ), where p θ (y t ) = µ θ (x t ) g θ (y t x t ) dx t. (MCQMC, 19/08/2016) 10 / 31

35 Likelihood Estimators For latent variable models, one has p θ (y 1:T ) = T t=1 p θ (y t ), where p θ (y t ) = µ θ (x t ) g θ (y t x t ) dx t. A non-negative unbiased estimator is given by { T T N 1 ( ) } p θ (y 1:T ) = p θ (y t ) = t=1 t=1 N g θ y t Xt k, X k i.i.d. t µ θ, k=1 i.e. U = ( X1 1,..., X 1 N,..., X T 1,..., X T N ) with m θ (u) = T N t=1 k=1 µ θ (x k t ). (MCQMC, 19/08/2016) 10 / 31

36 Likelihood Estimators For latent variable models, one has p θ (y 1:T ) = T t=1 p θ (y t ), where p θ (y t ) = µ θ (x t ) g θ (y t x t ) dx t. A non-negative unbiased estimator is given by { T T N 1 ( ) } p θ (y 1:T ) = p θ (y t ) = t=1 t=1 N g θ y t Xt k, X k i.i.d. t µ θ, k=1 i.e. U = ( X1 1,..., X 1 N,..., X T 1,..., X T N ) with m θ (u) = T N t=1 k=1 Computational complexity is O(NT ). µ θ (x k t ). (MCQMC, 19/08/2016) 10 / 31

37 Likelihood Estimators For state-space models, an alternative is to use particle filters where { T T N 1 ( ) } p θ (y 1:T ) = p θ (y 1 ) p θ (y t y 1:t 1 ) = t=2 t=1 N g θ y t Xt k k=1 where m θ (u) = N ) µ θ (x 1 k T k=1 t=2 with a k t 1 {1,..., N}, w j t g θ ( { N k=1 y t X j t ( w ak t 1 t f xt k x ak t 1 ), j w j t = 1. t 1 ) } (MCQMC, 19/08/2016) 11 / 31

38 Likelihood Estimators For state-space models, an alternative is to use particle filters where { T T N 1 ( ) } p θ (y 1:T ) = p θ (y 1 ) p θ (y t y 1:t 1 ) = t=2 t=1 N g θ y t Xt k k=1 where m θ (u) = N ) µ θ (x 1 k T k=1 t=2 with a k t 1 {1,..., N}, w j t g θ ( { N k=1 y t X j t ( w ak t 1 t f xt k x ak t 1 ), j w j t = 1. The estimator is unbiased (Del Moral, 1998), relative variance is bounded uniformly over T if N T (Cerou, Del Moral & Guyader, 2011), SQMC could also be used (Gerber & Chopin, JRSS B 2015). t 1 ) } (MCQMC, 19/08/2016) 11 / 31

39 Likelihood Estimators For state-space models, an alternative is to use particle filters where { T T N 1 ( ) } p θ (y 1:T ) = p θ (y 1 ) p θ (y t y 1:t 1 ) = t=2 t=1 N g θ y t Xt k k=1 where m θ (u) = N ) µ θ (x 1 k T k=1 t=2 with a k t 1 {1,..., N}, w j t g θ ( { N k=1 y t X j t ( w ak t 1 t f xt k x ak t 1 ), j w j t = 1. The estimator is unbiased (Del Moral, 1998), relative variance is bounded uniformly over T if N T (Cerou, Del Moral & Guyader, 2011), SQMC could also be used (Gerber & Chopin, JRSS B 2015). Computational complexity is O (NT ). t 1 ) } (MCQMC, 19/08/2016) 11 / 31

40 Examples of recent applications Measuring the impact of Ebola control measures in Sierra Leone, PNAS, (MCQMC, 19/08/2016) 12 / 31

41 Examples of recent applications Measuring the impact of Ebola control measures in Sierra Leone, PNAS, Dynamic prediction pools: an investigation of financial frictions and forecasting performance, J. Econometrics, (MCQMC, 19/08/2016) 12 / 31

42 Examples of recent applications Measuring the impact of Ebola control measures in Sierra Leone, PNAS, Dynamic prediction pools: an investigation of financial frictions and forecasting performance, J. Econometrics, Capturing the dynamics of pathogens with many strains, J. Mathematical Biology, (MCQMC, 19/08/2016) 12 / 31

43 Examples of recent applications Measuring the impact of Ebola control measures in Sierra Leone, PNAS, Dynamic prediction pools: an investigation of financial frictions and forecasting performance, J. Econometrics, Capturing the dynamics of pathogens with many strains, J. Mathematical Biology, Monte Carlo estimation of stage structured development from cohort data, Ecology, (MCQMC, 19/08/2016) 12 / 31

44 Examples of recent applications Measuring the impact of Ebola control measures in Sierra Leone, PNAS, Dynamic prediction pools: an investigation of financial frictions and forecasting performance, J. Econometrics, Capturing the dynamics of pathogens with many strains, J. Mathematical Biology, Monte Carlo estimation of stage structured development from cohort data, Ecology, Bayesian analysis of entry games: A simulated likelihood approach without simulation errors, Global Economic Review, (MCQMC, 19/08/2016) 12 / 31

45 Examples of recent applications Measuring the impact of Ebola control measures in Sierra Leone, PNAS, Dynamic prediction pools: an investigation of financial frictions and forecasting performance, J. Econometrics, Capturing the dynamics of pathogens with many strains, J. Mathematical Biology, Monte Carlo estimation of stage structured development from cohort data, Ecology, Bayesian analysis of entry games: A simulated likelihood approach without simulation errors, Global Economic Review, Controlling procedural modeling programs. SIGGRAPH, (MCQMC, 19/08/2016) 12 / 31

46 Empirical performance: Stochastic volatility model SV model with leverage (Huang & Tauchen, 2005): dv 1 (s) = k 1 {v 1 (s) µ 1 } ds + σ 1 dw 1 (s), dv 2 (s) = k 2 v 2 (s) ds + {1 + β 12 v 2 (s)} dw 2 (s), d log P (s) = µ y ds + s-exp [{v 1 (s) + β 2 v 2 (s)} /2] db (s), with φ 1 =corr{b (s), W 1 (s)} and φ 2 =corr{b (s), W 2 (s)}. (MCQMC, 19/08/2016) 13 / 31

47 Empirical performance: Stochastic volatility model SV model with leverage (Huang & Tauchen, 2005): dv 1 (s) = k 1 {v 1 (s) µ 1 } ds + σ 1 dw 1 (s), dv 2 (s) = k 2 v 2 (s) ds + {1 + β 12 v 2 (s)} dw 2 (s), d log P (s) = µ y ds + s-exp [{v 1 (s) + β 2 v 2 (s)} /2] db (s), with φ 1 =corr{b (s), W 1 (s)} and φ 2 =corr{b (s), W 2 (s)}. Euler discretization of the volatilities v 1 (s) and v 2 (s) provides closed form expression for Y t = log P ( t) log P ( (t 1)). (MCQMC, 19/08/2016) 13 / 31

48 Empirical performance: Stochastic volatility model SV model with leverage (Huang & Tauchen, 2005): dv 1 (s) = k 1 {v 1 (s) µ 1 } ds + σ 1 dw 1 (s), dv 2 (s) = k 2 v 2 (s) ds + {1 + β 12 v 2 (s)} dw 2 (s), d log P (s) = µ y ds + s-exp [{v 1 (s) + β 2 v 2 (s)} /2] db (s), with φ 1 =corr{b (s), W 1 (s)} and φ 2 =corr{b (s), W 2 (s)}. Euler discretization of the volatilities v 1 (s) and v 2 (s) provides closed form expression for Y t = log P ( t) log P ( (t 1)). Daily returns y = (y 1,..., y T ) of the S&P 500 index. (MCQMC, 19/08/2016) 13 / 31

49 Empirical performance: Stochastic volatility model SV model with leverage (Huang & Tauchen, 2005): dv 1 (s) = k 1 {v 1 (s) µ 1 } ds + σ 1 dw 1 (s), dv 2 (s) = k 2 v 2 (s) ds + {1 + β 12 v 2 (s)} dw 2 (s), d log P (s) = µ y ds + s-exp [{v 1 (s) + β 2 v 2 (s)} /2] db (s), with φ 1 =corr{b (s), W 1 (s)} and φ 2 =corr{b (s), W 2 (s)}. Euler discretization of the volatilities v 1 (s) and v 2 (s) provides closed form expression for Y t = log P ( t) log P ( (t 1)). Daily returns y = (y 1,..., y T ) of the S&P 500 index. ) Bayesian Inference on θ = (k 1, µ 1, σ 1, k 2, β 12, β 2, µ y, φ 1, φ 2. (MCQMC, 19/08/2016) 13 / 31

50 Empirical performance: Stochastic volatility model SV model with leverage (Huang & Tauchen, 2005): dv 1 (s) = k 1 {v 1 (s) µ 1 } ds + σ 1 dw 1 (s), dv 2 (s) = k 2 v 2 (s) ds + {1 + β 12 v 2 (s)} dw 2 (s), d log P (s) = µ y ds + s-exp [{v 1 (s) + β 2 v 2 (s)} /2] db (s), with φ 1 =corr{b (s), W 1 (s)} and φ 2 =corr{b (s), W 2 (s)}. Euler discretization of the volatilities v 1 (s) and v 2 (s) provides closed form expression for Y t = log P ( t) log P ( (t 1)). Daily returns y = (y 1,..., y T ) of the S&P 500 index. ) Bayesian Inference on θ = (k 1, µ 1, σ 1, k 2, β 12, β 2, µ y, φ 1, φ 2. Performance of the pseudo-marginal MH for RW proposal w.r.t σ, standard deviation of log p θ (y) at posterior mean θ. (MCQMC, 19/08/2016) 13 / 31

51 Integrated Autocorrelation Time of Pseudo-Marginal MH Figure: Average over the 9 parameter components of the log-integrated autocorrelation time of pseudo-marginal chain as a function of σ for T = 300. (MCQMC, 19/08/2016) 14 / 31

52 How precise should the log-likelihood estimator be? Aim: Minimize the computational time CT (Q, h) = IACT (Q, h) /σ 2 as σ 2 1/N and computational efforts proportional to N, where IACT (Q, h) = Integrated Autocorrelation Time of {h (ϑ i )} i 1 (MCQMC, 19/08/2016) 15 / 31

53 How precise should the log-likelihood estimator be? Aim: Minimize the computational time CT (Q, h) = IACT (Q, h) /σ 2 as σ 2 1/N and computational efforts proportional to N, where IACT (Q, h) The IACT is = Integrated Autocorrelation Time of {h (ϑ i )} i 1 IACT (Q, h) = τ=1 corr π,q {h (θ 0 ), h (θ τ )} where, for (θ, z) = (ϑ, w), the PM kernel Q is { Q {(θ, z), (dϑ, dw)} = q(ϑ θ)g ϑ (w) 1 π (ϑ) } exp (w z) dϑdw π (θ) with z = log{ p θ (y 1:T )/p θ (y 1:T )}, w = log{ p ϑ (y 1:T )/p ϑ (y 1:T )}. (MCQMC, 19/08/2016) 15 / 31

54 Computational time for the SV model Figure: Computational time as a function of σ (MCQMC, 19/08/2016) 16 / 31

55 Large sample analysis of the pseudo-marginal kernel Standard asymptotic study of MCMC relies on d and independence assumption on the target, interested here in fixed d, large T. (MCQMC, 19/08/2016) 17 / 31

56 Large sample analysis of the pseudo-marginal kernel Standard asymptotic study of MCMC relies on d and independence assumption on the target, interested here in fixed d, large T. Assumption 1 - Asymptotic Normality: We have p ( θ Y1:T ) φ(θ; θ T, Σ/T ) dθ P 0, where θ T P θ and Σ is a p.d. matrix. (MCQMC, 19/08/2016) 17 / 31

57 Large sample analysis of the pseudo-marginal kernel Standard asymptotic study of MCMC relies on d and independence assumption on the target, interested here in fixed d, large T. Assumption 1 - Asymptotic Normality: We have p ( θ Y1:T ) φ(θ; θ T, Σ/T ) dθ P 0, where θ T P θ and Σ is a p.d. matrix. Assumption 2 - Proposal: ϑ = θ + ε/ T where ε υ ( ) with υ (ε) = υ ( ε). (MCQMC, 19/08/2016) 17 / 31

58 Large sample analysis of the pseudo-marginal kernel Standard asymptotic study of MCMC relies on d and independence assumption on the target, interested here in fixed d, large T. Assumption 1 - Asymptotic Normality: We have p ( θ Y1:T ) φ(θ; θ T, Σ/T ) dθ P 0, where θ T P θ and Σ is a p.d. matrix. Assumption 2 - Proposal: ϑ = θ + ε/ T where ε υ ( ) with υ (ε) = υ ( ε). Proposition (Berard, Del Moral & D., EJP, 2014): Under regularity conditions, we have if N T log p θ (Y 1:T )/p θ (Y 1:T ) Y T N ( σ 2 (θ) /2, σ 2 (θ) ) (MCQMC, 19/08/2016) 17 / 31

59 Large sample analysis of the pseudo-marginal kernel Let {ϑ T i, Zi T := log p ϑ T (Y i 1:T )/p ϑ T (Y i 1:T )} i 0 the stationary PM Markov chain targetting π T (θ, z) = p ( θ Y 1:T ) exp (z) gθ T (z). (MCQMC, 19/08/2016) 18 / 31

60 Large sample analysis of the pseudo-marginal kernel Let {ϑ T i, Zi T := log p ϑ T (Y i 1:T )/p ϑ T (Y i 1:T )} i 0 the stationary PM Markov chain targetting π T (θ, z) = p ( θ Y 1:T ) exp (z) gθ T (z). Proposition: The F.D.D. of { ϑ T i = T (ϑ T i θ T ), Zi T } i 0 converge weakly as T ) to those of a stationary Markov chain of invariant density φ ( θ; 0, Σ φ ( z; σ 2 ( θ ) /2, σ 2 ( θ )) and kernel Q{(θ, z), (dϑ, dw)} = υ( ϑ θ)φ ( w; σ 2 ( θ ) /2, σ 2 ( θ )) { 1 φ( ϑ; 0, Σ) exp (w z) φ( θ; 0, Σ) } d ϑdw (MCQMC, 19/08/2016) 18 / 31

61 Large sample analysis of the pseudo-marginal kernel Let {ϑ T i, Zi T := log p ϑ T (Y i 1:T )/p ϑ T (Y i 1:T )} i 0 the stationary PM Markov chain targetting π T (θ, z) = p ( θ Y 1:T ) exp (z) gθ T (z). Proposition: The F.D.D. of { ϑ T i = T (ϑ T i θ T ), Zi T } i 0 converge weakly as T ) to those of a stationary Markov chain of invariant density φ ( θ; 0, Σ φ ( z; σ 2 ( θ ) /2, σ 2 ( θ )) and kernel Q{(θ, z), (dϑ, dw)} = υ( ϑ θ)φ ( w; σ 2 ( θ ) /2, σ 2 ( θ )) { 1 φ( ϑ; 0, Σ) exp (w z) φ( θ; 0, Σ) Let σ 2 = σ 2 ( θ ), it suggests a simplified analysis based on } d ϑdw Q σ {(θ, z), (dϑ, dw)} = q ( ϑ θ) φ ( w; σ 2 /2, σ 2) { 1 π (ϑ) } exp (w z) dϑdw π (θ) (MCQMC, 19/08/2016) 18 / 31

62 Available Results Aim: Minimize the computational cost CT( Q σ, h) = IACT( Q σ, h) σ 2. (MCQMC, 19/08/2016) 19 / 31

63 Available Results Aim: Minimize the computational cost Special cases: CT( Q σ, h) = IACT( Q σ, h) σ 2. (MCQMC, 19/08/2016) 19 / 31

64 Available Results Aim: Minimize the computational cost Special cases: CT( Q σ, h) = IACT( Q σ, h) σ 2. 1 When q(ϑ θ) = p ( ϑ y), σ opt = 0.9 (Pitt et al., 2012). (MCQMC, 19/08/2016) 19 / 31

65 Available Results Aim: Minimize the computational cost Special cases: CT( Q σ, h) = IACT( Q σ, h) σ 2. 1 When q(ϑ θ) = p ( ϑ y), σ opt = 0.9 (Pitt et al., 2012). 2 When π (θ) = d i=1 f (θ i ) and q(ϑ θ) is an isotropic Gaussian random walk then, as d, diffusion limit suggests σ opt = 1.8 (Sherlock et al., 2015). (MCQMC, 19/08/2016) 19 / 31

66 Sketch of the Analysis For general proposals and targets, direct minimization of CT( Q σ, h) impossible so minimize an upper bound over it. (MCQMC, 19/08/2016) 20 / 31

67 Sketch of the Analysis For general proposals and targets, direct minimization of CT( Q σ, h) impossible so minimize an upper bound over it. Theoretical study relies on π invariant kernel Qσ given for (θ, z) = (ϑ, w) by q(ϑ θ)φ ( w; σ 2 /2, σ 2) { min 1, π (ϑ) } min {1, exp (w z)} dϑdw, π (θ) instead of q(ϑ θ)φ ( w; σ 2 /2, σ 2) min { 1, π (ϑ) } exp (w z) dϑdw. π (θ) (MCQMC, 19/08/2016) 20 / 31

68 Sketch of the Analysis For general proposals and targets, direct minimization of CT( Q σ, h) impossible so minimize an upper bound over it. Theoretical study relies on π invariant kernel Qσ given for (θ, z) = (ϑ, w) by q(ϑ θ)φ ( w; σ 2 /2, σ 2) { min 1, π (ϑ) } min {1, exp (w z)} dϑdw, π (θ) instead of q(ϑ θ)φ ( w; σ 2 /2, σ 2) min { 1, π (ϑ) } exp (w z) dϑdw. π (θ) We have IACT( Q σ, h) IACT(Q σ, h) by Peskun (1973) so CT( Q σ, h) CT(Q σ, h) and exact expression of IACT(Q σ, h) is available. (MCQMC, 19/08/2016) 20 / 31

69 Practical Guidelines For good proposals, select σ 1.0 whereas for poor proposals, select σ 1.7. (MCQMC, 19/08/2016) 21 / 31

70 Practical Guidelines For good proposals, select σ 1.0 whereas for poor proposals, select σ 1.7. When you have no clue about the proposal effi ciency, (MCQMC, 19/08/2016) 21 / 31

71 Practical Guidelines For good proposals, select σ 1.0 whereas for poor proposals, select σ 1.7. When you have no clue about the proposal effi ciency, 1 If σ opt = 1.0 and you pick σ = 1.7, computing time increases by 150%. (MCQMC, 19/08/2016) 21 / 31

72 Practical Guidelines For good proposals, select σ 1.0 whereas for poor proposals, select σ 1.7. When you have no clue about the proposal effi ciency, 1 If σ opt = 1.0 and you pick σ = 1.7, computing time increases by 150%. 2 If σ opt = 1.7 and you pick σ = 1.0, computing time increases by 50%. (MCQMC, 19/08/2016) 21 / 31

73 Practical Guidelines For good proposals, select σ 1.0 whereas for poor proposals, select σ 1.7. When you have no clue about the proposal effi ciency, 1 If σ opt = 1.0 and you pick σ = 1.7, computing time increases by 150%. 2 If σ opt = 1.7 and you pick σ = 1.0, computing time increases by 50%. 3 If σ opt = 1.0 or σ opt = 1.7 and you pick σ = , computing time increases by 15%. (MCQMC, 19/08/2016) 21 / 31

74 Discussion Many sharp qualitative results for the pseudo-marginal MH algorithm have been obtained (Andrieu & Roberts, AoS 2009; Andrieu & Vihola, AAP 2015). (MCQMC, 19/08/2016) 22 / 31

75 Discussion Many sharp qualitative results for the pseudo-marginal MH algorithm have been obtained (Andrieu & Roberts, AoS 2009; Andrieu & Vihola, AAP 2015). Simplified quantitative analysis of the pseudo-marginal MH algorithm, useful in large data regime. (MCQMC, 19/08/2016) 22 / 31

76 Discussion Many sharp qualitative results for the pseudo-marginal MH algorithm have been obtained (Andrieu & Roberts, AoS 2009; Andrieu & Vihola, AAP 2015). Simplified quantitative analysis of the pseudo-marginal MH algorithm, useful in large data regime. Optimal σ depends on effi ciency of the ideal MH algorithm but σ is a sweet spot. (MCQMC, 19/08/2016) 22 / 31

77 Discussion Many sharp qualitative results for the pseudo-marginal MH algorithm have been obtained (Andrieu & Roberts, AoS 2009; Andrieu & Vihola, AAP 2015). Simplified quantitative analysis of the pseudo-marginal MH algorithm, useful in large data regime. Optimal σ depends on effi ciency of the ideal MH algorithm but σ is a sweet spot. Pseudo-marginal MH scales in O ( T 2) at each iteration as we require N T. (MCQMC, 19/08/2016) 22 / 31

78 Discussion Many sharp qualitative results for the pseudo-marginal MH algorithm have been obtained (Andrieu & Roberts, AoS 2009; Andrieu & Vihola, AAP 2015). Simplified quantitative analysis of the pseudo-marginal MH algorithm, useful in large data regime. Optimal σ depends on effi ciency of the ideal MH algorithm but σ is a sweet spot. Pseudo-marginal MH scales in O ( T 2) at each iteration as we require N T. Problem: the ratio p ϑ (y 1:T ) /p θ (y 1:T ) is estimated by estimating independently the numerator and denominator. (MCQMC, 19/08/2016) 22 / 31

79 The Correlated Pseudo-Marginal Algorithm Reparameterize the likelihood estimator p θ (y 1:T ) as a function of normal variates U N (0, I ) p θ (y 1:T ) = p θ (y 1:T ; U) (MCQMC, 19/08/2016) 23 / 31

80 The Correlated Pseudo-Marginal Algorithm Reparameterize the likelihood estimator p θ (y 1:T ) as a function of normal variates U N (0, I ) p θ (y 1:T ) = p θ (y 1:T ; U) Correlate estimators of p θ (y 1:T ) and p ϑ (y 1:T ) by setting p ϑ (y 1:T ) = p ϑ (y 1:T ; V ) where V = ρu + 1 ρ 2 ε, ε N (0, I ) for ρ ( 1, 1). (MCQMC, 19/08/2016) 23 / 31

81 The Correlated Pseudo-Marginal Algorithm Reparameterize the likelihood estimator p θ (y 1:T ) as a function of normal variates U N (0, I ) p θ (y 1:T ) = p θ (y 1:T ; U) Correlate estimators of p θ (y 1:T ) and p ϑ (y 1:T ) by setting p ϑ (y 1:T ) = p ϑ (y 1:T ; V ) where V = ρu + 1 ρ 2 ε, ε N (0, I ) for ρ ( 1, 1). In practice, ρ will be selected close to 1. (MCQMC, 19/08/2016) 23 / 31

82 Correlated Pseudo-Marginal Metropolis Hastings algorithm (MCQMC, 19/08/2016) 24 / 31

83 Correlated Pseudo-Marginal Metropolis Hastings algorithm At iteration i Sample ϑ q ( ϑ i 1 ) and V = ρu i ρ 2 ε, ε N (0, I ). (MCQMC, 19/08/2016) 24 / 31

84 Correlated Pseudo-Marginal Metropolis Hastings algorithm At iteration i Sample ϑ q ( ϑ i 1 ) and V = ρu i ρ 2 ε, ε N (0, I ). Compute the estimate p ϑ (y 1:T ; V ) of p ϑ (y 1:T ). (MCQMC, 19/08/2016) 24 / 31

85 Correlated Pseudo-Marginal Metropolis Hastings algorithm At iteration i Sample ϑ q ( ϑ i 1 ) and V = ρu i ρ 2 ε, ε N (0, I ). Compute the estimate p ϑ (y 1:T ; V ) of p ϑ (y 1:T ). With probability min{1, p ϑ (y 1:T ; V ) p (ϑ) q ( ϑ i 1 ϑ) p ϑi 1 (y 1:T ; U i 1 ) p (ϑ i 1 ) q ( ϑ ϑ i 1 ) } set ϑ i = ϑ, U i = V, otherwise set ϑ i = ϑ i 1, U i = U i 1. (MCQMC, 19/08/2016) 24 / 31

86 Large sample analysis of the correlated PM - i.i.d. case log Proposition. Let N T as T with N T = o(t ). When 1 ρ 2 T εt with ρ T = exp U T π( θ) and V T = ρ T U T + then as T ( { pθ+ξ/ T y1:t ; V T ) p θ (y 1:T ; U T ) / p θ+ξ/ T (y } 1:T ) p θ (y 1:T ) ( ) ψ N T T Y T, U T N ( κ2 (θ), κ 2 (θ)) 2 This CLT is conditional on the observation sequence and the current auxiliary variables. (MCQMC, 19/08/2016) 25 / 31

87 Large sample analysis of the correlated PM - i.i.d. case log Proposition. Let N T as T with N T = o(t ). When 1 ρ 2 T εt with ρ T = exp U T π( θ) and V T = ρ T U T + then as T ( { pθ+ξ/ T y1:t ; V T ) p θ (y 1:T ; U T ) / p θ+ξ/ T (y } 1:T ) p θ (y 1:T ) ( ) ψ N T T Y T, U T N ( κ2 (θ), κ 2 (θ)) 2 This CLT is conditional on the observation sequence and the current auxiliary variables. Asymptotically the distribution of the log-ratio decouples from the current location of the Markov chain. (MCQMC, 19/08/2016) 25 / 31

88 Large sample analysis of the correlated PM - i.i.d. case log Proposition. Let N T as T with N T = o(t ). When 1 ρ 2 T εt with ρ T = exp U T π( θ) and V T = ρ T U T + then as T ( { pθ+ξ/ T y1:t ; V T ) p θ (y 1:T ; U T ) / p θ+ξ/ T (y } 1:T ) p θ (y 1:T ) ( ) ψ N T T Y T, U T N ( κ2 (θ), κ 2 (θ)) 2 This CLT is conditional on the observation sequence and the current auxiliary variables. Asymptotically the distribution of the log-ratio decouples from the current location of the Markov chain. The asymptotic variance is O (1) even for N log(t ). (MCQMC, 19/08/2016) 25 / 31

89 Large sample analysis of the correlated PM - i.i.d. case Let {ϑ T i } i 0 the stationary non-markovian sequence of the correlated PM of invariant density p ( θ Y 1:T ). (MCQMC, 19/08/2016) 26 / 31

90 Large sample analysis of the correlated PM - i.i.d. case Let {ϑ T i } i 0 the stationary non-markovian sequence of the correlated PM of invariant density p ( θ Y 1:T ). Proposition. The F.D.D. of { ϑ T i = T (ϑ T i θ T )} i 0 converge weakly as T ) to those of a stationary Markov chain of invariant density φ ( θ; 0, Σ and kernel given for θ = ϑ by [ { }] Q( θ, d ϑ) = υ( ϑ θ)e R N ( κ2 min 1, φ( ϑ; 0, Σ) (θ)/2,κ 2 (θ)) φ( θ; 0, Σ) R d ϑ. (MCQMC, 19/08/2016) 26 / 31

91 Large sample analysis of the correlated PM - i.i.d. case Let {ϑ T i } i 0 the stationary non-markovian sequence of the correlated PM of invariant density p ( θ Y 1:T ). Proposition. The F.D.D. of { ϑ T i = T (ϑ T i θ T )} i 0 converge weakly as T ) to those of a stationary Markov chain of invariant density φ ( θ; 0, Σ and kernel given for θ = ϑ by [ { }] Q( θ, d ϑ) = υ( ϑ θ)e R N ( κ2 min 1, φ( ϑ; 0, Σ) (θ)/2,κ 2 (θ)) φ( θ; 0, Σ) R d ϑ. These results suggests that a simplified analysis of the CPM chain can be performed by looking at [ { Q(θ, dϑ) = q ( ϑ θ) E R N ( κ2 min 1, π (ϑ) }] (θ)/2,κ 2 (θ)) π (θ) R dϑ and κ 2.1 minimizes computational time. (MCQMC, 19/08/2016) 26 / 31

92 Limitations of Weak Convergence Previous result does not imply convergence of IACT. (MCQMC, 19/08/2016) 27 / 31

93 Limitations of Weak Convergence Previous result does not imply convergence of IACT. In the CPM context, the auxiliary variables U T evolves according to an AR scheme with persistency 1 N T /T. (MCQMC, 19/08/2016) 27 / 31

94 Limitations of Weak Convergence Previous result does not imply convergence of IACT. In the CPM context, the auxiliary variables U T evolves according to an AR scheme with persistency 1 N T /T. When N T grows to slowly with T, correlations decay too slowy leading to increasing IACT. (MCQMC, 19/08/2016) 27 / 31

95 Limitations of Weak Convergence Previous result does not imply convergence of IACT. In the CPM context, the auxiliary variables U T evolves according to an AR scheme with persistency 1 N T /T. When N T grows to slowly with T, correlations decay too slowy leading to increasing IACT. For the i.i.d. latent variable case, IACT(Q, h) is driven by the IACT of { ( ) Ψ T U T i := θ log p ( Y 1:T ; Ui T )} i 1. (MCQMC, 19/08/2016) 27 / 31

96 Limitations of Weak Convergence Previous result does not imply convergence of IACT. In the CPM context, the auxiliary variables U T evolves according to an AR scheme with persistency 1 N T /T. When N T grows to slowly with T, correlations decay too slowy leading to increasing IACT. For the i.i.d. latent variable case, IACT(Q, h) is driven by the IACT of { ( ) Ψ T U T i := θ log p ( Y 1:T ; Ui T )} i 1. Proposition. Let N T T α for 0 < α < 1 then IACT(Q, h) T 1 2α. (MCQMC, 19/08/2016) 27 / 31

97 Limitations of Weak Convergence Previous result does not imply convergence of IACT. In the CPM context, the auxiliary variables U T evolves according to an AR scheme with persistency 1 N T /T. When N T grows to slowly with T, correlations decay too slowy leading to increasing IACT. For the i.i.d. latent variable case, IACT(Q, h) is driven by the IACT of { ( ) Ψ T U T i := θ log p ( Y 1:T ; Ui T )} i 1. Proposition. Let N T T α for 0 < α < 1 then IACT(Q, h) T 1 2α. This result suggests we need at least T /N T = O (1). (MCQMC, 19/08/2016) 27 / 31

98 Example: Gaussian Latent Variable Model MH (T = 8192) IACT(θ) 15.6 PM (ρ = 0.0) N RIACT(θ) RCT(θ) CPM (ρ = ) N κ RIACT(θ) RCT(θ) Here RIACT = IACT/IACT MH and RCT = N RIACT. Improvement by 180 fold. (MCQMC, 19/08/2016) 28 / 31

99 Example: Noisy Autoregressive Model MH (T = 16, 000) IACT(θ) 5.8 PM (ρ = 0.0) N RIACT(θ) RCT(θ) CPM (ρ = ) N κ RIACT(θ) RCT(θ) Improvement by 100 fold. (MCQMC, 19/08/2016) 29 / 31

100 Discussion Large sample analysis of pseudo-marginal algorithm provides useful guidelines, overall complexity is O ( T 2). (MCQMC, 19/08/2016) 30 / 31

101 Discussion Large sample analysis of pseudo-marginal algorithm provides useful guidelines, overall complexity is O ( T 2). Correlated pseudo-marginal can achieve very substantial improvement. (MCQMC, 19/08/2016) 30 / 31

102 Discussion Large sample analysis of pseudo-marginal algorithm provides useful guidelines, overall complexity is O ( T 2). Correlated pseudo-marginal can achieve very substantial improvement. Analysis shows T /N T = O(1) is necessary and we conjecture it is suffi cient leading to complexity O(T 3/2 ) vs O ( T 2). (MCQMC, 19/08/2016) 30 / 31

103 Discussion Large sample analysis of pseudo-marginal algorithm provides useful guidelines, overall complexity is O ( T 2). Correlated pseudo-marginal can achieve very substantial improvement. Analysis shows T /N T = O(1) is necessary and we conjecture it is suffi cient leading to complexity O(T 3/2 ) vs O ( T 2). Implementation for state-space models in state dimension n > 1 relies on non-standard particle scheme (Gerber & Chopin, 2015): our analysis does not capture these cases, experimental results suggest O(T 1+ n+1 n ). (MCQMC, 19/08/2016) 30 / 31

104 Some References C. Andrieu & M Vihola, Convergence properties of pseudo-margninal MCMC, Ann. Applied Proba., J. Berard, P. Del Moral & A.D., A Lognormal CLT for Particle Approximations of Normalizing Constants, Elec. J. Proba., G. Deligiannidis, A.D. and M.K. Pitt, The Correlated Pseudo-marginal Method, arxiv: , A.D., M.K. Pitt, G. Deligiannidis and R. Kohn, Effi cient Implementation of Markov Chain Monte Carlo when Using an Unbiased Likelihood Estimator, Biometrika, L. Lin, K. Lin & J. Sloan, A Noisy Monte Carlo Algorithm, Phys. Rev. D, C. Sherlock, A. Thiery, G.O. Roberts & J.S. Rosenthal, On the Effi ciency of the RW Pseudo-Marginal MH, Ann. Stat., (MCQMC, 19/08/2016) 31 / 31

Controlled sequential Monte Carlo

Controlled sequential Monte Carlo Controlled sequential Monte Carlo Jeremy Heng, Department of Statistics, Harvard University Joint work with Adrian Bishop (UTS, CSIRO), George Deligiannidis & Arnaud Doucet (Oxford) Bayesian Computation

More information

Pseudo-marginal Metropolis-Hastings: a simple explanation and (partial) review of theory

Pseudo-marginal Metropolis-Hastings: a simple explanation and (partial) review of theory Pseudo-arginal Metropolis-Hastings: a siple explanation and (partial) review of theory Chris Sherlock Motivation Iagine a stochastic process V which arises fro soe distribution with density p(v θ ). Iagine

More information

A Review of Pseudo-Marginal Markov Chain Monte Carlo

A Review of Pseudo-Marginal Markov Chain Monte Carlo A Review of Pseudo-Marginal Markov Chain Monte Carlo Discussed by: Yizhe Zhang October 21, 2016 Outline 1 Overview 2 Paper review 3 experiment 4 conclusion Motivation & overview Notation: θ denotes the

More information

Inference in state-space models with multiple paths from conditional SMC

Inference in state-space models with multiple paths from conditional SMC Inference in state-space models with multiple paths from conditional SMC Sinan Yıldırım (Sabancı) joint work with Christophe Andrieu (Bristol), Arnaud Doucet (Oxford) and Nicolas Chopin (ENSAE) September

More information

Kernel Sequential Monte Carlo

Kernel Sequential Monte Carlo Kernel Sequential Monte Carlo Ingmar Schuster (Paris Dauphine) Heiko Strathmann (University College London) Brooks Paige (Oxford) Dino Sejdinovic (Oxford) * equal contribution April 25, 2016 1 / 37 Section

More information

Kernel adaptive Sequential Monte Carlo

Kernel adaptive Sequential Monte Carlo Kernel adaptive Sequential Monte Carlo Ingmar Schuster (Paris Dauphine) Heiko Strathmann (University College London) Brooks Paige (Oxford) Dino Sejdinovic (Oxford) December 7, 2015 1 / 36 Section 1 Outline

More information

Sequential Monte Carlo Samplers for Applications in High Dimensions

Sequential Monte Carlo Samplers for Applications in High Dimensions Sequential Monte Carlo Samplers for Applications in High Dimensions Alexandros Beskos National University of Singapore KAUST, 26th February 2014 Joint work with: Dan Crisan, Ajay Jasra, Nik Kantas, Alex

More information

PSEUDO-MARGINAL METROPOLIS-HASTINGS APPROACH AND ITS APPLICATION TO BAYESIAN COPULA MODEL

PSEUDO-MARGINAL METROPOLIS-HASTINGS APPROACH AND ITS APPLICATION TO BAYESIAN COPULA MODEL PSEUDO-MARGINAL METROPOLIS-HASTINGS APPROACH AND ITS APPLICATION TO BAYESIAN COPULA MODEL Xuebin Zheng Supervisor: Associate Professor Josef Dick Co-Supervisor: Dr. David Gunawan School of Mathematics

More information

A Note on Auxiliary Particle Filters

A Note on Auxiliary Particle Filters A Note on Auxiliary Particle Filters Adam M. Johansen a,, Arnaud Doucet b a Department of Mathematics, University of Bristol, UK b Departments of Statistics & Computer Science, University of British Columbia,

More information

Bayesian Estimation of DSGE Models 1 Chapter 3: A Crash Course in Bayesian Inference

Bayesian Estimation of DSGE Models 1 Chapter 3: A Crash Course in Bayesian Inference 1 The views expressed in this paper are those of the authors and do not necessarily reflect the views of the Federal Reserve Board of Governors or the Federal Reserve System. Bayesian Estimation of DSGE

More information

Likelihood-free MCMC

Likelihood-free MCMC Bayesian inference for stable distributions with applications in finance Department of Mathematics University of Leicester September 2, 2011 MSc project final presentation Outline 1 2 3 4 Classical Monte

More information

List of projects. FMS020F NAMS002 Statistical inference for partially observed stochastic processes, 2016

List of projects. FMS020F NAMS002 Statistical inference for partially observed stochastic processes, 2016 List of projects FMS020F NAMS002 Statistical inference for partially observed stochastic processes, 206 Work in groups of two (if this is absolutely not possible for some reason, please let the lecturers

More information

Flexible Particle Markov chain Monte Carlo methods with an application to a factor stochastic volatility model

Flexible Particle Markov chain Monte Carlo methods with an application to a factor stochastic volatility model Flexible Particle Markov chain Monte Carlo methods with an application to a factor stochastic volatility model arxiv:1401.1667v4 [stat.co] 23 Jan 2018 Eduardo F. Mendes School of Applied Mathematics Fundação

More information

SMC 2 : an efficient algorithm for sequential analysis of state-space models

SMC 2 : an efficient algorithm for sequential analysis of state-space models SMC 2 : an efficient algorithm for sequential analysis of state-space models N. CHOPIN 1, P.E. JACOB 2, & O. PAPASPILIOPOULOS 3 1 ENSAE-CREST 2 CREST & Université Paris Dauphine, 3 Universitat Pompeu Fabra

More information

Adaptive Monte Carlo methods

Adaptive Monte Carlo methods Adaptive Monte Carlo methods Jean-Michel Marin Projet Select, INRIA Futurs, Université Paris-Sud joint with Randal Douc (École Polytechnique), Arnaud Guillin (Université de Marseille) and Christian Robert

More information

One Pseudo-Sample is Enough in Approximate Bayesian Computation MCMC

One Pseudo-Sample is Enough in Approximate Bayesian Computation MCMC Biometrika (?), 99, 1, pp. 1 10 Printed in Great Britain Submitted to Biometrika One Pseudo-Sample is Enough in Approximate Bayesian Computation MCMC BY LUKE BORNN, NATESH PILLAI Department of Statistics,

More information

Infinite-State Markov-switching for Dynamic. Volatility Models : Web Appendix

Infinite-State Markov-switching for Dynamic. Volatility Models : Web Appendix Infinite-State Markov-switching for Dynamic Volatility Models : Web Appendix Arnaud Dufays 1 Centre de Recherche en Economie et Statistique March 19, 2014 1 Comparison of the two MS-GARCH approximations

More information

Markov Chain Monte Carlo methods

Markov Chain Monte Carlo methods Markov Chain Monte Carlo methods Tomas McKelvey and Lennart Svensson Signal Processing Group Department of Signals and Systems Chalmers University of Technology, Sweden November 26, 2012 Today s learning

More information

MCMC for big data. Geir Storvik. BigInsight lunch - May Geir Storvik MCMC for big data BigInsight lunch - May / 17

MCMC for big data. Geir Storvik. BigInsight lunch - May Geir Storvik MCMC for big data BigInsight lunch - May / 17 MCMC for big data Geir Storvik BigInsight lunch - May 2 2018 Geir Storvik MCMC for big data BigInsight lunch - May 2 2018 1 / 17 Outline Why ordinary MCMC is not scalable Different approaches for making

More information

Exercises Tutorial at ICASSP 2016 Learning Nonlinear Dynamical Models Using Particle Filters

Exercises Tutorial at ICASSP 2016 Learning Nonlinear Dynamical Models Using Particle Filters Exercises Tutorial at ICASSP 216 Learning Nonlinear Dynamical Models Using Particle Filters Andreas Svensson, Johan Dahlin and Thomas B. Schön March 18, 216 Good luck! 1 [Bootstrap particle filter for

More information

Particle Metropolis-adjusted Langevin algorithms

Particle Metropolis-adjusted Langevin algorithms Particle Metropolis-adjusted Langevin algorithms Christopher Nemeth, Chris Sherlock and Paul Fearnhead arxiv:1412.7299v3 [stat.me] 27 May 2016 Department of Mathematics and Statistics, Lancaster University,

More information

Riemann Manifold Methods in Bayesian Statistics

Riemann Manifold Methods in Bayesian Statistics Ricardo Ehlers ehlers@icmc.usp.br Applied Maths and Stats University of São Paulo, Brazil Working Group in Statistical Learning University College Dublin September 2015 Bayesian inference is based on Bayes

More information

Lecture 7 and 8: Markov Chain Monte Carlo

Lecture 7 and 8: Markov Chain Monte Carlo Lecture 7 and 8: Markov Chain Monte Carlo 4F13: Machine Learning Zoubin Ghahramani and Carl Edward Rasmussen Department of Engineering University of Cambridge http://mlg.eng.cam.ac.uk/teaching/4f13/ Ghahramani

More information

Bayesian Inference for DSGE Models. Lawrence J. Christiano

Bayesian Inference for DSGE Models. Lawrence J. Christiano Bayesian Inference for DSGE Models Lawrence J. Christiano Outline State space-observer form. convenient for model estimation and many other things. Bayesian inference Bayes rule. Monte Carlo integation.

More information

Sequential Monte Carlo Methods for Bayesian Computation

Sequential Monte Carlo Methods for Bayesian Computation Sequential Monte Carlo Methods for Bayesian Computation A. Doucet Kyoto Sept. 2012 A. Doucet (MLSS Sept. 2012) Sept. 2012 1 / 136 Motivating Example 1: Generic Bayesian Model Let X be a vector parameter

More information

arxiv: v5 [stat.co] 16 Feb 2016

arxiv: v5 [stat.co] 16 Feb 2016 The Use of a Single Pseudo-Sample in Approximate Bayesian Computation Luke Bornn 1,2, Natesh S. Pillai 2, Aaron Smith 3, and Dawn Woodard 4 arxiv:1404.6298v5 [stat.co] 16 Feb 2016 1 Department of Statistics

More information

Markov Chain Monte Carlo Methods

Markov Chain Monte Carlo Methods Markov Chain Monte Carlo Methods John Geweke University of Iowa, USA 2005 Institute on Computational Economics University of Chicago - Argonne National Laboaratories July 22, 2005 The problem p (θ, ω I)

More information

On some properties of Markov chain Monte Carlo simulation methods based on the particle filter

On some properties of Markov chain Monte Carlo simulation methods based on the particle filter On some properties of Markov chain Monte Carlo simulation methods based on the particle filter Michael K. Pitt Economics Department University of Warwick m.pitt@warwick.ac.uk Ralph S. Silva School of Economics

More information

Computational statistics

Computational statistics Computational statistics Markov Chain Monte Carlo methods Thierry Denœux March 2017 Thierry Denœux Computational statistics March 2017 1 / 71 Contents of this chapter When a target density f can be evaluated

More information

Sequential Monte Carlo Methods in High Dimensions

Sequential Monte Carlo Methods in High Dimensions Sequential Monte Carlo Methods in High Dimensions Alexandros Beskos Statistical Science, UCL Oxford, 24th September 2012 Joint work with: Dan Crisan, Ajay Jasra, Nik Kantas, Andrew Stuart Imperial College,

More information

SUPPLEMENT TO MARKET ENTRY COSTS, PRODUCER HETEROGENEITY, AND EXPORT DYNAMICS (Econometrica, Vol. 75, No. 3, May 2007, )

SUPPLEMENT TO MARKET ENTRY COSTS, PRODUCER HETEROGENEITY, AND EXPORT DYNAMICS (Econometrica, Vol. 75, No. 3, May 2007, ) Econometrica Supplementary Material SUPPLEMENT TO MARKET ENTRY COSTS, PRODUCER HETEROGENEITY, AND EXPORT DYNAMICS (Econometrica, Vol. 75, No. 3, May 2007, 653 710) BY SANGHAMITRA DAS, MARK ROBERTS, AND

More information

Weak convergence of Markov chain Monte Carlo II

Weak convergence of Markov chain Monte Carlo II Weak convergence of Markov chain Monte Carlo II KAMATANI, Kengo Mar 2011 at Le Mans Background Markov chain Monte Carlo (MCMC) method is widely used in Statistical Science. It is easy to use, but difficult

More information

arxiv: v1 [stat.co] 1 Jun 2015

arxiv: v1 [stat.co] 1 Jun 2015 arxiv:1506.00570v1 [stat.co] 1 Jun 2015 Towards automatic calibration of the number of state particles within the SMC 2 algorithm N. Chopin J. Ridgway M. Gerber O. Papaspiliopoulos CREST-ENSAE, Malakoff,

More information

Control Variates for Markov Chain Monte Carlo

Control Variates for Markov Chain Monte Carlo Control Variates for Markov Chain Monte Carlo Dellaportas, P., Kontoyiannis, I., and Tsourti, Z. Dept of Statistics, AUEB Dept of Informatics, AUEB 1st Greek Stochastics Meeting Monte Carlo: Probability

More information

Paul Karapanagiotidis ECO4060

Paul Karapanagiotidis ECO4060 Paul Karapanagiotidis ECO4060 The way forward 1) Motivate why Markov-Chain Monte Carlo (MCMC) is useful for econometric modeling 2) Introduce Markov-Chain Monte Carlo (MCMC) - Metropolis-Hastings (MH)

More information

Pattern Recognition and Machine Learning. Bishop Chapter 11: Sampling Methods

Pattern Recognition and Machine Learning. Bishop Chapter 11: Sampling Methods Pattern Recognition and Machine Learning Chapter 11: Sampling Methods Elise Arnaud Jakob Verbeek May 22, 2008 Outline of the chapter 11.1 Basic Sampling Algorithms 11.2 Markov Chain Monte Carlo 11.3 Gibbs

More information

Inference for partially observed stochastic dynamic systems: A new algorithm, its theory and applications

Inference for partially observed stochastic dynamic systems: A new algorithm, its theory and applications Inference for partially observed stochastic dynamic systems: A new algorithm, its theory and applications Edward Ionides Department of Statistics, University of Michigan ionides@umich.edu Statistics Department

More information

17 : Markov Chain Monte Carlo

17 : Markov Chain Monte Carlo 10-708: Probabilistic Graphical Models, Spring 2015 17 : Markov Chain Monte Carlo Lecturer: Eric P. Xing Scribes: Heran Lin, Bin Deng, Yun Huang 1 Review of Monte Carlo Methods 1.1 Overview Monte Carlo

More information

Bayesian Inference for DSGE Models. Lawrence J. Christiano

Bayesian Inference for DSGE Models. Lawrence J. Christiano Bayesian Inference for DSGE Models Lawrence J. Christiano Outline State space-observer form. convenient for model estimation and many other things. Preliminaries. Probabilities. Maximum Likelihood. Bayesian

More information

Markov Chain Monte Carlo (MCMC)

Markov Chain Monte Carlo (MCMC) Markov Chain Monte Carlo (MCMC Dependent Sampling Suppose we wish to sample from a density π, and we can evaluate π as a function but have no means to directly generate a sample. Rejection sampling can

More information

Stat 451 Lecture Notes Markov Chain Monte Carlo. Ryan Martin UIC

Stat 451 Lecture Notes Markov Chain Monte Carlo. Ryan Martin UIC Stat 451 Lecture Notes 07 12 Markov Chain Monte Carlo Ryan Martin UIC www.math.uic.edu/~rgmartin 1 Based on Chapters 8 9 in Givens & Hoeting, Chapters 25 27 in Lange 2 Updated: April 4, 2016 1 / 42 Outline

More information

A new iterated filtering algorithm

A new iterated filtering algorithm A new iterated filtering algorithm Edward Ionides University of Michigan, Ann Arbor ionides@umich.edu Statistics and Nonlinear Dynamics in Biology and Medicine Thursday July 31, 2014 Overview 1 Introduction

More information

Monte Carlo in Bayesian Statistics

Monte Carlo in Bayesian Statistics Monte Carlo in Bayesian Statistics Matthew Thomas SAMBa - University of Bath m.l.thomas@bath.ac.uk December 4, 2014 Matthew Thomas (SAMBa) Monte Carlo in Bayesian Statistics December 4, 2014 1 / 16 Overview

More information

Notes on pseudo-marginal methods, variational Bayes and ABC

Notes on pseudo-marginal methods, variational Bayes and ABC Notes on pseudo-marginal methods, variational Bayes and ABC Christian Andersson Naesseth October 3, 2016 The Pseudo-Marginal Framework Assume we are interested in sampling from the posterior distribution

More information

Monte Carlo methods for sampling-based Stochastic Optimization

Monte Carlo methods for sampling-based Stochastic Optimization Monte Carlo methods for sampling-based Stochastic Optimization Gersende FORT LTCI CNRS & Telecom ParisTech Paris, France Joint works with B. Jourdain, T. Lelièvre, G. Stoltz from ENPC and E. Kuhn from

More information

Online appendix to On the stability of the excess sensitivity of aggregate consumption growth in the US

Online appendix to On the stability of the excess sensitivity of aggregate consumption growth in the US Online appendix to On the stability of the excess sensitivity of aggregate consumption growth in the US Gerdie Everaert 1, Lorenzo Pozzi 2, and Ruben Schoonackers 3 1 Ghent University & SHERPPA 2 Erasmus

More information

Data-driven Particle Filters for Particle Markov Chain Monte Carlo

Data-driven Particle Filters for Particle Markov Chain Monte Carlo Data-driven Particle Filters for Particle Markov Chain Monte Carlo Patrick Leung, Catherine S. Forbes, Gael M. Martin and Brendan McCabe September 13, 2016 Abstract This paper proposes new automated proposal

More information

Stat 535 C - Statistical Computing & Monte Carlo Methods. Lecture 18-16th March Arnaud Doucet

Stat 535 C - Statistical Computing & Monte Carlo Methods. Lecture 18-16th March Arnaud Doucet Stat 535 C - Statistical Computing & Monte Carlo Methods Lecture 18-16th March 2006 Arnaud Doucet Email: arnaud@cs.ubc.ca 1 1.1 Outline Trans-dimensional Markov chain Monte Carlo. Bayesian model for autoregressions.

More information

Sequential Monte Carlo methods for system identification

Sequential Monte Carlo methods for system identification Technical report arxiv:1503.06058v3 [stat.co] 10 Mar 2016 Sequential Monte Carlo methods for system identification Thomas B. Schön, Fredrik Lindsten, Johan Dahlin, Johan Wågberg, Christian A. Naesseth,

More information

Lecture 8: The Metropolis-Hastings Algorithm

Lecture 8: The Metropolis-Hastings Algorithm 30.10.2008 What we have seen last time: Gibbs sampler Key idea: Generate a Markov chain by updating the component of (X 1,..., X p ) in turn by drawing from the full conditionals: X (t) j Two drawbacks:

More information

MCMC Sampling for Bayesian Inference using L1-type Priors

MCMC Sampling for Bayesian Inference using L1-type Priors MÜNSTER MCMC Sampling for Bayesian Inference using L1-type Priors (what I do whenever the ill-posedness of EEG/MEG is just not frustrating enough!) AG Imaging Seminar Felix Lucka 26.06.2012 , MÜNSTER Sampling

More information

Introduction to Machine Learning CMU-10701

Introduction to Machine Learning CMU-10701 Introduction to Machine Learning CMU-10701 Markov Chain Monte Carlo Methods Barnabás Póczos & Aarti Singh Contents Markov Chain Monte Carlo Methods Goal & Motivation Sampling Rejection Importance Markov

More information

Stat 535 C - Statistical Computing & Monte Carlo Methods. Lecture February Arnaud Doucet

Stat 535 C - Statistical Computing & Monte Carlo Methods. Lecture February Arnaud Doucet Stat 535 C - Statistical Computing & Monte Carlo Methods Lecture 13-28 February 2006 Arnaud Doucet Email: arnaud@cs.ubc.ca 1 1.1 Outline Limitations of Gibbs sampling. Metropolis-Hastings algorithm. Proof

More information

Advanced Monte Carlo integration methods. P. Del Moral (INRIA team ALEA) INRIA & Bordeaux Mathematical Institute & X CMAP

Advanced Monte Carlo integration methods. P. Del Moral (INRIA team ALEA) INRIA & Bordeaux Mathematical Institute & X CMAP Advanced Monte Carlo integration methods P. Del Moral (INRIA team ALEA) INRIA & Bordeaux Mathematical Institute & X CMAP MCQMC 2012, Sydney, Sunday Tutorial 12-th 2012 Some hyper-refs Feynman-Kac formulae,

More information

Sequential Monte Carlo Methods

Sequential Monte Carlo Methods University of Pennsylvania Bradley Visitor Lectures October 23, 2017 Introduction Unfortunately, standard MCMC can be inaccurate, especially in medium and large-scale DSGE models: disentangling importance

More information

Particle Metropolis-Hastings using gradient and Hessian information

Particle Metropolis-Hastings using gradient and Hessian information Particle Metropolis-Hastings using gradient and Hessian information Johan Dahlin, Fredrik Lindsten and Thomas B. Schön September 19, 2014 Abstract Particle Metropolis-Hastings (PMH) allows for Bayesian

More information

Introduction to Rare Event Simulation

Introduction to Rare Event Simulation Introduction to Rare Event Simulation Brown University: Summer School on Rare Event Simulation Jose Blanchet Columbia University. Department of Statistics, Department of IEOR. Blanchet (Columbia) 1 / 31

More information

MONTE CARLO METHODS. Hedibert Freitas Lopes

MONTE CARLO METHODS. Hedibert Freitas Lopes MONTE CARLO METHODS Hedibert Freitas Lopes The University of Chicago Booth School of Business 5807 South Woodlawn Avenue, Chicago, IL 60637 http://faculty.chicagobooth.edu/hedibert.lopes hlopes@chicagobooth.edu

More information

The University of Auckland Applied Mathematics Bayesian Methods for Inverse Problems : why and how Colin Fox Tiangang Cui, Mike O Sullivan (Auckland),

The University of Auckland Applied Mathematics Bayesian Methods for Inverse Problems : why and how Colin Fox Tiangang Cui, Mike O Sullivan (Auckland), The University of Auckland Applied Mathematics Bayesian Methods for Inverse Problems : why and how Colin Fox Tiangang Cui, Mike O Sullivan (Auckland), Geoff Nicholls (Statistics, Oxford) fox@math.auckland.ac.nz

More information

Hastings-within-Gibbs Algorithm: Introduction and Application on Hierarchical Model

Hastings-within-Gibbs Algorithm: Introduction and Application on Hierarchical Model UNIVERSITY OF TEXAS AT SAN ANTONIO Hastings-within-Gibbs Algorithm: Introduction and Application on Hierarchical Model Liang Jing April 2010 1 1 ABSTRACT In this paper, common MCMC algorithms are introduced

More information

University of Toronto Department of Statistics

University of Toronto Department of Statistics Norm Comparisons for Data Augmentation by James P. Hobert Department of Statistics University of Florida and Jeffrey S. Rosenthal Department of Statistics University of Toronto Technical Report No. 0704

More information

arxiv: v5 [stat.co] 10 Apr 2018

arxiv: v5 [stat.co] 10 Apr 2018 THE BLOCK-POISSON ESTIMATOR FOR OPTIMALLY TUNED EXACT SUBSAMPLING MCMC MATIAS QUIROZ 1,2, MINH-NGOC TRAN 3, MATTIAS VILLANI 4, ROBERT KOHN 1 AND KHUE-DUNG DANG 1 arxiv:1603.08232v5 [stat.co] 10 Apr 2018

More information

The Metropolis-Hastings Algorithm. June 8, 2012

The Metropolis-Hastings Algorithm. June 8, 2012 The Metropolis-Hastings Algorithm June 8, 22 The Plan. Understand what a simulated distribution is 2. Understand why the Metropolis-Hastings algorithm works 3. Learn how to apply the Metropolis-Hastings

More information

Calibration of Stochastic Volatility Models using Particle Markov Chain Monte Carlo Methods

Calibration of Stochastic Volatility Models using Particle Markov Chain Monte Carlo Methods Calibration of Stochastic Volatility Models using Particle Markov Chain Monte Carlo Methods Jonas Hallgren 1 1 Department of Mathematics KTH Royal Institute of Technology Stockholm, Sweden BFS 2012 June

More information

STAT 425: Introduction to Bayesian Analysis

STAT 425: Introduction to Bayesian Analysis STAT 425: Introduction to Bayesian Analysis Marina Vannucci Rice University, USA Fall 2017 Marina Vannucci (Rice University, USA) Bayesian Analysis (Part 2) Fall 2017 1 / 19 Part 2: Markov chain Monte

More information

General Construction of Irreversible Kernel in Markov Chain Monte Carlo

General Construction of Irreversible Kernel in Markov Chain Monte Carlo General Construction of Irreversible Kernel in Markov Chain Monte Carlo Metropolis heat bath Suwa Todo Department of Applied Physics, The University of Tokyo Department of Physics, Boston University (from

More information

Tutorial on ABC Algorithms

Tutorial on ABC Algorithms Tutorial on ABC Algorithms Dr Chris Drovandi Queensland University of Technology, Australia c.drovandi@qut.edu.au July 3, 2014 Notation Model parameter θ with prior π(θ) Likelihood is f(ý θ) with observed

More information

Stat 535 C - Statistical Computing & Monte Carlo Methods. Lecture 15-7th March Arnaud Doucet

Stat 535 C - Statistical Computing & Monte Carlo Methods. Lecture 15-7th March Arnaud Doucet Stat 535 C - Statistical Computing & Monte Carlo Methods Lecture 15-7th March 2006 Arnaud Doucet Email: arnaud@cs.ubc.ca 1 1.1 Outline Mixture and composition of kernels. Hybrid algorithms. Examples Overview

More information

Marginal Specifications and a Gaussian Copula Estimation

Marginal Specifications and a Gaussian Copula Estimation Marginal Specifications and a Gaussian Copula Estimation Kazim Azam Abstract Multivariate analysis involving random variables of different type like count, continuous or mixture of both is frequently required

More information

Session 3A: Markov chain Monte Carlo (MCMC)

Session 3A: Markov chain Monte Carlo (MCMC) Session 3A: Markov chain Monte Carlo (MCMC) John Geweke Bayesian Econometrics and its Applications August 15, 2012 ohn Geweke Bayesian Econometrics and its Session Applications 3A: Markov () chain Monte

More information

An introduction to Sequential Monte Carlo

An introduction to Sequential Monte Carlo An introduction to Sequential Monte Carlo Thang Bui Jes Frellsen Department of Engineering University of Cambridge Research and Communication Club 6 February 2014 1 Sequential Monte Carlo (SMC) methods

More information

Zig-Zag Monte Carlo. Delft University of Technology. Joris Bierkens February 7, 2017

Zig-Zag Monte Carlo. Delft University of Technology. Joris Bierkens February 7, 2017 Zig-Zag Monte Carlo Delft University of Technology Joris Bierkens February 7, 2017 Joris Bierkens (TU Delft) Zig-Zag Monte Carlo February 7, 2017 1 / 33 Acknowledgements Collaborators Andrew Duncan Paul

More information

Data-driven Particle Filters for Particle Markov Chain Monte Carlo

Data-driven Particle Filters for Particle Markov Chain Monte Carlo Data-driven Particle Filters for Particle Markov Chain Monte Carlo Patrick Leung, Catherine S. Forbes, Gael M. Martin and Brendan McCabe December 21, 2016 Abstract This paper proposes new automated proposal

More information

SC7/SM6 Bayes Methods HT18 Lecturer: Geoff Nicholls Lecture 2: Monte Carlo Methods Notes and Problem sheets are available at http://www.stats.ox.ac.uk/~nicholls/bayesmethods/ and via the MSc weblearn pages.

More information

Sequential Monte Carlo Methods (for DSGE Models)

Sequential Monte Carlo Methods (for DSGE Models) Sequential Monte Carlo Methods (for DSGE Models) Frank Schorfheide University of Pennsylvania, PIER, CEPR, and NBER October 23, 2017 Some References These lectures use material from our joint work: Tempered

More information

Practical unbiased Monte Carlo for Uncertainty Quantification

Practical unbiased Monte Carlo for Uncertainty Quantification Practical unbiased Monte Carlo for Uncertainty Quantification Sergios Agapiou Department of Statistics, University of Warwick MiR@W day: Uncertainty in Complex Computer Models, 2nd February 2015, University

More information

Inexact approximations for doubly and triply intractable problems

Inexact approximations for doubly and triply intractable problems Inexact approximations for doubly and triply intractable problems March 27th, 2014 Markov random fields Interacting objects Markov random fields (MRFs) are used for modelling (often large numbers of) interacting

More information

Markov Chain Monte Carlo, Numerical Integration

Markov Chain Monte Carlo, Numerical Integration Markov Chain Monte Carlo, Numerical Integration (See Statistics) Trevor Gallen Fall 2015 1 / 1 Agenda Numerical Integration: MCMC methods Estimating Markov Chains Estimating latent variables 2 / 1 Numerical

More information

Approximate Bayesian Computation: a simulation based approach to inference

Approximate Bayesian Computation: a simulation based approach to inference Approximate Bayesian Computation: a simulation based approach to inference Richard Wilkinson Simon Tavaré 2 Department of Probability and Statistics University of Sheffield 2 Department of Applied Mathematics

More information

Introduction to Markov Chain Monte Carlo & Gibbs Sampling

Introduction to Markov Chain Monte Carlo & Gibbs Sampling Introduction to Markov Chain Monte Carlo & Gibbs Sampling Prof. Nicholas Zabaras Sibley School of Mechanical and Aerospace Engineering 101 Frank H. T. Rhodes Hall Ithaca, NY 14853-3801 Email: zabaras@cornell.edu

More information

16 : Approximate Inference: Markov Chain Monte Carlo

16 : Approximate Inference: Markov Chain Monte Carlo 10-708: Probabilistic Graphical Models 10-708, Spring 2017 16 : Approximate Inference: Markov Chain Monte Carlo Lecturer: Eric P. Xing Scribes: Yuan Yang, Chao-Ming Yen 1 Introduction As the target distribution

More information

GMM, HAC estimators, & Standard Errors for Business Cycle Statistics

GMM, HAC estimators, & Standard Errors for Business Cycle Statistics GMM, HAC estimators, & Standard Errors for Business Cycle Statistics Wouter J. Den Haan London School of Economics c Wouter J. Den Haan Overview Generic GMM problem Estimation Heteroskedastic and Autocorrelation

More information

STA 4273H: Statistical Machine Learning

STA 4273H: Statistical Machine Learning STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Computer Science! Department of Statistical Sciences! rsalakhu@cs.toronto.edu! h0p://www.cs.utoronto.ca/~rsalakhu/ Lecture 7 Approximate

More information

Modeling conditional distributions with mixture models: Applications in finance and financial decision-making

Modeling conditional distributions with mixture models: Applications in finance and financial decision-making Modeling conditional distributions with mixture models: Applications in finance and financial decision-making John Geweke University of Iowa, USA Journal of Applied Econometrics Invited Lecture Università

More information

Bayesian Monte Carlo Filtering for Stochastic Volatility Models

Bayesian Monte Carlo Filtering for Stochastic Volatility Models Bayesian Monte Carlo Filtering for Stochastic Volatility Models Roberto Casarin CEREMADE University Paris IX (Dauphine) and Dept. of Economics University Ca Foscari, Venice Abstract Modelling of the financial

More information

The Block Pseudo-Marginal Sampler

The Block Pseudo-Marginal Sampler The Block Pseudo-Marginal Sampler M.-N. Tran R. Kohn M. Quiroz M. Villani September 12, 2017 arxiv:1603.02485v5 [stat.me] 10 Sep 2017 Abstract The pseudo-marginal (PM) approach is increasingly used for

More information

LECTURE 15 Markov chain Monte Carlo

LECTURE 15 Markov chain Monte Carlo LECTURE 15 Markov chain Monte Carlo There are many settings when posterior computation is a challenge in that one does not have a closed form expression for the posterior distribution. Markov chain Monte

More information

Non-homogeneous Markov Mixture of Periodic Autoregressions for the Analysis of Air Pollution in the Lagoon of Venice

Non-homogeneous Markov Mixture of Periodic Autoregressions for the Analysis of Air Pollution in the Lagoon of Venice Non-homogeneous Markov Mixture of Periodic Autoregressions for the Analysis of Air Pollution in the Lagoon of Venice Roberta Paroli 1, Silvia Pistollato, Maria Rosa, and Luigi Spezia 3 1 Istituto di Statistica

More information

Computer intensive statistical methods

Computer intensive statistical methods Lecture 13 MCMC, Hybrid chains October 13, 2015 Jonas Wallin jonwal@chalmers.se Chalmers, Gothenburg university MH algorithm, Chap:6.3 The metropolis hastings requires three objects, the distribution of

More information

1 / 32 Summer school on Foundations and advances in stochastic filtering (FASF) Barcelona, Spain, June 22, 2015.

1 / 32 Summer school on Foundations and advances in stochastic filtering (FASF) Barcelona, Spain, June 22, 2015. Outline Part 4 Nonlinear system identification using sequential Monte Carlo methods Part 4 Identification strategy 2 (Data augmentation) Aim: Show how SMC can be used to implement identification strategy

More information

Metropolis-Hastings Algorithm

Metropolis-Hastings Algorithm Strength of the Gibbs sampler Metropolis-Hastings Algorithm Easy algorithm to think about. Exploits the factorization properties of the joint probability distribution. No difficult choices to be made to

More information

POMP inference via iterated filtering

POMP inference via iterated filtering POMP inference via iterated filtering Edward Ionides University of Michigan, Department of Statistics Lecture 3 at Wharton Statistics Department Thursday 27th April, 2017 Slides are online at http://dept.stat.lsa.umich.edu/~ionides/talks/upenn

More information

Bayesian Estimation with Sparse Grids

Bayesian Estimation with Sparse Grids Bayesian Estimation with Sparse Grids Kenneth L. Judd and Thomas M. Mertens Institute on Computational Economics August 7, 27 / 48 Outline Introduction 2 Sparse grids Construction Integration with sparse

More information

arxiv: v1 [stat.co] 1 Jan 2014

arxiv: v1 [stat.co] 1 Jan 2014 Approximate Bayesian Computation for a Class of Time Series Models BY AJAY JASRA Department of Statistics & Applied Probability, National University of Singapore, Singapore, 117546, SG. E-Mail: staja@nus.edu.sg

More information

Generating the Sample

Generating the Sample STAT 80: Mathematical Statistics Monte Carlo Suppose you are given random variables X,..., X n whose joint density f (or distribution) is specified and a statistic T (X,..., X n ) whose distribution you

More information

Introduction. log p θ (y k y 1:k 1 ), k=1

Introduction. log p θ (y k y 1:k 1 ), k=1 ESAIM: PROCEEDINGS, September 2007, Vol.19, 115-120 Christophe Andrieu & Dan Crisan, Editors DOI: 10.1051/proc:071915 PARTICLE FILTER-BASED APPROXIMATE MAXIMUM LIKELIHOOD INFERENCE ASYMPTOTICS IN STATE-SPACE

More information

Session 5B: A worked example EGARCH model

Session 5B: A worked example EGARCH model Session 5B: A worked example EGARCH model John Geweke Bayesian Econometrics and its Applications August 7, worked example EGARCH model August 7, / 6 EGARCH Exponential generalized autoregressive conditional

More information

Bayesian Inference and MCMC

Bayesian Inference and MCMC Bayesian Inference and MCMC Aryan Arbabi Partly based on MCMC slides from CSC412 Fall 2018 1 / 18 Bayesian Inference - Motivation Consider we have a data set D = {x 1,..., x n }. E.g each x i can be the

More information

27 : Distributed Monte Carlo Markov Chain. 1 Recap of MCMC and Naive Parallel Gibbs Sampling

27 : Distributed Monte Carlo Markov Chain. 1 Recap of MCMC and Naive Parallel Gibbs Sampling 10-708: Probabilistic Graphical Models 10-708, Spring 2014 27 : Distributed Monte Carlo Markov Chain Lecturer: Eric P. Xing Scribes: Pengtao Xie, Khoa Luu In this scribe, we are going to review the Parallel

More information

MCMC algorithms for fitting Bayesian models

MCMC algorithms for fitting Bayesian models MCMC algorithms for fitting Bayesian models p. 1/1 MCMC algorithms for fitting Bayesian models Sudipto Banerjee sudiptob@biostat.umn.edu University of Minnesota MCMC algorithms for fitting Bayesian models

More information