Control Variates for Markov Chain Monte Carlo

Size: px
Start display at page:

Download "Control Variates for Markov Chain Monte Carlo"

Transcription

1 Control Variates for Markov Chain Monte Carlo Dellaportas, P., Kontoyiannis, I., and Tsourti, Z. Dept of Statistics, AUEB Dept of Informatics, AUEB 1st Greek Stochastics Meeting Monte Carlo: Probability and Methods Lefkada, August 2009 () August 29, / 42

2 Outline of presentation 1 MCMC and variance considerations 2 Theoretical framework for Control Variates to MCMC 3 Control variates for Random-Walk Metropolis-Hastings algorithms 4 Application of control variates to RW-MH algorithms 5 Summary and Further Research () August 29, / 42

3 MCMC and variance considerations 1 MCMC and variance considerations 2 Theoretical framework for Control Variates to MCMC 3 Control variates for Random-Walk Metropolis-Hastings algorithms 4 Application of control variates to RW-MH algorithms 5 Summary and Further Research () August 29, / 42

4 MCMC and variance considerations MCMC framework Let π( ) probability measure on (X, B) F : X R function of interest π(f ) := E π [F (X )] := Fdπ the quantity to be estimated In the MCMC setting, we have (X n ) Markov chain on X, with: P transition kernel π stationary probability measure π(f ) is estimated by the ergodic average: µ n (F ) = 1 n n F (X i ) i=1 () August 29, / 42

5 MCMC and variance considerations Main Theorems for Markov Chains Ergodic Theorem, for Ergodic Markov Chains and appropriate F, (π( F ) < ), lim µ n(f ) = π(f ), a.s. n Central Limit Theorem (CLT) for Markov Chains Under appropriate additional conditions: n[µn (F ) π(f )] = 1 n n [F (X i ) π(f )] D N(0, σf 2 ) i=1 where σf 2, the asymptotic variance of F, is: σ 2 F := lim n Varπ[ nµ n(f )] = n= Cov π[f (X 0), F (X n)] () August 29, / 42

6 MCMC and variance considerations Approaches to variance reduction Importance sampling Antithetic sampling/variates Control variates Rao-Blackwellization () August 29, / 42

7 Theoretical framework for Control Variates to MCMC 1 MCMC and variance considerations 2 Theoretical framework for Control Variates to MCMC Introduction of control variates to Markov chains Optimal results for reversible chains Extension to multiple control variates Control variates for MCMC algorithms 3 Control variates for Random-Walk Metropolis-Hastings algorithms 4 Application of control variates to RW-MH algorithms 5 Summary and Further Research () August 29, / 42

8 Theoretical framework for Control Variates to MCMC Introduction of control variates to Markov chains Introduction of control variates to Markov chains Let π( ) target probability measure on (X, B) F : X R function of interest π(f ) := E π [F (X )] := Fdπ the quantity to be estimated (X n ) Markov chain on X, with: P transition kernel π stationary probability measure () August 29, / 42

9 Theoretical framework for Control Variates to MCMC Introduction of control variates to Markov chains Introduction of control variates to Markov chains Let further: π( ) target probability measure on (X, B) F : X R function of interest π(f ) := E π [F (X )] := Fdπ the quantity to be estimated (X n ) Markov chain on X, with: P transition kernel π stationary probability measure function U : X R, with π(u) = 0 modified function F θ = F θu modified estimator µ n (F θ ) = µ n (F ) θµ n (U) () August 29, / 42

10 Theoretical framework for Control Variates to MCMC Introduction of control variates to Markov chains Introduction of control variates to Markov chains All the regularity properties of F also hold for F θ : lim µ 1 n(f θ ) := lim n n n n F θ (X i ) = π(f θ ) = π(f ), i=1 a.s. and n[µn (F θ ) π(f )] = 1 n n [F θ (X i ) π(f )] D N(0, σf 2 θ ) i=1 () August 29, / 42

11 Theoretical framework for Control Variates to MCMC Introduction of control variates to Markov chains Introduction of control variates to Markov chains All the regularity properties of F also hold for F θ : lim µ 1 n(f θ ) := lim n n n n F θ (X i ) = π(f θ ) = π(f ), i=1 a.s. and n[µn (F θ ) π(f )] = 1 n n [F θ (X i ) π(f )] D N(0, σf 2 θ ) i=1 Research interest: () August 29, / 42

12 Theoretical framework for Control Variates to MCMC Introduction of control variates to Markov chains Introduction of control variates to Markov chains All the regularity properties of F also hold for F θ : lim µ 1 n(f θ ) := lim n n n n F θ (X i ) = π(f θ ) = π(f ), i=1 a.s. and n[µn (F θ ) π(f )] = 1 n n [F θ (X i ) π(f )] D N(0, σf 2 θ ) i=1 Research interest: Find appropriate function U parameter θ so as to significantly reduce (σf 2 θ ) compared to (σf 2 ). () August 29, / 42

13 Theoretical framework for Control Variates to MCMC Choice of function U Introduction of control variates to Markov chains In this setting we use: U = G PG, for arbitrary G, with π( G ) < () August 29, / 42

14 Theoretical framework for Control Variates to MCMC Choice of function U Introduction of control variates to Markov chains In this setting we use: U = G PG, for arbitrary G, with π( G ) < PG(x) = E [ G(X 1 ) X 0 = x ] () August 29, / 42

15 Theoretical framework for Control Variates to MCMC Choice of function U Introduction of control variates to Markov chains In this setting we use: U = G PG, for arbitrary G, with π( G ) < PG(x) = E [ G(X 1 ) X 0 = x ] Then, Which U? Which G? () August 29, / 42

16 Theoretical framework for Control Variates to MCMC Introduction of control variates to Markov chains Elaboration on variance of the modified estimator (σ 2 F θ ) An alternative expression of the variance in the CLT is: σ 2 F = π(ˆf 2 (P ˆF ) 2) ˆF : the solution of Poisson s equation: P ˆF ˆF = F + π(f ) Analogously σ 2 θ := σ2 F θ = π (ˆF 2 θ (P ˆF θ ) 2) () August 29, / 42

17 Theoretical framework for Control Variates to MCMC Choice of function U Introduction of control variates to Markov chains In this setting we use: Then, U = G PG, for arbitrary G, with π( G ) < PG(x) = E [ G(X 1 ) X 0 = x ] Which U? Which G? Findings from σ 2 θ elaboration () August 29, / 42

18 Theoretical framework for Control Variates to MCMC Choice of function U Introduction of control variates to Markov chains In this setting we use: Then, U = G PG, for arbitrary G, with π( G ) < PG(x) = E [ G(X 1 ) X 0 = x ] Which U? Which G? Findings from σ 2 θ elaboration If G = ˆF (σ 2 F θ ) = 0 () August 29, / 42

19 Theoretical framework for Control Variates to MCMC Choice of function U Introduction of control variates to Markov chains In this setting we use: Then, U = G PG, for arbitrary G, with π( G ) < PG(x) = E [ G(X 1 ) X 0 = x ] Which U? Which G? Findings from σ 2 θ elaboration If G = ˆF (σ 2 F θ ) = 0 Guidelines for choosing G as close as possible to ˆF () August 29, / 42

20 Theoretical framework for Control Variates to MCMC Choice of function U Introduction of control variates to Markov chains In this setting we use: Then, U = G PG, for arbitrary G, with π( G ) < PG(x) = E [ G(X 1 ) X 0 = x ] Which U? Which G? Findings from σ 2 θ elaboration If G = ˆF (σ 2 F θ ) = 0 The higher the correlation between F,U the smaller the (σ 2 F θ ) Guidelines for choosing G as close as possible to ˆF () August 29, / 42

21 Theoretical framework for Control Variates to MCMC Choice of function U Introduction of control variates to Markov chains In this setting we use: Then, U = G PG, for arbitrary G, with π( G ) < PG(x) = E [ G(X 1 ) X 0 = x ] Which U? Which G? Findings from σ 2 θ elaboration If G = ˆF (σ 2 F θ ) = 0 The higher the correlation between F,U the smaller the (σ 2 F θ ) Guidelines for choosing G as close as possible to ˆF leading to U = G PG highly correlated to F () August 29, / 42

22 Theoretical framework for Control Variates to MCMC Optimal value of parameter θ Introduction of control variates to Markov chains It can be derived that: σ 2 θ =... quadratic function with respect to θ. () August 29, / 42

23 Theoretical framework for Control Variates to MCMC Optimal value of parameter θ Introduction of control variates to Markov chains It can be derived that: σ 2 θ =... quadratic function with respect to θ. Thus, θ = π(ˆf G (P ˆF )(PG) ) π ( G 2 (PG) 2) Alternative expression (proved by Dellaportas and Kontoyiannis, 2008) with practical usefulness: θ = π(ˆf G (P ˆF )(PG) ) [ E π (G(X 1 ) PG(X 0 )) 2] () August 29, / 42

24 Theoretical framework for Control Variates to MCMC Optimal results for reversible chains Optimal empirical estimate of θ for reversible chains If the chain (X n ) is reversible, Dellaportas and Kontoyiannis (2008) prove: Theorem ) π (ˆF G (P ˆF )(PG) = π ((F π(f ))(G + PG)) () August 29, / 42

25 Theoretical framework for Control Variates to MCMC Optimal results for reversible chains Optimal empirical estimate of θ for reversible chains If the chain (X n ) is reversible, Dellaportas and Kontoyiannis (2008) prove: Theorem ) π (ˆF G (P ˆF )(PG) = π ((F π(f ))(G + PG)) So, the optimal value of θ (for reversible chains) can be expressed as θ rev = π ((F π(f ))(G + PG)) E π [ (G(X 1 ) PG(X 0 )) 2] () August 29, / 42

26 Theoretical framework for Control Variates to MCMC Optimal results for reversible chains Optimal empirical estimate of θ for reversible chains If the chain (X n ) is reversible, Dellaportas and Kontoyiannis (2008) prove: Theorem ) π (ˆF G (P ˆF )(PG) = π ((F π(f ))(G + PG)) So, the optimal value of θ (for reversible chains) can be expressed as θ rev = π ((F π(f ))(G + PG)) E π [ (G(X 1 ) PG(X 0 )) 2] In this case, θ can be adaptively estimated as ˆθ n = µ n(f (G + PG)) µ n (F )µ n (G + PG) n i=1 [G(X i) PG(X i 1 )] 2 1 n () August 29, / 42

27 Theoretical framework for Control Variates to MCMC Extension to multiple control variates Extension to multiple control variates Let s further assume: k functions U j (= G j PG j ) : X R, with π(u j ) = 0 Notation with vectors: G 1 U 1 G = G 2... : X Rk U = U 2... : X Rk θ = G k U k modified function F θ = F θ, U = F k j=1 θ ju j θ 1 θ 2... θ k Rk () August 29, / 42

28 Theoretical framework for Control Variates to MCMC Extension to multiple control variates Extension to multiple control variates Analogously to the one-dimensional case, we have that: σ 2 F θ = σ 2 F 2π(ˆF θ, G P ˆF θ, PG ) + π ( θ, G 2 θ, PG 2) ) θ = K(G) 1 π (ˆF G (P ˆF )(PG) where K(G) ij = E π [( Gi (X 1 ) PG i (X 0 ) )( G j (X 1 ) PG i (X 0 ) )] For reversible chains, Dellaportas and Kontoyiannis (2008) prove: Theorem θ = K(G) 1 π ( (F π(f )) (G + PG) ) () August 29, / 42

29 Theoretical framework for Control Variates to MCMC Control variates for MCMC algorithms Use of control variates for the reduction of variance of MCMC algorithms Control Variate methodology directly applicable to reversible MCMC algorithms Derivation of µ n (F θ ) straightforward. Quantities needed: F (X i ), G j (X i ) for i = 1,..., n, j = 1,..., k: OK, readily available BUT PG j (X i )? This issue may require further attention... Most popular MCMC algorithms: Gibbs sampler Treated in detail in Dellaportas and Kontoyiannis (2008) Metropolis-Hastings algorithm Random Walk (RW-MH) () August 29, / 42

30 Theoretical framework for Control Variates to MCMC Control variates for MCMC algorithms Use of control variates for the reduction of variance of MCMC algorithms Control Variate methodology directly applicable to reversible MCMC algorithms Derivation of µ n (F θ ) straightforward. Quantities needed: F (X i ), G j (X i ) for i = 1,..., n, j = 1,..., k: OK, readily available BUT PG j (X i )? This issue may require further attention... Most popular MCMC algorithms: Gibbs sampler Treated in detail in Dellaportas and Kontoyiannis (2008) Metropolis-Hastings algorithm Random Walk (RW-MH) () August 29, / 42

31 Control variates for Random-Walk Metropolis-Hastings algorithms 1 MCMC and variance considerations 2 Theoretical framework for Control Variates to MCMC 3 Control variates for Random-Walk Metropolis-Hastings algorithms Introduction to the algorithm Elaboration on PG(X ) 4 Application of control variates to RW-MH algorithms 5 Summary and Further Research () August 29, / 42

32 Control variates for Random-Walk Metropolis-Hastings algorithms Setting of RW-MH Introduction to the algorithm Assume the following: Our target distribution is a d-dimensional probability density π(x) The Random Walk Metropolis-Hastings (RW-MH) algorithm used for the simulation of π can be described as: Assume initial value X 0 and Y 0 = X 0 + 0, where 0 P (d-variate symmetrical distribution) At step t + 1, given X t = x t, Y t = y t = x t + t, we have that: X t+1 = where { x t + t = y t w.pr. ρ(x t, y t) x t w.pr. 1 ρ(x t, y t) we simulate t+1 P { ρ(x t, y t)= min 1, π(yt) } π(x t) () August 29, / 42

33 Control variates for Random-Walk Metropolis-Hastings algorithms Elaboration on PG(X ) Use of control variates for RW-MH - PG(x) elaboration PG(x) = E X1 [ G (X1 ) X 0 = x ] () August 29, / 42

34 Control variates for Random-Walk Metropolis-Hastings algorithms Elaboration on PG(X ) Use of control variates for RW-MH - PG(x) elaboration PG(x) = E X1 [ G (X1 ) X 0 = x ] = E Y0 { EX1 [ G (X1 ) X 0 = x, Y 0 = y ]} () August 29, / 42

35 Control variates for Random-Walk Metropolis-Hastings algorithms Elaboration on PG(X ) Use of control variates for RW-MH - PG(x) elaboration PG(x) = E X1 [ G (X1 ) X 0 = x ] = E Y0 { EX1 [ G (X1 ) X 0 = x, Y 0 = y ]} = E Y0 {ρ (x, y) G (y) + (1 ρ (x, y)) G (x)} () August 29, / 42

36 Control variates for Random-Walk Metropolis-Hastings algorithms Elaboration on PG(X ) Use of control variates for RW-MH - PG(x) elaboration [ PG(x) = E X1 G (X1 ) X 0 = x ] { [ = E Y0 EX1 G (X1 ) X 0 = x, Y 0 = y ]} = E Y0 {ρ (x, y) G (y) + (1 ρ (x, y)) G (x)} = G(x) + E Y0 {ρ(x, y) (G(y) G(x))} () August 29, / 42

37 Control variates for Random-Walk Metropolis-Hastings algorithms Elaboration on PG(X ) Use of control variates for RW-MH - PG(x) elaboration [ PG(x) = E X1 G (X1 ) X 0 = x ] { [ = E Y0 EX1 G (X1 ) X 0 = x, Y 0 = y ]} = E Y0 {ρ (x, y) G (y) + (1 ρ (x, y)) G (x)} = G(x) + E Y0 {ρ(x, y) (G(y) G(x))} i.e. PG(x) = G(x) + E Y0 {ρ(x, y) (G(y) G(x))}, where Y 0 q(y x) () August 29, / 42

38 Control variates for Random-Walk Metropolis-Hastings algorithms Elaboration on PG(X ) Use of control variates for RW-MH - PG(x) elaboration [ PG(x) = E X1 G (X1 ) X 0 = x ] { [ = E Y0 EX1 G (X1 ) X 0 = x, Y 0 = y ]} = E Y0 {ρ (x, y) G (y) + (1 ρ (x, y)) G (x)} = G(x) + E Y0 {ρ(x, y) (G(y) G(x))} i.e. PG(x) = G(x) + E Y0 {ρ(x, y) (G(y) G(x))}, where Y 0 q(y x) Several approaches may be considered for the estimation of PG(x). Monte Carlo estimation based on P Importance sampling based on the proposed values y () August 29, / 42

39 Application of control variates to RW-MH algorithms 1 MCMC and variance considerations 2 Theoretical framework for Control Variates to MCMC 3 Control variates for Random-Walk Metropolis-Hastings algorithms 4 Application of control variates to RW-MH algorithms The simple case of Univariate Normal Case-study of a survival analysis Poisson generation Heavy tailed distributions 5 Summary and Further Research () August 29, / 42

40 Application of control variates to RW-MH algorithms The simple case of Univariate Normal Application of control variates to RW-MH algorithms - Univariate Normal Target : N ( 0, σ 2 t = 10 ) Proposal : N ( 0, σ 2 pr = 160 ) We use F (x) = x and G(x) = x () August 29, / 42

41 Application of control variates to RW-MH algorithms The simple case of Univariate Normal Application of control variates to RW-MH algorithms - Univariate Normal Target : N ( 0, σ 2 t = 10 ) Proposal : N ( 0, σ 2 pr = 160 ) We use F (x) = x and G(x) = x Framework for the evaluation of variance reduction: Optimal value of θ based on Dellaportas and Kontoyiannis (2008) approach for reversible chains Terms of PG(x) assessed using Monte Carlo estimates from the proposal distribution T = 100 repetitions for each simulation scenario. Variance Reduction Factors : VRF = S 2 µ n(f ) S 2 µ n(f θ ) () August 29, / 42

42 Figure: Ergodic means for different realizations with n = 100, 000 and n PG = 50 () August 29, / 42 Application of control variates to RW-MH algorithms The simple case of Univariate Normal Application of control variates to RW-MH algorithms - Univariate Normal

43 Application of control variates to RW-MH algorithms The simple case of Univariate Normal Application of control variates to RW-MH algorithms - Univariate Normal Table: VRF s in Univariate Normal case - Normal proposal - Simplest G(x) = x Length of Markov chain (number of iterations n) n PG 1,000 2,000 5,000 10,000 20,000 50, , , () August 29, / 42

44 Application of control variates to RW-MH algorithms The simple case of Univariate Normal Application of control variates to RW-MH algorithms - Univariate Normal Study of the effect of G functions 1. G(x) = x 2. G 1 (x) = x, G 2 (x) = x 2 k. G 1 (x) = x, G 2 (x) = x 2,, G k (x) = x k () August 29, / 42

45 Figure: Plot of VRF by the order of polynomial G functions () August 29, / 42 Application of control variates to RW-MH algorithms The simple case of Univariate Normal Application of control variates to RW-MH algorithms - Univariate Normal

46 Application of control variates to RW-MH algorithms The simple case of Univariate Normal Application of control variates to RW-MH algorithms - Univariate Normal Table: VRF s in Univariate Normal case - Normal proposal - Polynomial G functions of order 4, i.e. G 1 (x) = x, G 2 (x) = x 2, G 3 (x) = x 3, G 4 (x) = x 4 Length of Markov chain (number of iterations n) n PG 1,000 2,000 5,000 10,000 20,000 50, , , () August 29, / 42

47 Figure: Plot of VRF by different acceptance rates () August 29, / 42 Application of control variates to RW-MH algorithms The simple case of Univariate Normal Application of control variates to RW-MH algorithms - Univariate Normal

48 Application of control variates to RW-MH algorithms The simple case of Univariate Normal Application of control variates to RW-MH algorithms - Univariate Normal Study of the effect of the proposal distribution Table: VRF s in Univariate Normal case - T-student proposal (3 df s, variance 160) - Simplest function G(x) = x Length of Markov chain (number of iterations n) n PG 1,000 2,000 5,000 10,000 20,000 50, () August 29, / 42

49 Application of control variates to RW-MH algorithms The simple case of Univariate Normal Application of control variates to RW-MH algorithms - Univariate Normal Study of the effect of the proposal distribution Table: VRF s in Univariate Normal case - Uniform proposal ( 5.5 σ t, +5.5 σ t ) - Simplest function G(x) = x Length of Markov chain (number of iterations n) n PG 1,000 2,000 5,000 10,000 20,000 50, () August 29, / 42

50 Application of control variates to RW-MH algorithms Case-study of a survival analysis Application of control variates to RW-MH algorithms - Survival analysis Source: Albert, J. (2007). Bayesian Computation with R, Springer. Data: Lifetime of a number of patients some of which had a heart transplant. Model assumptions: Non-transplant patients: X i Exponential with mean 1/η Transplant patients: X i Exponential with mean 1/(τη) Parameter η Gamma(p, λ), i.e. f (η) = λp Γ(p) ηp 1 exp( λη) Unknown parameter vector: (τ, λ, p) (all positive) Notation: N non-transplant patients: n: died N n: censored x i survival time. M transplant patients: m: died M m: censored y i time to transplant z i survival time () August 29, / 42

51 Application of control variates to RW-MH algorithms Case-study of a survival analysis Application of control variates to RW-MH algorithms - Survival analysis Likelihood: L(τ, λ, p) = n i=1 pλ p (λ+x i ) p+1 N i=n+1 ( λ λ+x i ) p m j=1 τpλ p ( M (λ+y j +τz j ) p+1 j=m+1 ) p λ λ+y j +τz j Prior distribution of parameters: g(τ, λ, p) 1 Posterior distribution of parameters: g(τ, λ, p data) L(τ, λ, p) Transformation φ = (φ 1 := logτ, φ 2 := logλ, φ 3 := logp, ) ( ) g(φ data) L e φ 1, e φ 2, e φ 3 e 3 i=1 φ i () August 29, / 42

52 Application of control variates to RW-MH algorithms Case-study of a survival analysis Application of control variates to RW-MH algorithms - Survival analysis Table: Function F (τ, λ, p) = log(τ) = φ 1, n = 10, 000, n PG = 50 Form of G A. G 1 = φ 1 G 2 = φ 2 G 3 = φ 3 B. G 1 = φ 2 1 G 2 = φ 2 2 G 3 = φ 2 3 [ ] G1 = φ 1 C. G 2 = φ 2 1 D. [ G1 = φ 1 G 2 = e φ 1 VRF Form of G E. G 1 = φ 1 G 2 = φ 2 1 G 3 = φ 2 G 4 = φ 2 2 G 5 = φ 3 G 6 = φ 2 3 F. G 1 = φ 1 G 2 = φ 2 1 G 3 = φ 2 G 4 = φ 2 2 G. G 1 = φ 1 G 2 = φ 2 1 G 3 = φ 3 G 4 = φ 2 3 VRF () August 29, / 42

53 Figure: Plot of VRF by n, function F (τ, λ, p) = log(τ) = φ 1, G = [ G = φ, G = φ 2] () August 29, / 42 Application of control variates to RW-MH algorithms Case-study of a survival analysis Application of control variates to RW-MH algorithms - Survival analysis

54 Application of control variates to RW-MH algorithms Poisson generation Application of control variates to RW-MH algorithms - Poisson generation Setting: Target distribution π: Poisson(λ), Proposal: discrete bell-shaped: P M M + 1 δ ( = δ) = M(M + 1), δ { M, M+1,..., 1,+1,...,M 1,M} Inference is focused on: F (λ) = λ To enhance estimation of π(f ): F G,θ (λ) = F (λ) θ U(λ) where U = G PG () August 29, / 42

55 Application of control variates to RW-MH algorithms Poisson generation Application of control variates to RW-MH algorithms - Poisson generation The form of function G used here is In the present setting: G(λ) = λ PG(x) = E π [G(λ t+1 ) λ t = x] =... = x + k [ j= k, 0 P k (j) j ρ(x, x + j) ] The terms of PG(x) are assessed using two approaches: (i) Using the exact formula (ii) Using Monte Carlo estimates from P k distribution () August 29, / 42

56 Application of control variates to RW-MH algorithms Poisson generation Application of control variates to RW-MH algorithms - Poisson generation Table: VRF s for F (x) = x, G(x) = x for n = 5, 000 (analytic formula for PG(x)) Proposal distribution P M Target P(λ) λ = λ = λ = () August 29, / 42

57 Application of control variates to RW-MH algorithms Heavy tailed distributions Application of control variates to RW-MH algorithms - Heavy tailed distributions Source: Jarner, S.F. and Roberts, G.O. (2007). Convergence of Heavy-tailed Monte Carlo Markov Chain Algorithms. Scandinavian Journal of Statistics. 34, Main result for RW-MH: Heavy tailed proposals lead to higher rates of convergence Focus: Polynomially ergodic Markov chains For polynomial target distributions: they derive polynomial rates of convergence () August 29, / 42

58 Application of control variates to RW-MH algorithms Heavy tailed distributions Application of control variates to RW-MH algorithms - Heavy tailed distributions Table: Existence of central limit theorems for RW-MH algorithm (from Jarner and Roberts, 2007) Target distribution Proposal distr. t(2.5) t(3) t(4) t(5) t(6) t(7) Uniform L C C C Normal L C C C Cauchy L C C C C t(0.5) L L C C C C C: CLT holds for x, L: CLT holds for x s, s < 1 () August 29, / 42

59 Application of control variates to RW-MH algorithms Heavy tailed distributions Application of control variates to RW-MH algorithms - Heavy tailed distributions Table: VRF s for F (x) = x, with G i = x i, i = 1, 2, 3 (n = 200, 000, n PG = 50) Target distribution Proposal distr. t(2.5) t(3) t(4) t(5) t(6) t(7) Uniform Normal Cauchy t(0.5) () August 29, / 42

60 Summary and Further Research Summary and Further Research Summary A solid methodological framework has been provided for the development and use of control variates in MCMC For given G function, consistent estimates for optimal coefficients θ are defined Main reference: Dellaportas, P. and Kontoyiannis, I. (2008). Control Variates for Reversible MCMC samplers. Submitted to JRSS(Series B) () August 29, / 42

61 Summary and Further Research Summary and Further Research Summary A solid methodological framework has been provided for the development and use of control variates in MCMC For given G function, consistent estimates for optimal coefficients θ are defined Further research Approaches for elicitation of G functions and decision on k Techniques for more efficient derivation of PG(x) Extension to Reversible-Jump MCMC algorithms Main reference: Dellaportas, P. and Kontoyiannis, I. (2008). Control Variates for Reversible MCMC samplers. Submitted to JRSS(Series B) () August 29, / 42

62 Summary and Further Research Summary and Further Research Thank you for listening :) () August 29, / 42

Computational statistics

Computational statistics Computational statistics Markov Chain Monte Carlo methods Thierry Denœux March 2017 Thierry Denœux Computational statistics March 2017 1 / 71 Contents of this chapter When a target density f can be evaluated

More information

eqr094: Hierarchical MCMC for Bayesian System Reliability

eqr094: Hierarchical MCMC for Bayesian System Reliability eqr094: Hierarchical MCMC for Bayesian System Reliability Alyson G. Wilson Statistical Sciences Group, Los Alamos National Laboratory P.O. Box 1663, MS F600 Los Alamos, NM 87545 USA Phone: 505-667-9167

More information

Adaptive Monte Carlo methods

Adaptive Monte Carlo methods Adaptive Monte Carlo methods Jean-Michel Marin Projet Select, INRIA Futurs, Université Paris-Sud joint with Randal Douc (École Polytechnique), Arnaud Guillin (Université de Marseille) and Christian Robert

More information

Markov Chain Monte Carlo, Numerical Integration

Markov Chain Monte Carlo, Numerical Integration Markov Chain Monte Carlo, Numerical Integration (See Statistics) Trevor Gallen Fall 2015 1 / 1 Agenda Numerical Integration: MCMC methods Estimating Markov Chains Estimating latent variables 2 / 1 Numerical

More information

Markov Chain Monte Carlo (MCMC)

Markov Chain Monte Carlo (MCMC) Markov Chain Monte Carlo (MCMC Dependent Sampling Suppose we wish to sample from a density π, and we can evaluate π as a function but have no means to directly generate a sample. Rejection sampling can

More information

Markov Chain Monte Carlo methods

Markov Chain Monte Carlo methods Markov Chain Monte Carlo methods By Oleg Makhnin 1 Introduction a b c M = d e f g h i 0 f(x)dx 1.1 Motivation 1.1.1 Just here Supresses numbering 1.1.2 After this 1.2 Literature 2 Method 2.1 New math As

More information

Bayesian Methods for Machine Learning

Bayesian Methods for Machine Learning Bayesian Methods for Machine Learning CS 584: Big Data Analytics Material adapted from Radford Neal s tutorial (http://ftp.cs.utoronto.ca/pub/radford/bayes-tut.pdf), Zoubin Ghahramni (http://hunch.net/~coms-4771/zoubin_ghahramani_bayesian_learning.pdf),

More information

Lecture 8: The Metropolis-Hastings Algorithm

Lecture 8: The Metropolis-Hastings Algorithm 30.10.2008 What we have seen last time: Gibbs sampler Key idea: Generate a Markov chain by updating the component of (X 1,..., X p ) in turn by drawing from the full conditionals: X (t) j Two drawbacks:

More information

I. Bayesian econometrics

I. Bayesian econometrics I. Bayesian econometrics A. Introduction B. Bayesian inference in the univariate regression model C. Statistical decision theory D. Large sample results E. Diffuse priors F. Numerical Bayesian methods

More information

ABC methods for phase-type distributions with applications in insurance risk problems

ABC methods for phase-type distributions with applications in insurance risk problems ABC methods for phase-type with applications problems Concepcion Ausin, Department of Statistics, Universidad Carlos III de Madrid Joint work with: Pedro Galeano, Universidad Carlos III de Madrid Simon

More information

A generalization of the Multiple-try Metropolis algorithm for Bayesian estimation and model selection

A generalization of the Multiple-try Metropolis algorithm for Bayesian estimation and model selection A generalization of the Multiple-try Metropolis algorithm for Bayesian estimation and model selection Silvia Pandolfi Francesco Bartolucci Nial Friel University of Perugia, IT University of Perugia, IT

More information

Introduction to Machine Learning CMU-10701

Introduction to Machine Learning CMU-10701 Introduction to Machine Learning CMU-10701 Markov Chain Monte Carlo Methods Barnabás Póczos & Aarti Singh Contents Markov Chain Monte Carlo Methods Goal & Motivation Sampling Rejection Importance Markov

More information

Likelihood-free MCMC

Likelihood-free MCMC Bayesian inference for stable distributions with applications in finance Department of Mathematics University of Leicester September 2, 2011 MSc project final presentation Outline 1 2 3 4 Classical Monte

More information

MH I. Metropolis-Hastings (MH) algorithm is the most popular method of getting dependent samples from a probability distribution

MH I. Metropolis-Hastings (MH) algorithm is the most popular method of getting dependent samples from a probability distribution MH I Metropolis-Hastings (MH) algorithm is the most popular method of getting dependent samples from a probability distribution a lot of Bayesian mehods rely on the use of MH algorithm and it s famous

More information

MCMC algorithms for fitting Bayesian models

MCMC algorithms for fitting Bayesian models MCMC algorithms for fitting Bayesian models p. 1/1 MCMC algorithms for fitting Bayesian models Sudipto Banerjee sudiptob@biostat.umn.edu University of Minnesota MCMC algorithms for fitting Bayesian models

More information

Pattern Recognition and Machine Learning. Bishop Chapter 11: Sampling Methods

Pattern Recognition and Machine Learning. Bishop Chapter 11: Sampling Methods Pattern Recognition and Machine Learning Chapter 11: Sampling Methods Elise Arnaud Jakob Verbeek May 22, 2008 Outline of the chapter 11.1 Basic Sampling Algorithms 11.2 Markov Chain Monte Carlo 11.3 Gibbs

More information

Hierarchical models. Dr. Jarad Niemi. August 31, Iowa State University. Jarad Niemi (Iowa State) Hierarchical models August 31, / 31

Hierarchical models. Dr. Jarad Niemi. August 31, Iowa State University. Jarad Niemi (Iowa State) Hierarchical models August 31, / 31 Hierarchical models Dr. Jarad Niemi Iowa State University August 31, 2017 Jarad Niemi (Iowa State) Hierarchical models August 31, 2017 1 / 31 Normal hierarchical model Let Y ig N(θ g, σ 2 ) for i = 1,...,

More information

The Bayesian Choice. Christian P. Robert. From Decision-Theoretic Foundations to Computational Implementation. Second Edition.

The Bayesian Choice. Christian P. Robert. From Decision-Theoretic Foundations to Computational Implementation. Second Edition. Christian P. Robert The Bayesian Choice From Decision-Theoretic Foundations to Computational Implementation Second Edition With 23 Illustrations ^Springer" Contents Preface to the Second Edition Preface

More information

Stat 535 C - Statistical Computing & Monte Carlo Methods. Lecture February Arnaud Doucet

Stat 535 C - Statistical Computing & Monte Carlo Methods. Lecture February Arnaud Doucet Stat 535 C - Statistical Computing & Monte Carlo Methods Lecture 13-28 February 2006 Arnaud Doucet Email: arnaud@cs.ubc.ca 1 1.1 Outline Limitations of Gibbs sampling. Metropolis-Hastings algorithm. Proof

More information

6 Markov Chain Monte Carlo (MCMC)

6 Markov Chain Monte Carlo (MCMC) 6 Markov Chain Monte Carlo (MCMC) The underlying idea in MCMC is to replace the iid samples of basic MC methods, with dependent samples from an ergodic Markov chain, whose limiting (stationary) distribution

More information

Stat 516, Homework 1

Stat 516, Homework 1 Stat 516, Homework 1 Due date: October 7 1. Consider an urn with n distinct balls numbered 1,..., n. We sample balls from the urn with replacement. Let N be the number of draws until we encounter a ball

More information

Weak convergence of Markov chain Monte Carlo II

Weak convergence of Markov chain Monte Carlo II Weak convergence of Markov chain Monte Carlo II KAMATANI, Kengo Mar 2011 at Le Mans Background Markov chain Monte Carlo (MCMC) method is widely used in Statistical Science. It is easy to use, but difficult

More information

Hastings-within-Gibbs Algorithm: Introduction and Application on Hierarchical Model

Hastings-within-Gibbs Algorithm: Introduction and Application on Hierarchical Model UNIVERSITY OF TEXAS AT SAN ANTONIO Hastings-within-Gibbs Algorithm: Introduction and Application on Hierarchical Model Liang Jing April 2010 1 1 ABSTRACT In this paper, common MCMC algorithms are introduced

More information

Monte Carlo in Bayesian Statistics

Monte Carlo in Bayesian Statistics Monte Carlo in Bayesian Statistics Matthew Thomas SAMBa - University of Bath m.l.thomas@bath.ac.uk December 4, 2014 Matthew Thomas (SAMBa) Monte Carlo in Bayesian Statistics December 4, 2014 1 / 16 Overview

More information

Markov Chain Monte Carlo methods

Markov Chain Monte Carlo methods Markov Chain Monte Carlo methods Tomas McKelvey and Lennart Svensson Signal Processing Group Department of Signals and Systems Chalmers University of Technology, Sweden November 26, 2012 Today s learning

More information

Metropolis-Hastings Algorithm

Metropolis-Hastings Algorithm Strength of the Gibbs sampler Metropolis-Hastings Algorithm Easy algorithm to think about. Exploits the factorization properties of the joint probability distribution. No difficult choices to be made to

More information

Stat 451 Lecture Notes Markov Chain Monte Carlo. Ryan Martin UIC

Stat 451 Lecture Notes Markov Chain Monte Carlo. Ryan Martin UIC Stat 451 Lecture Notes 07 12 Markov Chain Monte Carlo Ryan Martin UIC www.math.uic.edu/~rgmartin 1 Based on Chapters 8 9 in Givens & Hoeting, Chapters 25 27 in Lange 2 Updated: April 4, 2016 1 / 42 Outline

More information

Bayesian Estimation of DSGE Models 1 Chapter 3: A Crash Course in Bayesian Inference

Bayesian Estimation of DSGE Models 1 Chapter 3: A Crash Course in Bayesian Inference 1 The views expressed in this paper are those of the authors and do not necessarily reflect the views of the Federal Reserve Board of Governors or the Federal Reserve System. Bayesian Estimation of DSGE

More information

Winter 2019 Math 106 Topics in Applied Mathematics. Lecture 9: Markov Chain Monte Carlo

Winter 2019 Math 106 Topics in Applied Mathematics. Lecture 9: Markov Chain Monte Carlo Winter 2019 Math 106 Topics in Applied Mathematics Data-driven Uncertainty Quantification Yoonsang Lee (yoonsang.lee@dartmouth.edu) Lecture 9: Markov Chain Monte Carlo 9.1 Markov Chain A Markov Chain Monte

More information

Sampling Algorithms for Probabilistic Graphical models

Sampling Algorithms for Probabilistic Graphical models Sampling Algorithms for Probabilistic Graphical models Vibhav Gogate University of Washington References: Chapter 12 of Probabilistic Graphical models: Principles and Techniques by Daphne Koller and Nir

More information

Markov Chain Monte Carlo

Markov Chain Monte Carlo Markov Chain Monte Carlo Recall: To compute the expectation E ( h(y ) ) we use the approximation E(h(Y )) 1 n n h(y ) t=1 with Y (1),..., Y (n) h(y). Thus our aim is to sample Y (1),..., Y (n) from f(y).

More information

A Review of Pseudo-Marginal Markov Chain Monte Carlo

A Review of Pseudo-Marginal Markov Chain Monte Carlo A Review of Pseudo-Marginal Markov Chain Monte Carlo Discussed by: Yizhe Zhang October 21, 2016 Outline 1 Overview 2 Paper review 3 experiment 4 conclusion Motivation & overview Notation: θ denotes the

More information

The Metropolis-Hastings Algorithm. June 8, 2012

The Metropolis-Hastings Algorithm. June 8, 2012 The Metropolis-Hastings Algorithm June 8, 22 The Plan. Understand what a simulated distribution is 2. Understand why the Metropolis-Hastings algorithm works 3. Learn how to apply the Metropolis-Hastings

More information

Sequential Monte Carlo Samplers for Applications in High Dimensions

Sequential Monte Carlo Samplers for Applications in High Dimensions Sequential Monte Carlo Samplers for Applications in High Dimensions Alexandros Beskos National University of Singapore KAUST, 26th February 2014 Joint work with: Dan Crisan, Ajay Jasra, Nik Kantas, Alex

More information

Zig-Zag Monte Carlo. Delft University of Technology. Joris Bierkens February 7, 2017

Zig-Zag Monte Carlo. Delft University of Technology. Joris Bierkens February 7, 2017 Zig-Zag Monte Carlo Delft University of Technology Joris Bierkens February 7, 2017 Joris Bierkens (TU Delft) Zig-Zag Monte Carlo February 7, 2017 1 / 33 Acknowledgements Collaborators Andrew Duncan Paul

More information

Kernel adaptive Sequential Monte Carlo

Kernel adaptive Sequential Monte Carlo Kernel adaptive Sequential Monte Carlo Ingmar Schuster (Paris Dauphine) Heiko Strathmann (University College London) Brooks Paige (Oxford) Dino Sejdinovic (Oxford) December 7, 2015 1 / 36 Section 1 Outline

More information

CSC 2541: Bayesian Methods for Machine Learning

CSC 2541: Bayesian Methods for Machine Learning CSC 2541: Bayesian Methods for Machine Learning Radford M. Neal, University of Toronto, 2011 Lecture 3 More Markov Chain Monte Carlo Methods The Metropolis algorithm isn t the only way to do MCMC. We ll

More information

MARKOV CHAIN MONTE CARLO

MARKOV CHAIN MONTE CARLO MARKOV CHAIN MONTE CARLO RYAN WANG Abstract. This paper gives a brief introduction to Markov Chain Monte Carlo methods, which offer a general framework for calculating difficult integrals. We start with

More information

STA 4273H: Statistical Machine Learning

STA 4273H: Statistical Machine Learning STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Computer Science! Department of Statistical Sciences! rsalakhu@cs.toronto.edu! h0p://www.cs.utoronto.ca/~rsalakhu/ Lecture 7 Approximate

More information

Computer intensive statistical methods

Computer intensive statistical methods Lecture 13 MCMC, Hybrid chains October 13, 2015 Jonas Wallin jonwal@chalmers.se Chalmers, Gothenburg university MH algorithm, Chap:6.3 The metropolis hastings requires three objects, the distribution of

More information

BAYESIAN METHODS FOR VARIABLE SELECTION WITH APPLICATIONS TO HIGH-DIMENSIONAL DATA

BAYESIAN METHODS FOR VARIABLE SELECTION WITH APPLICATIONS TO HIGH-DIMENSIONAL DATA BAYESIAN METHODS FOR VARIABLE SELECTION WITH APPLICATIONS TO HIGH-DIMENSIONAL DATA Intro: Course Outline and Brief Intro to Marina Vannucci Rice University, USA PASI-CIMAT 04/28-30/2010 Marina Vannucci

More information

Deblurring Jupiter (sampling in GLIP faster than regularized inversion) Colin Fox Richard A. Norton, J.

Deblurring Jupiter (sampling in GLIP faster than regularized inversion) Colin Fox Richard A. Norton, J. Deblurring Jupiter (sampling in GLIP faster than regularized inversion) Colin Fox fox@physics.otago.ac.nz Richard A. Norton, J. Andrés Christen Topics... Backstory (?) Sampling in linear-gaussian hierarchical

More information

Computer Vision Group Prof. Daniel Cremers. 10a. Markov Chain Monte Carlo

Computer Vision Group Prof. Daniel Cremers. 10a. Markov Chain Monte Carlo Group Prof. Daniel Cremers 10a. Markov Chain Monte Carlo Markov Chain Monte Carlo In high-dimensional spaces, rejection sampling and importance sampling are very inefficient An alternative is Markov Chain

More information

Brief introduction to Markov Chain Monte Carlo

Brief introduction to Markov Chain Monte Carlo Brief introduction to Department of Probability and Mathematical Statistics seminar Stochastic modeling in economics and finance November 7, 2011 Brief introduction to Content 1 and motivation Classical

More information

Markov Chain Monte Carlo Using the Ratio-of-Uniforms Transformation. Luke Tierney Department of Statistics & Actuarial Science University of Iowa

Markov Chain Monte Carlo Using the Ratio-of-Uniforms Transformation. Luke Tierney Department of Statistics & Actuarial Science University of Iowa Markov Chain Monte Carlo Using the Ratio-of-Uniforms Transformation Luke Tierney Department of Statistics & Actuarial Science University of Iowa Basic Ratio of Uniforms Method Introduced by Kinderman and

More information

Markov Chain Monte Carlo Methods

Markov Chain Monte Carlo Methods Markov Chain Monte Carlo Methods John Geweke University of Iowa, USA 2005 Institute on Computational Economics University of Chicago - Argonne National Laboaratories July 22, 2005 The problem p (θ, ω I)

More information

Bayesian model selection in graphs by using BDgraph package

Bayesian model selection in graphs by using BDgraph package Bayesian model selection in graphs by using BDgraph package A. Mohammadi and E. Wit March 26, 2013 MOTIVATION Flow cytometry data with 11 proteins from Sachs et al. (2005) RESULT FOR CELL SIGNALING DATA

More information

Simulation of truncated normal variables. Christian P. Robert LSTA, Université Pierre et Marie Curie, Paris

Simulation of truncated normal variables. Christian P. Robert LSTA, Université Pierre et Marie Curie, Paris Simulation of truncated normal variables Christian P. Robert LSTA, Université Pierre et Marie Curie, Paris Abstract arxiv:0907.4010v1 [stat.co] 23 Jul 2009 We provide in this paper simulation algorithms

More information

Generating the Sample

Generating the Sample STAT 80: Mathematical Statistics Monte Carlo Suppose you are given random variables X,..., X n whose joint density f (or distribution) is specified and a statistic T (X,..., X n ) whose distribution you

More information

18 : Advanced topics in MCMC. 1 Gibbs Sampling (Continued from the last lecture)

18 : Advanced topics in MCMC. 1 Gibbs Sampling (Continued from the last lecture) 10-708: Probabilistic Graphical Models 10-708, Spring 2014 18 : Advanced topics in MCMC Lecturer: Eric P. Xing Scribes: Jessica Chemali, Seungwhan Moon 1 Gibbs Sampling (Continued from the last lecture)

More information

Computer Vision Group Prof. Daniel Cremers. 11. Sampling Methods: Markov Chain Monte Carlo

Computer Vision Group Prof. Daniel Cremers. 11. Sampling Methods: Markov Chain Monte Carlo Group Prof. Daniel Cremers 11. Sampling Methods: Markov Chain Monte Carlo Markov Chain Monte Carlo In high-dimensional spaces, rejection sampling and importance sampling are very inefficient An alternative

More information

Lecture 7 and 8: Markov Chain Monte Carlo

Lecture 7 and 8: Markov Chain Monte Carlo Lecture 7 and 8: Markov Chain Monte Carlo 4F13: Machine Learning Zoubin Ghahramani and Carl Edward Rasmussen Department of Engineering University of Cambridge http://mlg.eng.cam.ac.uk/teaching/4f13/ Ghahramani

More information

Statistics & Data Sciences: First Year Prelim Exam May 2018

Statistics & Data Sciences: First Year Prelim Exam May 2018 Statistics & Data Sciences: First Year Prelim Exam May 2018 Instructions: 1. Do not turn this page until instructed to do so. 2. Start each new question on a new sheet of paper. 3. This is a closed book

More information

MCMC: Markov Chain Monte Carlo

MCMC: Markov Chain Monte Carlo I529: Machine Learning in Bioinformatics (Spring 2013) MCMC: Markov Chain Monte Carlo Yuzhen Ye School of Informatics and Computing Indiana University, Bloomington Spring 2013 Contents Review of Markov

More information

Asymptotics and Simulation of Heavy-Tailed Processes

Asymptotics and Simulation of Heavy-Tailed Processes Asymptotics and Simulation of Heavy-Tailed Processes Department of Mathematics Stockholm, Sweden Workshop on Heavy-tailed Distributions and Extreme Value Theory ISI Kolkata January 14-17, 2013 Outline

More information

A quick introduction to Markov chains and Markov chain Monte Carlo (revised version)

A quick introduction to Markov chains and Markov chain Monte Carlo (revised version) A quick introduction to Markov chains and Markov chain Monte Carlo (revised version) Rasmus Waagepetersen Institute of Mathematical Sciences Aalborg University 1 Introduction These notes are intended to

More information

Principles of Bayesian Inference

Principles of Bayesian Inference Principles of Bayesian Inference Sudipto Banerjee University of Minnesota July 20th, 2008 1 Bayesian Principles Classical statistics: model parameters are fixed and unknown. A Bayesian thinks of parameters

More information

Kernel Sequential Monte Carlo

Kernel Sequential Monte Carlo Kernel Sequential Monte Carlo Ingmar Schuster (Paris Dauphine) Heiko Strathmann (University College London) Brooks Paige (Oxford) Dino Sejdinovic (Oxford) * equal contribution April 25, 2016 1 / 37 Section

More information

LECTURE 15 Markov chain Monte Carlo

LECTURE 15 Markov chain Monte Carlo LECTURE 15 Markov chain Monte Carlo There are many settings when posterior computation is a challenge in that one does not have a closed form expression for the posterior distribution. Markov chain Monte

More information

Computer Vision Group Prof. Daniel Cremers. 11. Sampling Methods

Computer Vision Group Prof. Daniel Cremers. 11. Sampling Methods Prof. Daniel Cremers 11. Sampling Methods Sampling Methods Sampling Methods are widely used in Computer Science as an approximation of a deterministic algorithm to represent uncertainty without a parametric

More information

MCMC Sampling for Bayesian Inference using L1-type Priors

MCMC Sampling for Bayesian Inference using L1-type Priors MÜNSTER MCMC Sampling for Bayesian Inference using L1-type Priors (what I do whenever the ill-posedness of EEG/MEG is just not frustrating enough!) AG Imaging Seminar Felix Lucka 26.06.2012 , MÜNSTER Sampling

More information

Monte Carlo Methods. Leon Gu CSD, CMU

Monte Carlo Methods. Leon Gu CSD, CMU Monte Carlo Methods Leon Gu CSD, CMU Approximate Inference EM: y-observed variables; x-hidden variables; θ-parameters; E-step: q(x) = p(x y, θ t 1 ) M-step: θ t = arg max E q(x) [log p(y, x θ)] θ Monte

More information

Lecture 6: Markov Chain Monte Carlo

Lecture 6: Markov Chain Monte Carlo Lecture 6: Markov Chain Monte Carlo D. Jason Koskinen koskinen@nbi.ku.dk Photo by Howard Jackman University of Copenhagen Advanced Methods in Applied Statistics Feb - Apr 2016 Niels Bohr Institute 2 Outline

More information

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A.

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A. 1. Let P be a probability measure on a collection of sets A. (a) For each n N, let H n be a set in A such that H n H n+1. Show that P (H n ) monotonically converges to P ( k=1 H k) as n. (b) For each n

More information

Slice Sampling Mixture Models

Slice Sampling Mixture Models Slice Sampling Mixture Models Maria Kalli, Jim E. Griffin & Stephen G. Walker Centre for Health Services Studies, University of Kent Institute of Mathematics, Statistics & Actuarial Science, University

More information

Bayesian Estimation with Sparse Grids

Bayesian Estimation with Sparse Grids Bayesian Estimation with Sparse Grids Kenneth L. Judd and Thomas M. Mertens Institute on Computational Economics August 7, 27 / 48 Outline Introduction 2 Sparse grids Construction Integration with sparse

More information

The Mixture Approach for Simulating New Families of Bivariate Distributions with Specified Correlations

The Mixture Approach for Simulating New Families of Bivariate Distributions with Specified Correlations The Mixture Approach for Simulating New Families of Bivariate Distributions with Specified Correlations John R. Michael, Significance, Inc. and William R. Schucany, Southern Methodist University The mixture

More information

(5) Multi-parameter models - Gibbs sampling. ST440/540: Applied Bayesian Analysis

(5) Multi-parameter models - Gibbs sampling. ST440/540: Applied Bayesian Analysis Summarizing a posterior Given the data and prior the posterior is determined Summarizing the posterior gives parameter estimates, intervals, and hypothesis tests Most of these computations are integrals

More information

Bayesian GLMs and Metropolis-Hastings Algorithm

Bayesian GLMs and Metropolis-Hastings Algorithm Bayesian GLMs and Metropolis-Hastings Algorithm We have seen that with conjugate or semi-conjugate prior distributions the Gibbs sampler can be used to sample from the posterior distribution. In situations,

More information

Introduction to Markov Chain Monte Carlo & Gibbs Sampling

Introduction to Markov Chain Monte Carlo & Gibbs Sampling Introduction to Markov Chain Monte Carlo & Gibbs Sampling Prof. Nicholas Zabaras Sibley School of Mechanical and Aerospace Engineering 101 Frank H. T. Rhodes Hall Ithaca, NY 14853-3801 Email: zabaras@cornell.edu

More information

PSEUDO-MARGINAL METROPOLIS-HASTINGS APPROACH AND ITS APPLICATION TO BAYESIAN COPULA MODEL

PSEUDO-MARGINAL METROPOLIS-HASTINGS APPROACH AND ITS APPLICATION TO BAYESIAN COPULA MODEL PSEUDO-MARGINAL METROPOLIS-HASTINGS APPROACH AND ITS APPLICATION TO BAYESIAN COPULA MODEL Xuebin Zheng Supervisor: Associate Professor Josef Dick Co-Supervisor: Dr. David Gunawan School of Mathematics

More information

CPSC 540: Machine Learning

CPSC 540: Machine Learning CPSC 540: Machine Learning MCMC and Non-Parametric Bayes Mark Schmidt University of British Columbia Winter 2016 Admin I went through project proposals: Some of you got a message on Piazza. No news is

More information

Markov Chain Monte Carlo

Markov Chain Monte Carlo Chapter 5 Markov Chain Monte Carlo MCMC is a kind of improvement of the Monte Carlo method By sampling from a Markov chain whose stationary distribution is the desired sampling distributuion, it is possible

More information

Surveying the Characteristics of Population Monte Carlo

Surveying the Characteristics of Population Monte Carlo International Research Journal of Applied and Basic Sciences 2013 Available online at www.irjabs.com ISSN 2251-838X / Vol, 7 (9): 522-527 Science Explorer Publications Surveying the Characteristics of

More information

Markov Chain Monte Carlo Lecture 4

Markov Chain Monte Carlo Lecture 4 The local-trap problem refers to that in simulations of a complex system whose energy landscape is rugged, the sampler gets trapped in a local energy minimum indefinitely, rendering the simulation ineffective.

More information

Markov chain Monte Carlo

Markov chain Monte Carlo Markov chain Monte Carlo Feng Li feng.li@cufe.edu.cn School of Statistics and Mathematics Central University of Finance and Economics Revised on April 24, 2017 Today we are going to learn... 1 Markov Chains

More information

STA 294: Stochastic Processes & Bayesian Nonparametrics

STA 294: Stochastic Processes & Bayesian Nonparametrics MARKOV CHAINS AND CONVERGENCE CONCEPTS Markov chains are among the simplest stochastic processes, just one step beyond iid sequences of random variables. Traditionally they ve been used in modelling a

More information

Bayesian Modelling of Extreme Rainfall Data

Bayesian Modelling of Extreme Rainfall Data Bayesian Modelling of Extreme Rainfall Data Elizabeth Smith A thesis submitted for the degree of Doctor of Philosophy at the University of Newcastle upon Tyne September 2005 UNIVERSITY OF NEWCASTLE Bayesian

More information

Computer Practical: Metropolis-Hastings-based MCMC

Computer Practical: Metropolis-Hastings-based MCMC Computer Practical: Metropolis-Hastings-based MCMC Andrea Arnold and Franz Hamilton North Carolina State University July 30, 2016 A. Arnold / F. Hamilton (NCSU) MH-based MCMC July 30, 2016 1 / 19 Markov

More information

Semi-Parametric Importance Sampling for Rare-event probability Estimation

Semi-Parametric Importance Sampling for Rare-event probability Estimation Semi-Parametric Importance Sampling for Rare-event probability Estimation Z. I. Botev and P. L Ecuyer IMACS Seminar 2011 Borovets, Bulgaria Semi-Parametric Importance Sampling for Rare-event probability

More information

Practical Bayesian Quantile Regression. Keming Yu University of Plymouth, UK

Practical Bayesian Quantile Regression. Keming Yu University of Plymouth, UK Practical Bayesian Quantile Regression Keming Yu University of Plymouth, UK (kyu@plymouth.ac.uk) A brief summary of some recent work of us (Keming Yu, Rana Moyeed and Julian Stander). Summary We develops

More information

An introduction to adaptive MCMC

An introduction to adaptive MCMC An introduction to adaptive MCMC Gareth Roberts MIRAW Day on Monte Carlo methods March 2011 Mainly joint work with Jeff Rosenthal. http://www2.warwick.ac.uk/fac/sci/statistics/crism/ Conferences and workshops

More information

Markov Chain Monte Carlo Inference. Siamak Ravanbakhsh Winter 2018

Markov Chain Monte Carlo Inference. Siamak Ravanbakhsh Winter 2018 Graphical Models Markov Chain Monte Carlo Inference Siamak Ravanbakhsh Winter 2018 Learning objectives Markov chains the idea behind Markov Chain Monte Carlo (MCMC) two important examples: Gibbs sampling

More information

Session 3A: Markov chain Monte Carlo (MCMC)

Session 3A: Markov chain Monte Carlo (MCMC) Session 3A: Markov chain Monte Carlo (MCMC) John Geweke Bayesian Econometrics and its Applications August 15, 2012 ohn Geweke Bayesian Econometrics and its Session Applications 3A: Markov () chain Monte

More information

Markov Chain Monte Carlo

Markov Chain Monte Carlo 1 Motivation 1.1 Bayesian Learning Markov Chain Monte Carlo Yale Chang In Bayesian learning, given data X, we make assumptions on the generative process of X by introducing hidden variables Z: p(z): prior

More information

April 20th, Advanced Topics in Machine Learning California Institute of Technology. Markov Chain Monte Carlo for Machine Learning

April 20th, Advanced Topics in Machine Learning California Institute of Technology. Markov Chain Monte Carlo for Machine Learning for for Advanced Topics in California Institute of Technology April 20th, 2017 1 / 50 Table of Contents for 1 2 3 4 2 / 50 History of methods for Enrico Fermi used to calculate incredibly accurate predictions

More information

MCMC and Gibbs Sampling. Kayhan Batmanghelich

MCMC and Gibbs Sampling. Kayhan Batmanghelich MCMC and Gibbs Sampling Kayhan Batmanghelich 1 Approaches to inference l Exact inference algorithms l l l The elimination algorithm Message-passing algorithm (sum-product, belief propagation) The junction

More information

Markov chain Monte Carlo

Markov chain Monte Carlo 1 / 26 Markov chain Monte Carlo Timothy Hanson 1 and Alejandro Jara 2 1 Division of Biostatistics, University of Minnesota, USA 2 Department of Statistics, Universidad de Concepción, Chile IAP-Workshop

More information

A Search and Jump Algorithm for Markov Chain Monte Carlo Sampling. Christopher Jennison. Adriana Ibrahim. Seminar at University of Kuwait

A Search and Jump Algorithm for Markov Chain Monte Carlo Sampling. Christopher Jennison. Adriana Ibrahim. Seminar at University of Kuwait A Search and Jump Algorithm for Markov Chain Monte Carlo Sampling Christopher Jennison Department of Mathematical Sciences, University of Bath, UK http://people.bath.ac.uk/mascj Adriana Ibrahim Institute

More information

Answers and expectations

Answers and expectations Answers and expectations For a function f(x) and distribution P(x), the expectation of f with respect to P is The expectation is the average of f, when x is drawn from the probability distribution P E

More information

Pseudo-marginal MCMC methods for inference in latent variable models

Pseudo-marginal MCMC methods for inference in latent variable models Pseudo-marginal MCMC methods for inference in latent variable models Arnaud Doucet Department of Statistics, Oxford University Joint work with George Deligiannidis (Oxford) & Mike Pitt (Kings) MCQMC, 19/08/2016

More information

Exercises Tutorial at ICASSP 2016 Learning Nonlinear Dynamical Models Using Particle Filters

Exercises Tutorial at ICASSP 2016 Learning Nonlinear Dynamical Models Using Particle Filters Exercises Tutorial at ICASSP 216 Learning Nonlinear Dynamical Models Using Particle Filters Andreas Svensson, Johan Dahlin and Thomas B. Schön March 18, 216 Good luck! 1 [Bootstrap particle filter for

More information

Stat 451 Lecture Notes Monte Carlo Integration

Stat 451 Lecture Notes Monte Carlo Integration Stat 451 Lecture Notes 06 12 Monte Carlo Integration Ryan Martin UIC www.math.uic.edu/~rgmartin 1 Based on Chapter 6 in Givens & Hoeting, Chapter 23 in Lange, and Chapters 3 4 in Robert & Casella 2 Updated:

More information

TEORIA BAYESIANA Ralph S. Silva

TEORIA BAYESIANA Ralph S. Silva TEORIA BAYESIANA Ralph S. Silva Departamento de Métodos Estatísticos Instituto de Matemática Universidade Federal do Rio de Janeiro Sumário Numerical Integration Polynomial quadrature is intended to approximate

More information

Parameter Estimation. William H. Jefferys University of Texas at Austin Parameter Estimation 7/26/05 1

Parameter Estimation. William H. Jefferys University of Texas at Austin Parameter Estimation 7/26/05 1 Parameter Estimation William H. Jefferys University of Texas at Austin bill@bayesrules.net Parameter Estimation 7/26/05 1 Elements of Inference Inference problems contain two indispensable elements: Data

More information

Simulation - Lectures - Part III Markov chain Monte Carlo

Simulation - Lectures - Part III Markov chain Monte Carlo Simulation - Lectures - Part III Markov chain Monte Carlo Julien Berestycki Part A Simulation and Statistical Programming Hilary Term 2018 Part A Simulation. HT 2018. J. Berestycki. 1 / 50 Outline Markov

More information

University of Toronto Department of Statistics

University of Toronto Department of Statistics Norm Comparisons for Data Augmentation by James P. Hobert Department of Statistics University of Florida and Jeffrey S. Rosenthal Department of Statistics University of Toronto Technical Report No. 0704

More information

Theory of Stochastic Processes 8. Markov chain Monte Carlo

Theory of Stochastic Processes 8. Markov chain Monte Carlo Theory of Stochastic Processes 8. Markov chain Monte Carlo Tomonari Sei sei@mist.i.u-tokyo.ac.jp Department of Mathematical Informatics, University of Tokyo June 8, 2017 http://www.stat.t.u-tokyo.ac.jp/~sei/lec.html

More information

Reminder of some Markov Chain properties:

Reminder of some Markov Chain properties: Reminder of some Markov Chain properties: 1. a transition from one state to another occurs probabilistically 2. only state that matters is where you currently are (i.e. given present, future is independent

More information

Metropolis Hastings. Rebecca C. Steorts Bayesian Methods and Modern Statistics: STA 360/601. Module 9

Metropolis Hastings. Rebecca C. Steorts Bayesian Methods and Modern Statistics: STA 360/601. Module 9 Metropolis Hastings Rebecca C. Steorts Bayesian Methods and Modern Statistics: STA 360/601 Module 9 1 The Metropolis-Hastings algorithm is a general term for a family of Markov chain simulation methods

More information