Randomized Quasi-Monte Carlo for MCMC

Size: px
Start display at page:

Download "Randomized Quasi-Monte Carlo for MCMC"

Transcription

1 Randomized Quasi-Monte Carlo for MCMC Radu Craiu 1 Christiane Lemieux 2 1 Department of Statistics, Toronto 2 Department of Statistics, Waterloo Third Workshop on Monte Carlo Methods Harvard, May 2007 Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27

2 Outline 1 Local Search/Optimization Algorithms: MTM 2 Multiple-Try Metropolis and variations Multi-distributed MTM Multiple-Correlated-Try Metropolis 3 Randomized Quasi-Monte Carlo Antithetic or Stratified Sampling? QMC and Randomized QMC: Discussion and Examples Latin Hypercube Sampling Korobov Rule Beyond uniformity 4 Examples 5 Bonus Algorithm - Multi-Annealed Metropolis Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27

3 Local Search and Optimization Algorithms (LSOA) Search locally a region of the state space to sample from. Adapt locally to improve the efficiency of the algorithm. Local Search and Optimization Algorithms (LSOA): Multiple-Try Metropolis (Liu, Liang and Wong, JASA 2000), Hit-and-Run (Chen and Schmeiser, JCGS 1993), Adaptive Direction Sampling (Gilks, Roberts and George, Stat. 1994), Delayed rejection (Green and Mira, Biomka, 2001), Directional Metropolis (Eidsvik and Tjelmeland, Statist. & Comput. 2006). Idea: Produce a more structured search using stratified sampling. Today s Topic: Discussion of RQMC implemented within LSOA. Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27

4 Original MTM (Liu, Liang and Wong, JASA 2000) Suppose T is a proposal distribution such that T(x y) > 0 T(y x) > 0 and λ(x, y) is a symmetric function. Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27

5 Original MTM (Liu, Liang and Wong, JASA 2000) Suppose T is a proposal distribution such that T(x y) > 0 T(y x) > 0 and λ(x, y) is a symmetric function. (i) Draw K independent trial proposals y 1,...,y K from T( x t ). Sample one with p i w(y i x t ) = π(y i )T(x t y i )λ(x t, y i ). Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27

6 Original MTM (Liu, Liang and Wong, JASA 2000) Suppose T is a proposal distribution such that T(x y) > 0 T(y x) > 0 and λ(x, y) is a symmetric function. (i) Draw K independent trial proposals y 1,...,y K from T( x t ). Sample one with p i w(y i x t ) = π(y i )T(x t y i )λ(x t, y i ). (ii) Generate x 1,..., x k 1 T( y) and put x k = x t. Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27

7 Original MTM (Liu, Liang and Wong, JASA 2000) Suppose T is a proposal distribution such that T(x y) > 0 T(y x) > 0 and λ(x, y) is a symmetric function. (i) Draw K independent trial proposals y 1,...,y K from T( x t ). Sample one with p i w(y i x t ) = π(y i )T(x t y i )λ(x t, y i ). (ii) Generate x 1,..., x k 1 T( y) and put x k = x t. { (iii) Accept y with probability min 1, P K i=1 w(y i x t ) P K i=1 w(x i y) }. Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27

8 Original MTM (Liu, Liang and Wong, JASA 2000) Suppose T is a proposal distribution such that T(x y) > 0 T(y x) > 0 and λ(x, y) is a symmetric function. (i) Draw K independent trial proposals y 1,...,y K from T( x t ). Sample one with p i w(y i x t ) = π(y i )T(x t y i )λ(x t, y i ). (ii) Generate x 1,..., x k 1 T( y) and put x k = x t. { (iii) Accept y with probability min 1, P K i=1 w(y i x t ) P K i=1 w(x i y) Do we better explore the sample space with K proposals? }. Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27

9 Original MTM (Liu, Liang and Wong, JASA 2000) Suppose T is a proposal distribution such that T(x y) > 0 T(y x) > 0 and λ(x, y) is a symmetric function. (i) Draw K independent trial proposals y 1,...,y K from T( x t ). Sample one with p i w(y i x t ) = π(y i )T(x t y i )λ(x t, y i ). (ii) Generate x 1,..., x k 1 T( y) and put x k = x t. { (iii) Accept y with probability min 1, P K i=1 w(y i x t ) P K i=1 w(x i y) Do we better explore the sample space with K proposals? As long as we take advantage of the flexibility provided by the MTM. }. Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27

10 Multi-distributed MTM (MD-MTM) The proposals do not have to be identically distributed: y j T j ( x) for 1 j k. Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27

11 Multi-distributed MTM (MD-MTM) The proposals do not have to be identically distributed: y j T j ( x) for 1 j k. If y = y j0 is selected than put xj 0 = x t and sample xj T j ( y). May accomodate different sampling regimes. Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27

12 Multi-distributed MTM (MD-MTM) The proposals do not have to be identically distributed: y j T j ( x) for 1 j k. If y = y j0 is selected than put xj 0 = x t and sample xj T j ( y). May accomodate different sampling regimes. π(x) = 0.3N(( 20, 20) T,Σ 1 )+0.4N((0, 0) T,Σ 2 )+0.3N((20, 20) T,Σ 3 ), where Σ 1 = diag(0.5, 0.5), Σ 2 = diag(2, 2) and Σ 3 = diag(0.4, 0.4). Use MD-MTM for a random walk Metropolis with Gaussian proposal and variances σ {5, 50, 100, 150, 200}. Compare with MTM for a random walk Metropolis constructed with only one of the σ s. Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27

13 MD MTM sigma= Acc. Rate = Acc. Rate = sigma=20 sigma= Acc. Rate = Acc. Rate = Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27

14 Multiple-Correlated-Try Metropolis (MCTM) The proposals do not have to be independent. Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27

15 Multiple-Correlated-Try Metropolis (MCTM) The proposals do not have to be independent. MCTM(I) Suppose we sample K trial proposals y 1,...,y K from T(y 1,..., y K x t ) where T(y 1,..., y K x t )dy [ i] = T(y i x t ), 1 i K. Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27

16 Multiple-Correlated-Try Metropolis (MCTM) The proposals do not have to be independent. MCTM(I) Suppose we sample K trial proposals y 1,...,y K from T(y 1,..., y K x t ) where T(y 1,..., y K x t )dy [ i] = T(y i x t ), 1 i K. The algorithm proceeds as in the independent case with one exception. Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27

17 Multiple-Correlated-Try Metropolis (MCTM) The proposals do not have to be independent. MCTM(I) Suppose we sample K trial proposals y 1,...,y K from T(y 1,..., y K x t ) where T(y 1,..., y K x t )dy [ i] = T(y i x t ), 1 i K. The algorithm proceeds as in the independent case with one exception. MCTM (II) Draw (x1,..., x K 1 ) variates from the conditional transition kernel T(x 1,..., x K 1 y, x K = x t ) and let xk = x t. Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27

18 Multiple-Correlated-Try Metropolis (MCTM) The proposals do not have to be independent. MCTM(I) Suppose we sample K trial proposals y 1,...,y K from T(y 1,..., y K x t ) where T(y 1,..., y K x t )dy [ i] = T(y i x t ), 1 i K. The algorithm proceeds as in the independent case with one exception. MCTM (II) Draw (x1,..., x K 1 ) variates from the conditional transition kernel T(x 1,..., x K 1 y, x K = x t ) and let xk = x t. We have freedom in choosing T as long as we can perform the conditional sampling step. Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27

19 Antithetic or Stratified Sampling? Y j = (Y j1,...,y jd ) R d for 1 j K. Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27

20 Antithetic or Stratified Sampling? Y j = (Y j1,...,y jd ) R d for 1 j K. Say K = 2, Var(Y) = Σ = diag(σ1 2,...,σ2 d ), Cov(Y 1h, Y 2h ) = σh 2ρ h. Then [ [ d ] E Y 1 Y 2 2] d = E (Y 1h Y 2h ) 2 = σh 2 (1 ρ h). h=1 h=1 Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27

21 Antithetic or Stratified Sampling? Y j = (Y j1,...,y jd ) R d for 1 j K. Say K = 2, Var(Y) = Σ = diag(σ1 2,...,σ2 d ), Cov(Y 1h, Y 2h ) = σh 2ρ h. Then [ [ d ] E Y 1 Y 2 2] d = E (Y 1h Y 2h ) 2 = σh 2 (1 ρ h). h=1 One possible choice is ρ h = 1/(K 1) (covariance matrix needs to be positive definite!). The optimal choice of ρ h depends on σ h, 1 h d. h=1 Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27

22 Antithetic or Stratified Sampling? Y j = (Y j1,...,y jd ) R d for 1 j K. Say K = 2, Var(Y) = Σ = diag(σ1 2,...,σ2 d ), Cov(Y 1h, Y 2h ) = σh 2ρ h. Then [ [ d ] E Y 1 Y 2 2] d = E (Y 1h Y 2h ) 2 = σh 2 (1 ρ h). h=1 One possible choice is ρ h = 1/(K 1) (covariance matrix needs to be positive definite!). The optimal choice of ρ h depends on σ h, 1 h d. Conditional sampling needed for MCTM is straightforward in the case of Gaussian proposals. h=1 Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27

23 Antithetic or Stratified Sampling? Y j = (Y j1,...,y jd ) R d for 1 j K. Say K = 2, Var(Y) = Σ = diag(σ1 2,...,σ2 d ), Cov(Y 1h, Y 2h ) = σh 2ρ h. Then [ [ d ] E Y 1 Y 2 2] d = E (Y 1h Y 2h ) 2 = σh 2 (1 ρ h). h=1 One possible choice is ρ h = 1/(K 1) (covariance matrix needs to be positive definite!). The optimal choice of ρ h depends on σ h, 1 h d. Conditional sampling needed for MCTM is straightforward in the case of Gaussian proposals. If Y j = ψ(x t, U j ) with ψ increasing in U j then we need to use U s which are negatively associated (NA - Craiu and Meng, AnnStat. 2005). Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27 h=1

24 Quasi-Monte Carlo (QMC) QMC is a de-randomized MC; the samples are constructed deterministically. QMC: highly uniform and stratified sampling in the unit hypercube. Interest: high-uniformity, a.k.a. low discrepancy. Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27

25 Quasi-Monte Carlo (QMC) QMC is a de-randomized MC; the samples are constructed deterministically. QMC: highly uniform and stratified sampling in the unit hypercube. Interest: high-uniformity, a.k.a. low discrepancy. For v (0, 1) d, B(v) = {u (0, 1) d : 0 < u j v j, 1 j d}. Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27

26 Quasi-Monte Carlo (QMC) QMC is a de-randomized MC; the samples are constructed deterministically. QMC: highly uniform and stratified sampling in the unit hypercube. Interest: high-uniformity, a.k.a. low discrepancy. For v (0, 1) d, B(v) = {u (0, 1) d : 0 < u j v j, 1 j d}. For a point set S n define α(s n, v) = card{s n B(v)}. Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27

27 Quasi-Monte Carlo (QMC) QMC is a de-randomized MC; the samples are constructed deterministically. QMC: highly uniform and stratified sampling in the unit hypercube. Interest: high-uniformity, a.k.a. low discrepancy. For v (0, 1) d, B(v) = {u (0, 1) d : 0 < u j v j, 1 j d}. For a point set S n define α(s n, v) = card{s n B(v)}. Star-discrepancy D n (S n) = sup v (0,1) d v 1... v d α(s n, v)/n. Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27

28 Randomized QMC (RQMC) Efficiency can be connected to discrepancy. Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27

29 Randomized QMC (RQMC) Efficiency can be connected to discrepancy. Koksma-Hlawka Theorem µ 1 f(u) n u S n D n V(f). Both V(f) and D are difficult to compute in general. Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27

30 Randomized QMC (RQMC) Efficiency can be connected to discrepancy. Koksma-Hlawka Theorem µ 1 f(u) n u S n D n V(f). Both V(f) and D are difficult to compute in general. Randomized versions of QMC (RQMC) algorithms are used to establish error estimates via Monte Carlo simulation. (Art Owen, Stanford; Christiane Lemieux, Waterloo; Pierre L Ecuyer, Montréal). Randomization is performed so that it preserves the low-discrepancy of the set point. Makes QMC set points suitable for MC as marginals are uniform. RQMC and MCMC (Owen and Tribble, PNAS 05; Lemieux and Sidorsky, Math. & Comp. Model., 2006). Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27

31 Latin Hypercube Sampling (LHS) Simultaneous stratification of the univariate marginals if U Uniform(0, 1) d. Construction: For 1 i K, 1 j d: Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27

32 Latin Hypercube Sampling (LHS) Simultaneous stratification of the univariate marginals if U Uniform(0, 1) d. Construction: For 1 i K, 1 j d: i) V ij Uniform(0, 1) are i.i.d Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27

33 Latin Hypercube Sampling (LHS) Simultaneous stratification of the univariate marginals if U Uniform(0, 1) d. Construction: For 1 i K, 1 j d: i) V ij Uniform(0, 1) are i.i.d ii) {τ j, 1 j d} are independent permutations of {0,..., K 1} Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27

34 Latin Hypercube Sampling (LHS) Simultaneous stratification of the univariate marginals if U Uniform(0, 1) d. Construction: For 1 i K, 1 j d: i) V ij Uniform(0, 1) are i.i.d ii) {τ j, 1 j d} are independent permutations of {0,..., K 1} U ij = τ j(i 1) + V ij. K Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27

35 Latin Hypercube Sampling (LHS) Simultaneous stratification of the univariate marginals if U Uniform(0, 1) d. Construction: For 1 i K, 1 j d: i) V ij Uniform(0, 1) are i.i.d ii) {τ j, 1 j d} are independent permutations of {0,..., K 1} U ij = τ j(i 1) + V ij. K U (i) is the ith row of the matrix U. Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27

36 Latin Hypercube Sampling (LHS) Simultaneous stratification of the univariate marginals if U Uniform(0, 1) d. Construction: For 1 i K, 1 j d: i) V ij Uniform(0, 1) are i.i.d ii) {τ j, 1 j d} are independent permutations of {0,..., K 1} U ij = τ j(i 1) + V ij. K U (i) is the ith row of the matrix U. Given U (i) = (U i1,...,u id ) (0, 1) d, 1 i K we can use: a) For 1 j d fixed: {U 1j,...,U (Kj) } as a set of NA Uniform(0, 1) r.v. s. b) (U (1),..., U (K) ) as a stratified sample of (0, 1) d. Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27

37 LHS and Gaussian sampling: K = 64, d = 2 Independent Bivariate normal y x LHS Bivariate normal y x Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27

38 LHS in MTM Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27

39 LHS in MTM V i U(0, 1), U i = i+v i K, Y i = F 1 X t (U i ), 0 i K 1 (no permutation is needed!). Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27

40 LHS in MTM V i U(0, 1), U i = i+v i K, Y i = F 1 X t (U i ), 0 i K 1 (no permutation is needed!). If Y = Y i0 is selected: Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27

41 LHS in MTM V i U(0, 1), U i = i+v i K, Y i = F 1 X t (U i ), 0 i K 1 (no permutation is needed!). If Y = Y i0 is selected: Intuitively, if Y has been generated using a uniform from the interval, ( i 0 K, i 0+1 K ), then we want to sample x 2,..., x k using all the K 1 uniforms which ARE NOT IN ( i 0 K, i 0+1 K ). Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27

42 LHS in MTM V i U(0, 1), U i = i+v i K, Y i = F 1 X t (U i ), 0 i K 1 (no permutation is needed!). If Y = Y i0 is selected: Intuitively, if Y has been generated using a uniform from the interval, ( i 0 K, i 0+1 K ), then we want to sample x 2,..., x k using all the K 1 uniforms which ARE NOT IN ( i 0 K, i 0+1 K ). i) Put j 0 = [KF y (X t )] (where [ ] is integer part) and construct a random permutation τ such that τ(j 0 ) = i 0. Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27

43 LHS in MTM V i U(0, 1), U i = i+v i K, Y i = F 1 X t (U i ), 0 i K 1 (no permutation is needed!). If Y = Y i0 is selected: Intuitively, if Y has been generated using a uniform from the interval, ( i 0 K, i 0+1 K ), then we want to sample x 2,..., x k using all the K 1 uniforms which ARE NOT IN ( i 0 K, i 0+1 K ). i) Put j 0 = [KF y (X t )] (where [ ] is integer part) and construct a random permutation τ such that τ(j 0 ) = i 0. ii) Put xj 0 = X t, and for i j 0 Ui = τ(i)+w i K and xi = Fy 1 (Ui ), with W i Uniform(0, 1) independent from the V s. Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27

44 LHS in MTM V i U(0, 1), U i = i+v i K, Y i = F 1 X t (U i ), 0 i K 1 (no permutation is needed!). If Y = Y i0 is selected: Intuitively, if Y has been generated using a uniform from the interval, ( i 0 K, i 0+1 K ), then we want to sample x 2,..., x k using all the K 1 uniforms which ARE NOT IN ( i 0 K, i 0+1 K ). i) Put j 0 = [KF y (X t )] (where [ ] is integer part) and construct a random permutation τ such that τ(j 0 ) = i 0. ii) Put xj 0 = X t, and for i j 0 Ui = τ(i)+w i K and xi = Fy 1 (Ui ), with W i Uniform(0, 1) independent from the V s. One can get away without permuting at all!! Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27

45 Korobov Rules K points in [0, 1] d are generated using a generator a {1,...,K 1}. Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27

46 Korobov Rules K points in [0, 1] d are generated using a generator a {1,...,K 1}. S K = { i K (1, a, a2 mod K,...,a d 1 mod K) mod 1 } 0 i K 1. Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27

47 Korobov Rules K points in [0, 1] d are generated using a generator a {1,...,K 1}. S K = { i K (1, a, a2 mod K,...,a d 1 mod K) mod 1 } 0 i K 1. Bad choices: a and K are not relatively prime or a = 1. The point set is fully projection regular, i.e. any lower-dimensional projection contains K distinct points. Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27

48 Korobov Rules K points in [0, 1] d are generated using a generator a {1,...,K 1}. S K = { i K (1, a, a2 mod K,...,a d 1 mod K) mod 1 } 0 i K 1. Bad choices: a and K are not relatively prime or a = 1. The point set is fully projection regular, i.e. any lower-dimensional projection contains K distinct points. To randomize add a vector U K Uniform(0, 1) d mod 1 i.e. S R K = S K + U K mod 1. Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27

49 Example: Korobov rule and Gaussian sampling Independent N=128, a=47 y x Randomized Korobov y Radu Craiu (fisher.utstat.toronto.edu/craiu/) x RQMC for MCMC 3rd WMCM - May, / 27

50 Going beyond uniformity Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27

51 Going beyond uniformity Sometimes we want to "favor" regions in the hypercube but uniformity is lost once we transform the QMC point set. Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27

52 Going beyond uniformity Sometimes we want to "favor" regions in the hypercube but uniformity is lost once we transform the QMC point set. g : [0, 1] [0, 1] is a bijection and Ũ = g(u) = (g(u 1),..., g(u d )). Y T( X) Y = ψ(x, U) so if Y = ψ(x, Ũ) then Y T( X). Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27

53 Going beyond uniformity Sometimes we want to "favor" regions in the hypercube but uniformity is lost once we transform the QMC point set. g : [0, 1] [0, 1] is a bijection and Ũ = g(u) = (g(u 1),..., g(u d )). Y T( X) Y = ψ(x, U) so if Y = ψ(x, Ũ) then Y T( X). If λ(x, Y) = T(Y X)/ T(Y X) is symmetric, i.e. λ(x, Y) = λ(y, X), then MCTM can be implemented. without computing T. Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27

54 Going beyond uniformity Sometimes we want to "favor" regions in the hypercube but uniformity is lost once we transform the QMC point set. g : [0, 1] [0, 1] is a bijection and Ũ = g(u) = (g(u 1),..., g(u d )). Y T( X) Y = ψ(x, U) so if Y = ψ(x, Ũ) then Y T( X). If λ(x, Y) = T(Y X)/ T(Y X) is symmetric, i.e. λ(x, Y) = λ(y, X), then MCTM can be implemented. without computing T. Example: sin[π(u 0.5)] + 1 g(u) =. 2 i) g is bijective. ii) g(x) = x, x {0, 1/2, 1}, g(u) + g(1 u) = 1, u (0, 1). Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27

55 Example Randomized Korobov y x Transformed Korobov y Radu Craiu (fisher.utstat.toronto.edu/craiu/) x RQMC for MCMC 3rd WMCM - May, / 27

56 Gaussian Proposals Suppose T(y x) is a multivariate Gaussian with no correlation. Then T(y x) = d j=1 e (y j x j ) 2 /2σ 2 j 2πσ 2 j T(y x) = d j=1 e (y j x j ) 2 /2σ 2 j 2πσ 2 j π 1 [ ]. Φ( y j x j σ j ) 1 Φ( y j x j σ j ) λ(y, x) = d j=1 1 [ π 1 Φ( x j y j σ j ) ] = λ(x, y). Φ( x j y j σ j ) Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27

57 Bimodal density Use MCTM combined with random-ray Monte Carlo to sample from (Gelman and Meng, Am. Stat. 91) π(x 1, x 2 ) exp{ (9x 2 1 x x x 2 2 8x 1 8x 2 )/2} Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27

58 Efficiency Comparison At each iteration a random direction v is generated; Along direction v we sample proposals y 1,..., y k around the current state x t using y i = x t + r i with r i Uniform( σ,σ). Table: Values of the MSE reduction factor R = MSE anti MSE ind. σ\k Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27

59 Lupus Data Lupus Data (van Dyk and Meng, JCGS 2001): 55 patients, 2 covariates; logistic model. The posterior density is proportional to 2 π(β x, y) j=0 e 0.5β j/ π 55 i=1 [ ] exp(x T yi [ ] 1 yi i β) exp(xi T β) 1 + exp(xi T. β) Sample using MCTM with antithetically coupled proposals or stratified proposals via a randomized Korobov set shifted using the transformation g. Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27

60 Efficiency comparison for Lupus Data Table: Values of R for β 1 and p 25 = 1 {β1 >25} in the logit example. Antithetic QMC k\σ / / / / / / / / / / / / / / / / / / / / / / / /0.75 Table: Computation Times k = 8 Sample Size MTM MCTM (A) MCTM (S) Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27

61 Multi-Annealed Metropolis (MAM) Inspired by Neal s (Stat. & Comp., 1996) tempered transition method. Uses a series of distributions {π t } t {1=T0,T 1,...,T M }, constructed between the distribution of interest, π, and π TM. Relies on a "path" that crosses all the intermediate distributions. If 1 = T 0 < T 1 < T 2 = T and Q Ti is a proposal distribution for π Ti π 1/T i. Suppose that Q Ti (X Y) = Q Ti (Y X) (e.g., random walk Metropolis). Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27

62 MAM - The algorithm Π 0 X t Y 1 Π 1 X 1 Y 2 Π 2 X 2 Given the chain is in state X: i) Sample X 1 Q 1 ( X), X 2 Q 2 {( X 1 ), Y 2 Q 2 ( X 2 ),} Y 1 Q 1 ( Y 2 ). ii) Accept Y 1 with probability min 1, π 0(Y 1 ) π 0 (X) π2(x 1 )π 1 (Y 2 ) π 1 (X 1 )π 2 (Y 2 ). Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27

63 Example MAM(4) T=(2,8,10,15,20, 30) sigma=(2,3,4,5,8,15) Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for0mcmc 10 3rd WMCM 20 - May, / 27

64 Multiple-try extension of the MAM Π 0 X t Y 1 Y 1 Π 1 X 1 X 1 Y 2 Y 2 Π 2 X 2 X 2 Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27

65 Multiple-try extension of the MAM Π 0 X t Y 1 Y 1 Π 1 X 1 X 1 Y 2 Y 2 Π 2 X 2 X 2 MTM can be implemented by generating multiple paths and compensate using the same number of reverse paths. We have freedom in choosing the "weight of a path". Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27

66 Multiple-try extension of the MAM Π 0 X t Y 1 Y 1 Π 1 X 1 X 1 Y 2 Y 2 Π 2 X 2 X 2 MTM can be implemented by generating multiple paths and compensate using the same number of reverse paths. We have freedom in choosing the "weight of a path". MAM Pros : increases the acceptance rate >10 fold, does not require running the parallel chains all the time; Cons: Requires symmetric proposals. Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27

67 Negative Association for Multiple-Try MAM Suppose that X i Q i ( X i 1 ) X i = ψ i (X i 1, U i ) with U i Uniform[0, 1] d and ψ i is monotone in each component of U i. Y 1 = ψ 1 (ψ 2 (ψ 2 (ψ 1 (X, U 1 ), U 2 ), V 2 ), V 1 ) and Y 1 = ψ 1(ψ 2 (ψ 2 (ψ 1 (X, U 1 ), U 2 ), V 2 ), V 1 ). The paths do not need to be generated independently. If (U i, U i ) and (V i, V i ) are negatively associated then Corr(Y 1, Y 1 ) < 0 so the average distance between the paths is larger than under independence. Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27

Acceleration of the Multiple-Try Metropolis using Antithetic and Stratified Sampling

Acceleration of the Multiple-Try Metropolis using Antithetic and Stratified Sampling Acceleration of the Multiple-Try Metropolis using Antithetic and Stratified Sampling Radu V Craiu 1 Christiane Lemieux 2 October 4, 2005 Abstract The Multiple-Try Metropolis is a recent extension of the

More information

Markov Chain Monte Carlo

Markov Chain Monte Carlo Markov Chain Monte Carlo Recall: To compute the expectation E ( h(y ) ) we use the approximation E(h(Y )) 1 n n h(y ) t=1 with Y (1),..., Y (n) h(y). Thus our aim is to sample Y (1),..., Y (n) from f(y).

More information

A Search and Jump Algorithm for Markov Chain Monte Carlo Sampling. Christopher Jennison. Adriana Ibrahim. Seminar at University of Kuwait

A Search and Jump Algorithm for Markov Chain Monte Carlo Sampling. Christopher Jennison. Adriana Ibrahim. Seminar at University of Kuwait A Search and Jump Algorithm for Markov Chain Monte Carlo Sampling Christopher Jennison Department of Mathematical Sciences, University of Bath, UK http://people.bath.ac.uk/mascj Adriana Ibrahim Institute

More information

Markov Chain Monte Carlo Lecture 1

Markov Chain Monte Carlo Lecture 1 What are Monte Carlo Methods? The subject of Monte Carlo methods can be viewed as a branch of experimental mathematics in which one uses random numbers to conduct experiments. Typically the experiments

More information

CTDL-Positive Stable Frailty Model

CTDL-Positive Stable Frailty Model CTDL-Positive Stable Frailty Model M. Blagojevic 1, G. MacKenzie 2 1 Department of Mathematics, Keele University, Staffordshire ST5 5BG,UK and 2 Centre of Biostatistics, University of Limerick, Ireland

More information

Randomized Quasi-Monte Carlo Simulation of Markov Chains with an Ordered State Space

Randomized Quasi-Monte Carlo Simulation of Markov Chains with an Ordered State Space Randomized Quasi-Monte Carlo Simulation of Markov Chains with an Ordered State Space Pierre L Ecuyer 1, Christian Lécot 2, and Bruno Tuffin 3 1 Département d informatique et de recherche opérationnelle,

More information

A generalization of the Multiple-try Metropolis algorithm for Bayesian estimation and model selection

A generalization of the Multiple-try Metropolis algorithm for Bayesian estimation and model selection A generalization of the Multiple-try Metropolis algorithm for Bayesian estimation and model selection Silvia Pandolfi Francesco Bartolucci Nial Friel University of Perugia, IT University of Perugia, IT

More information

University of Toronto Department of Statistics

University of Toronto Department of Statistics Norm Comparisons for Data Augmentation by James P. Hobert Department of Statistics University of Florida and Jeffrey S. Rosenthal Department of Statistics University of Toronto Technical Report No. 0704

More information

Metropolis-Hastings Algorithm

Metropolis-Hastings Algorithm Strength of the Gibbs sampler Metropolis-Hastings Algorithm Easy algorithm to think about. Exploits the factorization properties of the joint probability distribution. No difficult choices to be made to

More information

Tutorial on quasi-monte Carlo methods

Tutorial on quasi-monte Carlo methods Tutorial on quasi-monte Carlo methods Josef Dick School of Mathematics and Statistics, UNSW, Sydney, Australia josef.dick@unsw.edu.au Comparison: MCMC, MC, QMC Roughly speaking: Markov chain Monte Carlo

More information

Brief introduction to Markov Chain Monte Carlo

Brief introduction to Markov Chain Monte Carlo Brief introduction to Department of Probability and Mathematical Statistics seminar Stochastic modeling in economics and finance November 7, 2011 Brief introduction to Content 1 and motivation Classical

More information

Computer Practical: Metropolis-Hastings-based MCMC

Computer Practical: Metropolis-Hastings-based MCMC Computer Practical: Metropolis-Hastings-based MCMC Andrea Arnold and Franz Hamilton North Carolina State University July 30, 2016 A. Arnold / F. Hamilton (NCSU) MH-based MCMC July 30, 2016 1 / 19 Markov

More information

Control Variates for Markov Chain Monte Carlo

Control Variates for Markov Chain Monte Carlo Control Variates for Markov Chain Monte Carlo Dellaportas, P., Kontoyiannis, I., and Tsourti, Z. Dept of Statistics, AUEB Dept of Informatics, AUEB 1st Greek Stochastics Meeting Monte Carlo: Probability

More information

Scenario Generation and Sampling Methods

Scenario Generation and Sampling Methods Scenario Generation and Sampling Methods Tito Homem-de-Mello Güzin Bayraksan SVAN 2016 IMPA 9 13 May 2106 Let s Start with a Recap Sample Average Approximation We want to solve the true problem min {g(x)

More information

Bayesian Inference in GLMs. Frequentists typically base inferences on MLEs, asymptotic confidence

Bayesian Inference in GLMs. Frequentists typically base inferences on MLEs, asymptotic confidence Bayesian Inference in GLMs Frequentists typically base inferences on MLEs, asymptotic confidence limits, and log-likelihood ratio tests Bayesians base inferences on the posterior distribution of the unknowns

More information

Constructions of Weakly CUD Sequences for MCMC

Constructions of Weakly CUD Sequences for MCMC Constructions of Weakly CUD Sequences for MCMC Seth D. Tribble Stanford University Art B. Owen Stanford University November 2005 Revised February 2006 Abstract In Markov chain Monte Carlo (MCMC) sampling

More information

Multivariate Normal & Wishart

Multivariate Normal & Wishart Multivariate Normal & Wishart Hoff Chapter 7 October 21, 2010 Reading Comprehesion Example Twenty-two children are given a reading comprehsion test before and after receiving a particular instruction method.

More information

A Repelling-Attracting Metropolis Algorithm for Multimodality

A Repelling-Attracting Metropolis Algorithm for Multimodality A Repelling-Attracting Metropolis Algorithm for Multimodality arxiv:1601.05633v5 [stat.me] 20 Oct 2017 Hyungsuk Tak Statistical and Applied Mathematical Sciences Institute Xiao-Li Meng Department of Statistics,

More information

Mathematical Methods for Neurosciences. ENS - Master MVA Paris 6 - Master Maths-Bio ( )

Mathematical Methods for Neurosciences. ENS - Master MVA Paris 6 - Master Maths-Bio ( ) Mathematical Methods for Neurosciences. ENS - Master MVA Paris 6 - Master Maths-Bio (2014-2015) Etienne Tanré - Olivier Faugeras INRIA - Team Tosca October 22nd, 2014 E. Tanré (INRIA - Team Tosca) Mathematical

More information

Winter 2019 Math 106 Topics in Applied Mathematics. Lecture 9: Markov Chain Monte Carlo

Winter 2019 Math 106 Topics in Applied Mathematics. Lecture 9: Markov Chain Monte Carlo Winter 2019 Math 106 Topics in Applied Mathematics Data-driven Uncertainty Quantification Yoonsang Lee (yoonsang.lee@dartmouth.edu) Lecture 9: Markov Chain Monte Carlo 9.1 Markov Chain A Markov Chain Monte

More information

Monte Carlo methods for sampling-based Stochastic Optimization

Monte Carlo methods for sampling-based Stochastic Optimization Monte Carlo methods for sampling-based Stochastic Optimization Gersende FORT LTCI CNRS & Telecom ParisTech Paris, France Joint works with B. Jourdain, T. Lelièvre, G. Stoltz from ENPC and E. Kuhn from

More information

Computational statistics

Computational statistics Computational statistics Markov Chain Monte Carlo methods Thierry Denœux March 2017 Thierry Denœux Computational statistics March 2017 1 / 71 Contents of this chapter When a target density f can be evaluated

More information

QUASI-MONTE CARLO VIA LINEAR SHIFT-REGISTER SEQUENCES. Pierre L Ecuyer Christiane Lemieux

QUASI-MONTE CARLO VIA LINEAR SHIFT-REGISTER SEQUENCES. Pierre L Ecuyer Christiane Lemieux Proceedings of the 1999 Winter Simulation Conference P. A. Farrington, H. B. Nembhard, D. T. Sturrock, and G. W. Evans, eds. QUASI-MONTE CARLO VIA LINEAR SHIFT-REGISTER SEQUENCES Pierre L Ecuyer Christiane

More information

April 20th, Advanced Topics in Machine Learning California Institute of Technology. Markov Chain Monte Carlo for Machine Learning

April 20th, Advanced Topics in Machine Learning California Institute of Technology. Markov Chain Monte Carlo for Machine Learning for for Advanced Topics in California Institute of Technology April 20th, 2017 1 / 50 Table of Contents for 1 2 3 4 2 / 50 History of methods for Enrico Fermi used to calculate incredibly accurate predictions

More information

On the Optimal Scaling of the Modified Metropolis-Hastings algorithm

On the Optimal Scaling of the Modified Metropolis-Hastings algorithm On the Optimal Scaling of the Modified Metropolis-Hastings algorithm K. M. Zuev & J. L. Beck Division of Engineering and Applied Science California Institute of Technology, MC 4-44, Pasadena, CA 925, USA

More information

Kernel adaptive Sequential Monte Carlo

Kernel adaptive Sequential Monte Carlo Kernel adaptive Sequential Monte Carlo Ingmar Schuster (Paris Dauphine) Heiko Strathmann (University College London) Brooks Paige (Oxford) Dino Sejdinovic (Oxford) December 7, 2015 1 / 36 Section 1 Outline

More information

Recent Advances in Regional Adaptation for MCMC

Recent Advances in Regional Adaptation for MCMC Recent Advances in Regional Adaptation for MCMC Radu Craiu Department of Statistics University of Toronto Collaborators: Yan Bai (Statistics, Toronto) Antonio Fabio di Narzo (Statistics, Bologna) Jeffrey

More information

1. Introduction. In this paper we consider stochastic optimization problems of the form

1. Introduction. In this paper we consider stochastic optimization problems of the form O RATES OF COVERGECE FOR STOCHASTIC OPTIMIZATIO PROBLEMS UDER O-I.I.D. SAMPLIG TITO HOMEM-DE-MELLO Abstract. In this paper we discuss the issue of solving stochastic optimization problems by means of sample

More information

Convergence Rate of Markov Chains

Convergence Rate of Markov Chains Convergence Rate of Markov Chains Will Perkins April 16, 2013 Convergence Last class we saw that if X n is an irreducible, aperiodic, positive recurrent Markov chain, then there exists a stationary distribution

More information

Resampling Markov chain Monte Carlo algorithms: Basic analysis and empirical comparisons. Zhiqiang Tan 1. February 2014

Resampling Markov chain Monte Carlo algorithms: Basic analysis and empirical comparisons. Zhiqiang Tan 1. February 2014 Resampling Markov chain Monte Carlo s: Basic analysis and empirical comparisons Zhiqiang Tan 1 February 2014 Abstract. Sampling from complex distributions is an important but challenging topic in scientific

More information

A Randomized Quasi-Monte Carlo Simulation Method for Markov Chains

A Randomized Quasi-Monte Carlo Simulation Method for Markov Chains A Randomized Quasi-Monte Carlo Simulation Method for Markov Chains Pierre L Ecuyer GERAD and Département d Informatique et de Recherche Opérationnelle Université de Montréal, C.P. 6128, Succ. Centre-Ville,

More information

Markov Chain Monte Carlo methods

Markov Chain Monte Carlo methods Markov Chain Monte Carlo methods Tomas McKelvey and Lennart Svensson Signal Processing Group Department of Signals and Systems Chalmers University of Technology, Sweden November 26, 2012 Today s learning

More information

Markov Chain Monte Carlo (MCMC)

Markov Chain Monte Carlo (MCMC) Markov Chain Monte Carlo (MCMC Dependent Sampling Suppose we wish to sample from a density π, and we can evaluate π as a function but have no means to directly generate a sample. Rejection sampling can

More information

6 Markov Chain Monte Carlo (MCMC)

6 Markov Chain Monte Carlo (MCMC) 6 Markov Chain Monte Carlo (MCMC) The underlying idea in MCMC is to replace the iid samples of basic MC methods, with dependent samples from an ergodic Markov chain, whose limiting (stationary) distribution

More information

Doing Bayesian Integrals

Doing Bayesian Integrals ASTR509-13 Doing Bayesian Integrals The Reverend Thomas Bayes (c.1702 1761) Philosopher, theologian, mathematician Presbyterian (non-conformist) minister Tunbridge Wells, UK Elected FRS, perhaps due to

More information

arxiv: v1 [stat.me] 10 Jul 2009

arxiv: v1 [stat.me] 10 Jul 2009 6th St.Petersburg Workshop on Simulation (2009) 1091-1096 Improvement of random LHD for high dimensions arxiv:0907.1823v1 [stat.me] 10 Jul 2009 Andrey Pepelyshev 1 Abstract Designs of experiments for multivariate

More information

STA 294: Stochastic Processes & Bayesian Nonparametrics

STA 294: Stochastic Processes & Bayesian Nonparametrics MARKOV CHAINS AND CONVERGENCE CONCEPTS Markov chains are among the simplest stochastic processes, just one step beyond iid sequences of random variables. Traditionally they ve been used in modelling a

More information

CSCI-6971 Lecture Notes: Monte Carlo integration

CSCI-6971 Lecture Notes: Monte Carlo integration CSCI-6971 Lecture otes: Monte Carlo integration Kristopher R. Beevers Department of Computer Science Rensselaer Polytechnic Institute beevek@cs.rpi.edu February 21, 2006 1 Overview Consider the following

More information

On the flexibility of Metropolis-Hastings acceptance probabilities in auxiliary variable proposal generation

On the flexibility of Metropolis-Hastings acceptance probabilities in auxiliary variable proposal generation On the flexibility of Metropolis-Hastings acceptance probabilities in auxiliary variable proposal generation Geir Storvik Department of Mathematics and (sfi) 2 Statistics for Innovation, University of

More information

Monte Carlo Methods. Leon Gu CSD, CMU

Monte Carlo Methods. Leon Gu CSD, CMU Monte Carlo Methods Leon Gu CSD, CMU Approximate Inference EM: y-observed variables; x-hidden variables; θ-parameters; E-step: q(x) = p(x y, θ t 1 ) M-step: θ t = arg max E q(x) [log p(y, x θ)] θ Monte

More information

MCMC algorithms for Subset Simulation

MCMC algorithms for Subset Simulation To appear in Probabilistic Engineering Mechanics June 2015 MCMC algorithms for Subset Simulation Iason Papaioannou *, Wolfgang Betz, Kilian Zwirglmaier, Daniel Straub Engineering Risk Analysis Group, Technische

More information

Directional Metropolis Hastings algorithms on hyperplanes

Directional Metropolis Hastings algorithms on hyperplanes Directional Metropolis Hastings algorithms on hyperplanes Hugo Hammer and Håon Tjelmeland Department of Mathematical Sciences Norwegian University of Science and Technology Trondheim, Norway Abstract In

More information

Stat 451 Lecture Notes Markov Chain Monte Carlo. Ryan Martin UIC

Stat 451 Lecture Notes Markov Chain Monte Carlo. Ryan Martin UIC Stat 451 Lecture Notes 07 12 Markov Chain Monte Carlo Ryan Martin UIC www.math.uic.edu/~rgmartin 1 Based on Chapters 8 9 in Givens & Hoeting, Chapters 25 27 in Lange 2 Updated: April 4, 2016 1 / 42 Outline

More information

Bayesian Multivariate Logistic Regression

Bayesian Multivariate Logistic Regression Bayesian Multivariate Logistic Regression Sean M. O Brien and David B. Dunson Biostatistics Branch National Institute of Environmental Health Sciences Research Triangle Park, NC 1 Goals Brief review of

More information

Tutorial on ABC Algorithms

Tutorial on ABC Algorithms Tutorial on ABC Algorithms Dr Chris Drovandi Queensland University of Technology, Australia c.drovandi@qut.edu.au July 3, 2014 Notation Model parameter θ with prior π(θ) Likelihood is f(ý θ) with observed

More information

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A.

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A. 1. Let P be a probability measure on a collection of sets A. (a) For each n N, let H n be a set in A such that H n H n+1. Show that P (H n ) monotonically converges to P ( k=1 H k) as n. (b) For each n

More information

Partial factor modeling: predictor-dependent shrinkage for linear regression

Partial factor modeling: predictor-dependent shrinkage for linear regression modeling: predictor-dependent shrinkage for linear Richard Hahn, Carlos Carvalho and Sayan Mukherjee JASA 2013 Review by Esther Salazar Duke University December, 2013 Factor framework The factor framework

More information

Probabilistic Graphical Models Lecture 17: Markov chain Monte Carlo

Probabilistic Graphical Models Lecture 17: Markov chain Monte Carlo Probabilistic Graphical Models Lecture 17: Markov chain Monte Carlo Andrew Gordon Wilson www.cs.cmu.edu/~andrewgw Carnegie Mellon University March 18, 2015 1 / 45 Resources and Attribution Image credits,

More information

A quick introduction to Markov chains and Markov chain Monte Carlo (revised version)

A quick introduction to Markov chains and Markov chain Monte Carlo (revised version) A quick introduction to Markov chains and Markov chain Monte Carlo (revised version) Rasmus Waagepetersen Institute of Mathematical Sciences Aalborg University 1 Introduction These notes are intended to

More information

SAMPLING ALGORITHMS. In general. Inference in Bayesian models

SAMPLING ALGORITHMS. In general. Inference in Bayesian models SAMPLING ALGORITHMS SAMPLING ALGORITHMS In general A sampling algorithm is an algorithm that outputs samples x 1, x 2,... from a given distribution P or density p. Sampling algorithms can for example be

More information

The square root rule for adaptive importance sampling

The square root rule for adaptive importance sampling The square root rule for adaptive importance sampling Art B. Owen Stanford University Yi Zhou January 2019 Abstract In adaptive importance sampling, and other contexts, we have unbiased and uncorrelated

More information

A Randomized Quasi-Monte Carlo Simulation Method for Markov Chains

A Randomized Quasi-Monte Carlo Simulation Method for Markov Chains A Randomized Quasi-Monte Carlo Simulation Method for Markov Chains Pierre L Ecuyer GERAD and Département d Informatique et de Recherche Opérationnelle Université de Montréal, C.P. 6128, Succ. Centre-Ville,

More information

University of Toronto Department of Statistics

University of Toronto Department of Statistics Interacting Multiple Try Algorithms with Different Proposal Distributions by Roberto Casarin Advanced School of Economics, Venice University of Brescia and Radu Craiu University of Toronto and Fabrizio

More information

Introduction to Machine Learning CMU-10701

Introduction to Machine Learning CMU-10701 Introduction to Machine Learning CMU-10701 Markov Chain Monte Carlo Methods Barnabás Póczos & Aarti Singh Contents Markov Chain Monte Carlo Methods Goal & Motivation Sampling Rejection Importance Markov

More information

Scaling analysis of multiple-try MCMC methods

Scaling analysis of multiple-try MCMC methods Scaling analysis of multiple-try MCMC methods Mylène Bédard a,1,, Randal Douc b,2, Eric Moulines c,2 a Département de mathématiques et de statistique, Université de Montréal, C.P. 6128, succ. Centre-ville,

More information

Uniform Random Number Generators

Uniform Random Number Generators JHU 553.633/433: Monte Carlo Methods J. C. Spall 25 September 2017 CHAPTER 2 RANDOM NUMBER GENERATION Motivation and criteria for generators Linear generators (e.g., linear congruential generators) Multiple

More information

Perfect simulation for image analysis

Perfect simulation for image analysis Perfect simulation for image analysis Mark Huber Fletcher Jones Foundation Associate Professor of Mathematics and Statistics and George R. Roberts Fellow Mathematical Sciences Claremont McKenna College

More information

MH I. Metropolis-Hastings (MH) algorithm is the most popular method of getting dependent samples from a probability distribution

MH I. Metropolis-Hastings (MH) algorithm is the most popular method of getting dependent samples from a probability distribution MH I Metropolis-Hastings (MH) algorithm is the most popular method of getting dependent samples from a probability distribution a lot of Bayesian mehods rely on the use of MH algorithm and it s famous

More information

Calibrating Environmental Engineering Models and Uncertainty Analysis

Calibrating Environmental Engineering Models and Uncertainty Analysis Models and Cornell University Oct 14, 2008 Project Team Christine Shoemaker, co-pi, Professor of Civil and works in applied optimization, co-pi Nikolai Blizniouk, PhD student in Operations Research now

More information

Sequential Monte Carlo Samplers for Applications in High Dimensions

Sequential Monte Carlo Samplers for Applications in High Dimensions Sequential Monte Carlo Samplers for Applications in High Dimensions Alexandros Beskos National University of Singapore KAUST, 26th February 2014 Joint work with: Dan Crisan, Ajay Jasra, Nik Kantas, Alex

More information

17 : Markov Chain Monte Carlo

17 : Markov Chain Monte Carlo 10-708: Probabilistic Graphical Models, Spring 2015 17 : Markov Chain Monte Carlo Lecturer: Eric P. Xing Scribes: Heran Lin, Bin Deng, Yun Huang 1 Review of Monte Carlo Methods 1.1 Overview Monte Carlo

More information

Introduction to Bayesian methods in inverse problems

Introduction to Bayesian methods in inverse problems Introduction to Bayesian methods in inverse problems Ville Kolehmainen 1 1 Department of Applied Physics, University of Eastern Finland, Kuopio, Finland March 4 2013 Manchester, UK. Contents Introduction

More information

Monte-Carlo MMD-MA, Université Paris-Dauphine. Xiaolu Tan

Monte-Carlo MMD-MA, Université Paris-Dauphine. Xiaolu Tan Monte-Carlo MMD-MA, Université Paris-Dauphine Xiaolu Tan tan@ceremade.dauphine.fr Septembre 2015 Contents 1 Introduction 1 1.1 The principle.................................. 1 1.2 The error analysis

More information

Package TruncatedNormal

Package TruncatedNormal Package TruncatedNormal November 10, 2015 Type Package Title Truncated Multivariate Normal Version 1.0 Date 2015-11-08 Author Zdravko I. Botev Maintainer Zdravko Botev A collection

More information

Bayesian Computation via Markov chain Monte Carlo

Bayesian Computation via Markov chain Monte Carlo Bayesian Computation via Markov chain Monte Carlo Radu V. Craiu Department of Statistics University of Toronto Jeffrey S. Rosenthal Department of Statistics University of Toronto (Version of June 6, 2013.)

More information

Sequential Quasi-Monte Carlo

Sequential Quasi-Monte Carlo Sequential Quasi-Monte Carlo Mathieu Gerber icolas Chopin arxiv:1402.4039v5 [stat.co] 28 ov 2014 We derive and study SQMC Sequential Quasi-Monte Carlo), a class of algorithms obtained by introducing QMC

More information

University of Toronto Department of Statistics

University of Toronto Department of Statistics Learn From Thy Neighbor: Parallel-Chain and Regional Adaptive MCMC by V. Radu Craiu Department of Statistics University of Toronto and Jeffrey S. Rosenthal Department of Statistics University of Toronto

More information

Deblurring Jupiter (sampling in GLIP faster than regularized inversion) Colin Fox Richard A. Norton, J.

Deblurring Jupiter (sampling in GLIP faster than regularized inversion) Colin Fox Richard A. Norton, J. Deblurring Jupiter (sampling in GLIP faster than regularized inversion) Colin Fox fox@physics.otago.ac.nz Richard A. Norton, J. Andrés Christen Topics... Backstory (?) Sampling in linear-gaussian hierarchical

More information

Why experimenters should not randomize, and what they should do instead

Why experimenters should not randomize, and what they should do instead Why experimenters should not randomize, and what they should do instead Maximilian Kasy Department of Economics, Harvard University Maximilian Kasy (Harvard) Experimental design 1 / 42 project STAR Introduction

More information

Adaptive Monte Carlo methods

Adaptive Monte Carlo methods Adaptive Monte Carlo methods Jean-Michel Marin Projet Select, INRIA Futurs, Université Paris-Sud joint with Randal Douc (École Polytechnique), Arnaud Guillin (Université de Marseille) and Christian Robert

More information

MCMC Sampling for Bayesian Inference using L1-type Priors

MCMC Sampling for Bayesian Inference using L1-type Priors MÜNSTER MCMC Sampling for Bayesian Inference using L1-type Priors (what I do whenever the ill-posedness of EEG/MEG is just not frustrating enough!) AG Imaging Seminar Felix Lucka 26.06.2012 , MÜNSTER Sampling

More information

Markov chain Monte Carlo

Markov chain Monte Carlo Markov chain Monte Carlo Karl Oskar Ekvall Galin L. Jones University of Minnesota March 12, 2019 Abstract Practically relevant statistical models often give rise to probability distributions that are analytically

More information

eqr094: Hierarchical MCMC for Bayesian System Reliability

eqr094: Hierarchical MCMC for Bayesian System Reliability eqr094: Hierarchical MCMC for Bayesian System Reliability Alyson G. Wilson Statistical Sciences Group, Los Alamos National Laboratory P.O. Box 1663, MS F600 Los Alamos, NM 87545 USA Phone: 505-667-9167

More information

Partially Collapsed Gibbs Samplers: Theory and Methods. Ever increasing computational power along with ever more sophisticated statistical computing

Partially Collapsed Gibbs Samplers: Theory and Methods. Ever increasing computational power along with ever more sophisticated statistical computing Partially Collapsed Gibbs Samplers: Theory and Methods David A. van Dyk 1 and Taeyoung Park Ever increasing computational power along with ever more sophisticated statistical computing techniques is making

More information

Lecture 8: The Metropolis-Hastings Algorithm

Lecture 8: The Metropolis-Hastings Algorithm 30.10.2008 What we have seen last time: Gibbs sampler Key idea: Generate a Markov chain by updating the component of (X 1,..., X p ) in turn by drawing from the full conditionals: X (t) j Two drawbacks:

More information

Low Discrepancy Sequences in High Dimensions: How Well Are Their Projections Distributed?

Low Discrepancy Sequences in High Dimensions: How Well Are Their Projections Distributed? Low Discrepancy Sequences in High Dimensions: How Well Are Their Projections Distributed? Xiaoqun Wang 1,2 and Ian H. Sloan 2 1 Department of Mathematical Sciences, Tsinghua University, Beijing 100084,

More information

MONTE CARLO METHODS. Hedibert Freitas Lopes

MONTE CARLO METHODS. Hedibert Freitas Lopes MONTE CARLO METHODS Hedibert Freitas Lopes The University of Chicago Booth School of Business 5807 South Woodlawn Avenue, Chicago, IL 60637 http://faculty.chicagobooth.edu/hedibert.lopes hlopes@chicagobooth.edu

More information

16 : Approximate Inference: Markov Chain Monte Carlo

16 : Approximate Inference: Markov Chain Monte Carlo 10-708: Probabilistic Graphical Models 10-708, Spring 2017 16 : Approximate Inference: Markov Chain Monte Carlo Lecturer: Eric P. Xing Scribes: Yuan Yang, Chao-Ming Yen 1 Introduction As the target distribution

More information

Stat 890 Design of computer experiments

Stat 890 Design of computer experiments Stat 890 Design of computer experiments Will introduce design concepts for computer experiments Will look at more elaborate constructions next day Experiment design In computer experiments, as in many

More information

Markov Chain Monte Carlo (MCMC) and Model Evaluation. August 15, 2017

Markov Chain Monte Carlo (MCMC) and Model Evaluation. August 15, 2017 Markov Chain Monte Carlo (MCMC) and Model Evaluation August 15, 2017 Frequentist Linking Frequentist and Bayesian Statistics How can we estimate model parameters and what does it imply? Want to find the

More information

Extensible grids: uniform sampling on a space-filling curve

Extensible grids: uniform sampling on a space-filling curve Extensible grids: uniform sampling on a space-filling curve Zhijian He Tsinghua University Art B. Owen Stanford University June 2014 Abstract We study the properties of points in [0, 1] d generated by

More information

Fitting Narrow Emission Lines in X-ray Spectra

Fitting Narrow Emission Lines in X-ray Spectra Outline Fitting Narrow Emission Lines in X-ray Spectra Taeyoung Park Department of Statistics, University of Pittsburgh October 11, 2007 Outline of Presentation Outline This talk has three components:

More information

2 Random Variable Generation

2 Random Variable Generation 2 Random Variable Generation Most Monte Carlo computations require, as a starting point, a sequence of i.i.d. random variables with given marginal distribution. We describe here some of the basic methods

More information

Learning the hyper-parameters. Luca Martino

Learning the hyper-parameters. Luca Martino Learning the hyper-parameters Luca Martino 2017 2017 1 / 28 Parameters and hyper-parameters 1. All the described methods depend on some choice of hyper-parameters... 2. For instance, do you recall λ (bandwidth

More information

Wrapped Gaussian processes: a short review and some new results

Wrapped Gaussian processes: a short review and some new results Wrapped Gaussian processes: a short review and some new results Giovanna Jona Lasinio 1, Gianluca Mastrantonio 2 and Alan Gelfand 3 1-Università Sapienza di Roma 2- Università RomaTRE 3- Duke University

More information

Lecture 7 and 8: Markov Chain Monte Carlo

Lecture 7 and 8: Markov Chain Monte Carlo Lecture 7 and 8: Markov Chain Monte Carlo 4F13: Machine Learning Zoubin Ghahramani and Carl Edward Rasmussen Department of Engineering University of Cambridge http://mlg.eng.cam.ac.uk/teaching/4f13/ Ghahramani

More information

MALA versus Random Walk Metropolis Dootika Vats June 4, 2017

MALA versus Random Walk Metropolis Dootika Vats June 4, 2017 MALA versus Random Walk Metropolis Dootika Vats June 4, 2017 Introduction My research thus far has predominantly been on output analysis for Markov chain Monte Carlo. The examples on which I have implemented

More information

Statistics 360/601 Modern Bayesian Theory

Statistics 360/601 Modern Bayesian Theory Statistics 360/601 Modern Bayesian Theory Alexander Volfovsky Lecture 8 - Sept 25, 2018 Monte Carlo 1 2 Monte Carlo approximation Want to compute Z E[ y] = p( y)d without worrying about actual integration...

More information

ORTHOGONAL PARALLEL MCMC METHODS FOR SAMPLING AND OPTIMIZATION

ORTHOGONAL PARALLEL MCMC METHODS FOR SAMPLING AND OPTIMIZATION ORTHOGONAL PARALLEL MCMC METHODS FOR SAMPLING AND OPTIMIZATION L Martino, V Elvira, D Luengo, J Corander, F Louzada Institute of Mathematical Sciences and Computing, Universidade de São Paulo, São Carlos

More information

Semi-Parametric Importance Sampling for Rare-event probability Estimation

Semi-Parametric Importance Sampling for Rare-event probability Estimation Semi-Parametric Importance Sampling for Rare-event probability Estimation Z. I. Botev and P. L Ecuyer IMACS Seminar 2011 Borovets, Bulgaria Semi-Parametric Importance Sampling for Rare-event probability

More information

Chapter 12 PAWL-Forced Simulated Tempering

Chapter 12 PAWL-Forced Simulated Tempering Chapter 12 PAWL-Forced Simulated Tempering Luke Bornn Abstract In this short note, we show how the parallel adaptive Wang Landau (PAWL) algorithm of Bornn et al. (J Comput Graph Stat, to appear) can be

More information

Ph.D. Qualifying Exam Friday Saturday, January 6 7, 2017

Ph.D. Qualifying Exam Friday Saturday, January 6 7, 2017 Ph.D. Qualifying Exam Friday Saturday, January 6 7, 2017 Put your solution to each problem on a separate sheet of paper. Problem 1. (5106) Let X 1, X 2,, X n be a sequence of i.i.d. observations from a

More information

Econ 673: Microeconometrics

Econ 673: Microeconometrics Econ 673: Microeconometrics Chapter 2: Simulation Tools for Estimation and Inference Fall 2008 Herriges (ISU) Chapter 1: Simulation Tools Fall 2008 1 / 63 Outline 1 The Role of Simulation in Estimation

More information

Simulating Random Variables

Simulating Random Variables Simulating Random Variables Timothy Hanson Department of Statistics, University of South Carolina Stat 740: Statistical Computing 1 / 23 R has many built-in random number generators... Beta, gamma (also

More information

Kernel Sequential Monte Carlo

Kernel Sequential Monte Carlo Kernel Sequential Monte Carlo Ingmar Schuster (Paris Dauphine) Heiko Strathmann (University College London) Brooks Paige (Oxford) Dino Sejdinovic (Oxford) * equal contribution April 25, 2016 1 / 37 Section

More information

Sampling from complex probability distributions

Sampling from complex probability distributions Sampling from complex probability distributions Louis J. M. Aslett (louis.aslett@durham.ac.uk) Department of Mathematical Sciences Durham University UTOPIAE Training School II 4 July 2017 1/37 Motivation

More information

Default Priors and Effcient Posterior Computation in Bayesian

Default Priors and Effcient Posterior Computation in Bayesian Default Priors and Effcient Posterior Computation in Bayesian Factor Analysis January 16, 2010 Presented by Eric Wang, Duke University Background and Motivation A Brief Review of Parameter Expansion Literature

More information

Three examples of a Practical Exact Markov Chain Sampling

Three examples of a Practical Exact Markov Chain Sampling Three examples of a Practical Exact Markov Chain Sampling Zdravko Botev November 2007 Abstract We present three examples of exact sampling from complex multidimensional densities using Markov Chain theory

More information

Metropolis-Hastings sampling

Metropolis-Hastings sampling Metropolis-Hastings sampling Gibbs sampling requires that a sample from each full conditional distribution. In all the cases we have looked at so far the conditional distributions were conjugate so sampling

More information

Markov chain Monte Carlo methods in atmospheric remote sensing

Markov chain Monte Carlo methods in atmospheric remote sensing 1 / 45 Markov chain Monte Carlo methods in atmospheric remote sensing Johanna Tamminen johanna.tamminen@fmi.fi ESA Summer School on Earth System Monitoring and Modeling July 3 Aug 11, 212, Frascati July,

More information