Randomized Quasi-Monte Carlo for MCMC
|
|
- Mitchell Walton
- 6 years ago
- Views:
Transcription
1 Randomized Quasi-Monte Carlo for MCMC Radu Craiu 1 Christiane Lemieux 2 1 Department of Statistics, Toronto 2 Department of Statistics, Waterloo Third Workshop on Monte Carlo Methods Harvard, May 2007 Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27
2 Outline 1 Local Search/Optimization Algorithms: MTM 2 Multiple-Try Metropolis and variations Multi-distributed MTM Multiple-Correlated-Try Metropolis 3 Randomized Quasi-Monte Carlo Antithetic or Stratified Sampling? QMC and Randomized QMC: Discussion and Examples Latin Hypercube Sampling Korobov Rule Beyond uniformity 4 Examples 5 Bonus Algorithm - Multi-Annealed Metropolis Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27
3 Local Search and Optimization Algorithms (LSOA) Search locally a region of the state space to sample from. Adapt locally to improve the efficiency of the algorithm. Local Search and Optimization Algorithms (LSOA): Multiple-Try Metropolis (Liu, Liang and Wong, JASA 2000), Hit-and-Run (Chen and Schmeiser, JCGS 1993), Adaptive Direction Sampling (Gilks, Roberts and George, Stat. 1994), Delayed rejection (Green and Mira, Biomka, 2001), Directional Metropolis (Eidsvik and Tjelmeland, Statist. & Comput. 2006). Idea: Produce a more structured search using stratified sampling. Today s Topic: Discussion of RQMC implemented within LSOA. Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27
4 Original MTM (Liu, Liang and Wong, JASA 2000) Suppose T is a proposal distribution such that T(x y) > 0 T(y x) > 0 and λ(x, y) is a symmetric function. Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27
5 Original MTM (Liu, Liang and Wong, JASA 2000) Suppose T is a proposal distribution such that T(x y) > 0 T(y x) > 0 and λ(x, y) is a symmetric function. (i) Draw K independent trial proposals y 1,...,y K from T( x t ). Sample one with p i w(y i x t ) = π(y i )T(x t y i )λ(x t, y i ). Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27
6 Original MTM (Liu, Liang and Wong, JASA 2000) Suppose T is a proposal distribution such that T(x y) > 0 T(y x) > 0 and λ(x, y) is a symmetric function. (i) Draw K independent trial proposals y 1,...,y K from T( x t ). Sample one with p i w(y i x t ) = π(y i )T(x t y i )λ(x t, y i ). (ii) Generate x 1,..., x k 1 T( y) and put x k = x t. Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27
7 Original MTM (Liu, Liang and Wong, JASA 2000) Suppose T is a proposal distribution such that T(x y) > 0 T(y x) > 0 and λ(x, y) is a symmetric function. (i) Draw K independent trial proposals y 1,...,y K from T( x t ). Sample one with p i w(y i x t ) = π(y i )T(x t y i )λ(x t, y i ). (ii) Generate x 1,..., x k 1 T( y) and put x k = x t. { (iii) Accept y with probability min 1, P K i=1 w(y i x t ) P K i=1 w(x i y) }. Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27
8 Original MTM (Liu, Liang and Wong, JASA 2000) Suppose T is a proposal distribution such that T(x y) > 0 T(y x) > 0 and λ(x, y) is a symmetric function. (i) Draw K independent trial proposals y 1,...,y K from T( x t ). Sample one with p i w(y i x t ) = π(y i )T(x t y i )λ(x t, y i ). (ii) Generate x 1,..., x k 1 T( y) and put x k = x t. { (iii) Accept y with probability min 1, P K i=1 w(y i x t ) P K i=1 w(x i y) Do we better explore the sample space with K proposals? }. Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27
9 Original MTM (Liu, Liang and Wong, JASA 2000) Suppose T is a proposal distribution such that T(x y) > 0 T(y x) > 0 and λ(x, y) is a symmetric function. (i) Draw K independent trial proposals y 1,...,y K from T( x t ). Sample one with p i w(y i x t ) = π(y i )T(x t y i )λ(x t, y i ). (ii) Generate x 1,..., x k 1 T( y) and put x k = x t. { (iii) Accept y with probability min 1, P K i=1 w(y i x t ) P K i=1 w(x i y) Do we better explore the sample space with K proposals? As long as we take advantage of the flexibility provided by the MTM. }. Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27
10 Multi-distributed MTM (MD-MTM) The proposals do not have to be identically distributed: y j T j ( x) for 1 j k. Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27
11 Multi-distributed MTM (MD-MTM) The proposals do not have to be identically distributed: y j T j ( x) for 1 j k. If y = y j0 is selected than put xj 0 = x t and sample xj T j ( y). May accomodate different sampling regimes. Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27
12 Multi-distributed MTM (MD-MTM) The proposals do not have to be identically distributed: y j T j ( x) for 1 j k. If y = y j0 is selected than put xj 0 = x t and sample xj T j ( y). May accomodate different sampling regimes. π(x) = 0.3N(( 20, 20) T,Σ 1 )+0.4N((0, 0) T,Σ 2 )+0.3N((20, 20) T,Σ 3 ), where Σ 1 = diag(0.5, 0.5), Σ 2 = diag(2, 2) and Σ 3 = diag(0.4, 0.4). Use MD-MTM for a random walk Metropolis with Gaussian proposal and variances σ {5, 50, 100, 150, 200}. Compare with MTM for a random walk Metropolis constructed with only one of the σ s. Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27
13 MD MTM sigma= Acc. Rate = Acc. Rate = sigma=20 sigma= Acc. Rate = Acc. Rate = Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27
14 Multiple-Correlated-Try Metropolis (MCTM) The proposals do not have to be independent. Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27
15 Multiple-Correlated-Try Metropolis (MCTM) The proposals do not have to be independent. MCTM(I) Suppose we sample K trial proposals y 1,...,y K from T(y 1,..., y K x t ) where T(y 1,..., y K x t )dy [ i] = T(y i x t ), 1 i K. Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27
16 Multiple-Correlated-Try Metropolis (MCTM) The proposals do not have to be independent. MCTM(I) Suppose we sample K trial proposals y 1,...,y K from T(y 1,..., y K x t ) where T(y 1,..., y K x t )dy [ i] = T(y i x t ), 1 i K. The algorithm proceeds as in the independent case with one exception. Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27
17 Multiple-Correlated-Try Metropolis (MCTM) The proposals do not have to be independent. MCTM(I) Suppose we sample K trial proposals y 1,...,y K from T(y 1,..., y K x t ) where T(y 1,..., y K x t )dy [ i] = T(y i x t ), 1 i K. The algorithm proceeds as in the independent case with one exception. MCTM (II) Draw (x1,..., x K 1 ) variates from the conditional transition kernel T(x 1,..., x K 1 y, x K = x t ) and let xk = x t. Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27
18 Multiple-Correlated-Try Metropolis (MCTM) The proposals do not have to be independent. MCTM(I) Suppose we sample K trial proposals y 1,...,y K from T(y 1,..., y K x t ) where T(y 1,..., y K x t )dy [ i] = T(y i x t ), 1 i K. The algorithm proceeds as in the independent case with one exception. MCTM (II) Draw (x1,..., x K 1 ) variates from the conditional transition kernel T(x 1,..., x K 1 y, x K = x t ) and let xk = x t. We have freedom in choosing T as long as we can perform the conditional sampling step. Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27
19 Antithetic or Stratified Sampling? Y j = (Y j1,...,y jd ) R d for 1 j K. Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27
20 Antithetic or Stratified Sampling? Y j = (Y j1,...,y jd ) R d for 1 j K. Say K = 2, Var(Y) = Σ = diag(σ1 2,...,σ2 d ), Cov(Y 1h, Y 2h ) = σh 2ρ h. Then [ [ d ] E Y 1 Y 2 2] d = E (Y 1h Y 2h ) 2 = σh 2 (1 ρ h). h=1 h=1 Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27
21 Antithetic or Stratified Sampling? Y j = (Y j1,...,y jd ) R d for 1 j K. Say K = 2, Var(Y) = Σ = diag(σ1 2,...,σ2 d ), Cov(Y 1h, Y 2h ) = σh 2ρ h. Then [ [ d ] E Y 1 Y 2 2] d = E (Y 1h Y 2h ) 2 = σh 2 (1 ρ h). h=1 One possible choice is ρ h = 1/(K 1) (covariance matrix needs to be positive definite!). The optimal choice of ρ h depends on σ h, 1 h d. h=1 Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27
22 Antithetic or Stratified Sampling? Y j = (Y j1,...,y jd ) R d for 1 j K. Say K = 2, Var(Y) = Σ = diag(σ1 2,...,σ2 d ), Cov(Y 1h, Y 2h ) = σh 2ρ h. Then [ [ d ] E Y 1 Y 2 2] d = E (Y 1h Y 2h ) 2 = σh 2 (1 ρ h). h=1 One possible choice is ρ h = 1/(K 1) (covariance matrix needs to be positive definite!). The optimal choice of ρ h depends on σ h, 1 h d. Conditional sampling needed for MCTM is straightforward in the case of Gaussian proposals. h=1 Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27
23 Antithetic or Stratified Sampling? Y j = (Y j1,...,y jd ) R d for 1 j K. Say K = 2, Var(Y) = Σ = diag(σ1 2,...,σ2 d ), Cov(Y 1h, Y 2h ) = σh 2ρ h. Then [ [ d ] E Y 1 Y 2 2] d = E (Y 1h Y 2h ) 2 = σh 2 (1 ρ h). h=1 One possible choice is ρ h = 1/(K 1) (covariance matrix needs to be positive definite!). The optimal choice of ρ h depends on σ h, 1 h d. Conditional sampling needed for MCTM is straightforward in the case of Gaussian proposals. If Y j = ψ(x t, U j ) with ψ increasing in U j then we need to use U s which are negatively associated (NA - Craiu and Meng, AnnStat. 2005). Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27 h=1
24 Quasi-Monte Carlo (QMC) QMC is a de-randomized MC; the samples are constructed deterministically. QMC: highly uniform and stratified sampling in the unit hypercube. Interest: high-uniformity, a.k.a. low discrepancy. Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27
25 Quasi-Monte Carlo (QMC) QMC is a de-randomized MC; the samples are constructed deterministically. QMC: highly uniform and stratified sampling in the unit hypercube. Interest: high-uniformity, a.k.a. low discrepancy. For v (0, 1) d, B(v) = {u (0, 1) d : 0 < u j v j, 1 j d}. Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27
26 Quasi-Monte Carlo (QMC) QMC is a de-randomized MC; the samples are constructed deterministically. QMC: highly uniform and stratified sampling in the unit hypercube. Interest: high-uniformity, a.k.a. low discrepancy. For v (0, 1) d, B(v) = {u (0, 1) d : 0 < u j v j, 1 j d}. For a point set S n define α(s n, v) = card{s n B(v)}. Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27
27 Quasi-Monte Carlo (QMC) QMC is a de-randomized MC; the samples are constructed deterministically. QMC: highly uniform and stratified sampling in the unit hypercube. Interest: high-uniformity, a.k.a. low discrepancy. For v (0, 1) d, B(v) = {u (0, 1) d : 0 < u j v j, 1 j d}. For a point set S n define α(s n, v) = card{s n B(v)}. Star-discrepancy D n (S n) = sup v (0,1) d v 1... v d α(s n, v)/n. Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27
28 Randomized QMC (RQMC) Efficiency can be connected to discrepancy. Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27
29 Randomized QMC (RQMC) Efficiency can be connected to discrepancy. Koksma-Hlawka Theorem µ 1 f(u) n u S n D n V(f). Both V(f) and D are difficult to compute in general. Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27
30 Randomized QMC (RQMC) Efficiency can be connected to discrepancy. Koksma-Hlawka Theorem µ 1 f(u) n u S n D n V(f). Both V(f) and D are difficult to compute in general. Randomized versions of QMC (RQMC) algorithms are used to establish error estimates via Monte Carlo simulation. (Art Owen, Stanford; Christiane Lemieux, Waterloo; Pierre L Ecuyer, Montréal). Randomization is performed so that it preserves the low-discrepancy of the set point. Makes QMC set points suitable for MC as marginals are uniform. RQMC and MCMC (Owen and Tribble, PNAS 05; Lemieux and Sidorsky, Math. & Comp. Model., 2006). Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27
31 Latin Hypercube Sampling (LHS) Simultaneous stratification of the univariate marginals if U Uniform(0, 1) d. Construction: For 1 i K, 1 j d: Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27
32 Latin Hypercube Sampling (LHS) Simultaneous stratification of the univariate marginals if U Uniform(0, 1) d. Construction: For 1 i K, 1 j d: i) V ij Uniform(0, 1) are i.i.d Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27
33 Latin Hypercube Sampling (LHS) Simultaneous stratification of the univariate marginals if U Uniform(0, 1) d. Construction: For 1 i K, 1 j d: i) V ij Uniform(0, 1) are i.i.d ii) {τ j, 1 j d} are independent permutations of {0,..., K 1} Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27
34 Latin Hypercube Sampling (LHS) Simultaneous stratification of the univariate marginals if U Uniform(0, 1) d. Construction: For 1 i K, 1 j d: i) V ij Uniform(0, 1) are i.i.d ii) {τ j, 1 j d} are independent permutations of {0,..., K 1} U ij = τ j(i 1) + V ij. K Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27
35 Latin Hypercube Sampling (LHS) Simultaneous stratification of the univariate marginals if U Uniform(0, 1) d. Construction: For 1 i K, 1 j d: i) V ij Uniform(0, 1) are i.i.d ii) {τ j, 1 j d} are independent permutations of {0,..., K 1} U ij = τ j(i 1) + V ij. K U (i) is the ith row of the matrix U. Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27
36 Latin Hypercube Sampling (LHS) Simultaneous stratification of the univariate marginals if U Uniform(0, 1) d. Construction: For 1 i K, 1 j d: i) V ij Uniform(0, 1) are i.i.d ii) {τ j, 1 j d} are independent permutations of {0,..., K 1} U ij = τ j(i 1) + V ij. K U (i) is the ith row of the matrix U. Given U (i) = (U i1,...,u id ) (0, 1) d, 1 i K we can use: a) For 1 j d fixed: {U 1j,...,U (Kj) } as a set of NA Uniform(0, 1) r.v. s. b) (U (1),..., U (K) ) as a stratified sample of (0, 1) d. Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27
37 LHS and Gaussian sampling: K = 64, d = 2 Independent Bivariate normal y x LHS Bivariate normal y x Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27
38 LHS in MTM Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27
39 LHS in MTM V i U(0, 1), U i = i+v i K, Y i = F 1 X t (U i ), 0 i K 1 (no permutation is needed!). Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27
40 LHS in MTM V i U(0, 1), U i = i+v i K, Y i = F 1 X t (U i ), 0 i K 1 (no permutation is needed!). If Y = Y i0 is selected: Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27
41 LHS in MTM V i U(0, 1), U i = i+v i K, Y i = F 1 X t (U i ), 0 i K 1 (no permutation is needed!). If Y = Y i0 is selected: Intuitively, if Y has been generated using a uniform from the interval, ( i 0 K, i 0+1 K ), then we want to sample x 2,..., x k using all the K 1 uniforms which ARE NOT IN ( i 0 K, i 0+1 K ). Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27
42 LHS in MTM V i U(0, 1), U i = i+v i K, Y i = F 1 X t (U i ), 0 i K 1 (no permutation is needed!). If Y = Y i0 is selected: Intuitively, if Y has been generated using a uniform from the interval, ( i 0 K, i 0+1 K ), then we want to sample x 2,..., x k using all the K 1 uniforms which ARE NOT IN ( i 0 K, i 0+1 K ). i) Put j 0 = [KF y (X t )] (where [ ] is integer part) and construct a random permutation τ such that τ(j 0 ) = i 0. Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27
43 LHS in MTM V i U(0, 1), U i = i+v i K, Y i = F 1 X t (U i ), 0 i K 1 (no permutation is needed!). If Y = Y i0 is selected: Intuitively, if Y has been generated using a uniform from the interval, ( i 0 K, i 0+1 K ), then we want to sample x 2,..., x k using all the K 1 uniforms which ARE NOT IN ( i 0 K, i 0+1 K ). i) Put j 0 = [KF y (X t )] (where [ ] is integer part) and construct a random permutation τ such that τ(j 0 ) = i 0. ii) Put xj 0 = X t, and for i j 0 Ui = τ(i)+w i K and xi = Fy 1 (Ui ), with W i Uniform(0, 1) independent from the V s. Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27
44 LHS in MTM V i U(0, 1), U i = i+v i K, Y i = F 1 X t (U i ), 0 i K 1 (no permutation is needed!). If Y = Y i0 is selected: Intuitively, if Y has been generated using a uniform from the interval, ( i 0 K, i 0+1 K ), then we want to sample x 2,..., x k using all the K 1 uniforms which ARE NOT IN ( i 0 K, i 0+1 K ). i) Put j 0 = [KF y (X t )] (where [ ] is integer part) and construct a random permutation τ such that τ(j 0 ) = i 0. ii) Put xj 0 = X t, and for i j 0 Ui = τ(i)+w i K and xi = Fy 1 (Ui ), with W i Uniform(0, 1) independent from the V s. One can get away without permuting at all!! Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27
45 Korobov Rules K points in [0, 1] d are generated using a generator a {1,...,K 1}. Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27
46 Korobov Rules K points in [0, 1] d are generated using a generator a {1,...,K 1}. S K = { i K (1, a, a2 mod K,...,a d 1 mod K) mod 1 } 0 i K 1. Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27
47 Korobov Rules K points in [0, 1] d are generated using a generator a {1,...,K 1}. S K = { i K (1, a, a2 mod K,...,a d 1 mod K) mod 1 } 0 i K 1. Bad choices: a and K are not relatively prime or a = 1. The point set is fully projection regular, i.e. any lower-dimensional projection contains K distinct points. Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27
48 Korobov Rules K points in [0, 1] d are generated using a generator a {1,...,K 1}. S K = { i K (1, a, a2 mod K,...,a d 1 mod K) mod 1 } 0 i K 1. Bad choices: a and K are not relatively prime or a = 1. The point set is fully projection regular, i.e. any lower-dimensional projection contains K distinct points. To randomize add a vector U K Uniform(0, 1) d mod 1 i.e. S R K = S K + U K mod 1. Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27
49 Example: Korobov rule and Gaussian sampling Independent N=128, a=47 y x Randomized Korobov y Radu Craiu (fisher.utstat.toronto.edu/craiu/) x RQMC for MCMC 3rd WMCM - May, / 27
50 Going beyond uniformity Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27
51 Going beyond uniformity Sometimes we want to "favor" regions in the hypercube but uniformity is lost once we transform the QMC point set. Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27
52 Going beyond uniformity Sometimes we want to "favor" regions in the hypercube but uniformity is lost once we transform the QMC point set. g : [0, 1] [0, 1] is a bijection and Ũ = g(u) = (g(u 1),..., g(u d )). Y T( X) Y = ψ(x, U) so if Y = ψ(x, Ũ) then Y T( X). Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27
53 Going beyond uniformity Sometimes we want to "favor" regions in the hypercube but uniformity is lost once we transform the QMC point set. g : [0, 1] [0, 1] is a bijection and Ũ = g(u) = (g(u 1),..., g(u d )). Y T( X) Y = ψ(x, U) so if Y = ψ(x, Ũ) then Y T( X). If λ(x, Y) = T(Y X)/ T(Y X) is symmetric, i.e. λ(x, Y) = λ(y, X), then MCTM can be implemented. without computing T. Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27
54 Going beyond uniformity Sometimes we want to "favor" regions in the hypercube but uniformity is lost once we transform the QMC point set. g : [0, 1] [0, 1] is a bijection and Ũ = g(u) = (g(u 1),..., g(u d )). Y T( X) Y = ψ(x, U) so if Y = ψ(x, Ũ) then Y T( X). If λ(x, Y) = T(Y X)/ T(Y X) is symmetric, i.e. λ(x, Y) = λ(y, X), then MCTM can be implemented. without computing T. Example: sin[π(u 0.5)] + 1 g(u) =. 2 i) g is bijective. ii) g(x) = x, x {0, 1/2, 1}, g(u) + g(1 u) = 1, u (0, 1). Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27
55 Example Randomized Korobov y x Transformed Korobov y Radu Craiu (fisher.utstat.toronto.edu/craiu/) x RQMC for MCMC 3rd WMCM - May, / 27
56 Gaussian Proposals Suppose T(y x) is a multivariate Gaussian with no correlation. Then T(y x) = d j=1 e (y j x j ) 2 /2σ 2 j 2πσ 2 j T(y x) = d j=1 e (y j x j ) 2 /2σ 2 j 2πσ 2 j π 1 [ ]. Φ( y j x j σ j ) 1 Φ( y j x j σ j ) λ(y, x) = d j=1 1 [ π 1 Φ( x j y j σ j ) ] = λ(x, y). Φ( x j y j σ j ) Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27
57 Bimodal density Use MCTM combined with random-ray Monte Carlo to sample from (Gelman and Meng, Am. Stat. 91) π(x 1, x 2 ) exp{ (9x 2 1 x x x 2 2 8x 1 8x 2 )/2} Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27
58 Efficiency Comparison At each iteration a random direction v is generated; Along direction v we sample proposals y 1,..., y k around the current state x t using y i = x t + r i with r i Uniform( σ,σ). Table: Values of the MSE reduction factor R = MSE anti MSE ind. σ\k Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27
59 Lupus Data Lupus Data (van Dyk and Meng, JCGS 2001): 55 patients, 2 covariates; logistic model. The posterior density is proportional to 2 π(β x, y) j=0 e 0.5β j/ π 55 i=1 [ ] exp(x T yi [ ] 1 yi i β) exp(xi T β) 1 + exp(xi T. β) Sample using MCTM with antithetically coupled proposals or stratified proposals via a randomized Korobov set shifted using the transformation g. Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27
60 Efficiency comparison for Lupus Data Table: Values of R for β 1 and p 25 = 1 {β1 >25} in the logit example. Antithetic QMC k\σ / / / / / / / / / / / / / / / / / / / / / / / /0.75 Table: Computation Times k = 8 Sample Size MTM MCTM (A) MCTM (S) Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27
61 Multi-Annealed Metropolis (MAM) Inspired by Neal s (Stat. & Comp., 1996) tempered transition method. Uses a series of distributions {π t } t {1=T0,T 1,...,T M }, constructed between the distribution of interest, π, and π TM. Relies on a "path" that crosses all the intermediate distributions. If 1 = T 0 < T 1 < T 2 = T and Q Ti is a proposal distribution for π Ti π 1/T i. Suppose that Q Ti (X Y) = Q Ti (Y X) (e.g., random walk Metropolis). Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27
62 MAM - The algorithm Π 0 X t Y 1 Π 1 X 1 Y 2 Π 2 X 2 Given the chain is in state X: i) Sample X 1 Q 1 ( X), X 2 Q 2 {( X 1 ), Y 2 Q 2 ( X 2 ),} Y 1 Q 1 ( Y 2 ). ii) Accept Y 1 with probability min 1, π 0(Y 1 ) π 0 (X) π2(x 1 )π 1 (Y 2 ) π 1 (X 1 )π 2 (Y 2 ). Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27
63 Example MAM(4) T=(2,8,10,15,20, 30) sigma=(2,3,4,5,8,15) Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for0mcmc 10 3rd WMCM 20 - May, / 27
64 Multiple-try extension of the MAM Π 0 X t Y 1 Y 1 Π 1 X 1 X 1 Y 2 Y 2 Π 2 X 2 X 2 Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27
65 Multiple-try extension of the MAM Π 0 X t Y 1 Y 1 Π 1 X 1 X 1 Y 2 Y 2 Π 2 X 2 X 2 MTM can be implemented by generating multiple paths and compensate using the same number of reverse paths. We have freedom in choosing the "weight of a path". Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27
66 Multiple-try extension of the MAM Π 0 X t Y 1 Y 1 Π 1 X 1 X 1 Y 2 Y 2 Π 2 X 2 X 2 MTM can be implemented by generating multiple paths and compensate using the same number of reverse paths. We have freedom in choosing the "weight of a path". MAM Pros : increases the acceptance rate >10 fold, does not require running the parallel chains all the time; Cons: Requires symmetric proposals. Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27
67 Negative Association for Multiple-Try MAM Suppose that X i Q i ( X i 1 ) X i = ψ i (X i 1, U i ) with U i Uniform[0, 1] d and ψ i is monotone in each component of U i. Y 1 = ψ 1 (ψ 2 (ψ 2 (ψ 1 (X, U 1 ), U 2 ), V 2 ), V 1 ) and Y 1 = ψ 1(ψ 2 (ψ 2 (ψ 1 (X, U 1 ), U 2 ), V 2 ), V 1 ). The paths do not need to be generated independently. If (U i, U i ) and (V i, V i ) are negatively associated then Corr(Y 1, Y 1 ) < 0 so the average distance between the paths is larger than under independence. Radu Craiu (fisher.utstat.toronto.edu/craiu/) RQMC for MCMC 3rd WMCM - May, / 27
Acceleration of the Multiple-Try Metropolis using Antithetic and Stratified Sampling
Acceleration of the Multiple-Try Metropolis using Antithetic and Stratified Sampling Radu V Craiu 1 Christiane Lemieux 2 October 4, 2005 Abstract The Multiple-Try Metropolis is a recent extension of the
More informationMarkov Chain Monte Carlo
Markov Chain Monte Carlo Recall: To compute the expectation E ( h(y ) ) we use the approximation E(h(Y )) 1 n n h(y ) t=1 with Y (1),..., Y (n) h(y). Thus our aim is to sample Y (1),..., Y (n) from f(y).
More informationA Search and Jump Algorithm for Markov Chain Monte Carlo Sampling. Christopher Jennison. Adriana Ibrahim. Seminar at University of Kuwait
A Search and Jump Algorithm for Markov Chain Monte Carlo Sampling Christopher Jennison Department of Mathematical Sciences, University of Bath, UK http://people.bath.ac.uk/mascj Adriana Ibrahim Institute
More informationMarkov Chain Monte Carlo Lecture 1
What are Monte Carlo Methods? The subject of Monte Carlo methods can be viewed as a branch of experimental mathematics in which one uses random numbers to conduct experiments. Typically the experiments
More informationCTDL-Positive Stable Frailty Model
CTDL-Positive Stable Frailty Model M. Blagojevic 1, G. MacKenzie 2 1 Department of Mathematics, Keele University, Staffordshire ST5 5BG,UK and 2 Centre of Biostatistics, University of Limerick, Ireland
More informationRandomized Quasi-Monte Carlo Simulation of Markov Chains with an Ordered State Space
Randomized Quasi-Monte Carlo Simulation of Markov Chains with an Ordered State Space Pierre L Ecuyer 1, Christian Lécot 2, and Bruno Tuffin 3 1 Département d informatique et de recherche opérationnelle,
More informationA generalization of the Multiple-try Metropolis algorithm for Bayesian estimation and model selection
A generalization of the Multiple-try Metropolis algorithm for Bayesian estimation and model selection Silvia Pandolfi Francesco Bartolucci Nial Friel University of Perugia, IT University of Perugia, IT
More informationUniversity of Toronto Department of Statistics
Norm Comparisons for Data Augmentation by James P. Hobert Department of Statistics University of Florida and Jeffrey S. Rosenthal Department of Statistics University of Toronto Technical Report No. 0704
More informationMetropolis-Hastings Algorithm
Strength of the Gibbs sampler Metropolis-Hastings Algorithm Easy algorithm to think about. Exploits the factorization properties of the joint probability distribution. No difficult choices to be made to
More informationTutorial on quasi-monte Carlo methods
Tutorial on quasi-monte Carlo methods Josef Dick School of Mathematics and Statistics, UNSW, Sydney, Australia josef.dick@unsw.edu.au Comparison: MCMC, MC, QMC Roughly speaking: Markov chain Monte Carlo
More informationBrief introduction to Markov Chain Monte Carlo
Brief introduction to Department of Probability and Mathematical Statistics seminar Stochastic modeling in economics and finance November 7, 2011 Brief introduction to Content 1 and motivation Classical
More informationComputer Practical: Metropolis-Hastings-based MCMC
Computer Practical: Metropolis-Hastings-based MCMC Andrea Arnold and Franz Hamilton North Carolina State University July 30, 2016 A. Arnold / F. Hamilton (NCSU) MH-based MCMC July 30, 2016 1 / 19 Markov
More informationControl Variates for Markov Chain Monte Carlo
Control Variates for Markov Chain Monte Carlo Dellaportas, P., Kontoyiannis, I., and Tsourti, Z. Dept of Statistics, AUEB Dept of Informatics, AUEB 1st Greek Stochastics Meeting Monte Carlo: Probability
More informationScenario Generation and Sampling Methods
Scenario Generation and Sampling Methods Tito Homem-de-Mello Güzin Bayraksan SVAN 2016 IMPA 9 13 May 2106 Let s Start with a Recap Sample Average Approximation We want to solve the true problem min {g(x)
More informationBayesian Inference in GLMs. Frequentists typically base inferences on MLEs, asymptotic confidence
Bayesian Inference in GLMs Frequentists typically base inferences on MLEs, asymptotic confidence limits, and log-likelihood ratio tests Bayesians base inferences on the posterior distribution of the unknowns
More informationConstructions of Weakly CUD Sequences for MCMC
Constructions of Weakly CUD Sequences for MCMC Seth D. Tribble Stanford University Art B. Owen Stanford University November 2005 Revised February 2006 Abstract In Markov chain Monte Carlo (MCMC) sampling
More informationMultivariate Normal & Wishart
Multivariate Normal & Wishart Hoff Chapter 7 October 21, 2010 Reading Comprehesion Example Twenty-two children are given a reading comprehsion test before and after receiving a particular instruction method.
More informationA Repelling-Attracting Metropolis Algorithm for Multimodality
A Repelling-Attracting Metropolis Algorithm for Multimodality arxiv:1601.05633v5 [stat.me] 20 Oct 2017 Hyungsuk Tak Statistical and Applied Mathematical Sciences Institute Xiao-Li Meng Department of Statistics,
More informationMathematical Methods for Neurosciences. ENS - Master MVA Paris 6 - Master Maths-Bio ( )
Mathematical Methods for Neurosciences. ENS - Master MVA Paris 6 - Master Maths-Bio (2014-2015) Etienne Tanré - Olivier Faugeras INRIA - Team Tosca October 22nd, 2014 E. Tanré (INRIA - Team Tosca) Mathematical
More informationWinter 2019 Math 106 Topics in Applied Mathematics. Lecture 9: Markov Chain Monte Carlo
Winter 2019 Math 106 Topics in Applied Mathematics Data-driven Uncertainty Quantification Yoonsang Lee (yoonsang.lee@dartmouth.edu) Lecture 9: Markov Chain Monte Carlo 9.1 Markov Chain A Markov Chain Monte
More informationMonte Carlo methods for sampling-based Stochastic Optimization
Monte Carlo methods for sampling-based Stochastic Optimization Gersende FORT LTCI CNRS & Telecom ParisTech Paris, France Joint works with B. Jourdain, T. Lelièvre, G. Stoltz from ENPC and E. Kuhn from
More informationComputational statistics
Computational statistics Markov Chain Monte Carlo methods Thierry Denœux March 2017 Thierry Denœux Computational statistics March 2017 1 / 71 Contents of this chapter When a target density f can be evaluated
More informationQUASI-MONTE CARLO VIA LINEAR SHIFT-REGISTER SEQUENCES. Pierre L Ecuyer Christiane Lemieux
Proceedings of the 1999 Winter Simulation Conference P. A. Farrington, H. B. Nembhard, D. T. Sturrock, and G. W. Evans, eds. QUASI-MONTE CARLO VIA LINEAR SHIFT-REGISTER SEQUENCES Pierre L Ecuyer Christiane
More informationApril 20th, Advanced Topics in Machine Learning California Institute of Technology. Markov Chain Monte Carlo for Machine Learning
for for Advanced Topics in California Institute of Technology April 20th, 2017 1 / 50 Table of Contents for 1 2 3 4 2 / 50 History of methods for Enrico Fermi used to calculate incredibly accurate predictions
More informationOn the Optimal Scaling of the Modified Metropolis-Hastings algorithm
On the Optimal Scaling of the Modified Metropolis-Hastings algorithm K. M. Zuev & J. L. Beck Division of Engineering and Applied Science California Institute of Technology, MC 4-44, Pasadena, CA 925, USA
More informationKernel adaptive Sequential Monte Carlo
Kernel adaptive Sequential Monte Carlo Ingmar Schuster (Paris Dauphine) Heiko Strathmann (University College London) Brooks Paige (Oxford) Dino Sejdinovic (Oxford) December 7, 2015 1 / 36 Section 1 Outline
More informationRecent Advances in Regional Adaptation for MCMC
Recent Advances in Regional Adaptation for MCMC Radu Craiu Department of Statistics University of Toronto Collaborators: Yan Bai (Statistics, Toronto) Antonio Fabio di Narzo (Statistics, Bologna) Jeffrey
More information1. Introduction. In this paper we consider stochastic optimization problems of the form
O RATES OF COVERGECE FOR STOCHASTIC OPTIMIZATIO PROBLEMS UDER O-I.I.D. SAMPLIG TITO HOMEM-DE-MELLO Abstract. In this paper we discuss the issue of solving stochastic optimization problems by means of sample
More informationConvergence Rate of Markov Chains
Convergence Rate of Markov Chains Will Perkins April 16, 2013 Convergence Last class we saw that if X n is an irreducible, aperiodic, positive recurrent Markov chain, then there exists a stationary distribution
More informationResampling Markov chain Monte Carlo algorithms: Basic analysis and empirical comparisons. Zhiqiang Tan 1. February 2014
Resampling Markov chain Monte Carlo s: Basic analysis and empirical comparisons Zhiqiang Tan 1 February 2014 Abstract. Sampling from complex distributions is an important but challenging topic in scientific
More informationA Randomized Quasi-Monte Carlo Simulation Method for Markov Chains
A Randomized Quasi-Monte Carlo Simulation Method for Markov Chains Pierre L Ecuyer GERAD and Département d Informatique et de Recherche Opérationnelle Université de Montréal, C.P. 6128, Succ. Centre-Ville,
More informationMarkov Chain Monte Carlo methods
Markov Chain Monte Carlo methods Tomas McKelvey and Lennart Svensson Signal Processing Group Department of Signals and Systems Chalmers University of Technology, Sweden November 26, 2012 Today s learning
More informationMarkov Chain Monte Carlo (MCMC)
Markov Chain Monte Carlo (MCMC Dependent Sampling Suppose we wish to sample from a density π, and we can evaluate π as a function but have no means to directly generate a sample. Rejection sampling can
More information6 Markov Chain Monte Carlo (MCMC)
6 Markov Chain Monte Carlo (MCMC) The underlying idea in MCMC is to replace the iid samples of basic MC methods, with dependent samples from an ergodic Markov chain, whose limiting (stationary) distribution
More informationDoing Bayesian Integrals
ASTR509-13 Doing Bayesian Integrals The Reverend Thomas Bayes (c.1702 1761) Philosopher, theologian, mathematician Presbyterian (non-conformist) minister Tunbridge Wells, UK Elected FRS, perhaps due to
More informationarxiv: v1 [stat.me] 10 Jul 2009
6th St.Petersburg Workshop on Simulation (2009) 1091-1096 Improvement of random LHD for high dimensions arxiv:0907.1823v1 [stat.me] 10 Jul 2009 Andrey Pepelyshev 1 Abstract Designs of experiments for multivariate
More informationSTA 294: Stochastic Processes & Bayesian Nonparametrics
MARKOV CHAINS AND CONVERGENCE CONCEPTS Markov chains are among the simplest stochastic processes, just one step beyond iid sequences of random variables. Traditionally they ve been used in modelling a
More informationCSCI-6971 Lecture Notes: Monte Carlo integration
CSCI-6971 Lecture otes: Monte Carlo integration Kristopher R. Beevers Department of Computer Science Rensselaer Polytechnic Institute beevek@cs.rpi.edu February 21, 2006 1 Overview Consider the following
More informationOn the flexibility of Metropolis-Hastings acceptance probabilities in auxiliary variable proposal generation
On the flexibility of Metropolis-Hastings acceptance probabilities in auxiliary variable proposal generation Geir Storvik Department of Mathematics and (sfi) 2 Statistics for Innovation, University of
More informationMonte Carlo Methods. Leon Gu CSD, CMU
Monte Carlo Methods Leon Gu CSD, CMU Approximate Inference EM: y-observed variables; x-hidden variables; θ-parameters; E-step: q(x) = p(x y, θ t 1 ) M-step: θ t = arg max E q(x) [log p(y, x θ)] θ Monte
More informationMCMC algorithms for Subset Simulation
To appear in Probabilistic Engineering Mechanics June 2015 MCMC algorithms for Subset Simulation Iason Papaioannou *, Wolfgang Betz, Kilian Zwirglmaier, Daniel Straub Engineering Risk Analysis Group, Technische
More informationDirectional Metropolis Hastings algorithms on hyperplanes
Directional Metropolis Hastings algorithms on hyperplanes Hugo Hammer and Håon Tjelmeland Department of Mathematical Sciences Norwegian University of Science and Technology Trondheim, Norway Abstract In
More informationStat 451 Lecture Notes Markov Chain Monte Carlo. Ryan Martin UIC
Stat 451 Lecture Notes 07 12 Markov Chain Monte Carlo Ryan Martin UIC www.math.uic.edu/~rgmartin 1 Based on Chapters 8 9 in Givens & Hoeting, Chapters 25 27 in Lange 2 Updated: April 4, 2016 1 / 42 Outline
More informationBayesian Multivariate Logistic Regression
Bayesian Multivariate Logistic Regression Sean M. O Brien and David B. Dunson Biostatistics Branch National Institute of Environmental Health Sciences Research Triangle Park, NC 1 Goals Brief review of
More informationTutorial on ABC Algorithms
Tutorial on ABC Algorithms Dr Chris Drovandi Queensland University of Technology, Australia c.drovandi@qut.edu.au July 3, 2014 Notation Model parameter θ with prior π(θ) Likelihood is f(ý θ) with observed
More informationFall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A.
1. Let P be a probability measure on a collection of sets A. (a) For each n N, let H n be a set in A such that H n H n+1. Show that P (H n ) monotonically converges to P ( k=1 H k) as n. (b) For each n
More informationPartial factor modeling: predictor-dependent shrinkage for linear regression
modeling: predictor-dependent shrinkage for linear Richard Hahn, Carlos Carvalho and Sayan Mukherjee JASA 2013 Review by Esther Salazar Duke University December, 2013 Factor framework The factor framework
More informationProbabilistic Graphical Models Lecture 17: Markov chain Monte Carlo
Probabilistic Graphical Models Lecture 17: Markov chain Monte Carlo Andrew Gordon Wilson www.cs.cmu.edu/~andrewgw Carnegie Mellon University March 18, 2015 1 / 45 Resources and Attribution Image credits,
More informationA quick introduction to Markov chains and Markov chain Monte Carlo (revised version)
A quick introduction to Markov chains and Markov chain Monte Carlo (revised version) Rasmus Waagepetersen Institute of Mathematical Sciences Aalborg University 1 Introduction These notes are intended to
More informationSAMPLING ALGORITHMS. In general. Inference in Bayesian models
SAMPLING ALGORITHMS SAMPLING ALGORITHMS In general A sampling algorithm is an algorithm that outputs samples x 1, x 2,... from a given distribution P or density p. Sampling algorithms can for example be
More informationThe square root rule for adaptive importance sampling
The square root rule for adaptive importance sampling Art B. Owen Stanford University Yi Zhou January 2019 Abstract In adaptive importance sampling, and other contexts, we have unbiased and uncorrelated
More informationA Randomized Quasi-Monte Carlo Simulation Method for Markov Chains
A Randomized Quasi-Monte Carlo Simulation Method for Markov Chains Pierre L Ecuyer GERAD and Département d Informatique et de Recherche Opérationnelle Université de Montréal, C.P. 6128, Succ. Centre-Ville,
More informationUniversity of Toronto Department of Statistics
Interacting Multiple Try Algorithms with Different Proposal Distributions by Roberto Casarin Advanced School of Economics, Venice University of Brescia and Radu Craiu University of Toronto and Fabrizio
More informationIntroduction to Machine Learning CMU-10701
Introduction to Machine Learning CMU-10701 Markov Chain Monte Carlo Methods Barnabás Póczos & Aarti Singh Contents Markov Chain Monte Carlo Methods Goal & Motivation Sampling Rejection Importance Markov
More informationScaling analysis of multiple-try MCMC methods
Scaling analysis of multiple-try MCMC methods Mylène Bédard a,1,, Randal Douc b,2, Eric Moulines c,2 a Département de mathématiques et de statistique, Université de Montréal, C.P. 6128, succ. Centre-ville,
More informationUniform Random Number Generators
JHU 553.633/433: Monte Carlo Methods J. C. Spall 25 September 2017 CHAPTER 2 RANDOM NUMBER GENERATION Motivation and criteria for generators Linear generators (e.g., linear congruential generators) Multiple
More informationPerfect simulation for image analysis
Perfect simulation for image analysis Mark Huber Fletcher Jones Foundation Associate Professor of Mathematics and Statistics and George R. Roberts Fellow Mathematical Sciences Claremont McKenna College
More informationMH I. Metropolis-Hastings (MH) algorithm is the most popular method of getting dependent samples from a probability distribution
MH I Metropolis-Hastings (MH) algorithm is the most popular method of getting dependent samples from a probability distribution a lot of Bayesian mehods rely on the use of MH algorithm and it s famous
More informationCalibrating Environmental Engineering Models and Uncertainty Analysis
Models and Cornell University Oct 14, 2008 Project Team Christine Shoemaker, co-pi, Professor of Civil and works in applied optimization, co-pi Nikolai Blizniouk, PhD student in Operations Research now
More informationSequential Monte Carlo Samplers for Applications in High Dimensions
Sequential Monte Carlo Samplers for Applications in High Dimensions Alexandros Beskos National University of Singapore KAUST, 26th February 2014 Joint work with: Dan Crisan, Ajay Jasra, Nik Kantas, Alex
More information17 : Markov Chain Monte Carlo
10-708: Probabilistic Graphical Models, Spring 2015 17 : Markov Chain Monte Carlo Lecturer: Eric P. Xing Scribes: Heran Lin, Bin Deng, Yun Huang 1 Review of Monte Carlo Methods 1.1 Overview Monte Carlo
More informationIntroduction to Bayesian methods in inverse problems
Introduction to Bayesian methods in inverse problems Ville Kolehmainen 1 1 Department of Applied Physics, University of Eastern Finland, Kuopio, Finland March 4 2013 Manchester, UK. Contents Introduction
More informationMonte-Carlo MMD-MA, Université Paris-Dauphine. Xiaolu Tan
Monte-Carlo MMD-MA, Université Paris-Dauphine Xiaolu Tan tan@ceremade.dauphine.fr Septembre 2015 Contents 1 Introduction 1 1.1 The principle.................................. 1 1.2 The error analysis
More informationPackage TruncatedNormal
Package TruncatedNormal November 10, 2015 Type Package Title Truncated Multivariate Normal Version 1.0 Date 2015-11-08 Author Zdravko I. Botev Maintainer Zdravko Botev A collection
More informationBayesian Computation via Markov chain Monte Carlo
Bayesian Computation via Markov chain Monte Carlo Radu V. Craiu Department of Statistics University of Toronto Jeffrey S. Rosenthal Department of Statistics University of Toronto (Version of June 6, 2013.)
More informationSequential Quasi-Monte Carlo
Sequential Quasi-Monte Carlo Mathieu Gerber icolas Chopin arxiv:1402.4039v5 [stat.co] 28 ov 2014 We derive and study SQMC Sequential Quasi-Monte Carlo), a class of algorithms obtained by introducing QMC
More informationUniversity of Toronto Department of Statistics
Learn From Thy Neighbor: Parallel-Chain and Regional Adaptive MCMC by V. Radu Craiu Department of Statistics University of Toronto and Jeffrey S. Rosenthal Department of Statistics University of Toronto
More informationDeblurring Jupiter (sampling in GLIP faster than regularized inversion) Colin Fox Richard A. Norton, J.
Deblurring Jupiter (sampling in GLIP faster than regularized inversion) Colin Fox fox@physics.otago.ac.nz Richard A. Norton, J. Andrés Christen Topics... Backstory (?) Sampling in linear-gaussian hierarchical
More informationWhy experimenters should not randomize, and what they should do instead
Why experimenters should not randomize, and what they should do instead Maximilian Kasy Department of Economics, Harvard University Maximilian Kasy (Harvard) Experimental design 1 / 42 project STAR Introduction
More informationAdaptive Monte Carlo methods
Adaptive Monte Carlo methods Jean-Michel Marin Projet Select, INRIA Futurs, Université Paris-Sud joint with Randal Douc (École Polytechnique), Arnaud Guillin (Université de Marseille) and Christian Robert
More informationMCMC Sampling for Bayesian Inference using L1-type Priors
MÜNSTER MCMC Sampling for Bayesian Inference using L1-type Priors (what I do whenever the ill-posedness of EEG/MEG is just not frustrating enough!) AG Imaging Seminar Felix Lucka 26.06.2012 , MÜNSTER Sampling
More informationMarkov chain Monte Carlo
Markov chain Monte Carlo Karl Oskar Ekvall Galin L. Jones University of Minnesota March 12, 2019 Abstract Practically relevant statistical models often give rise to probability distributions that are analytically
More informationeqr094: Hierarchical MCMC for Bayesian System Reliability
eqr094: Hierarchical MCMC for Bayesian System Reliability Alyson G. Wilson Statistical Sciences Group, Los Alamos National Laboratory P.O. Box 1663, MS F600 Los Alamos, NM 87545 USA Phone: 505-667-9167
More informationPartially Collapsed Gibbs Samplers: Theory and Methods. Ever increasing computational power along with ever more sophisticated statistical computing
Partially Collapsed Gibbs Samplers: Theory and Methods David A. van Dyk 1 and Taeyoung Park Ever increasing computational power along with ever more sophisticated statistical computing techniques is making
More informationLecture 8: The Metropolis-Hastings Algorithm
30.10.2008 What we have seen last time: Gibbs sampler Key idea: Generate a Markov chain by updating the component of (X 1,..., X p ) in turn by drawing from the full conditionals: X (t) j Two drawbacks:
More informationLow Discrepancy Sequences in High Dimensions: How Well Are Their Projections Distributed?
Low Discrepancy Sequences in High Dimensions: How Well Are Their Projections Distributed? Xiaoqun Wang 1,2 and Ian H. Sloan 2 1 Department of Mathematical Sciences, Tsinghua University, Beijing 100084,
More informationMONTE CARLO METHODS. Hedibert Freitas Lopes
MONTE CARLO METHODS Hedibert Freitas Lopes The University of Chicago Booth School of Business 5807 South Woodlawn Avenue, Chicago, IL 60637 http://faculty.chicagobooth.edu/hedibert.lopes hlopes@chicagobooth.edu
More information16 : Approximate Inference: Markov Chain Monte Carlo
10-708: Probabilistic Graphical Models 10-708, Spring 2017 16 : Approximate Inference: Markov Chain Monte Carlo Lecturer: Eric P. Xing Scribes: Yuan Yang, Chao-Ming Yen 1 Introduction As the target distribution
More informationStat 890 Design of computer experiments
Stat 890 Design of computer experiments Will introduce design concepts for computer experiments Will look at more elaborate constructions next day Experiment design In computer experiments, as in many
More informationMarkov Chain Monte Carlo (MCMC) and Model Evaluation. August 15, 2017
Markov Chain Monte Carlo (MCMC) and Model Evaluation August 15, 2017 Frequentist Linking Frequentist and Bayesian Statistics How can we estimate model parameters and what does it imply? Want to find the
More informationExtensible grids: uniform sampling on a space-filling curve
Extensible grids: uniform sampling on a space-filling curve Zhijian He Tsinghua University Art B. Owen Stanford University June 2014 Abstract We study the properties of points in [0, 1] d generated by
More informationFitting Narrow Emission Lines in X-ray Spectra
Outline Fitting Narrow Emission Lines in X-ray Spectra Taeyoung Park Department of Statistics, University of Pittsburgh October 11, 2007 Outline of Presentation Outline This talk has three components:
More information2 Random Variable Generation
2 Random Variable Generation Most Monte Carlo computations require, as a starting point, a sequence of i.i.d. random variables with given marginal distribution. We describe here some of the basic methods
More informationLearning the hyper-parameters. Luca Martino
Learning the hyper-parameters Luca Martino 2017 2017 1 / 28 Parameters and hyper-parameters 1. All the described methods depend on some choice of hyper-parameters... 2. For instance, do you recall λ (bandwidth
More informationWrapped Gaussian processes: a short review and some new results
Wrapped Gaussian processes: a short review and some new results Giovanna Jona Lasinio 1, Gianluca Mastrantonio 2 and Alan Gelfand 3 1-Università Sapienza di Roma 2- Università RomaTRE 3- Duke University
More informationLecture 7 and 8: Markov Chain Monte Carlo
Lecture 7 and 8: Markov Chain Monte Carlo 4F13: Machine Learning Zoubin Ghahramani and Carl Edward Rasmussen Department of Engineering University of Cambridge http://mlg.eng.cam.ac.uk/teaching/4f13/ Ghahramani
More informationMALA versus Random Walk Metropolis Dootika Vats June 4, 2017
MALA versus Random Walk Metropolis Dootika Vats June 4, 2017 Introduction My research thus far has predominantly been on output analysis for Markov chain Monte Carlo. The examples on which I have implemented
More informationStatistics 360/601 Modern Bayesian Theory
Statistics 360/601 Modern Bayesian Theory Alexander Volfovsky Lecture 8 - Sept 25, 2018 Monte Carlo 1 2 Monte Carlo approximation Want to compute Z E[ y] = p( y)d without worrying about actual integration...
More informationORTHOGONAL PARALLEL MCMC METHODS FOR SAMPLING AND OPTIMIZATION
ORTHOGONAL PARALLEL MCMC METHODS FOR SAMPLING AND OPTIMIZATION L Martino, V Elvira, D Luengo, J Corander, F Louzada Institute of Mathematical Sciences and Computing, Universidade de São Paulo, São Carlos
More informationSemi-Parametric Importance Sampling for Rare-event probability Estimation
Semi-Parametric Importance Sampling for Rare-event probability Estimation Z. I. Botev and P. L Ecuyer IMACS Seminar 2011 Borovets, Bulgaria Semi-Parametric Importance Sampling for Rare-event probability
More informationChapter 12 PAWL-Forced Simulated Tempering
Chapter 12 PAWL-Forced Simulated Tempering Luke Bornn Abstract In this short note, we show how the parallel adaptive Wang Landau (PAWL) algorithm of Bornn et al. (J Comput Graph Stat, to appear) can be
More informationPh.D. Qualifying Exam Friday Saturday, January 6 7, 2017
Ph.D. Qualifying Exam Friday Saturday, January 6 7, 2017 Put your solution to each problem on a separate sheet of paper. Problem 1. (5106) Let X 1, X 2,, X n be a sequence of i.i.d. observations from a
More informationEcon 673: Microeconometrics
Econ 673: Microeconometrics Chapter 2: Simulation Tools for Estimation and Inference Fall 2008 Herriges (ISU) Chapter 1: Simulation Tools Fall 2008 1 / 63 Outline 1 The Role of Simulation in Estimation
More informationSimulating Random Variables
Simulating Random Variables Timothy Hanson Department of Statistics, University of South Carolina Stat 740: Statistical Computing 1 / 23 R has many built-in random number generators... Beta, gamma (also
More informationKernel Sequential Monte Carlo
Kernel Sequential Monte Carlo Ingmar Schuster (Paris Dauphine) Heiko Strathmann (University College London) Brooks Paige (Oxford) Dino Sejdinovic (Oxford) * equal contribution April 25, 2016 1 / 37 Section
More informationSampling from complex probability distributions
Sampling from complex probability distributions Louis J. M. Aslett (louis.aslett@durham.ac.uk) Department of Mathematical Sciences Durham University UTOPIAE Training School II 4 July 2017 1/37 Motivation
More informationDefault Priors and Effcient Posterior Computation in Bayesian
Default Priors and Effcient Posterior Computation in Bayesian Factor Analysis January 16, 2010 Presented by Eric Wang, Duke University Background and Motivation A Brief Review of Parameter Expansion Literature
More informationThree examples of a Practical Exact Markov Chain Sampling
Three examples of a Practical Exact Markov Chain Sampling Zdravko Botev November 2007 Abstract We present three examples of exact sampling from complex multidimensional densities using Markov Chain theory
More informationMetropolis-Hastings sampling
Metropolis-Hastings sampling Gibbs sampling requires that a sample from each full conditional distribution. In all the cases we have looked at so far the conditional distributions were conjugate so sampling
More informationMarkov chain Monte Carlo methods in atmospheric remote sensing
1 / 45 Markov chain Monte Carlo methods in atmospheric remote sensing Johanna Tamminen johanna.tamminen@fmi.fi ESA Summer School on Earth System Monitoring and Modeling July 3 Aug 11, 212, Frascati July,
More information