Computer intensive statistical methods

Size: px
Start display at page:

Download "Computer intensive statistical methods"

Transcription

1 Lecture 13 MCMC, Hybrid chains October 13, 2015 Jonas Wallin Chalmers, Gothenburg university

2 MH algorithm, Chap:6.3 The metropolis hastings requires three objects, the distribution of interest π, and a conditional density q (does not need to depend on π) and a initial value x (0) (or distribution).

3 MH algorithm, Chap:6.3 The metropolis hastings requires three objects, the distribution of interest π, and a conditional density q (does not need to depend on π) and a initial value x (0) (or distribution).the first iteration goes as follows (generation of X (1) ) 1 Generate X q(x x (0) ). 2 Take X (1) = { X with probability ρ(x (0), X ), x (0) with probability 1 ρ(x (0), X ). where ( ) π(y)q(x y) ρ(x, y) = min π(x)q(y x), 1.

4 The transition kernel of the Metropolis-Hastings algorithm Recall P(X (t) A X (t 1) = x) = A K(y x)dy. Lemma (MH transition kernel) The MH algorithm is a Markov chain with the following transition kernel: K(y x) = ρ(x, y)q(y x) + p R (x)δ x (y), where p R (x) = 1 ρ(x, z)q(z x)dz with ρ(x, y) = 1 π(y)q(x y) π(x)q(y x). Note that δ x (y) = δ(y x) = δ(x y) where δ is the so called Dirac delta function, δ x (y) = I(x = y).

5 Convergence of the MH algorithm Theorem (stationary distribution of MH*) The chain (X (k) ) generated by the MH sampler has π as stationary distribution. π-irreducibility depends on q, but as we shall see is often very easy to verify

6 different types of proposal kernels There are a number of different ways of constructing the proposal kernel q. The three main classes are independent proposals, multiplicative proposals, and symmetric proposals.

7 Independent proposal Independent proposals are characterized as follows. (recall the accept reject method) Draw the candidates from q(y), i.e. independently of the current state x. onas Wallin (Chalmers)

8 Independent proposal Independent proposals are characterized as follows. (recall the accept reject method) Draw the candidates from q(y), i.e. independently of the current state x. The acceptance probability reduces to ρ(x, y) = 1 π(y)q(x) π(x)q(y). onas Wallin (Chalmers)

9 Independent proposal Independent proposals are characterized as follows. (recall the accept reject method) Draw the candidates from q(y), i.e. independently of the current state x. The acceptance probability reduces to ρ(x, y) = 1 π(y)q(x) π(x)q(y). Here it is required that {x : π(x) > 0} {x : q(x) > 0} to ensure convergence. (π-irreducibility) onas Wallin (Chalmers)

10 Independent proposal Independent proposals are characterized as follows. (recall the accept reject method) Draw the candidates from q(y), i.e. independently of the current state x. The acceptance probability reduces to ρ(x, y) = 1 π(y)q(x) π(x)q(y). Here it is required that {x : π(x) > 0} {x : q(x) > 0} to ensure convergence. (π-irreducibility) If we take q(x) = π(x), which is of course infeasible in general, the acceptance probability reduces to 1 and we get independent samples from π. onas Wallin (Chalmers)

11 Multiplicative proposals An easy way to obtain an asymmetric proposal where the size of the jump depends on the current state x k is to take X = ɛx k, where ɛ is drawn from some density p. The proposal kernel now becomes q(y x) = p(y/x)/x and the acceptance probability becomes ρ(x, y) = 1 π(y)p(x/y)/y π(x)p(y/x)/x.

12 Symmetric proposal Symmetric proposals are characterized by the following. It holds that q(y x) = q(x y).

13 Symmetric proposal Symmetric proposals are characterized by the following. It holds that q(y x) = q(x y). In this case the acceptance probability reduces to ρ(x, y) = 1 π(y) π(x).

14 Symmetric proposal Symmetric proposals are characterized by the following. It holds that q(y x) = q(x y). In this case the acceptance probability reduces to ρ(x, y) = 1 π(y) π(x). Commonly this corresponds to X = x k + ɛ (random walk proposal) with, e.g., ɛ N (0, σ 2 ), ɛ σt(1). ɛ U[ a, a] or ɛ U{ a, a + 1,..., a 1, a}.

15 Symmetric proposal for multivariate distribution, algorithm Choose an initial distribution π 0 (π 0 (x) = δ(x) for fixed start point), and use ɛ N (0, Σ), draw X (0) f 0 for l = 1 N do draw X N (x; X (l 1), Σ) draw U U[0, 1] if U π(x ) π(x l 1 ) then set X (l) = X else set X (l) = X (l 1) end if end for return X = (X (0),..., X (N) )

16 A simple example Using the MH symmetric algorithm for sampling from a bivariate Gaussian distribution. That is π(x) = N (x; µ, Σ) where [ ] µ = 0, Σ = We use a symmetric proposal with ɛ N (0, σ 2 MH I).

17 Mark-capture-recapture Estimation of populations is a tricky problem.

18 Mark-capture-recapture Estimation of populations is a tricky problem. Suppose there exists N animals, and one observe n animals. If you observe each animal independently then the likelihood is a binomial distribution where p is unknown. n Bin(N, p),

19 Mark-capture-recapture Estimation of populations is a tricky problem. Suppose there exists N animals, and one observe n animals. If you observe each animal independently then the likelihood is a binomial distribution where p is unknown. n Bin(N, p), The issue is that this likelihood does not contain information about N, p jointly.

20 Mark-capture-recapture Estimation of populations is a tricky problem. Suppose there exists N animals, and one observe n animals. If you observe each animal independently then the likelihood is a binomial distribution where p is unknown. n Bin(N, p), The issue is that this likelihood does not contain information about N, p jointly. To remedy this issue one mark captured animals, then release them again, and see how many animals are captured next time.

21 Mark-capture-recapture Suppose one tries to capture a set of animals k times. Each time the animals are marked, and resealed.

22 Mark-capture-recapture Suppose one tries to capture a set of animals k times. Each time the animals are marked, and resealed. Let the number of time each captured animal is captured denote y i.

23 Mark-capture-recapture Suppose one tries to capture a set of animals k times. Each time the animals are marked, and resealed. Let the number of time each captured animal is captured denote y i. Let N denote the total number of animals, n denote the number of unique animals captured and p the probability of capturing an animal at one instance.

24 Mark-capture-recapture Suppose one tries to capture a set of animals k times. Each time the animals are marked, and resealed. Let the number of time each captured animal is captured denote y i. Let N denote the total number of animals, n denote the number of unique animals captured and p the probability of capturing an animal at one instance. Then the likelihood can be derived as follows: onas Wallin (Chalmers)

25 Mark-capture-recapture Suppose one tries to capture a set of animals k times. Each time the animals are marked, and resealed. Let the number of time each captured animal is captured denote y i. Let N denote the total number of animals, n denote the number of unique animals captured and p the probability of capturing an animal at one instance. Then the likelihood can be derived as follows: 1 The probability of detecting n animals is: where p k = 1 (1 p) k. n Bin(N, p k ), onas Wallin (Chalmers)

26 Mark-capture-recapture Suppose one tries to capture a set of animals k times. Each time the animals are marked, and resealed. Let the number of time each captured animal is captured denote y i. Let N denote the total number of animals, n denote the number of unique animals captured and p the probability of capturing an animal at one instance. Then the likelihood can be derived as follows: 1 The probability of detecting n animals is: n Bin(N, p k ), where p k = 1 (1 p) k. 2 Given N, n the probability of y i is: y i 1 Bin(k 1, p).

27 Mark-capture-recapture Thus the posterior distribution of N, p is ( ) N n π(n, p y) (p c ) n (1 p c ) N n p y i 1 (1 p) k y i π(n, p). n i=1

28 Mark-capture-recapture Thus the posterior distribution of N, p is ( ) N n π(n, p y) (p c ) n (1 p c ) N n p y i 1 (1 p) k y i π(n, p). n To sample from this distribution, one can use a MH algorithm. There exists a large number of possbile proposal distribution, we are going to see two: and i=1 q 1 (N, p N, p) = U(N a,..., N + a)i [0,1] (p ). q 2 (N, p N, p) = U(N a,..., N + a)n (p, σ 2 p).

29 Mark-capture-recapture We apply the model of a real data sets where hares where captured and marked six times iter p

30 Mark-capture-recapture We apply the model of a real data sets where hares where captured and marked six times iter p

31 Mark-capture-recapture We apply the model of a real data sets where hares where captured and marked six times. Density N

32 Mark-capture-recapture We apply the model of a real data sets where hares where captured and marked six times. p Counts N

33 Multivariate Gaussian example By far the most used proposal distribution is the MH random walk with normal proposal. If there is a lot of dependence in the posterior distribution it is important that the MH proposal captures this.

34 Poisson Regerssion A classical data (first studied in 1898) is number of deaths in the Prussian Army due to kicks from horses: A question of interest is whether there is a trend in the deaths over time. Bayesian model (using uninformative priors): β j N (0, 10 2 ), Po(y i ; λ = exp(x i β)) where X i = [1, t i ] where t i is the year the observation was made.

35 Poisson Regerssion cont.. The posterior distribution of β is, through Bayes formula, proportional to π(β y, X) e βt β } {{ } π(β) n exp(y i (X i β) exp(x i β)). i=1 } {{ } π(y β)

36 Poisson Regerssion acf Σ = d I.

37 Poisson Regerssion acf Σ = d I. Series Theta[, 1] ACF Lag acf (1) =

38 Poisson Regerssion hist β Counts β

39 Poisson Regerssion acf 2 Σ = d Diag(C π[β]).

40 Poisson Regerssion acf 2 Σ = d Diag(C π[β]). Series Theta[, 1] ACF acf (1) = Lag 0.9.

41 Poisson Regerssion acf 3 Σ = d C π[β].

42 Poisson Regerssion acf 3 Σ = d C π[β]. Series Theta[, 1] ACF acf (1) = Lag

43 Choosing Σ For large dimensions Σ plays a crucial roll for convergence. onas Wallin (Chalmers)

44 Choosing Σ For large dimensions Σ plays a crucial roll for convergence. So far I have manually tuned Σ. onas Wallin (Chalmers)

45 Choosing Σ For large dimensions Σ plays a crucial roll for convergence. So far I have manually tuned Σ. Annoying to impossible for higher dimension. onas Wallin (Chalmers)

46 Choosing Σ For large dimensions Σ plays a crucial roll for convergence. So far I have manually tuned Σ. Annoying to impossible for higher dimension. Optimal scaling: Σ = d C π[x]. onas Wallin (Chalmers)

47 Choosing Σ But we don t have C π [X]. onas Wallin (Chalmers)

48 Choosing Σ But we don t have C π [X]. Lets estimate it! ˆΣ N = 1 N N (x i x i ) (x i x i ) T. i=1 onas Wallin (Chalmers)

49 Symmetric proposal for multivariate distribution, algorithm Choose a Σ 0 and a X draw X (0) π 0 for l = 1 N do draw X N (y; X (l 1), d draw U U[0, 1] if U π(y ) then π(x l 1 ) set X (l) = X else set X (l) = X (l 1) end if update X = l 1 l X + l l X(l) update Σ l = l 1 l Σ l 1 + l l end for return X = (X (0),..., X (N) ) Σ l 1) ( ) ( ) T X (l) X X (l) X

50 Mathematical issues Issue, no longer a Markov chain. But converges to one. However, can sometimes fail (in rare occasions6)!

51 Monte Carlo and credibility interval the confidence interval for different E[β 2 y 1,... y 10 ] number MCMC samples N: N = (black line π(β 2 y 1,... y 10 ), blue lines β 95% credibility interval, and red lines 95% confidence interval of the mean)

52 Monte Carlo and credibility interval the confidence interval for different E[β 2 y 1,... y 10 ] number MCMC samples N: N = (black line π(β 2 y 1,... y 10 ), blue lines β 95% credibility interval, and red lines 95% confidence interval of the mean)

53 Monte Carlo and credibility interval the confidence interval for different E[β 2 y 1,... y 10 ] number MCMC samples N: N = (black line π(β 2 y 1,... y 10 ), blue lines β 95% credibility interval, and red lines 95% confidence interval of the mean)

54 Monte Carlo and credibility interval When the data increase the posterior distribution changes: (blue line π(β 2 y 1,... y 10 ), red line β π(β 2 y 1,... y 100 ), and black line π(β 2 y 1,... y 1000 ), green line π(β 2 y 1,... y 5000 ))

55 MH within Gibbs Going back to Gibbs, Often Gibbs samplers has better convergence then MH.

56 MH within Gibbs Going back to Gibbs, Often Gibbs samplers has better convergence then MH. However, in larger models there typically is one or many parameters that the conditional distribution is not explicit.

57 MH within Gibbs Going back to Gibbs, Often Gibbs samplers has better convergence then MH. However, in larger models there typically is one or many parameters that the conditional distribution is not explicit. An idea is to combine, Gibbs with MH, where the intractable conditional sampling is replaced with MH.

58 Hybrid chains (theoretical motivation) Assume without loss of generality that we have two blocks and want to sample from π(x 1, x 2 ). The hybrid MCMC goes like this, given X (t 1) = x = (x (t 1) 1, x (t 1) 2 ). 1 Draw X (t) 1 q(x 1 x (t 1) 1, x (t 1) 2 ) with corresponding K MH as MH kernel for π(x 1 x 2 = x (t 1) 2 ). onas Wallin (Chalmers)

59 Hybrid chains (theoretical motivation) Assume without loss of generality that we have two blocks and want to sample from π(x 1, x 2 ). The hybrid MCMC goes like this, given X (t 1) = x = (x (t 1) 1, x (t 1) 2 ). 1 Draw X (t) 1 q(x 1 x (t 1) 1, x (t 1) 2 ) with corresponding K MH as MH kernel for π(x 1 x 2 = x (t 1) 2 ). 2 Draw X (t) 2 π(x 2 x 1 = x (t) 1 ) i.e. standard Gibbs. onas Wallin (Chalmers)

60 Hybrid chains (theoretical motivation) Theorem (stationary distribution of Hybrid chain*) The chain (X (k) ) generated by the Hybrid sampler has π as stationary distribution.

61 Global balance equation K(z x) = π(z 2 z 1 )K MH (z 1 x 2, x 1 ) Proof. K(z 2, z 1 x 1, x 2 )π(x 1, x 2 )dx 1 dx 2 = π(z 2 z 1 )K MH (z 1 x 2, x 1 )π(x 1 x 2 )π(x 2 )dx 1 dx 2 = π(z 2 z 1 ) K MH (z 1 x 2, x 1 )π(x 1 x 2 )dx 1 π(x 2 )dx 2 Global Balance MH = π(z 2 z 1 ) π(z 1 x 2 )π(x 2 )dx 2 Law of total prob = π(z 2 z 1 )π(z 1 ) = π(z 1, z 2 ).

62 Example:rats The joint posterior distribution is given by n n ( π(θ, α, β y) θ y i i (1 θ i ) n i y i Γ(α + β) Γ(α)Γ(β) i=1 i=1 }{{} π(y θ) ) θ α i (1 θ i ) β } {{ } π(θ α,β) e α β }{{} π(α,β)

63 Example:rats The joint posterior distribution is given by n n ( π(θ, α, β y) θ y i i (1 θ i ) n i y i Γ(α + β) Γ(α)Γ(β) i=1 i=1 }{{} π(y θ) ) θ α i (1 θ i ) β } {{ } π(θ α,β) The t + 1 iterations of the Gibbs sampler follows as θ (t+1) j... Beta(α (t) + y j, β (t) + n j y j ) ( ) n α (t+1) Γ(α + β (t) ) j... π(α...) e Γ(α) β (t+1) j... π(β...) ( Γ(α (t+1) + β) Γ(β) ) n e (t+1) log(θ j )α e α (t+1) log(θ j 1)β e β The densities of α and β are both log concave. For log concave densities there are good rejection algorithms. e α β }{{} π(α,β)

64 Example:rats (acf) However it turns out that acf for α, β (worth 1 samples) 171, regular MC α β ACF ACF lag lag

65 Hybrid chains ratdata Recall π(θ, α, β y) n i=1 θ y i i (1 θ i ) n i y i } {{ } π(y θ) n ( ) Γ(α + β) θi α (1 θ i ) β Γ(α)Γ(β) i=1 }{{} π(θ α,β) The t + 1 iterations of the Hybrid sampler follows as θ (t+1) j (α (t), β (t) ) Beta(α (t) + y j, β (t) + n j y j ) ( ) Γ(α + β) n (α (t+1), β (t+1) ) θ (t+1) Γ(α)Γ(β) (t+1) e log(θ j )α+ log(θ (t+1) j 1)β (α + β) 5 2 Where the second step is updated through a MH random walk. (α + β) 5 2 }{{} π(α,β)

66 R-code R code for the log likelihood. post_dens_alpha_beta <- function(alpha_beta, theta) { if(min(alpha_beta) <=0){return(-Inf)} alpha <- alpha_beta[1] beta <- alpha_beta[2] n <- length(theta) log_f <- -5/2*log(alpha+beta)# prior log_f <- log_f + n * (lgamma(alpha+beta) -lgamma(alpha) -lgamma(beta)) log_f <- log_f + alpha * sum(log(theta)) + beta * sum(log(1 - theta)) return(log_f) }

67 R-code the MH within Gibbs for(i in 1:sim){ theta <- rbeta(n, ab[1]+dat[,1], ab[2] + dat[,2]-dat[,1]) lik_old <- post_dens_alpha_beta(ab, theta = theta) ab_star <- rmvnorm(1, mean = ab, sigma = Sigma) lik_star <- post_dens_alpha_beta(ab_star, theta = theta) if( log(runif(1)) < lik_star - lik_old) ab = ab_star

68 Example:rats (acf) The acf for α, β (worth 1 40, 1 38 regular MC samples) α β ACF ACF lag lag

69 β hist Counts α

70 Gibbs within Metropolis We have seen that Metropolis within Gibbs works, and also often is quite easy to implement.

71 Gibbs within Metropolis We have seen that Metropolis within Gibbs works, and also often is quite easy to implement. However, if there is strong dependence between parameters it is often better to use Gibbs within Metropolis, or block updating. (Although often a bit harder to implement.)

72 Gibbs within Metropolis(Block sampler) Suppose we want to sample (α, θ), we can sample directly from θ α. Then at t iteration of the Gibbs within MH goes as follows: Sample α N (α, σ 2 MH ) onas Wallin (Chalmers)

73 Gibbs within Metropolis(Block sampler) Suppose we want to sample (α, θ), we can sample directly from θ α. Then at t iteration of the Gibbs within MH goes as follows: Sample α N (α, σ 2 MH ) Sample θ π(. α ) onas Wallin (Chalmers)

74 Gibbs within Metropolis(Block sampler) Suppose we want to sample (α, θ), we can sample directly from θ α. Then at t iteration of the Gibbs within MH goes as follows: Sample α N (α, σ 2 MH ) Sample θ π(. α ) Accept reject (θ, α ) jointly. { (α, θ) (t) (α, θ ) w probability ρ((α, θ) (t 1), (α, θ )), = (α, θ) (t 1) w probability 1 ρ((α, θ) (t 1), (α, θ )). where ( π(α ρ((α, θ), (α, θ, θ ) )π(θ α) )) = min π(α, θ)π(θ α ), 1. onas Wallin (Chalmers)

75 Gibbs within Metropolis(Block sampler) Suppose we want to sample (α, θ), we can sample directly from θ α. Then at t iteration of the Gibbs within MH goes as follows: Sample α N (α, σmh 2 ) Sample θ π(. α ) Accept reject (θ, α ) jointly. { (α, θ) (t) (α, θ ) w probability ρ((α, θ) (t 1), (α, θ )), = (α, θ) (t 1) w probability 1 ρ((α, θ) (t 1), (α, θ )). where ( π(α ρ((α, θ), (α, θ, θ ) )π(θ α) )) = min π(α, θ)π(θ α ), 1. Of course one can generalize this beyond just one dimension and MH-RW kernel. Since all we do is construct a complicated proposal kernel. onas Wallin (Chalmers)

76 R-code gibbs_within_mh.r

77 Example:rats (acf) again However it turns out that acf for α, β, worth 1 14, 1 14 samples, big difference from 1 171, regular MC α β ACF ACF lag lag

78 Theorem (Central limit theorem) Let X (t) be geometric ergodic Markov chain, with stationary distribution π. Then for any h s.t E π [ h(x ) 2+ɛ ] <, where ɛ > 0, T (τt E π (h(x ))) N (0, σ 2 h), Under some additional assumption: σh 2 = V π [h(x 0 )] + 2 C π [h(x 0 ), h(x i )]. i=1

79 Effective Sample size MCMC style 1 σ 2 h is typically not known. 2 Sometimes an AR(1) is a good approximation of the behavior of MCMC chain. 3 Approximate the Markov chain h(x t ) with an AR(1) process, where C(h(X (t) ), h(x (t+h) )) = ρ h (here C is the correlation function). 4 Then the variance of the estimator V[ 1 T T h(x (i) )] 1 + ρ 1 1 ρ T V[h(X (t) )] i=1 Then T 1 ρ 1+ρ is the effective sample size.

80 Estimating asymptotic variance using blocking An alternative is to block the data as follows: τ N = 1 T T h(x k ) = 1 n k=1 n Y l, l=1 where G = T n and def. Y l = 1 G lg m=(l 1)G+1 h(x m ), l = 1, 2,..., n. If the blocks are large enough we can view these as close to independent and identically distributed.

81 Estimating asymptotic variance using blocking (cont.) We may thus expect the CLT to hold at least approximately, implying that ( ) 1 n V(τ N ) = V Y l V(Y 1), n n l=1 where V(Y 1 ) can be estimated using the standard estimator V(Y 1 ) 1 n 1 n (Y m Y n ) 2, m=1 with Y n = n m=1 Y m/n denoting the sample mean.

82 Estimating asymptotic variance using blocking in R T <- length ( X) G = 100 n = floor (T/G) Y = sapply (1:n, function (k) return ( mean (X [(( k - 1) * G + 1):( k * G )])))

83 blocking rats(cont.) G = 100 for the Gibbs within MH for α in the rats example: α ACF Lag

Computer intensive statistical methods

Computer intensive statistical methods Lecture 11 Markov Chain Monte Carlo cont. October 6, 2015 Jonas Wallin jonwal@chalmers.se Chalmers, Gothenburg university The two stage Gibbs sampler If the conditional distributions are easy to sample

More information

INTRODUCTION TO BAYESIAN STATISTICS

INTRODUCTION TO BAYESIAN STATISTICS INTRODUCTION TO BAYESIAN STATISTICS Sarat C. Dass Department of Statistics & Probability Department of Computer Science & Engineering Michigan State University TOPICS The Bayesian Framework Different Types

More information

Markov Chain Monte Carlo (MCMC)

Markov Chain Monte Carlo (MCMC) Markov Chain Monte Carlo (MCMC Dependent Sampling Suppose we wish to sample from a density π, and we can evaluate π as a function but have no means to directly generate a sample. Rejection sampling can

More information

Lecture 8: The Metropolis-Hastings Algorithm

Lecture 8: The Metropolis-Hastings Algorithm 30.10.2008 What we have seen last time: Gibbs sampler Key idea: Generate a Markov chain by updating the component of (X 1,..., X p ) in turn by drawing from the full conditionals: X (t) j Two drawbacks:

More information

Reminder of some Markov Chain properties:

Reminder of some Markov Chain properties: Reminder of some Markov Chain properties: 1. a transition from one state to another occurs probabilistically 2. only state that matters is where you currently are (i.e. given present, future is independent

More information

Monte Carlo Methods. Leon Gu CSD, CMU

Monte Carlo Methods. Leon Gu CSD, CMU Monte Carlo Methods Leon Gu CSD, CMU Approximate Inference EM: y-observed variables; x-hidden variables; θ-parameters; E-step: q(x) = p(x y, θ t 1 ) M-step: θ t = arg max E q(x) [log p(y, x θ)] θ Monte

More information

Adaptive Monte Carlo methods

Adaptive Monte Carlo methods Adaptive Monte Carlo methods Jean-Michel Marin Projet Select, INRIA Futurs, Université Paris-Sud joint with Randal Douc (École Polytechnique), Arnaud Guillin (Université de Marseille) and Christian Robert

More information

Metropolis-Hastings Algorithm

Metropolis-Hastings Algorithm Strength of the Gibbs sampler Metropolis-Hastings Algorithm Easy algorithm to think about. Exploits the factorization properties of the joint probability distribution. No difficult choices to be made to

More information

Markov Chain Monte Carlo methods

Markov Chain Monte Carlo methods Markov Chain Monte Carlo methods Tomas McKelvey and Lennart Svensson Signal Processing Group Department of Signals and Systems Chalmers University of Technology, Sweden November 26, 2012 Today s learning

More information

Computer intensive statistical methods Lecture 1

Computer intensive statistical methods Lecture 1 Computer intensive statistical methods Lecture 1 Jonas Wallin Chalmers, Gothenburg. Jonas Wallin - jonwal@chalmers.se 1/27 People and literature The following people are involved in the course: Function

More information

Computational statistics

Computational statistics Computational statistics Markov Chain Monte Carlo methods Thierry Denœux March 2017 Thierry Denœux Computational statistics March 2017 1 / 71 Contents of this chapter When a target density f can be evaluated

More information

April 20th, Advanced Topics in Machine Learning California Institute of Technology. Markov Chain Monte Carlo for Machine Learning

April 20th, Advanced Topics in Machine Learning California Institute of Technology. Markov Chain Monte Carlo for Machine Learning for for Advanced Topics in California Institute of Technology April 20th, 2017 1 / 50 Table of Contents for 1 2 3 4 2 / 50 History of methods for Enrico Fermi used to calculate incredibly accurate predictions

More information

Bayesian Methods for Machine Learning

Bayesian Methods for Machine Learning Bayesian Methods for Machine Learning CS 584: Big Data Analytics Material adapted from Radford Neal s tutorial (http://ftp.cs.utoronto.ca/pub/radford/bayes-tut.pdf), Zoubin Ghahramni (http://hunch.net/~coms-4771/zoubin_ghahramani_bayesian_learning.pdf),

More information

Metropolis Hastings. Rebecca C. Steorts Bayesian Methods and Modern Statistics: STA 360/601. Module 9

Metropolis Hastings. Rebecca C. Steorts Bayesian Methods and Modern Statistics: STA 360/601. Module 9 Metropolis Hastings Rebecca C. Steorts Bayesian Methods and Modern Statistics: STA 360/601 Module 9 1 The Metropolis-Hastings algorithm is a general term for a family of Markov chain simulation methods

More information

Markov Chain Monte Carlo Using the Ratio-of-Uniforms Transformation. Luke Tierney Department of Statistics & Actuarial Science University of Iowa

Markov Chain Monte Carlo Using the Ratio-of-Uniforms Transformation. Luke Tierney Department of Statistics & Actuarial Science University of Iowa Markov Chain Monte Carlo Using the Ratio-of-Uniforms Transformation Luke Tierney Department of Statistics & Actuarial Science University of Iowa Basic Ratio of Uniforms Method Introduced by Kinderman and

More information

Stat 451 Lecture Notes Markov Chain Monte Carlo. Ryan Martin UIC

Stat 451 Lecture Notes Markov Chain Monte Carlo. Ryan Martin UIC Stat 451 Lecture Notes 07 12 Markov Chain Monte Carlo Ryan Martin UIC www.math.uic.edu/~rgmartin 1 Based on Chapters 8 9 in Givens & Hoeting, Chapters 25 27 in Lange 2 Updated: April 4, 2016 1 / 42 Outline

More information

Bayesian Inference in GLMs. Frequentists typically base inferences on MLEs, asymptotic confidence

Bayesian Inference in GLMs. Frequentists typically base inferences on MLEs, asymptotic confidence Bayesian Inference in GLMs Frequentists typically base inferences on MLEs, asymptotic confidence limits, and log-likelihood ratio tests Bayesians base inferences on the posterior distribution of the unknowns

More information

Markov Chain Monte Carlo methods

Markov Chain Monte Carlo methods Markov Chain Monte Carlo methods By Oleg Makhnin 1 Introduction a b c M = d e f g h i 0 f(x)dx 1.1 Motivation 1.1.1 Just here Supresses numbering 1.1.2 After this 1.2 Literature 2 Method 2.1 New math As

More information

Lecture 6: Markov Chain Monte Carlo

Lecture 6: Markov Chain Monte Carlo Lecture 6: Markov Chain Monte Carlo D. Jason Koskinen koskinen@nbi.ku.dk Photo by Howard Jackman University of Copenhagen Advanced Methods in Applied Statistics Feb - Apr 2016 Niels Bohr Institute 2 Outline

More information

Bayesian Graphical Models

Bayesian Graphical Models Graphical Models and Inference, Lecture 16, Michaelmas Term 2009 December 4, 2009 Parameter θ, data X = x, likelihood L(θ x) p(x θ). Express knowledge about θ through prior distribution π on θ. Inference

More information

Bayesian GLMs and Metropolis-Hastings Algorithm

Bayesian GLMs and Metropolis-Hastings Algorithm Bayesian GLMs and Metropolis-Hastings Algorithm We have seen that with conjugate or semi-conjugate prior distributions the Gibbs sampler can be used to sample from the posterior distribution. In situations,

More information

A short diversion into the theory of Markov chains, with a view to Markov chain Monte Carlo methods

A short diversion into the theory of Markov chains, with a view to Markov chain Monte Carlo methods A short diversion into the theory of Markov chains, with a view to Markov chain Monte Carlo methods by Kasper K. Berthelsen and Jesper Møller June 2004 2004-01 DEPARTMENT OF MATHEMATICAL SCIENCES AALBORG

More information

Bayesian philosophy Bayesian computation Bayesian software. Bayesian Statistics. Petter Mostad. Chalmers. April 6, 2017

Bayesian philosophy Bayesian computation Bayesian software. Bayesian Statistics. Petter Mostad. Chalmers. April 6, 2017 Chalmers April 6, 2017 Bayesian philosophy Bayesian philosophy Bayesian statistics versus classical statistics: War or co-existence? Classical statistics: Models have variables and parameters; these are

More information

CPSC 540: Machine Learning

CPSC 540: Machine Learning CPSC 540: Machine Learning MCMC and Non-Parametric Bayes Mark Schmidt University of British Columbia Winter 2016 Admin I went through project proposals: Some of you got a message on Piazza. No news is

More information

eqr094: Hierarchical MCMC for Bayesian System Reliability

eqr094: Hierarchical MCMC for Bayesian System Reliability eqr094: Hierarchical MCMC for Bayesian System Reliability Alyson G. Wilson Statistical Sciences Group, Los Alamos National Laboratory P.O. Box 1663, MS F600 Los Alamos, NM 87545 USA Phone: 505-667-9167

More information

Introduction to Machine Learning CMU-10701

Introduction to Machine Learning CMU-10701 Introduction to Machine Learning CMU-10701 Markov Chain Monte Carlo Methods Barnabás Póczos & Aarti Singh Contents Markov Chain Monte Carlo Methods Goal & Motivation Sampling Rejection Importance Markov

More information

SC7/SM6 Bayes Methods HT18 Lecturer: Geoff Nicholls Lecture 2: Monte Carlo Methods Notes and Problem sheets are available at http://www.stats.ox.ac.uk/~nicholls/bayesmethods/ and via the MSc weblearn pages.

More information

Computer Vision Group Prof. Daniel Cremers. 10a. Markov Chain Monte Carlo

Computer Vision Group Prof. Daniel Cremers. 10a. Markov Chain Monte Carlo Group Prof. Daniel Cremers 10a. Markov Chain Monte Carlo Markov Chain Monte Carlo In high-dimensional spaces, rejection sampling and importance sampling are very inefficient An alternative is Markov Chain

More information

Markov chain Monte Carlo Lecture 9

Markov chain Monte Carlo Lecture 9 Markov chain Monte Carlo Lecture 9 David Sontag New York University Slides adapted from Eric Xing and Qirong Ho (CMU) Limitations of Monte Carlo Direct (unconditional) sampling Hard to get rare events

More information

Monte Carlo-based statistical methods (MASM11/FMS091)

Monte Carlo-based statistical methods (MASM11/FMS091) Monte Carlo-based statistical methods (MASM11/FMS091) Jimmy Olsson Centre for Mathematical Sciences Lund University, Sweden Lecture 12 MCMC for Bayesian computation II March 1, 2013 J. Olsson Monte Carlo-based

More information

Markov Chain Monte Carlo (MCMC) and Model Evaluation. August 15, 2017

Markov Chain Monte Carlo (MCMC) and Model Evaluation. August 15, 2017 Markov Chain Monte Carlo (MCMC) and Model Evaluation August 15, 2017 Frequentist Linking Frequentist and Bayesian Statistics How can we estimate model parameters and what does it imply? Want to find the

More information

17 : Markov Chain Monte Carlo

17 : Markov Chain Monte Carlo 10-708: Probabilistic Graphical Models, Spring 2015 17 : Markov Chain Monte Carlo Lecturer: Eric P. Xing Scribes: Heran Lin, Bin Deng, Yun Huang 1 Review of Monte Carlo Methods 1.1 Overview Monte Carlo

More information

Computer Vision Group Prof. Daniel Cremers. 14. Sampling Methods

Computer Vision Group Prof. Daniel Cremers. 14. Sampling Methods Prof. Daniel Cremers 14. Sampling Methods Sampling Methods Sampling Methods are widely used in Computer Science as an approximation of a deterministic algorithm to represent uncertainty without a parametric

More information

Pattern Recognition and Machine Learning. Bishop Chapter 11: Sampling Methods

Pattern Recognition and Machine Learning. Bishop Chapter 11: Sampling Methods Pattern Recognition and Machine Learning Chapter 11: Sampling Methods Elise Arnaud Jakob Verbeek May 22, 2008 Outline of the chapter 11.1 Basic Sampling Algorithms 11.2 Markov Chain Monte Carlo 11.3 Gibbs

More information

Computer Practical: Metropolis-Hastings-based MCMC

Computer Practical: Metropolis-Hastings-based MCMC Computer Practical: Metropolis-Hastings-based MCMC Andrea Arnold and Franz Hamilton North Carolina State University July 30, 2016 A. Arnold / F. Hamilton (NCSU) MH-based MCMC July 30, 2016 1 / 19 Markov

More information

LECTURE 15 Markov chain Monte Carlo

LECTURE 15 Markov chain Monte Carlo LECTURE 15 Markov chain Monte Carlo There are many settings when posterior computation is a challenge in that one does not have a closed form expression for the posterior distribution. Markov chain Monte

More information

CSC 2541: Bayesian Methods for Machine Learning

CSC 2541: Bayesian Methods for Machine Learning CSC 2541: Bayesian Methods for Machine Learning Radford M. Neal, University of Toronto, 2011 Lecture 3 More Markov Chain Monte Carlo Methods The Metropolis algorithm isn t the only way to do MCMC. We ll

More information

Deblurring Jupiter (sampling in GLIP faster than regularized inversion) Colin Fox Richard A. Norton, J.

Deblurring Jupiter (sampling in GLIP faster than regularized inversion) Colin Fox Richard A. Norton, J. Deblurring Jupiter (sampling in GLIP faster than regularized inversion) Colin Fox fox@physics.otago.ac.nz Richard A. Norton, J. Andrés Christen Topics... Backstory (?) Sampling in linear-gaussian hierarchical

More information

Introduction to Markov Chain Monte Carlo & Gibbs Sampling

Introduction to Markov Chain Monte Carlo & Gibbs Sampling Introduction to Markov Chain Monte Carlo & Gibbs Sampling Prof. Nicholas Zabaras Sibley School of Mechanical and Aerospace Engineering 101 Frank H. T. Rhodes Hall Ithaca, NY 14853-3801 Email: zabaras@cornell.edu

More information

Markov chain Monte Carlo methods in atmospheric remote sensing

Markov chain Monte Carlo methods in atmospheric remote sensing 1 / 45 Markov chain Monte Carlo methods in atmospheric remote sensing Johanna Tamminen johanna.tamminen@fmi.fi ESA Summer School on Earth System Monitoring and Modeling July 3 Aug 11, 212, Frascati July,

More information

Convergence Rate of Markov Chains

Convergence Rate of Markov Chains Convergence Rate of Markov Chains Will Perkins April 16, 2013 Convergence Last class we saw that if X n is an irreducible, aperiodic, positive recurrent Markov chain, then there exists a stationary distribution

More information

Computer Vision Group Prof. Daniel Cremers. 11. Sampling Methods: Markov Chain Monte Carlo

Computer Vision Group Prof. Daniel Cremers. 11. Sampling Methods: Markov Chain Monte Carlo Group Prof. Daniel Cremers 11. Sampling Methods: Markov Chain Monte Carlo Markov Chain Monte Carlo In high-dimensional spaces, rejection sampling and importance sampling are very inefficient An alternative

More information

Markov Chain Monte Carlo

Markov Chain Monte Carlo Markov Chain Monte Carlo Recall: To compute the expectation E ( h(y ) ) we use the approximation E(h(Y )) 1 n n h(y ) t=1 with Y (1),..., Y (n) h(y). Thus our aim is to sample Y (1),..., Y (n) from f(y).

More information

27 : Distributed Monte Carlo Markov Chain. 1 Recap of MCMC and Naive Parallel Gibbs Sampling

27 : Distributed Monte Carlo Markov Chain. 1 Recap of MCMC and Naive Parallel Gibbs Sampling 10-708: Probabilistic Graphical Models 10-708, Spring 2014 27 : Distributed Monte Carlo Markov Chain Lecturer: Eric P. Xing Scribes: Pengtao Xie, Khoa Luu In this scribe, we are going to review the Parallel

More information

Stat 535 C - Statistical Computing & Monte Carlo Methods. Lecture February Arnaud Doucet

Stat 535 C - Statistical Computing & Monte Carlo Methods. Lecture February Arnaud Doucet Stat 535 C - Statistical Computing & Monte Carlo Methods Lecture 13-28 February 2006 Arnaud Doucet Email: arnaud@cs.ubc.ca 1 1.1 Outline Limitations of Gibbs sampling. Metropolis-Hastings algorithm. Proof

More information

MARKOV CHAIN MONTE CARLO

MARKOV CHAIN MONTE CARLO MARKOV CHAIN MONTE CARLO RYAN WANG Abstract. This paper gives a brief introduction to Markov Chain Monte Carlo methods, which offer a general framework for calculating difficult integrals. We start with

More information

Lecture 7 and 8: Markov Chain Monte Carlo

Lecture 7 and 8: Markov Chain Monte Carlo Lecture 7 and 8: Markov Chain Monte Carlo 4F13: Machine Learning Zoubin Ghahramani and Carl Edward Rasmussen Department of Engineering University of Cambridge http://mlg.eng.cam.ac.uk/teaching/4f13/ Ghahramani

More information

STA 4273H: Statistical Machine Learning

STA 4273H: Statistical Machine Learning STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Computer Science! Department of Statistical Sciences! rsalakhu@cs.toronto.edu! h0p://www.cs.utoronto.ca/~rsalakhu/ Lecture 7 Approximate

More information

Markov chain Monte Carlo

Markov chain Monte Carlo Markov chain Monte Carlo Karl Oskar Ekvall Galin L. Jones University of Minnesota March 12, 2019 Abstract Practically relevant statistical models often give rise to probability distributions that are analytically

More information

Markov Chain Monte Carlo Inference. Siamak Ravanbakhsh Winter 2018

Markov Chain Monte Carlo Inference. Siamak Ravanbakhsh Winter 2018 Graphical Models Markov Chain Monte Carlo Inference Siamak Ravanbakhsh Winter 2018 Learning objectives Markov chains the idea behind Markov Chain Monte Carlo (MCMC) two important examples: Gibbs sampling

More information

Hastings-within-Gibbs Algorithm: Introduction and Application on Hierarchical Model

Hastings-within-Gibbs Algorithm: Introduction and Application on Hierarchical Model UNIVERSITY OF TEXAS AT SAN ANTONIO Hastings-within-Gibbs Algorithm: Introduction and Application on Hierarchical Model Liang Jing April 2010 1 1 ABSTRACT In this paper, common MCMC algorithms are introduced

More information

Answers and expectations

Answers and expectations Answers and expectations For a function f(x) and distribution P(x), the expectation of f with respect to P is The expectation is the average of f, when x is drawn from the probability distribution P E

More information

MCMC and Gibbs Sampling. Kayhan Batmanghelich

MCMC and Gibbs Sampling. Kayhan Batmanghelich MCMC and Gibbs Sampling Kayhan Batmanghelich 1 Approaches to inference l Exact inference algorithms l l l The elimination algorithm Message-passing algorithm (sum-product, belief propagation) The junction

More information

Beta statistics. Keywords. Bayes theorem. Bayes rule

Beta statistics. Keywords. Bayes theorem. Bayes rule Keywords Beta statistics Tommy Norberg tommy@chalmers.se Mathematical Sciences Chalmers University of Technology Gothenburg, SWEDEN Bayes s formula Prior density Likelihood Posterior density Conjugate

More information

The Particle Filter. PD Dr. Rudolph Triebel Computer Vision Group. Machine Learning for Computer Vision

The Particle Filter. PD Dr. Rudolph Triebel Computer Vision Group. Machine Learning for Computer Vision The Particle Filter Non-parametric implementation of Bayes filter Represents the belief (posterior) random state samples. by a set of This representation is approximate. Can represent distributions that

More information

I. Bayesian econometrics

I. Bayesian econometrics I. Bayesian econometrics A. Introduction B. Bayesian inference in the univariate regression model C. Statistical decision theory D. Large sample results E. Diffuse priors F. Numerical Bayesian methods

More information

ST 740: Markov Chain Monte Carlo

ST 740: Markov Chain Monte Carlo ST 740: Markov Chain Monte Carlo Alyson Wilson Department of Statistics North Carolina State University October 14, 2012 A. Wilson (NCSU Stsatistics) MCMC October 14, 2012 1 / 20 Convergence Diagnostics:

More information

Computer Vision Group Prof. Daniel Cremers. 11. Sampling Methods

Computer Vision Group Prof. Daniel Cremers. 11. Sampling Methods Prof. Daniel Cremers 11. Sampling Methods Sampling Methods Sampling Methods are widely used in Computer Science as an approximation of a deterministic algorithm to represent uncertainty without a parametric

More information

Markov Chains and MCMC

Markov Chains and MCMC Markov Chains and MCMC CompSci 590.02 Instructor: AshwinMachanavajjhala Lecture 4 : 590.02 Spring 13 1 Recap: Monte Carlo Method If U is a universe of items, and G is a subset satisfying some property,

More information

MCMC: Markov Chain Monte Carlo

MCMC: Markov Chain Monte Carlo I529: Machine Learning in Bioinformatics (Spring 2013) MCMC: Markov Chain Monte Carlo Yuzhen Ye School of Informatics and Computing Indiana University, Bloomington Spring 2013 Contents Review of Markov

More information

MCMC Methods: Gibbs and Metropolis

MCMC Methods: Gibbs and Metropolis MCMC Methods: Gibbs and Metropolis Patrick Breheny February 28 Patrick Breheny BST 701: Bayesian Modeling in Biostatistics 1/30 Introduction As we have seen, the ability to sample from the posterior distribution

More information

Control Variates for Markov Chain Monte Carlo

Control Variates for Markov Chain Monte Carlo Control Variates for Markov Chain Monte Carlo Dellaportas, P., Kontoyiannis, I., and Tsourti, Z. Dept of Statistics, AUEB Dept of Informatics, AUEB 1st Greek Stochastics Meeting Monte Carlo: Probability

More information

Introduction to Bayesian Methods

Introduction to Bayesian Methods Introduction to Bayesian Methods Jessi Cisewski Department of Statistics Yale University Sagan Summer Workshop 2016 Our goal: introduction to Bayesian methods Likelihoods Priors: conjugate priors, non-informative

More information

Introduction to Bayesian Statistics 1

Introduction to Bayesian Statistics 1 Introduction to Bayesian Statistics 1 STA 442/2101 Fall 2018 1 This slide show is an open-source document. See last slide for copyright information. 1 / 42 Thomas Bayes (1701-1761) Image from the Wikipedia

More information

SAMPLING ALGORITHMS. In general. Inference in Bayesian models

SAMPLING ALGORITHMS. In general. Inference in Bayesian models SAMPLING ALGORITHMS SAMPLING ALGORITHMS In general A sampling algorithm is an algorithm that outputs samples x 1, x 2,... from a given distribution P or density p. Sampling algorithms can for example be

More information

A Search and Jump Algorithm for Markov Chain Monte Carlo Sampling. Christopher Jennison. Adriana Ibrahim. Seminar at University of Kuwait

A Search and Jump Algorithm for Markov Chain Monte Carlo Sampling. Christopher Jennison. Adriana Ibrahim. Seminar at University of Kuwait A Search and Jump Algorithm for Markov Chain Monte Carlo Sampling Christopher Jennison Department of Mathematical Sciences, University of Bath, UK http://people.bath.ac.uk/mascj Adriana Ibrahim Institute

More information

16 : Approximate Inference: Markov Chain Monte Carlo

16 : Approximate Inference: Markov Chain Monte Carlo 10-708: Probabilistic Graphical Models 10-708, Spring 2017 16 : Approximate Inference: Markov Chain Monte Carlo Lecturer: Eric P. Xing Scribes: Yuan Yang, Chao-Ming Yen 1 Introduction As the target distribution

More information

Stat 516, Homework 1

Stat 516, Homework 1 Stat 516, Homework 1 Due date: October 7 1. Consider an urn with n distinct balls numbered 1,..., n. We sample balls from the urn with replacement. Let N be the number of draws until we encounter a ball

More information

CSC 2541: Bayesian Methods for Machine Learning

CSC 2541: Bayesian Methods for Machine Learning CSC 2541: Bayesian Methods for Machine Learning Radford M. Neal, University of Toronto, 2011 Lecture 4 Problem: Density Estimation We have observed data, y 1,..., y n, drawn independently from some unknown

More information

Markov Chain Monte Carlo Methods

Markov Chain Monte Carlo Methods Markov Chain Monte Carlo Methods John Geweke University of Iowa, USA 2005 Institute on Computational Economics University of Chicago - Argonne National Laboaratories July 22, 2005 The problem p (θ, ω I)

More information

Bayesian Methods and Uncertainty Quantification for Nonlinear Inverse Problems

Bayesian Methods and Uncertainty Quantification for Nonlinear Inverse Problems Bayesian Methods and Uncertainty Quantification for Nonlinear Inverse Problems John Bardsley, University of Montana Collaborators: H. Haario, J. Kaipio, M. Laine, Y. Marzouk, A. Seppänen, A. Solonen, Z.

More information

Down by the Bayes, where the Watermelons Grow

Down by the Bayes, where the Watermelons Grow Down by the Bayes, where the Watermelons Grow A Bayesian example using SAS SUAVe: Victoria SAS User Group Meeting November 21, 2017 Peter K. Ott, M.Sc., P.Stat. Strategic Analysis 1 Outline 1. Motivating

More information

Markov Chain Monte Carlo, Numerical Integration

Markov Chain Monte Carlo, Numerical Integration Markov Chain Monte Carlo, Numerical Integration (See Statistics) Trevor Gallen Fall 2015 1 / 1 Agenda Numerical Integration: MCMC methods Estimating Markov Chains Estimating latent variables 2 / 1 Numerical

More information

The Recycling Gibbs Sampler for Efficient Learning

The Recycling Gibbs Sampler for Efficient Learning The Recycling Gibbs Sampler for Efficient Learning L. Martino, V. Elvira, G. Camps-Valls Universidade de São Paulo, São Carlos (Brazil). Télécom ParisTech, Université Paris-Saclay. (France), Universidad

More information

Eco517 Fall 2013 C. Sims MCMC. October 8, 2013

Eco517 Fall 2013 C. Sims MCMC. October 8, 2013 Eco517 Fall 2013 C. Sims MCMC October 8, 2013 c 2013 by Christopher A. Sims. This document may be reproduced for educational and research purposes, so long as the copies contain this notice and are retained

More information

Bayesian Inference and MCMC

Bayesian Inference and MCMC Bayesian Inference and MCMC Aryan Arbabi Partly based on MCMC slides from CSC412 Fall 2018 1 / 18 Bayesian Inference - Motivation Consider we have a data set D = {x 1,..., x n }. E.g each x i can be the

More information

6 Markov Chain Monte Carlo (MCMC)

6 Markov Chain Monte Carlo (MCMC) 6 Markov Chain Monte Carlo (MCMC) The underlying idea in MCMC is to replace the iid samples of basic MC methods, with dependent samples from an ergodic Markov chain, whose limiting (stationary) distribution

More information

Likelihood-free inference and approximate Bayesian computation for stochastic modelling

Likelihood-free inference and approximate Bayesian computation for stochastic modelling Likelihood-free inference and approximate Bayesian computation for stochastic modelling Master Thesis April of 2013 September of 2013 Written by Oskar Nilsson Supervised by Umberto Picchini Centre for

More information

MSc MT15. Further Statistical Methods: MCMC. Lecture 5-6: Markov chains; Metropolis Hastings MCMC. Notes and Practicals available at

MSc MT15. Further Statistical Methods: MCMC. Lecture 5-6: Markov chains; Metropolis Hastings MCMC. Notes and Practicals available at MSc MT15. Further Statistical Methods: MCMC Lecture 5-6: Markov chains; Metropolis Hastings MCMC Notes and Practicals available at www.stats.ox.ac.uk\ nicholls\mscmcmc15 Markov chain Monte Carlo Methods

More information

Advanced Statistical Modelling

Advanced Statistical Modelling Markov chain Monte Carlo (MCMC) Methods and Their Applications in Bayesian Statistics School of Technology and Business Studies/Statistics Dalarna University Borlänge, Sweden. Feb. 05, 2014. Outlines 1

More information

Brief introduction to Markov Chain Monte Carlo

Brief introduction to Markov Chain Monte Carlo Brief introduction to Department of Probability and Mathematical Statistics seminar Stochastic modeling in economics and finance November 7, 2011 Brief introduction to Content 1 and motivation Classical

More information

DAG models and Markov Chain Monte Carlo methods a short overview

DAG models and Markov Chain Monte Carlo methods a short overview DAG models and Markov Chain Monte Carlo methods a short overview Søren Højsgaard Institute of Genetics and Biotechnology University of Aarhus August 18, 2008 Printed: August 18, 2008 File: DAGMC-Lecture.tex

More information

Inference for a Population Proportion

Inference for a Population Proportion Al Nosedal. University of Toronto. November 11, 2015 Statistical inference is drawing conclusions about an entire population based on data in a sample drawn from that population. From both frequentist

More information

Lecture 8: Bayesian Estimation of Parameters in State Space Models

Lecture 8: Bayesian Estimation of Parameters in State Space Models in State Space Models March 30, 2016 Contents 1 Bayesian estimation of parameters in state space models 2 Computational methods for parameter estimation 3 Practical parameter estimation in state space

More information

PSEUDO-MARGINAL METROPOLIS-HASTINGS APPROACH AND ITS APPLICATION TO BAYESIAN COPULA MODEL

PSEUDO-MARGINAL METROPOLIS-HASTINGS APPROACH AND ITS APPLICATION TO BAYESIAN COPULA MODEL PSEUDO-MARGINAL METROPOLIS-HASTINGS APPROACH AND ITS APPLICATION TO BAYESIAN COPULA MODEL Xuebin Zheng Supervisor: Associate Professor Josef Dick Co-Supervisor: Dr. David Gunawan School of Mathematics

More information

Markov-Chain Monte Carlo

Markov-Chain Monte Carlo Markov-Chain Monte Carlo CSE586 Computer Vision II Spring 2010, Penn State Univ. References Recall: Sampling Motivation If we can generate random samples x i from a given distribution P(x), then we can

More information

F denotes cumulative density. denotes probability density function; (.)

F denotes cumulative density. denotes probability density function; (.) BAYESIAN ANALYSIS: FOREWORDS Notation. System means the real thing and a model is an assumed mathematical form for the system.. he probability model class M contains the set of the all admissible models

More information

Lattice Gaussian Sampling with Markov Chain Monte Carlo (MCMC)

Lattice Gaussian Sampling with Markov Chain Monte Carlo (MCMC) Lattice Gaussian Sampling with Markov Chain Monte Carlo (MCMC) Cong Ling Imperial College London aaa joint work with Zheng Wang (Huawei Technologies Shanghai) aaa September 20, 2016 Cong Ling (ICL) MCMC

More information

13 Notes on Markov Chain Monte Carlo

13 Notes on Markov Chain Monte Carlo 13 Notes on Markov Chain Monte Carlo Markov Chain Monte Carlo is a big, and currently very rapidly developing, subject in statistical computation. Many complex and multivariate types of random data, useful

More information

Stat 451 Lecture Notes Monte Carlo Integration

Stat 451 Lecture Notes Monte Carlo Integration Stat 451 Lecture Notes 06 12 Monte Carlo Integration Ryan Martin UIC www.math.uic.edu/~rgmartin 1 Based on Chapter 6 in Givens & Hoeting, Chapter 23 in Lange, and Chapters 3 4 in Robert & Casella 2 Updated:

More information

Stat 5102 Notes: Markov Chain Monte Carlo and Bayesian Inference

Stat 5102 Notes: Markov Chain Monte Carlo and Bayesian Inference Stat 5102 Notes: Markov Chain Monte Carlo and Bayesian Inference Charles J. Geyer April 6, 2009 1 The Problem This is an example of an application of Bayes rule that requires some form of computer analysis.

More information

Sampling from complex probability distributions

Sampling from complex probability distributions Sampling from complex probability distributions Louis J. M. Aslett (louis.aslett@durham.ac.uk) Department of Mathematical Sciences Durham University UTOPIAE Training School II 4 July 2017 1/37 Motivation

More information

References. Markov-Chain Monte Carlo. Recall: Sampling Motivation. Problem. Recall: Sampling Methods. CSE586 Computer Vision II

References. Markov-Chain Monte Carlo. Recall: Sampling Motivation. Problem. Recall: Sampling Methods. CSE586 Computer Vision II References Markov-Chain Monte Carlo CSE586 Computer Vision II Spring 2010, Penn State Univ. Recall: Sampling Motivation If we can generate random samples x i from a given distribution P(x), then we can

More information

COPYRIGHTED MATERIAL CONTENTS. Preface Preface to the First Edition

COPYRIGHTED MATERIAL CONTENTS. Preface Preface to the First Edition Preface Preface to the First Edition xi xiii 1 Basic Probability Theory 1 1.1 Introduction 1 1.2 Sample Spaces and Events 3 1.3 The Axioms of Probability 7 1.4 Finite Sample Spaces and Combinatorics 15

More information

Minicourse on: Markov Chain Monte Carlo: Simulation Techniques in Statistics

Minicourse on: Markov Chain Monte Carlo: Simulation Techniques in Statistics Minicourse on: Markov Chain Monte Carlo: Simulation Techniques in Statistics Eric Slud, Statistics Program Lecture 1: Metropolis-Hastings Algorithm, plus background in Simulation and Markov Chains. Lecture

More information

Principles of Bayesian Inference

Principles of Bayesian Inference Principles of Bayesian Inference Sudipto Banerjee University of Minnesota July 20th, 2008 1 Bayesian Principles Classical statistics: model parameters are fixed and unknown. A Bayesian thinks of parameters

More information

Likelihood-free MCMC

Likelihood-free MCMC Bayesian inference for stable distributions with applications in finance Department of Mathematics University of Leicester September 2, 2011 MSc project final presentation Outline 1 2 3 4 Classical Monte

More information

A quick introduction to Markov chains and Markov chain Monte Carlo (revised version)

A quick introduction to Markov chains and Markov chain Monte Carlo (revised version) A quick introduction to Markov chains and Markov chain Monte Carlo (revised version) Rasmus Waagepetersen Institute of Mathematical Sciences Aalborg University 1 Introduction These notes are intended to

More information

Simulation - Lectures - Part III Markov chain Monte Carlo

Simulation - Lectures - Part III Markov chain Monte Carlo Simulation - Lectures - Part III Markov chain Monte Carlo Julien Berestycki Part A Simulation and Statistical Programming Hilary Term 2018 Part A Simulation. HT 2018. J. Berestycki. 1 / 50 Outline Markov

More information

Introduction to Computational Biology Lecture # 14: MCMC - Markov Chain Monte Carlo

Introduction to Computational Biology Lecture # 14: MCMC - Markov Chain Monte Carlo Introduction to Computational Biology Lecture # 14: MCMC - Markov Chain Monte Carlo Assaf Weiner Tuesday, March 13, 2007 1 Introduction Today we will return to the motif finding problem, in lecture 10

More information