Robust MCMC Algorithms for Bayesian Inference in Stochastic Epidemic Models.

Size: px
Start display at page:

Download "Robust MCMC Algorithms for Bayesian Inference in Stochastic Epidemic Models."

Transcription

1 Robust MCMC Algorithms for Bayesian Inference in Stochastic Epidemic Models. An Application to the 2001 UK Foot-and-Mouth Outbreak Theodore University of Nottingham Gareth O. Lancaster University November, Cambridge

2 Introduction to Epidemic Models

3 Background Understanding the spread of an infectious disease is an important issue in order to prevent major outbreaks of an epidemic. In general, inference problems for disease outbreak data are complicated by the facts that the data are inherently dependent and the data are usually incomplete in the sense that the actual process of infection is not observed. However, it is often possible to formulate simple stochastic models which describe the key features of epidemic spread.

4 The SIR Model Principles Consider a closed population (no births/deaths) of size N individuals α initially infected individuals at any given time point, each individual i is in one of the three states: Susceptible Infected Removed S I R

5 The Stochastic Process Each of the infected individuals remains infectious for some period which is distributed according to the distribution of random variable D. While infectious, makes contacts with each of the N individuals in the population at times given by the points of a Poisson process of rate β > 0. Any such contact with a susceptible individual immediately makes the susceptible an infective. At the end of its infectious period an infective no longer makes any contacts and is said to be removed.

6 Infectious Period The infectious periods of different infectives are assumed to be independent and identically distributed according to the distribution of a random variable D, which can have any arbitrary but specified distribution. The special (Markovian) case where the infectious period follows an Exponential distribution is known as the General Stochastic Epidemic (GSE).

7 The GSE Model The GSE model has received a lot of attention in the probabilistic literature. Although assuming an Exponential infectious period is mathematically convenient (limit results etc) it is not biologically motivated. The GSE has been generalised in a large number of ways, such as: Gamma or Weibull distributed infectious periods (O Neill and Becker, 2001 & Streftaris and Gibson, 2004). Incorporate latent infectious periods (eg. SEIR models)

8 Modelling the 2001 UK Foot-and-Mouth Disease (Motivation)

9 Introduction Livestock diseases are becoming significantly important due to their welfare and economic consequences. The Foot-and-Mouth disease (FMD) is considered the most important since it can be spread rapidly between livestock species. In general, FMD is rarely lethal to adult livestock, but causes blister on the mouth and feet which often leads to a significant drop in milk production in dairy cattle. It also causes very slow weight gain in other livestock. The economic effects of infection within a country are dramatic; prevention of export of meat and milk to other countries eliminates a vital source of revenue.

10 Effects of the FMD to the country The UK Foot-and-Mouth epidemic of 2001 is thought to have entered the UK in early February of that year. The epidemic resulted in recorded known infections and culls at Infected Premises (IPs) across the UK. Over 10,000 other farms with unknown infection status, but determined to be at high risk, had animals slaughtered. Non-IP slaughter policy has also taken place; Contiguous Premises (CPs) and Dangerous Contacts (DC). In total over 4,000,000 animals were slaughtered as part of the control program to eradicate the disease.

11 Effects of the FMD to the country (cont.) The estimated costs to the UK economy of the 2001 FMD epidemic consisted of: US $4.7 billion to the food and farming industries US $4.5 billion to the leisure and tourism industry; as well as US $3 billion in indirect costs. Source: Thompson et al. (2002)

12 Population of Farms in the UK

13 FMD in Cumbria The resulting dataset consists of N = 5378 (initially susceptible) farms in total. At the end of the outbreak, n I = 1021 got infected. This data set directly gives the cull-dates of each farm on which animals were culled, but the infection-dates are not observed. Unfortunately, information on which farms have been culled without knowing their infection status (eg. dangerous contacts) is unavailable to us.

14 Modelling Issues The GSE does not seem to be appropriate since the following assumptions are questionable: Homogeneously mixing population; Exponential Infectious Period. Therefore, Associate the infection rate with the available risk factors such as the distance between the two farms and the number different species within each of them. Assume a more plausible distribution for the infectious period (eg. Gamma).

15 Heterogeneously Mixing Stochastic Epidemic Models (A General Framework)

16 A Heterogeneously Mixing Stochastic Epidemic Model Model Specification Infection Rate: Infectious Period: β ij := β 0 h(i, j) (1) D i Ga(α, γ) (2) The parameters of interest are β 0, γ > 0 (α assumed to be known). The (deterministic) function h(i, j) refers to individual s characteristics which might involve some other parameters as well.

17 Inference GOAL: Draw inference for the parameters (β 0, γ) Likelihood: f (I, R β 0, γ) where n I i=1,i k γ αn R exp { Y i = {j : I j < I i < R j } β ji exp j Yi n R i=1 k denotes the initial infective. γ(r i I i )} T I k i I t j S t I t and S t denote the set of infectives and susceptibles just prior time t. I = (I 1,..., I ni ) and R = (R 1,..., R nr ) β ij dt

18 Inference for Partially Observed Epidemics However the infections times are rarely observed and therefore the inference now becomes slightly more complicated since the likelihood of the observed data (R) given the parameters (β 0, γ) is intractable. In a Bayesian framework, we need to impute the missing data I. Multiplying the prior of the parameters times the likelihood we get the posterior distribution of the parameters given the removal times π(β 0, γ, I R).

19 Full Conditional Distributions of the Parameters Model Parameters: β 0 γ, I, R Ga λ β0 + n I 1, ν β0 + Missing Data: π(i R, β 0, γ) γ β 0, I, R Ga n I i=1,i k exp ( T I k λ γ + n R, ν γ + i I t j S t n R i=1 β ji exp j Yi { } γ n R i=1 (R i I i ) (R i I i ) T I k h(i, j) dt ) i I t j S t β ij dt (3)

20 MCMC Algorithms for Stochastic Epidemics Gibson and Renshaw (1998) & Roberts and O Neill (1999) were the first to apply MCMC methods for stochastic epidemics. Outline of the MCMC (centered) algorithm: 1. Initialise the algorithm; 2. Choose uniformly one (or more) infection time(s) I j,j = 1,..., n I and update it (them) individually using a Metropolis Hastings step by proposing a replacement infection time; 3. Update the parameters β and γ via a Gibbs step.

21 MCMC Implementation Terminology: Random Scan (RS): At each MCMC iteration we update only one of the infection times (chosen at random). δ% Deterministic Scan (δ% DS): At each MCMC iteration, we choose (at random) to update δ% of the infection times. Computation: The integral involved in the likelihood can be rewritten as follows: T I k i I t j S t β ij dt = n I i=1 j=1 N β ij (R i I j I j I i ). This reduces the computational cost of evaluating the likelihood significantly.

22 How to propose a new infection time?

23 Updating the Infection Times The Metropolis step can be implemented by choosing one of the following proposals: Independent Sampler I - O Neill & Roberts (1999): I j U(I k, R k ) Independent Sampler II - O Neill & Becker (2001) : R j I j Ga(α, γ) Random Walk Metropolis - Neal and Roberts (2003): I j N(I j, σ 2 ).

24 Remarks (Efficient) Block updating of the infection times is very hard especially when the size of the data (n I, N ) is big. Gibbs step for the infection times is possible, but difficult to construct. The independence sampler R i I i Ga(α, γ) seems to be the most efficient proposal among the previously mentioned sampler.

25 Random Scan vs Deterministic Scan

26 Updating the infection times I [C] RS: I ACF Lag

27 Updating the infection times I (cont.) [C] 10% DS: I ACF Lag

28 Updating the infection times I(cont.) [C] 50% DS: I ACF Lag

29 Updating the infection times I(cont.) [C] 100% DS: I ACF Lag

30 Results Ignoring the cpu time needed to run an algorithm, 100% Deterministic scan performs better than any other chosen algorithm (ie. δ < 1). In general, increasing δ improves the mixing of the MCMC algorithm, nevertheless is more computationally costly (slower in cpu time). It is of interest to quantify the optimal percentage of deterministic scan taking into account mixing and cpu time. The optimal δ will depend on the size of the epidemic (n I, N ). Easy to do it computationally hard to derive a theoretical result.

31 Alternate the target distribution

32 Alternative Target Distributions Since the full conditional posterior distributions of β 0 and γ are of a standard form (Gamma distr.) and they are conditionally independent. Therefore either or both can be integrated out from the target distribution. Therefore, we could construct (similar to the previous) MCMC algorithms having the following target distributions: π(β 0, I R) = π(β 0, γ, I R) dγ γ π(γ, I R) = π(β 0, γ, I R) dβ 0 β 0 π(i R) = π(β 0, γ, I R) dβ 0 dγ β 0 γ

33 Which is one is the best? It turns out that if δ is chosen to be relatively small, then pretty similar results (regarding efficiency) are obtained. However If we increase δ, and in particular we choose a 100% DS, integrating both model parameters out (i.e. β 0 and γ) and choose as target distribution π(i R) seems to increase (slightly) the algorithm s efficiency compared to the standard target distribution (π(i, β 0, γ R).

34 Remarks When the parameter γ is integrated out, for example in π(β 0, I R) or π(i R), the previously mentioned independence sampler needs to be modified: R i I i Ga(α, γ) A fixed value could be assigned to γ (obtained via a pilot study). Otherwise, a pseudo -MLE of γ can be used: αn R γ = nr i=1 (R i Ii c ) where I c denotes the vector of the current infection times.

35 Mixing of the MCMC Algorithms for different level of variability of the infectious period

36 Model s Equations A Working Example β ij = β 0 exp { ρ(i, j)} R i I i Gamma(α, γ) N = 501 (1 initially infective) where ρ(i, j) denotes the distance between the two individuals. Simulated Dataset D1 D2 D3 True β True α True γ n I = n R

37 A Working Example (cont.) Distributions of the Infectious Periods Density f(x) D1 D2 D

38 (α, γ) = (0.5, 0.25) Mixing of the MCMC Algorithm ACF plot for γ :(Dataset 1) ACF Lag

39 (α, γ) = (2, 1) Mixing of the MCMC Algorithms (cont.) ACF plot for γ :(Dataset 2) ACF Lag

40 Mixing of the MCMC Algorithms (cont.) (α, γ) = (5, 2.5) ACF plot for γ :(Dataset 3) ACF Lag

41 Reasons for Poor Mixing If we carefully look at one of the likelihood equations which describes the infectious period of each individual: R i I i Ga(α, γ). This reveals that a-priori γ and I are dependent and this causes problems to the MCMC mixing. R i I i Ga(α, γ), for i = 1,..., n I n I (R i I i ) Ga(αn I, γ) i=1

42 Reasons for Poor Mixing (cont.) Thus for large n I or α, the parameter γ and the sum of the infectious periods n I i=1 (R i I i ) are a-priori heavily dependent. If these two were the parameters of interest, then this a-priori correlation would have caused mixing problems in the case of a two-state Gibbs sampler [see Amit(1991), Roberts and Sahu (1997)]. Things are more complicated in practice since the MCMC schemes used so far involved deterministic scan update of the each of infection times I i, i = 1,..., n I.

43 Correlation between Infection times and γ Dataset 1 Dataset 2 Dataset 3 γ γ γ I I I

44 Aim: Construct more efficient algorithms

45 Bayesian Hierarchical Models In Bayesian hierarchical models when the missing data contain much more information than the observed data about the parameter of interest then the standard data-augmentation techniques perform poorly. In epidemic models β 0 and γ are conditionally independent on the infection times I. Nevertheless, each of them is very dependent with I. Therefore we are interested in applying a parameterisation where γ and I are a-priori independent

46 Non Centered Parameterisations (NCP) The non-centered methodology presented in Papaspiliopoulos et. al (2003) suggests that appropriate non centered parameterisations out perform the centered existing centered algorithms over a range of examples. Their findings can be summarized as follows: When the observed (missing) data are much more informative about the parameter of interest than the missing (observed) then is better to choose a centered (non centered) parameterisation.

47 NCP for Stochastic Epidemics Neal and Roberts (2005) have applied this methodology to the General Stochastic Epidemic model. We introduce a non centered parameterisation: U i = γ(r i I i ), i = 1,..., n I It easy to see that since U i Ga(α, 1),i = 1,..., n I, the random variables U = (U 1,..., U ni ) T and γ are apriori independent. Change in variables: (I, β 0, γ, R) (U, β 0, γ, R).

48 NC Algorithm s Implementation An outline of an implementation of a Non Centered algorithm is the following: Outline 1. Get a sample of (γ, I) via a centered algorithm; 2. Transform the I s to U s via U i = γ(r i I i ) 3. Update γ using Metropolis Hastings algorithm using π(γ U, β 0, R) ;

49 The Existing Approach Neal and Roberts (2005) draw samples of (γ, I) via MCMC on π(γ, I R) using the standard algorithm. They update γ via a Random Walk Metropolis step, i.e. γ N(γ, σ 2 γ)

50 An Alternative Approach We propose to draw samples of (γ, I) via the most efficient centered algorithm having as target distribution one of the following: π(β 0, I R), π(γ, I R), π(i R) Note: A modification of the previously used proposal [R i I i Γ(α, γ)] is needed when γ is integrated out from the target distribution π( ).

51 Updating γ Apart from a Multiplicative Random Walk, we also propose to use one of the 3 different proposals: Independent Sampler I ( Pseudo Gibbs ): In more detail: q 1 (γ ) π(γ I c, R) Given the current values of the infection times, Ii c, i = 1,..., n I propose a new value for γ by sampling from ( ) γ new I c, R Ga λ γ + n R, ν γ + n R i=1 (R i I c i )

52 Updating γ (cont.) Independent Sampler II (Normal Approximation): In more detail: q 1 (γ ) N(ˆγ MLE, ɛ σγ 2 MLE ˆ ), for some ɛ > 1, If the infection times were known, inference for γ is straightforward: γ MLE = αn R nr i=1 (R i I i ) σγ 2 MLE ˆ can be very easily derived ɛ is chosen in order to allow the MH algorithm for big jumps.

53 Updating γ (cont.) Independent sampler III (Adaptive Sampling): In more detail: q 1 (γ ) Ga(a, b) Run the centered algorithm for an adequate number of iterations (pilot study) Try to approximate the derived (marginal) posterior with an appropriate well known distribution Estimate the parameters of such a distribution (eg. methods of moments)

54 Partially Non Centered Parameterisations The set of the infected individuals in the epidemic, is partitioned into two groups, C and U. Each individual is assigned to one of the two groups (independently of each other) with probability µ For those individuals in the U, let U i = γ(r i I i ) (i U) We propose a change in variable from (I C, I U, β 0, γ, R) to (I C, U U, β 0, γ, R)

55 Results - Dataset D1 [C] [PNC] ACF ACF Lag Lag

56 Results - Dataset D2 [C] [PNC] ACF ACF Lag Lag

57 Results - Dataset D3 [C] [PNC] ACF ACF Lag Lag

58 Conclusions Although the implementation of the standard (centered) MCMC algorithms for SIR-type models is quite straightforward their efficiency deteriorates as the number of infected individuals (n I ) increases. Non Centered Parameterisations often offer much more robust algorithms there is relatively little extra cost in implementing such an algorithm.

59 Modelling the 2001 UK Foot-and-Mouth Disease (A Bayesian Analysis)

60 Population of Farms in Cumbria

61 A Recap... The resulting dataset consists of N = 5378 (initially susceptible) farms in total. At the end of the outbreak, n I = 1021 got infected. This data set directly gives the cull-dates of each farm on which animals were culled, but the infection-dates are not observed. Unfortunately, information on which farms have been culled without knowing their infection status (eg. dangerous contacts) is unavailable to us.

62 Modelling the Infection Rate The Model β ij = β 0 K(ρ(i, j); δ) Infection Kernel ( ɛ (nc) i ζ + (ns) i ζ) Farm s Infectivity ( ξ (nc) j ζ + (ns) j ζ) Farm s Susceptibility where ɛ and ξ represent the relative infectivity and susceptibility of cows to sheep. Parameter ζ allows for a non linear effect. ρ(i, j) denotes a measure of distance between farms i and j.

63 Infectious Kernel The environmental spread is modelled by assuming a Cauchy kernel K(ρ i,j, δ) and is associated with parameter δ: K(ρ(i, j); δ) = δ ρ 2 ij + δ2 Other Kernels such as Exponential, Geometric could be also used. The Cauchy is chosen such that we adjust for long-range infections.

64 Infectious Kernel (cont.) The Euclidean distance is used as a measure for the distance between the farms. Other metrics could be used such as minimum walking distance. It is questionable which is the most appropriate (see Savill et. al. 2006).

65 Infectious Period The infectious period is assumed to be Gamma distributed: We also assume that α = 4. R i I i Ga(α, γ) Note that the infection times I i are unknown and will be imputed (estimated) within our framework.

66 Results - Kernel Spatial Kernel K(i, j) distance (in Km)

67 Results - Infectivity Posterior Distribution of ε Posterior Density ε

68 Results - Susceptibility Posterior Distribution of ξ Posterior Density ξ

69 Results - Non Linearity Posterior Distribution of ζ Posterior Density ζ

70 Summary A Heterogeneously Mixing Stochastic Epidemic (HMSE) model was applied in order to describe the dynamics of the 2001 UK FMD in Cumbria. The effect of the spatial kernel is very little when the distance between the two farms is > 3Km. Cattle were found to be more infective and more susceptible than sheep. There is strong evidence that the effect of the number of animals to the infection rate is non linear.

71 Work in Progress Incorporate time dependent infection rates. Develop efficient NC parameterisations for more complex models (eg. latent periods) Providing a real-time risk assessment tool for a potential outbreak of an Avian Influenza (Bird Flu) outbreak in the poultry industry of the UK (Joint work with Chris Jewell). Perform a comprehensive analysis for the 2001 UK FMD by taking into account the total population of farms in the UK (extra information is needed).

72 Modelling Highly Pathogenic Avian Influenza (A Bayesian Approach) Joint work with Chris Jewell and Gareth Lancaster University

73 Statistical Analysis A comprehensive statistical analysis must address: The dynamics of an infection Transmission and removal rates Imputation of unobserved data: Infection times If used during an epidemic, we require a method to impute occult infections

74 The Model S I N R Infectivity Time

75 Infectivity and Time to Notification Functions Infectivity function We assume a time-dependent transmission rate where n I β j = β 0 + [β ij h(i j I i )] h(t) = i eνt µ + e νt µ > 0, ν > 0 Time to Notification P(D > d) = e { a(e{b d} 1)} Time between infection and notification, D I = N I where a > 0, b > 0. [ Note: Parameters µ, ν, a and b are assumed known]

76 Transmission rate The Model (cont.) β ij = β sp,j β 1 C FM + β 2 C SH + β 3 C CP + β 4 e δρ[i,j], i I, j S β 5 e δρ[i,j], i N, j S Parameters β 0 β 1, β 2, β 3 β 4, β 5, δ β sp,j Spread Eg. Wild Bird Population Network Spatial Species susceptibility C FM, C SH, C CP : denote the contact matrices; ρ[i, j]: denotes the Euclidean distance between farms and i n and j

77 Toolbox Prior Information for θ Expert s opinion Previous disease outbreak data Likelihood Probability of observing what we have observed. Prior MCMC Posterior Likelihood

78 Markov Chain Monte Carlo algorithm Repeat the following steps: 1. Simulate the model parameters θ 2. Simulate the missing data (infection times) 3. Simulate occult infections NB. Simulation of occult infections changes the dimension of the parameter space.

79 Posteriors Rigorous probability statements for quantifying: Effectiveness of protection zones (β 5 /β 4 ) Relative importance of Network and Environmental spread. Contribution of Network components: (Feed Mills, Slaughter Houses, Company Contacts) Transmission rate β ij = β sp,j β 1 C FM + β 2 C SH + β 3 C CP + β 4 e δρ[i,j], i I, j S β 5 e δρ[i,j], i N, j S

80 Risk Analysis Target Provide real-time risk analysis during a potential outbreak, taking into account uncertainty in our parameter estimates. Challenges Denote by R i, the expected number of farms each farm i might infect were it to be the initial infective in an otherwise susceptible population. How does R i for each single farm evolve during the outbreak? How does the risk to specific areas of the country change over the course of the epidemic? How does the parameter uncertainty change with the amount of available data?

81 Farm-specific R 0 : R i The expected number of farms each farm i might infect were it to be the initial infective in an otherwise susceptible population). (a) Day 7 (b) Day 14 (c) End Figure: Risk analyses for the three stages of the epidemic.

82 Risk Maps The probability of farms becoming infected given the current situation (a) Day 7 (b) Day 14 (c) End Figure: Risk analyses for the three stages of the epidemic.

83 Conclusions We have constructed a robust flexible MCMC-based Bayesian approach for real-time parameter estimation. In conjunction with forward simulation, this provides a powerful risk assessment tool for use during a disease epidemic in the UK.. Such an approach can be easily adopted for other model structures as well. Parallel computing is necessary when dealing with large datasets.

MCMC 2: Lecture 3 SIR models - more topics. Phil O Neill Theo Kypraios School of Mathematical Sciences University of Nottingham

MCMC 2: Lecture 3 SIR models - more topics. Phil O Neill Theo Kypraios School of Mathematical Sciences University of Nottingham MCMC 2: Lecture 3 SIR models - more topics Phil O Neill Theo Kypraios School of Mathematical Sciences University of Nottingham Contents 1. What can be estimated? 2. Reparameterisation 3. Marginalisation

More information

Statistical Inference for Stochastic Epidemic Models

Statistical Inference for Stochastic Epidemic Models Statistical Inference for Stochastic Epidemic Models George Streftaris 1 and Gavin J. Gibson 1 1 Department of Actuarial Mathematics & Statistics, Heriot-Watt University, Riccarton, Edinburgh EH14 4AS,

More information

MCMC 2: Lecture 2 Coding and output. Phil O Neill Theo Kypraios School of Mathematical Sciences University of Nottingham

MCMC 2: Lecture 2 Coding and output. Phil O Neill Theo Kypraios School of Mathematical Sciences University of Nottingham MCMC 2: Lecture 2 Coding and output Phil O Neill Theo Kypraios School of Mathematical Sciences University of Nottingham Contents 1. General (Markov) epidemic model 2. Non-Markov epidemic model 3. Debugging

More information

Bayesian inference and model selection for stochastic epidemics and other coupled hidden Markov models

Bayesian inference and model selection for stochastic epidemics and other coupled hidden Markov models Bayesian inference and model selection for stochastic epidemics and other coupled hidden Markov models (with special attention to epidemics of Escherichia coli O157:H7 in cattle) Simon Spencer 3rd May

More information

Statistical Analysis of Spatio-temporal Point Process Data. Peter J Diggle

Statistical Analysis of Spatio-temporal Point Process Data. Peter J Diggle Statistical Analysis of Spatio-temporal Point Process Data Peter J Diggle Department of Medicine, Lancaster University and Department of Biostatistics, Johns Hopkins University School of Public Health

More information

Bayesian Inference for Contact Networks Given Epidemic Data

Bayesian Inference for Contact Networks Given Epidemic Data Bayesian Inference for Contact Networks Given Epidemic Data Chris Groendyke, David Welch, Shweta Bansal, David Hunter Departments of Statistics and Biology Pennsylvania State University SAMSI, April 17,

More information

Markov Chain Monte Carlo

Markov Chain Monte Carlo 1 Motivation 1.1 Bayesian Learning Markov Chain Monte Carlo Yale Chang In Bayesian learning, given data X, we make assumptions on the generative process of X by introducing hidden variables Z: p(z): prior

More information

A tutorial introduction to Bayesian inference for stochastic epidemic models using Approximate Bayesian Computation

A tutorial introduction to Bayesian inference for stochastic epidemic models using Approximate Bayesian Computation A tutorial introduction to Bayesian inference for stochastic epidemic models using Approximate Bayesian Computation Theodore Kypraios 1, Peter Neal 2, Dennis Prangle 3 June 15, 2016 1 University of Nottingham,

More information

STA 4273H: Statistical Machine Learning

STA 4273H: Statistical Machine Learning STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Computer Science! Department of Statistical Sciences! rsalakhu@cs.toronto.edu! h0p://www.cs.utoronto.ca/~rsalakhu/ Lecture 7 Approximate

More information

Computational statistics

Computational statistics Computational statistics Markov Chain Monte Carlo methods Thierry Denœux March 2017 Thierry Denœux Computational statistics March 2017 1 / 71 Contents of this chapter When a target density f can be evaluated

More information

arxiv: v1 [stat.co] 13 Oct 2017

arxiv: v1 [stat.co] 13 Oct 2017 Bayes factors for partially observed stochastic epidemic models arxiv:1710.04977v1 [stat.co] 13 Oct 2017 Muteb Alharthi Theodore Kypraios Philip D. O Neill June 27, 2018 Abstract We consider the problem

More information

Markov Chain Monte Carlo methods

Markov Chain Monte Carlo methods Markov Chain Monte Carlo methods Tomas McKelvey and Lennart Svensson Signal Processing Group Department of Signals and Systems Chalmers University of Technology, Sweden November 26, 2012 Today s learning

More information

Introduction to Machine Learning CMU-10701

Introduction to Machine Learning CMU-10701 Introduction to Machine Learning CMU-10701 Markov Chain Monte Carlo Methods Barnabás Póczos & Aarti Singh Contents Markov Chain Monte Carlo Methods Goal & Motivation Sampling Rejection Importance Markov

More information

Bayesian Estimation of Input Output Tables for Russia

Bayesian Estimation of Input Output Tables for Russia Bayesian Estimation of Input Output Tables for Russia Oleg Lugovoy (EDF, RANE) Andrey Polbin (RANE) Vladimir Potashnikov (RANE) WIOD Conference April 24, 2012 Groningen Outline Motivation Objectives Bayesian

More information

Bayesian Inference in GLMs. Frequentists typically base inferences on MLEs, asymptotic confidence

Bayesian Inference in GLMs. Frequentists typically base inferences on MLEs, asymptotic confidence Bayesian Inference in GLMs Frequentists typically base inferences on MLEs, asymptotic confidence limits, and log-likelihood ratio tests Bayesians base inferences on the posterior distribution of the unknowns

More information

ABC methods for phase-type distributions with applications in insurance risk problems

ABC methods for phase-type distributions with applications in insurance risk problems ABC methods for phase-type with applications problems Concepcion Ausin, Department of Statistics, Universidad Carlos III de Madrid Joint work with: Pedro Galeano, Universidad Carlos III de Madrid Simon

More information

Modelling spatio-temporal patterns of disease

Modelling spatio-temporal patterns of disease Modelling spatio-temporal patterns of disease Peter J Diggle CHICAS combining health information, computation and statistics References AEGISS Brix, A. and Diggle, P.J. (2001). Spatio-temporal prediction

More information

Bayesian Methods for Machine Learning

Bayesian Methods for Machine Learning Bayesian Methods for Machine Learning CS 584: Big Data Analytics Material adapted from Radford Neal s tutorial (http://ftp.cs.utoronto.ca/pub/radford/bayes-tut.pdf), Zoubin Ghahramni (http://hunch.net/~coms-4771/zoubin_ghahramani_bayesian_learning.pdf),

More information

CPSC 540: Machine Learning

CPSC 540: Machine Learning CPSC 540: Machine Learning MCMC and Non-Parametric Bayes Mark Schmidt University of British Columbia Winter 2016 Admin I went through project proposals: Some of you got a message on Piazza. No news is

More information

Bayesian inference for stochastic multitype epidemics in structured populations using sample data

Bayesian inference for stochastic multitype epidemics in structured populations using sample data Bayesian inference for stochastic multitype epidemics in structured populations using sample data PHILIP D. O NEILL School of Mathematical Sciences, University of Nottingham, Nottingham, UK. SUMMARY This

More information

Introduction to Markov Chain Monte Carlo & Gibbs Sampling

Introduction to Markov Chain Monte Carlo & Gibbs Sampling Introduction to Markov Chain Monte Carlo & Gibbs Sampling Prof. Nicholas Zabaras Sibley School of Mechanical and Aerospace Engineering 101 Frank H. T. Rhodes Hall Ithaca, NY 14853-3801 Email: zabaras@cornell.edu

More information

eqr094: Hierarchical MCMC for Bayesian System Reliability

eqr094: Hierarchical MCMC for Bayesian System Reliability eqr094: Hierarchical MCMC for Bayesian System Reliability Alyson G. Wilson Statistical Sciences Group, Los Alamos National Laboratory P.O. Box 1663, MS F600 Los Alamos, NM 87545 USA Phone: 505-667-9167

More information

Likelihood-free MCMC

Likelihood-free MCMC Bayesian inference for stable distributions with applications in finance Department of Mathematics University of Leicester September 2, 2011 MSc project final presentation Outline 1 2 3 4 Classical Monte

More information

A Search and Jump Algorithm for Markov Chain Monte Carlo Sampling. Christopher Jennison. Adriana Ibrahim. Seminar at University of Kuwait

A Search and Jump Algorithm for Markov Chain Monte Carlo Sampling. Christopher Jennison. Adriana Ibrahim. Seminar at University of Kuwait A Search and Jump Algorithm for Markov Chain Monte Carlo Sampling Christopher Jennison Department of Mathematical Sciences, University of Bath, UK http://people.bath.ac.uk/mascj Adriana Ibrahim Institute

More information

A Review of Pseudo-Marginal Markov Chain Monte Carlo

A Review of Pseudo-Marginal Markov Chain Monte Carlo A Review of Pseudo-Marginal Markov Chain Monte Carlo Discussed by: Yizhe Zhang October 21, 2016 Outline 1 Overview 2 Paper review 3 experiment 4 conclusion Motivation & overview Notation: θ denotes the

More information

Bayesian estimation of the basic reproduction number in stochastic epidemic models

Bayesian estimation of the basic reproduction number in stochastic epidemic models Bayesian estimation of the basic reproduction number in stochastic epidemic models Damian Clancy Department of Mathematical Sciences, University of Liverpool Philip D. O Neill School of Mathematical Sciences,

More information

Lecture 16: Mixtures of Generalized Linear Models

Lecture 16: Mixtures of Generalized Linear Models Lecture 16: Mixtures of Generalized Linear Models October 26, 2006 Setting Outline Often, a single GLM may be insufficiently flexible to characterize the data Setting Often, a single GLM may be insufficiently

More information

Markov Chain Monte Carlo

Markov Chain Monte Carlo Markov Chain Monte Carlo Recall: To compute the expectation E ( h(y ) ) we use the approximation E(h(Y )) 1 n n h(y ) t=1 with Y (1),..., Y (n) h(y). Thus our aim is to sample Y (1),..., Y (n) from f(y).

More information

17 : Markov Chain Monte Carlo

17 : Markov Chain Monte Carlo 10-708: Probabilistic Graphical Models, Spring 2015 17 : Markov Chain Monte Carlo Lecturer: Eric P. Xing Scribes: Heran Lin, Bin Deng, Yun Huang 1 Review of Monte Carlo Methods 1.1 Overview Monte Carlo

More information

Enhancing Bayesian risk prediction for epidemics. using contact tracing

Enhancing Bayesian risk prediction for epidemics. using contact tracing Enhancing Bayesian risk prediction for epidemics using contact tracing arxiv:1203.3366v1 [stat.me] 15 Mar 2012 Chris P Jewell and Gareth O Roberts Department of Statistics University of Warwick Coventry

More information

27 : Distributed Monte Carlo Markov Chain. 1 Recap of MCMC and Naive Parallel Gibbs Sampling

27 : Distributed Monte Carlo Markov Chain. 1 Recap of MCMC and Naive Parallel Gibbs Sampling 10-708: Probabilistic Graphical Models 10-708, Spring 2014 27 : Distributed Monte Carlo Markov Chain Lecturer: Eric P. Xing Scribes: Pengtao Xie, Khoa Luu In this scribe, we are going to review the Parallel

More information

Pattern Recognition and Machine Learning. Bishop Chapter 11: Sampling Methods

Pattern Recognition and Machine Learning. Bishop Chapter 11: Sampling Methods Pattern Recognition and Machine Learning Chapter 11: Sampling Methods Elise Arnaud Jakob Verbeek May 22, 2008 Outline of the chapter 11.1 Basic Sampling Algorithms 11.2 Markov Chain Monte Carlo 11.3 Gibbs

More information

Default Priors and Effcient Posterior Computation in Bayesian

Default Priors and Effcient Posterior Computation in Bayesian Default Priors and Effcient Posterior Computation in Bayesian Factor Analysis January 16, 2010 Presented by Eric Wang, Duke University Background and Motivation A Brief Review of Parameter Expansion Literature

More information

Kernel adaptive Sequential Monte Carlo

Kernel adaptive Sequential Monte Carlo Kernel adaptive Sequential Monte Carlo Ingmar Schuster (Paris Dauphine) Heiko Strathmann (University College London) Brooks Paige (Oxford) Dino Sejdinovic (Oxford) December 7, 2015 1 / 36 Section 1 Outline

More information

Development of Stochastic Artificial Neural Networks for Hydrological Prediction

Development of Stochastic Artificial Neural Networks for Hydrological Prediction Development of Stochastic Artificial Neural Networks for Hydrological Prediction G. B. Kingston, M. F. Lambert and H. R. Maier Centre for Applied Modelling in Water Engineering, School of Civil and Environmental

More information

Monte Carlo in Bayesian Statistics

Monte Carlo in Bayesian Statistics Monte Carlo in Bayesian Statistics Matthew Thomas SAMBa - University of Bath m.l.thomas@bath.ac.uk December 4, 2014 Matthew Thomas (SAMBa) Monte Carlo in Bayesian Statistics December 4, 2014 1 / 16 Overview

More information

Assessing Regime Uncertainty Through Reversible Jump McMC

Assessing Regime Uncertainty Through Reversible Jump McMC Assessing Regime Uncertainty Through Reversible Jump McMC August 14, 2008 1 Introduction Background Research Question 2 The RJMcMC Method McMC RJMcMC Algorithm Dependent Proposals Independent Proposals

More information

An introduction to adaptive MCMC

An introduction to adaptive MCMC An introduction to adaptive MCMC Gareth Roberts MIRAW Day on Monte Carlo methods March 2011 Mainly joint work with Jeff Rosenthal. http://www2.warwick.ac.uk/fac/sci/statistics/crism/ Conferences and workshops

More information

Dynamic Factor Models and Factor Augmented Vector Autoregressions. Lawrence J. Christiano

Dynamic Factor Models and Factor Augmented Vector Autoregressions. Lawrence J. Christiano Dynamic Factor Models and Factor Augmented Vector Autoregressions Lawrence J Christiano Dynamic Factor Models and Factor Augmented Vector Autoregressions Problem: the time series dimension of data is relatively

More information

Inference for partially observed stochastic dynamic systems: A new algorithm, its theory and applications

Inference for partially observed stochastic dynamic systems: A new algorithm, its theory and applications Inference for partially observed stochastic dynamic systems: A new algorithm, its theory and applications Edward Ionides Department of Statistics, University of Michigan ionides@umich.edu Statistics Department

More information

STA 216, GLM, Lecture 16. October 29, 2007

STA 216, GLM, Lecture 16. October 29, 2007 STA 216, GLM, Lecture 16 October 29, 2007 Efficient Posterior Computation in Factor Models Underlying Normal Models Generalized Latent Trait Models Formulation Genetic Epidemiology Illustration Structural

More information

A Bayesian methodology for systemic risk assessment in financial networks

A Bayesian methodology for systemic risk assessment in financial networks A Bayesian methodology for systemic risk assessment in financial networks Axel Gandy Imperial College London 9/12/2015 Joint work with Luitgard Veraart (London School of Economics) LUH-Kolloquium Versicherungs-

More information

Bayesian model selection for computer model validation via mixture model estimation

Bayesian model selection for computer model validation via mixture model estimation Bayesian model selection for computer model validation via mixture model estimation Kaniav Kamary ATER, CNAM Joint work with É. Parent, P. Barbillon, M. Keller and N. Bousquet Outline Computer model validation

More information

Brief introduction to Markov Chain Monte Carlo

Brief introduction to Markov Chain Monte Carlo Brief introduction to Department of Probability and Mathematical Statistics seminar Stochastic modeling in economics and finance November 7, 2011 Brief introduction to Content 1 and motivation Classical

More information

A = {(x, u) : 0 u f(x)},

A = {(x, u) : 0 u f(x)}, Draw x uniformly from the region {x : f(x) u }. Markov Chain Monte Carlo Lecture 5 Slice sampler: Suppose that one is interested in sampling from a density f(x), x X. Recall that sampling x f(x) is equivalent

More information

CSC 2541: Bayesian Methods for Machine Learning

CSC 2541: Bayesian Methods for Machine Learning CSC 2541: Bayesian Methods for Machine Learning Radford M. Neal, University of Toronto, 2011 Lecture 3 More Markov Chain Monte Carlo Methods The Metropolis algorithm isn t the only way to do MCMC. We ll

More information

Lecture 7 and 8: Markov Chain Monte Carlo

Lecture 7 and 8: Markov Chain Monte Carlo Lecture 7 and 8: Markov Chain Monte Carlo 4F13: Machine Learning Zoubin Ghahramani and Carl Edward Rasmussen Department of Engineering University of Cambridge http://mlg.eng.cam.ac.uk/teaching/4f13/ Ghahramani

More information

Control Variates for Markov Chain Monte Carlo

Control Variates for Markov Chain Monte Carlo Control Variates for Markov Chain Monte Carlo Dellaportas, P., Kontoyiannis, I., and Tsourti, Z. Dept of Statistics, AUEB Dept of Informatics, AUEB 1st Greek Stochastics Meeting Monte Carlo: Probability

More information

Part 8: GLMs and Hierarchical LMs and GLMs

Part 8: GLMs and Hierarchical LMs and GLMs Part 8: GLMs and Hierarchical LMs and GLMs 1 Example: Song sparrow reproductive success Arcese et al., (1992) provide data on a sample from a population of 52 female song sparrows studied over the course

More information

MCMC Sampling for Bayesian Inference using L1-type Priors

MCMC Sampling for Bayesian Inference using L1-type Priors MÜNSTER MCMC Sampling for Bayesian Inference using L1-type Priors (what I do whenever the ill-posedness of EEG/MEG is just not frustrating enough!) AG Imaging Seminar Felix Lucka 26.06.2012 , MÜNSTER Sampling

More information

Computer intensive statistical methods

Computer intensive statistical methods Lecture 11 Markov Chain Monte Carlo cont. October 6, 2015 Jonas Wallin jonwal@chalmers.se Chalmers, Gothenburg university The two stage Gibbs sampler If the conditional distributions are easy to sample

More information

Multimodal Nested Sampling

Multimodal Nested Sampling Multimodal Nested Sampling Farhan Feroz Astrophysics Group, Cavendish Lab, Cambridge Inverse Problems & Cosmology Most obvious example: standard CMB data analysis pipeline But many others: object detection,

More information

Mixture models. Mixture models MCMC approaches Label switching MCMC for variable dimension models. 5 Mixture models

Mixture models. Mixture models MCMC approaches Label switching MCMC for variable dimension models. 5 Mixture models 5 MCMC approaches Label switching MCMC for variable dimension models 291/459 Missing variable models Complexity of a model may originate from the fact that some piece of information is missing Example

More information

Multilevel Statistical Models: 3 rd edition, 2003 Contents

Multilevel Statistical Models: 3 rd edition, 2003 Contents Multilevel Statistical Models: 3 rd edition, 2003 Contents Preface Acknowledgements Notation Two and three level models. A general classification notation and diagram Glossary Chapter 1 An introduction

More information

Markov Chain Monte Carlo (MCMC)

Markov Chain Monte Carlo (MCMC) Markov Chain Monte Carlo (MCMC Dependent Sampling Suppose we wish to sample from a density π, and we can evaluate π as a function but have no means to directly generate a sample. Rejection sampling can

More information

MARKOV CHAIN MONTE CARLO

MARKOV CHAIN MONTE CARLO MARKOV CHAIN MONTE CARLO RYAN WANG Abstract. This paper gives a brief introduction to Markov Chain Monte Carlo methods, which offer a general framework for calculating difficult integrals. We start with

More information

Online appendix to On the stability of the excess sensitivity of aggregate consumption growth in the US

Online appendix to On the stability of the excess sensitivity of aggregate consumption growth in the US Online appendix to On the stability of the excess sensitivity of aggregate consumption growth in the US Gerdie Everaert 1, Lorenzo Pozzi 2, and Ruben Schoonackers 3 1 Ghent University & SHERPPA 2 Erasmus

More information

Markov Chain Monte Carlo, Numerical Integration

Markov Chain Monte Carlo, Numerical Integration Markov Chain Monte Carlo, Numerical Integration (See Statistics) Trevor Gallen Fall 2015 1 / 1 Agenda Numerical Integration: MCMC methods Estimating Markov Chains Estimating latent variables 2 / 1 Numerical

More information

Computer intensive statistical methods

Computer intensive statistical methods Lecture 13 MCMC, Hybrid chains October 13, 2015 Jonas Wallin jonwal@chalmers.se Chalmers, Gothenburg university MH algorithm, Chap:6.3 The metropolis hastings requires three objects, the distribution of

More information

Stat 451 Lecture Notes Markov Chain Monte Carlo. Ryan Martin UIC

Stat 451 Lecture Notes Markov Chain Monte Carlo. Ryan Martin UIC Stat 451 Lecture Notes 07 12 Markov Chain Monte Carlo Ryan Martin UIC www.math.uic.edu/~rgmartin 1 Based on Chapters 8 9 in Givens & Hoeting, Chapters 25 27 in Lange 2 Updated: April 4, 2016 1 / 42 Outline

More information

A Generic Multivariate Distribution for Counting Data

A Generic Multivariate Distribution for Counting Data arxiv:1103.4866v1 [stat.ap] 24 Mar 2011 A Generic Multivariate Distribution for Counting Data Marcos Capistrán and J. Andrés Christen Centro de Investigación en Matemáticas, A. C. (CIMAT) Guanajuato, MEXICO.

More information

Dynamic models. Dependent data The AR(p) model The MA(q) model Hidden Markov models. 6 Dynamic models

Dynamic models. Dependent data The AR(p) model The MA(q) model Hidden Markov models. 6 Dynamic models 6 Dependent data The AR(p) model The MA(q) model Hidden Markov models Dependent data Dependent data Huge portion of real-life data involving dependent datapoints Example (Capture-recapture) capture histories

More information

Paul Karapanagiotidis ECO4060

Paul Karapanagiotidis ECO4060 Paul Karapanagiotidis ECO4060 The way forward 1) Motivate why Markov-Chain Monte Carlo (MCMC) is useful for econometric modeling 2) Introduce Markov-Chain Monte Carlo (MCMC) - Metropolis-Hastings (MH)

More information

Bayesian Inference for Discretely Sampled Diffusion Processes: A New MCMC Based Approach to Inference

Bayesian Inference for Discretely Sampled Diffusion Processes: A New MCMC Based Approach to Inference Bayesian Inference for Discretely Sampled Diffusion Processes: A New MCMC Based Approach to Inference Osnat Stramer 1 and Matthew Bognar 1 Department of Statistics and Actuarial Science, University of

More information

Markov Chain Monte Carlo methods

Markov Chain Monte Carlo methods Markov Chain Monte Carlo methods By Oleg Makhnin 1 Introduction a b c M = d e f g h i 0 f(x)dx 1.1 Motivation 1.1.1 Just here Supresses numbering 1.1.2 After this 1.2 Literature 2 Method 2.1 New math As

More information

Fitting Narrow Emission Lines in X-ray Spectra

Fitting Narrow Emission Lines in X-ray Spectra Outline Fitting Narrow Emission Lines in X-ray Spectra Taeyoung Park Department of Statistics, University of Pittsburgh October 11, 2007 Outline of Presentation Outline This talk has three components:

More information

An introduction to Sequential Monte Carlo

An introduction to Sequential Monte Carlo An introduction to Sequential Monte Carlo Thang Bui Jes Frellsen Department of Engineering University of Cambridge Research and Communication Club 6 February 2014 1 Sequential Monte Carlo (SMC) methods

More information

Hastings-within-Gibbs Algorithm: Introduction and Application on Hierarchical Model

Hastings-within-Gibbs Algorithm: Introduction and Application on Hierarchical Model UNIVERSITY OF TEXAS AT SAN ANTONIO Hastings-within-Gibbs Algorithm: Introduction and Application on Hierarchical Model Liang Jing April 2010 1 1 ABSTRACT In this paper, common MCMC algorithms are introduced

More information

Stat 535 C - Statistical Computing & Monte Carlo Methods. Arnaud Doucet.

Stat 535 C - Statistical Computing & Monte Carlo Methods. Arnaud Doucet. Stat 535 C - Statistical Computing & Monte Carlo Methods Arnaud Doucet Email: arnaud@cs.ubc.ca 1 1.1 Outline Introduction to Markov chain Monte Carlo The Gibbs Sampler Examples Overview of the Lecture

More information

On Reparametrization and the Gibbs Sampler

On Reparametrization and the Gibbs Sampler On Reparametrization and the Gibbs Sampler Jorge Carlos Román Department of Mathematics Vanderbilt University James P. Hobert Department of Statistics University of Florida March 2014 Brett Presnell Department

More information

High-dimensional Problems in Finance and Economics. Thomas M. Mertens

High-dimensional Problems in Finance and Economics. Thomas M. Mertens High-dimensional Problems in Finance and Economics Thomas M. Mertens NYU Stern Risk Economics Lab April 17, 2012 1 / 78 Motivation Many problems in finance and economics are high dimensional. Dynamic Optimization:

More information

Approximate Bayesian Computation

Approximate Bayesian Computation Approximate Bayesian Computation Michael Gutmann https://sites.google.com/site/michaelgutmann University of Helsinki and Aalto University 1st December 2015 Content Two parts: 1. The basics of approximate

More information

Sequential Monte Carlo Methods for Bayesian Model Selection in Positron Emission Tomography

Sequential Monte Carlo Methods for Bayesian Model Selection in Positron Emission Tomography Methods for Bayesian Model Selection in Positron Emission Tomography Yan Zhou John A.D. Aston and Adam M. Johansen 6th January 2014 Y. Zhou J. A. D. Aston and A. M. Johansen Outline Positron emission tomography

More information

Hmms with variable dimension structures and extensions

Hmms with variable dimension structures and extensions Hmm days/enst/january 21, 2002 1 Hmms with variable dimension structures and extensions Christian P. Robert Université Paris Dauphine www.ceremade.dauphine.fr/ xian Hmm days/enst/january 21, 2002 2 1 Estimating

More information

Modelling Operational Risk Using Bayesian Inference

Modelling Operational Risk Using Bayesian Inference Pavel V. Shevchenko Modelling Operational Risk Using Bayesian Inference 4y Springer 1 Operational Risk and Basel II 1 1.1 Introduction to Operational Risk 1 1.2 Defining Operational Risk 4 1.3 Basel II

More information

Lecture 6: Markov Chain Monte Carlo

Lecture 6: Markov Chain Monte Carlo Lecture 6: Markov Chain Monte Carlo D. Jason Koskinen koskinen@nbi.ku.dk Photo by Howard Jackman University of Copenhagen Advanced Methods in Applied Statistics Feb - Apr 2016 Niels Bohr Institute 2 Outline

More information

Afternoon Meeting on Bayesian Computation 2018 University of Reading

Afternoon Meeting on Bayesian Computation 2018 University of Reading Gabriele Abbati 1, Alessra Tosi 2, Seth Flaxman 3, Michael A Osborne 1 1 University of Oxford, 2 Mind Foundry Ltd, 3 Imperial College London Afternoon Meeting on Bayesian Computation 2018 University of

More information

Ages of stellar populations from color-magnitude diagrams. Paul Baines. September 30, 2008

Ages of stellar populations from color-magnitude diagrams. Paul Baines. September 30, 2008 Ages of stellar populations from color-magnitude diagrams Paul Baines Department of Statistics Harvard University September 30, 2008 Context & Example Welcome! Today we will look at using hierarchical

More information

Monte Carlo-based statistical methods (MASM11/FMS091)

Monte Carlo-based statistical methods (MASM11/FMS091) Monte Carlo-based statistical methods (MASM11/FMS091) Jimmy Olsson Centre for Mathematical Sciences Lund University, Sweden Lecture 12 MCMC for Bayesian computation II March 1, 2013 J. Olsson Monte Carlo-based

More information

Parameter Estimation. William H. Jefferys University of Texas at Austin Parameter Estimation 7/26/05 1

Parameter Estimation. William H. Jefferys University of Texas at Austin Parameter Estimation 7/26/05 1 Parameter Estimation William H. Jefferys University of Texas at Austin bill@bayesrules.net Parameter Estimation 7/26/05 1 Elements of Inference Inference problems contain two indispensable elements: Data

More information

(5) Multi-parameter models - Gibbs sampling. ST440/540: Applied Bayesian Analysis

(5) Multi-parameter models - Gibbs sampling. ST440/540: Applied Bayesian Analysis Summarizing a posterior Given the data and prior the posterior is determined Summarizing the posterior gives parameter estimates, intervals, and hypothesis tests Most of these computations are integrals

More information

ESTIMATING LIKELIHOODS FOR SPATIO-TEMPORAL MODELS USING IMPORTANCE SAMPLING

ESTIMATING LIKELIHOODS FOR SPATIO-TEMPORAL MODELS USING IMPORTANCE SAMPLING To appear in Statistics and Computing ESTIMATING LIKELIHOODS FOR SPATIO-TEMPORAL MODELS USING IMPORTANCE SAMPLING Glenn Marion Biomathematics & Statistics Scotland, James Clerk Maxwell Building, The King

More information

13 Notes on Markov Chain Monte Carlo

13 Notes on Markov Chain Monte Carlo 13 Notes on Markov Chain Monte Carlo Markov Chain Monte Carlo is a big, and currently very rapidly developing, subject in statistical computation. Many complex and multivariate types of random data, useful

More information

Efficient model comparison techniques for models requiring large scale data augmentation

Efficient model comparison techniques for models requiring large scale data augmentation Vol. 00 (0000) 1 DOI: 0000 Efficient model comparison techniques for models requiring large scale data augmentation Panayiota Touloupou 1, Naif Alzahrani 2, Peter Neal 2, Simon E.F. Spencer 1 and Trevelyan

More information

MFM Practitioner Module: Risk & Asset Allocation. John Dodson. February 3, 2010

MFM Practitioner Module: Risk & Asset Allocation. John Dodson. February 3, 2010 MFM Practitioner Module: Risk & Asset Allocation Estimator February 3, 2010 Estimator Estimator In estimation we do not endow the sample with a characterization; rather, we endow the parameters with a

More information

Machine Learning Summer School

Machine Learning Summer School Machine Learning Summer School Lecture 3: Learning parameters and structure Zoubin Ghahramani zoubin@eng.cam.ac.uk http://learning.eng.cam.ac.uk/zoubin/ Department of Engineering University of Cambridge,

More information

Tutorial on Approximate Bayesian Computation

Tutorial on Approximate Bayesian Computation Tutorial on Approximate Bayesian Computation Michael Gutmann https://sites.google.com/site/michaelgutmann University of Helsinki Aalto University Helsinki Institute for Information Technology 16 May 2016

More information

Scaling Neighbourhood Methods

Scaling Neighbourhood Methods Quick Recap Scaling Neighbourhood Methods Collaborative Filtering m = #items n = #users Complexity : m * m * n Comparative Scale of Signals ~50 M users ~25 M items Explicit Ratings ~ O(1M) (1 per billion)

More information

MCMC for big data. Geir Storvik. BigInsight lunch - May Geir Storvik MCMC for big data BigInsight lunch - May / 17

MCMC for big data. Geir Storvik. BigInsight lunch - May Geir Storvik MCMC for big data BigInsight lunch - May / 17 MCMC for big data Geir Storvik BigInsight lunch - May 2 2018 Geir Storvik MCMC for big data BigInsight lunch - May 2 2018 1 / 17 Outline Why ordinary MCMC is not scalable Different approaches for making

More information

Machine Learning. Probabilistic KNN.

Machine Learning. Probabilistic KNN. Machine Learning. Mark Girolami girolami@dcs.gla.ac.uk Department of Computing Science University of Glasgow June 21, 2007 p. 1/3 KNN is a remarkably simple algorithm with proven error-rates June 21, 2007

More information

STA 294: Stochastic Processes & Bayesian Nonparametrics

STA 294: Stochastic Processes & Bayesian Nonparametrics MARKOV CHAINS AND CONVERGENCE CONCEPTS Markov chains are among the simplest stochastic processes, just one step beyond iid sequences of random variables. Traditionally they ve been used in modelling a

More information

SUPPLEMENT TO MARKET ENTRY COSTS, PRODUCER HETEROGENEITY, AND EXPORT DYNAMICS (Econometrica, Vol. 75, No. 3, May 2007, )

SUPPLEMENT TO MARKET ENTRY COSTS, PRODUCER HETEROGENEITY, AND EXPORT DYNAMICS (Econometrica, Vol. 75, No. 3, May 2007, ) Econometrica Supplementary Material SUPPLEMENT TO MARKET ENTRY COSTS, PRODUCER HETEROGENEITY, AND EXPORT DYNAMICS (Econometrica, Vol. 75, No. 3, May 2007, 653 710) BY SANGHAMITRA DAS, MARK ROBERTS, AND

More information

Markov Chain Monte Carlo Lecture 4

Markov Chain Monte Carlo Lecture 4 The local-trap problem refers to that in simulations of a complex system whose energy landscape is rugged, the sampler gets trapped in a local energy minimum indefinitely, rendering the simulation ineffective.

More information

Sequential Monte Carlo Samplers for Applications in High Dimensions

Sequential Monte Carlo Samplers for Applications in High Dimensions Sequential Monte Carlo Samplers for Applications in High Dimensions Alexandros Beskos National University of Singapore KAUST, 26th February 2014 Joint work with: Dan Crisan, Ajay Jasra, Nik Kantas, Alex

More information

Local Likelihood Bayesian Cluster Modeling for small area health data. Andrew Lawson Arnold School of Public Health University of South Carolina

Local Likelihood Bayesian Cluster Modeling for small area health data. Andrew Lawson Arnold School of Public Health University of South Carolina Local Likelihood Bayesian Cluster Modeling for small area health data Andrew Lawson Arnold School of Public Health University of South Carolina Local Likelihood Bayesian Cluster Modelling for Small Area

More information

Bayesian Learning in Undirected Graphical Models

Bayesian Learning in Undirected Graphical Models Bayesian Learning in Undirected Graphical Models Zoubin Ghahramani Gatsby Computational Neuroscience Unit University College London, UK http://www.gatsby.ucl.ac.uk/ Work with: Iain Murray and Hyun-Chul

More information

MCMC: Markov Chain Monte Carlo

MCMC: Markov Chain Monte Carlo I529: Machine Learning in Bioinformatics (Spring 2013) MCMC: Markov Chain Monte Carlo Yuzhen Ye School of Informatics and Computing Indiana University, Bloomington Spring 2013 Contents Review of Markov

More information

Estimation of Operational Risk Capital Charge under Parameter Uncertainty

Estimation of Operational Risk Capital Charge under Parameter Uncertainty Estimation of Operational Risk Capital Charge under Parameter Uncertainty Pavel V. Shevchenko Principal Research Scientist, CSIRO Mathematical and Information Sciences, Sydney, Locked Bag 17, North Ryde,

More information

Point spread function reconstruction from the image of a sharp edge

Point spread function reconstruction from the image of a sharp edge DOE/NV/5946--49 Point spread function reconstruction from the image of a sharp edge John Bardsley, Kevin Joyce, Aaron Luttman The University of Montana National Security Technologies LLC Montana Uncertainty

More information

MH I. Metropolis-Hastings (MH) algorithm is the most popular method of getting dependent samples from a probability distribution

MH I. Metropolis-Hastings (MH) algorithm is the most popular method of getting dependent samples from a probability distribution MH I Metropolis-Hastings (MH) algorithm is the most popular method of getting dependent samples from a probability distribution a lot of Bayesian mehods rely on the use of MH algorithm and it s famous

More information