ABCME: Summary statistics selection for ABC inference in R

Similar documents
Non-linear Regression Approaches in ABC

Nunes MA, Prangle D. abctools: an R package for tuning approximate Bayesian computation analyses. The R Journal 2015, 7(2),

Tutorial on ABC Algorithms

Approximate Bayesian computation: methods and applications for complex systems

An introduction to Approximate Bayesian Computation methods

CONDITIONING ON PARAMETER POINT ESTIMATES IN APPROXIMATE BAYESIAN COMPUTATION

Fundamentals and Recent Developments in Approximate Bayesian Computation

arxiv: v3 [stat.me] 16 Mar 2017

Proceedings of the 2012 Winter Simulation Conference C. Laroque, J. Himmelspach, R. Pasupathy, O. Rose, and A. M. Uhrmacher, eds.

Monte Carlo in Bayesian Statistics

Addressing the Challenges of Luminosity Function Estimation via Likelihood-Free Inference

Approximate Bayesian Computation

Approximate Bayesian Computation: a simulation based approach to inference

VCMC: Variational Consensus Monte Carlo

Approximate Bayesian computation (ABC) and the challenge of big simulation

arxiv: v1 [stat.co] 17 Dec 2015

Coalescent based demographic inference. Daniel Wegmann University of Fribourg

Likelihood-free MCMC

STA 4273H: Statistical Machine Learning

Estimating Evolutionary Trees. Phylogenetic Methods

Bayes via forward simulation. Approximate Bayesian Computation

BAYESIAN METHODS FOR VARIABLE SELECTION WITH APPLICATIONS TO HIGH-DIMENSIONAL DATA

Advances and Applications in Perfect Sampling

Tutorial on Probabilistic Programming with PyMC3

Choosing the Summary Statistics and the Acceptance Rate in Approximate Bayesian Computation

arxiv: v1 [stat.co] 2 May 2011

(5) Multi-parameter models - Gibbs sampling. ST440/540: Applied Bayesian Analysis

Distance-Based Probability Distribution for Set Partitions with Applications to Bayesian Nonparametrics

New Insights into History Matching via Sequential Monte Carlo

Tutorial on Approximate Bayesian Computation

Risk Estimation and Uncertainty Quantification by Markov Chain Monte Carlo Methods

Approximate Bayesian computation (ABC) NIPS Tutorial

Closed-form sampling formulas for the coalescent with recombination

Machine Learning. Probabilistic KNN.

Markov Chain Monte Carlo, Numerical Integration

Fitting the Bartlett-Lewis rainfall model using Approximate Bayesian Computation

Using phylogenetics to estimate species divergence times... Basics and basic issues for Bayesian inference of divergence times (plus some digression)

Forward Problems and their Inverse Solutions

Computer Vision Group Prof. Daniel Cremers. 10a. Markov Chain Monte Carlo

Bayesian inference for multivariate extreme value distributions

Reminder of some Markov Chain properties:

A = {(x, u) : 0 u f(x)},

Approximate Bayesian Computation

Bayesian model selection: methodology, computation and applications

On Bayesian Computation

Learning the hyper-parameters. Luca Martino

Metropolis-Hastings Algorithm

Stat 451 Lecture Notes Monte Carlo Integration

Bayesian Estimation of Input Output Tables for Russia

spbayes: An R Package for Univariate and Multivariate Hierarchical Point-referenced Spatial Models

Bayesian Inference and MCMC

MARKOV CHAIN MONTE CARLO

Deblurring Jupiter (sampling in GLIP faster than regularized inversion) Colin Fox Richard A. Norton, J.

Pattern Recognition and Machine Learning

Markov Chain Monte Carlo methods

On Markov chain Monte Carlo methods for tall data

Bayesian Phylogenetics:

Nested Sampling. Brendon J. Brewer. brewer/ Department of Statistics The University of Auckland

Bayesian Methods for Machine Learning

Bayesian Sequential Design under Model Uncertainty using Sequential Monte Carlo

ABC random forest for parameter estimation. Jean-Michel Marin

XXV ENCONTRO BRASILEIRO DE ECONOMETRIA Porto Seguro - BA, 2003 REVISITING DISTRIBUTED LAG MODELS THROUGH A BAYESIAN PERSPECTIVE

Reverse Engineering Gene Networks Using Approximate Bayesian Computation (ABC)

Bayesian Inference. Chapter 1. Introduction and basic concepts

Challenges when applying stochastic models to reconstruct the demographic history of populations.

Two step estimation for Neyman-Scott point process with inhomogeneous cluster centers. May 2012

Bayesian Inference of Interactions and Associations

Multifidelity Approaches to Approximate Bayesian Computation

STA 4273H: Statistical Machine Learning

CS242: Probabilistic Graphical Models Lecture 7B: Markov Chain Monte Carlo & Gibbs Sampling

arxiv: v1 [stat.me] 30 Sep 2009

Approximate Bayesian computation (ABC) gives exact results under the assumption of. model error

MCMC notes by Mark Holder

Multi-level Approximate Bayesian Computation

STA 4273H: Statistical Machine Learning

Demographic Inference with Coalescent Hidden Markov Model

Markov Chain Monte Carlo Methods

ABC (Approximate Bayesian Computation) methods to make inference about population history from molecular data: principles and applications

Inference for partially observed stochastic dynamic systems: A new algorithm, its theory and applications

Bayesian hierarchical methods for sea turtle mixed stock analysis

An Introduction to Reversible Jump MCMC for Bayesian Networks, with Application

LECTURE 5 NOTES. n t. t Γ(a)Γ(b) pt+a 1 (1 p) n t+b 1. The marginal density of t is. Γ(t + a)γ(n t + b) Γ(n + a + b)

Stat 535 C - Statistical Computing & Monte Carlo Methods. Lecture February Arnaud Doucet

Reducing The Computational Cost of Bayesian Indoor Positioning Systems

Hierarchical Bayesian Modeling with Approximate Bayesian Computation

Bayesian Inference in GLMs. Frequentists typically base inferences on MLEs, asymptotic confidence

2 Inference for Multinomial Distribution

Likelihood-free inference and approximate Bayesian computation for stochastic modelling

Accelerating ABC methods using Gaussian processes

STAT 518 Intro Student Presentation

Analysis Methods for Supersaturated Design: Some Comparisons

Integrated Non-Factorized Variational Inference

One Pseudo-Sample is Enough in Approximate Bayesian Computation MCMC

MCMC for big data. Geir Storvik. BigInsight lunch - May Geir Storvik MCMC for big data BigInsight lunch - May / 17

Down by the Bayes, where the Watermelons Grow

Fast Likelihood-Free Inference via Bayesian Optimization

Bayesian philosophy Bayesian computation Bayesian software. Bayesian Statistics. Petter Mostad. Chalmers. April 6, 2017

Log Gaussian Cox Processes. Chi Group Meeting February 23, 2016

Bayesian Nonparametric Regression for Diabetes Deaths

Bayesian Networks in Educational Assessment

Transcription:

ABCME: Summary statistics selection for ABC inference in R Matt Nunes and David Balding Lancaster University UCL Genetics Institute

Outline Motivation: why the ABCME package? Description of the package Example of informative summary selection algorithms Implementation and examples of the ABCME package Summary

ABC in a nutshell For complex or high-dimensional data, inference about a parameter θ is often performed via Approximate Bayesian Computation (ABC), since likelihoods are often intractable/difficult to compute. In ABC, datasets X i are simulated under a model, M(θ) and then summary statistics of the simulated data are used to compare to the (summaries of the) observed dataset X for inference.

ABC in a nutshell For complex or high-dimensional data, inference about a parameter θ is often performed via Approximate Bayesian Computation (ABC), since likelihoods are often intractable/difficult to compute. In ABC, datasets X i are simulated under a model, M(θ) and then summary statistics of the simulated data are used to compare to the (summaries of the) observed dataset X for inference. e.g, for DNA sequences, the number of polymorphic sites is useful for global parameters such as the mutation rate.

Rejection-ABC Parameter 4 6 8 10 12 14 16 18 3 4 5 6 Data 0 2 4 6 8 10 8 10 12 14 Parameter

Motivation: which statistics are best? In some scientific areas (e.g. population genetics), there are many well-established summary statistics that investigators know and love, often selected by investigators on the basis of intuition and established practice in the field.

Motivation: which statistics are best? In some scientific areas (e.g. population genetics), there are many well-established summary statistics that investigators know and love, often selected by investigators on the basis of intuition and established practice in the field. However, the most informative summary statistics are difficult to judge, and depend on the application and observed dataset under investigation.

Motivation: which statistics are best? In some scientific areas (e.g. population genetics), there are many well-established summary statistics that investigators know and love, often selected by investigators on the basis of intuition and established practice in the field. However, the most informative summary statistics are difficult to judge, and depend on the application and observed dataset under investigation. Given a set of summary statistics, the ABCME package offers practitioners automatic algorithms for finding the best subset of statistics.

What the ABCME package does The ABCME package implementing procedures to select summary statistics for ABC inference, namely Approximate sufficiency ratio (Joyce and Marjoram, 2008) selection via PLS regression (Wegmann et al., 2009) summary selection using a minimum entropy criterion (Nunes and Balding, 2010) two-stage approximate error summary selection (Nunes and Balding, 2010) Code allows for greedy searches and reduction of the space of summary statistics. Designed to be as modular as possible for flexibility. Suggests: abc (Csillery et al., 2011) pls (Mevik and Wehrens, 2007)

Choosing summaries via minimum entropy ME algorithm: minimises sample-based entropy of posterior samples as a heuristic for selecting summary statistics. Motivation: Entropy measures concentration of the posterior sample ( spread, peakedness of distribution). Frequency 0 5 10 15 Frequency 0 5 10 15 20 25 0.0 0.2 0.4 0.6 0.8 1.0 x 3 2 1 0 1 2 3 y

Choosing summaries via minimum entropy ME algorithm: minimises sample-based entropy of posterior samples as a heuristic for selecting summary statistics. Motivation: Entropy measures concentration of the posterior sample ( spread, peakedness of distribution). For every subset of summary statistics S P(Ω): 1. Perform ABC inference to obtain posterior sample (e.g. rejection-abc using the abc function in the abc package). 2. Compute sample entropy of resulting posterior. The ABCME function nn.ent can be used to compute nearest neighbour entropy (Singh et al., 2003). Choose the subset S with the minimum entropy.

Two-stage procedure Stage 1: Find an initial good candidate S Ω via rejection-abc (e.g. using ME algorithm). Stage 2: Using S identified in Stage 1, take k simulated datasets nearest S 0. For each S Ω, perform rejection-abc and compute the error of Π(θ S) for each of the k datasets, then select the subset of Ω that minimises mean square error.

Two-stage procedure Stage 1: Find an initial good candidate S Ω via rejection-abc (e.g. using ME algorithm). Stage 2: Using S identified in Stage 1, take k simulated datasets nearest S 0. For each S Ω, perform rejection-abc and compute the error of Π(θ S) for each of the k datasets, then select the subset of Ω that minimises mean square error. The two-stage algorithm is only dependent on the subset chosen in Stage 1. Any preferred measure of accuracy can be used in place of MSE.

ME algorithm implementation: selectsumm() Function call: selectsumm(theta,stats,data,abcmethod,crit,...) Arguments: theta: parameter values used to generate data from a model stats: summary statistics computed from simulated datasets corresponding to theta data: the summary statistics for the observed dataset abcmethod: a function to perform an ABC algorithm (e.g. abc) crit: an entropy function to minimize using the approximate posterior samples (e.g nn.ent)

Two stage algorithm implementation: stage2() Function call: stage2(theta,stats,data,stage1,abcmethod,crit,dsets,...)

Two stage algorithm implementation: stage2() Function call: stage2(theta,stats,data,stage1,abcmethod,crit,dsets,...) Same arguments as selectsumm except: stage1: index of the best summary subset from stage 1 crit: an error criterion to minimize using simulated datasets (e.g the ABCME function mse) dsets: number of closest simulated datasets to minimize crit

Two stage algorithm implementation: stage2() Function call: stage2(theta,stats,data,stage1,abcmethod,crit,dsets,...) Same arguments as selectsumm except: stage1: index of the best summary subset from stage 1 crit: an error criterion to minimize using simulated datasets (e.g the ABCME function mse) dsets: number of closest simulated datasets to minimize crit There are other optional arguments to both algorithms including: limit, do.only, final.dens...

Example: coal Data Datasets consisted of sets of 50 haplotypes (DNA sequences) simulated under the coalescent model, summarized by 6 statistics, e.g. the number of segregating sites, the number of distinct haplotypes etc.

Example: coal Data Datasets consisted of sets of 50 haplotypes (DNA sequences) simulated under the coalescent model, summarized by 6 statistics, e.g. the number of segregating sites, the number of distinct haplotypes etc. The goal is inference about the scaled mutation and recombination parameters, θ and ρ, which were simulated under uniform priors for input into the model.

Example: coal Data Datasets consisted of sets of 50 haplotypes (DNA sequences) simulated under the coalescent model, summarized by 6 statistics, e.g. the number of segregating sites, the number of distinct haplotypes etc. The goal is inference about the scaled mutation and recombination parameters, θ and ρ, which were simulated under uniform priors for input into the model. coal consists of two columns of parameter values (θ, ρ), followed by the 6 columns of statistics (computed from 10 6 datasets simulated under the model).

Example (code snippet I) # load coalescent data: > data(coal) > data(coalobs) # use entropy to find subset guess (bivariate inference): > stage1<-selectsumm(coal[,1:2],coal[,3:8],coalobs, + abcmethod=abc,crit=nn.ent) doing statistics: 1 ( 1 / 63 ) doing statistics: 2 ( 2 / 63 ) doing statistics: 3 ( 3 / 63 ) doing statistics: 4 ( 4 / 63 ) doing statistics: 5 ( 5 / 63 ) doing statistics: 6 ( 6 / 63 ) doing statistics: 1 2 ( 7 / 63 )... > stage1$best C1 C4 1 4

Example (code snippet II) Now perform stage2 using stage1$best and mse as the error function, with some optional arguments: final.dens=true to retrieve approximate posterior sample limit=3 to limit the search to subsets of less than 4 statistics > st2<-stage2(coal[,1:2],coal[,3:9],coalobs,stage1=c(1,4), +abcmethod=abc,crit=mse,dsets=100,final.dens=t,limit=3) best guess subset is: 9 close datasets: 100 dataset... 1 dataset... 2... getting final posterior sample...done. Can then estimate and plot the density using e.g. kde2d + filled.contour/persp etc

Concluding remarks The ABCME package provides an implementation for recent ABC summary statistics selection methods. The package could be easily extended to use other ABC methods (e.g. MCMC-ABC, SMC-ABC). Other error/entropy criteria can also be implemented within the main routines of the package. Since the separate subset calculations are independent, the routines could benefit from parallelization (using e.g. snow). More information about ABCME can be found at: http://www.maths.lancs.ac.uk/ nunes/abcme.html

Csillery, K., M. Blum, and O. Francois (2011). abc: Tools for approximate bayesian computation (abc). R package version 1.3. Joyce, P. and P. Marjoram (2008). Approximately sufficient statistics and Bayesian computation. Stat. Appl. Gen. Mol. Biol. 7(1), 1 16. Mevik, B.-H. and R. Wehrens (2007). The PLS package: principal component and partial least squares regression in R. J. Stat. Soft. 18(2), 1 24. Nunes, M. A. and D. J. Balding (2010). On optimal selection of summary statistics for Approximate Bayesian Computation. Stat. Appl. Genet. Mol. Biol. 9(1). Singh, H., V. Misra, N.and Hnizdo, A. Fedorowicz, and E. Demchuk (2003). Nearest neighbor estimates of entropy. Am. J. Math. Man. Sci. 23, 301 321. Wegmann, D., C. Leuenberger, and L. Excoffier (2009). Efficient approximate Bayesian computation coupled with Markov chain Monte Carlo without likelihood. Genetics 182(4), 1207 1218.

Rejection-ABC & stage 2 Parameter 4 6 8 10 12 14 16 18 3 4 5 6 Data 0 2 4 6 8 10 8 10 12 14 Parameter