Bayesian model selection in graphs by using BDgraph package

Similar documents
Gaussian graphical model determination based on birth-death MCMC inference

Bayesian Model Determination in Complex Systems

Package BDgraph. June 21, 2018

Introduction to Machine Learning CMU-10701

STA 4273H: Statistical Machine Learning

Computer Vision Group Prof. Daniel Cremers. 10a. Markov Chain Monte Carlo

Bayesian Inference of Multiple Gaussian Graphical Models

Markov Chain Monte Carlo Inference. Siamak Ravanbakhsh Winter 2018

CS242: Probabilistic Graphical Models Lecture 7B: Markov Chain Monte Carlo & Gibbs Sampling

Hmms with variable dimension structures and extensions

Markov Chain Monte Carlo

Markov Chain Monte Carlo

Machine Learning for Data Science (CS4786) Lecture 24

(5) Multi-parameter models - Gibbs sampling. ST440/540: Applied Bayesian Analysis

Likelihood-free MCMC

Markov Networks.

CPSC 540: Machine Learning

April 20th, Advanced Topics in Machine Learning California Institute of Technology. Markov Chain Monte Carlo for Machine Learning

Learning the hyper-parameters. Luca Martino

The ratio of normalizing constants for Bayesian graphical Gaussian model selection Mohammadi, A.; Massam, Helene; Letac, Gerard

Supplement to A Hierarchical Approach for Fitting Curves to Response Time Measurements

Monte Carlo in Bayesian Statistics

Bayesian Methods for Machine Learning

Markov Chain Monte Carlo methods

Computer Vision Group Prof. Daniel Cremers. 11. Sampling Methods: Markov Chain Monte Carlo

Computational statistics

Bayesian Estimation with Sparse Grids

MCMC and Gibbs Sampling. Kayhan Batmanghelich

Stat 535 C - Statistical Computing & Monte Carlo Methods. Lecture 18-16th March Arnaud Doucet

Efficient adaptive covariate modelling for extremes

Bayesian spatial hierarchical modeling for temperature extremes

27 : Distributed Monte Carlo Markov Chain. 1 Recap of MCMC and Naive Parallel Gibbs Sampling

On Bayesian Computation

The Ising model and Markov chain Monte Carlo

An Introduction to Reversible Jump MCMC for Bayesian Networks, with Application

Bayesian Nonparametric Regression for Diabetes Deaths

Graphical Models and Kernel Methods

MH I. Metropolis-Hastings (MH) algorithm is the most popular method of getting dependent samples from a probability distribution

Control Variates for Markov Chain Monte Carlo

The Origin of Deep Learning. Lili Mou Jan, 2015

Stat 516, Homework 1

Bayesian Inference for Gaussian Mixed Graph Models

Inference in Bayesian Networks

Bayesian Inference for Gaussian Mixed Graph Models

Dynamic models. Dependent data The AR(p) model The MA(q) model Hidden Markov models. 6 Dynamic models

Bayesian (conditionally) conjugate inference for discrete data models. Jon Forster (University of Southampton)

The Particle Filter. PD Dr. Rudolph Triebel Computer Vision Group. Machine Learning for Computer Vision

Advanced Statistical Modelling

Advances and Applications in Perfect Sampling

Convex Optimization CMU-10725

Contents. Part I: Fundamentals of Bayesian Inference 1

Markov-Chain Monte Carlo

Session 5B: A worked example EGARCH model

Theory of Stochastic Processes 8. Markov chain Monte Carlo

Down by the Bayes, where the Watermelons Grow

A Review of Pseudo-Marginal Markov Chain Monte Carlo

VCMC: Variational Consensus Monte Carlo

Monte Carlo Methods. Leon Gu CSD, CMU

Bayesian Analysis of Mixture Models with an Unknown Number of Components an alternative to reversible jump methods

MCMC: Markov Chain Monte Carlo

Bayesian Image Segmentation Using MRF s Combined with Hierarchical Prior Models

Results: MCMC Dancers, q=10, n=500

Introduction to Bayesian methods in inverse problems

Marginal Specifications and a Gaussian Copula Estimation

Bayesian Model Comparison:

Bayesian Graphical Models for Structural Vector AutoregressiveMarch Processes 21, / 1

Sampling Methods (11/30/04)

Sampling Algorithms for Probabilistic Graphical models

Stochastic Proximal Gradient Algorithm

Sampling Rejection Sampling Importance Sampling Markov Chain Monte Carlo. Sampling Methods. Oliver Schulte - CMPT 419/726. Bishop PRML Ch.

The Bayesian Choice. Christian P. Robert. From Decision-Theoretic Foundations to Computational Implementation. Second Edition.

Estimating the marginal likelihood with Integrated nested Laplace approximation (INLA)

Bayesian Inference in GLMs. Frequentists typically base inferences on MLEs, asymptotic confidence

Kernel adaptive Sequential Monte Carlo

Calibration of Stochastic Volatility Models using Particle Markov Chain Monte Carlo Methods

ST 740: Markov Chain Monte Carlo

18 : Advanced topics in MCMC. 1 Gibbs Sampling (Continued from the last lecture)

Bayesian inference for multivariate extreme value distributions

An ABC interpretation of the multiple auxiliary variable method

LECTURE 15 Markov chain Monte Carlo

Lecture 7 and 8: Markov Chain Monte Carlo

17 : Markov Chain Monte Carlo

Markov Chain Monte Carlo

MCMC 2: Lecture 2 Coding and output. Phil O Neill Theo Kypraios School of Mathematical Sciences University of Nottingham

Markov Chains and MCMC

Modelling Operational Risk Using Bayesian Inference

Bayesian inference for general Gaussian graphical models with application to multivariate lattice data

Machine Learning. Probabilistic KNN.

Markov Chain Monte Carlo methods

eqr094: Hierarchical MCMC for Bayesian System Reliability

Markov Chain Monte Carlo, Numerical Integration

ABC methods for phase-type distributions with applications in insurance risk problems

Hamiltonian Monte Carlo for Scalable Deep Learning

Supplementary Notes: Segment Parameter Labelling in MCMC Change Detection

arxiv: v1 [stat.co] 2 Nov 2017

Markov Chain Monte Carlo Lecture 6

Log Gaussian Cox Processes. Chi Group Meeting February 23, 2016

16 : Markov Chain Monte Carlo (MCMC)

Perfect simulation of repulsive point processes

A comparison of reversible jump MCMC algorithms for DNA sequence segmentation using hidden Markov models

Transcription:

Bayesian model selection in graphs by using BDgraph package A. Mohammadi and E. Wit March 26, 2013

MOTIVATION Flow cytometry data with 11 proteins from Sachs et al. (2005)

RESULT FOR CELL SIGNALING DATA

Gaussian graphical model Respect to graph G = (V, E) as M G = { } N p (0, Σ) K = Σ 1 is positive definite based on G Pairwise Markov property X i X j X V\{i,j} k ij = 0,

BIRTH-DEATH PROCESS Spacial birth-death process: Preston (1976) Birth-death MCMC: Stephen (2000) in mixture models

BIRTH-DEATH MCMC DESIGN General birth-death process Continuous Markov process Birth and death events are independent Poisson processes Time of birth or death event is exponentially distributed Birth-death process in GGM Adding new edge in birth and deleting edge in death time

SIMPLE CASE

SIMPLE CASE

SIMPLE CASE

BALANCE CONDITION Preston (1976): Backward Kolmogorov Under Balance condition, process converges to unique stationary distribution. Mohammadi and Wit (2013): BDMCMC in GGM Stationary distribution = Posterior distribution of (G,K) So, relative sojourn time in graph G = posterior probability of G

PROPOSED BDMCMC ALGORITHM Step 1: (a). Calculate birth and death rates β ξ (K) = λ b, new link ξ = (i, j) δ ξ (K) = b ξ(k ξ )p(g ξ, K ξ x) λ b, existing link ξ = (i, j) p(g, K x) (b). Calculate waiting time, (c). Simulate type of jump, birth or death Step 2: Sampling new precision matrix: K + ξ or K ξ

PROPOSED PRIOR DISTRIBUTIONS Prior for graph Discrete Uniform Truncated Poisson according to number of links Prior for precision matrix G-Wishart: W G (b, D) p(k G) K (b 2)/2 exp { 1 } 2 tr(dk) I G (b, D) = K (b 2)/2 exp { 1 } P G 2 tr(dk) dk

SPECIFIC ELEMENT OF BDMCMC Sampling from G-Wishart distribution Block Gibbs sampler Edgewise block Gibbs sampler According to maximum cliques Metropolis-Hastings algorithm Computing death rates δ ξ (K) = p(g ξ, K ξ x) p(g, K x) γ bb ξ (k ξ ) = I G (b, D) ( K ξ (b, D) K I G ξ ) (b 2)/2 { exp 1 } 2 tr(d (K ξ K)) γ b

Ratio of normalizing constant I G (b, D) I G ξ(b, D) = 2 Γ πt ii t jj Γ ( ) b+νi 2 ( b+νi 1 2 ) E G [f T (ψ ν )] E G ξ [f T (ψ ν )] Plot for ratio of normalizing constants ratio of expectation 1.00 1.05 1.10 1.15 1.20 0 50 100 150 200 250 300 number of nodes

Death rates for high-dimensional cases δ ξ (K) = 2 πt ii t jj Γ ( ) b+νi 2 ( ) b+νi 1 2 Γ { exp 1 2 tr(d (K ξ K)) ) (b 2)/2 ( K ξ K } γ b b ξ (k ξ ) BDgraph package bdmcmc.high : for high-dimensional graphs bdmcmc.low : for low-dimensional graphs

SIMULATION: 8 NODES } M G = {N 8 (0, Σ) K = Σ 1 P G 1.5 0 0 0 0 0.4 1.5 0 0 0 0 0 1.5 0 0 0 0 K = 1.5 0 0 0 1.5 0 0 1.5 0 1.5 1

SOME RESULT Effect of Sample size Number of data 20 30 40 60 80 100 p(true graph data) 0.018 0.067 0.121 0.2 0.22 0.35 false positive 0 0 0 0 0 0 false negative 1 0 0 0 0 0

SIMULATION: 120 NODES M G = {N 120 (0, Σ) K = Σ 1 P G }, n = 2000 7260 Priors: K W G (3, I 120 ) and G TU(all possible graphs) 10000 iterations and 5000 iterations as burn-in Result Time 4 hours p(true graph data) = 0.09 which is most probable graph

SUMMARY

Thanks for your attention References MOHAMMADI, A. AND E. C. WIT (2013) Gaussian graphical model determination based on birth-death MCMC inference, arxiv preprint arxiv:1210.5371v4 WANG, H. AND S. LI (2012) Efficient Gaussian graphical model determination under G-Wishart prior distributions. Electronic Journal of Statistics, 6:168-198 ATAY-KAYIS, A. AND H. MASSAM (2005) A Monte Carlo method for computing the marginal likelihood in nondecomposable Gaussian graphical models. Biometrika Trust, 92(2):317-335 PRESTON, C. J. (1976) Special birth-and-death processes. Bull. Inst. Internat. Statist., 34:1436-1462