Inferring the origin of an epidemic with a dynamic message-passing algorithm

Similar documents
CS Lecture 8 & 9. Lagrange Multipliers & Varitional Bounds

arxiv: v3 [physics.soc-ph] 2 Jul 2014

Expectation Propagation performs smooth gradient descent GUILLAUME DEHAENE

Lecture 3. STAT161/261 Introduction to Pattern Recognition and Machine Learning Spring 2018 Prof. Allie Fletcher

Epidemics and information spreading

Social Influence in Online Social Networks. Epidemiological Models. Epidemic Process

CS249: ADVANCED DATA MINING

KINETICS OF COMPLEX SOCIAL CONTAGION. János Kertész Central European University. Pohang, May 27, 2016

Lecture 11. Kernel Methods

General Strong Polarization

SECTION 8: ROOT-LOCUS ANALYSIS. ESE 499 Feedback Control Systems

Uncertain Compression & Graph Coloring. Madhu Sudan Harvard

Advanced data analysis

SECTION 5: POWER FLOW. ESE 470 Energy Distribution Systems

A robust information source estimator with sparse observations

Support Vector Machines. CSE 4309 Machine Learning Vassilis Athitsos Computer Science and Engineering Department University of Texas at Arlington

SECTION 7: STEADY-STATE ERROR. ESE 499 Feedback Control Systems

A Step Towards the Cognitive Radar: Target Detection under Nonstationary Clutter

Angular Momentum, Electromagnetic Waves

CHAPTER 5 Wave Properties of Matter and Quantum Mechanics I

Locality in Coding Theory

General Strong Polarization

Variational algorithms for marginal MAP

Cold atoms in optical lattices

Work, Energy, and Power. Chapter 6 of Essential University Physics, Richard Wolfson, 3 rd Edition

Quantum state measurement

Network Infusion to Infer Information Sources in Networks Soheil Feizi, Ken Duffy, Manolis Kellis, and Muriel Medard

Wave Motion. Chapter 14 of Essential University Physics, Richard Wolfson, 3 rd Edition

PHY103A: Lecture # 4

1. The graph of a function f is given above. Answer the question: a. Find the value(s) of x where f is not differentiable. Ans: x = 4, x = 3, x = 2,

Worksheets for GCSE Mathematics. Quadratics. mr-mathematics.com Maths Resources for Teachers. Algebra

Bayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2014

10.4 The Cross Product

Variations. ECE 6540, Lecture 02 Multivariate Random Variables & Linear Algebra

14- Hardening Soil Model with Small Strain Stiffness - PLAXIS

(1) Correspondence of the density matrix to traditional method

CS224W: Analysis of Networks Jure Leskovec, Stanford University

Hopfield Network for Associative Memory

A new centrality measure for probabilistic diffusion in network

9 Forward-backward algorithm, sum-product on factor graphs

Lecture 6. Notes on Linear Algebra. Perceptron

(1) Introduction: a new basis set

Gradient expansion formalism for generic spin torques

CS249: ADVANCED DATA MINING

A linear operator of a new class of multivalent harmonic functions

Secondary 3H Unit = 1 = 7. Lesson 3.3 Worksheet. Simplify: Lesson 3.6 Worksheet

Time Domain Analysis of Linear Systems Ch2. University of Central Oklahoma Dr. Mohamed Bingabr

Variable Elimination: Algorithm

Learning with Temporal Point Processes

Algebraic Codes and Invariance

Network observability and localization of the source of diffusion based on a subset of nodes

A new procedure for sensitivity testing with two stress factors

Terms of Use. Copyright Embark on the Journey

Charge carrier density in metals and semiconductors

Epidemics on networks

Epidemics in Complex Networks and Phase Transitions

Review for Exam Hyunse Yoon, Ph.D. Assistant Research Scientist IIHR-Hydroscience & Engineering University of Iowa

MCMC 2: Lecture 2 Coding and output. Phil O Neill Theo Kypraios School of Mathematical Sciences University of Nottingham

Modeling, Analysis, and Control of Information Propagation in Multi-layer and Multiplex Networks. Osman Yağan

Bayesian Inference for Contact Networks Given Epidemic Data

Modeling Social Media Memes as a Contagious Process

Variable Elimination: Algorithm

Classical RSA algorithm

Chapter 22 : Electric potential

Lecture No. 5. For all weighted residual methods. For all (Bubnov) Galerkin methods. Summary of Conventional Galerkin Method

Lecture 3. Linear Regression

The willingness to pay for health improvements under comorbidity ambiguity 1

Susceptible-Infective-Removed Epidemics and Erdős-Rényi random

Patterns of soiling in the Old Library Trinity College Dublin. Allyson Smith, Robbie Goodhue, Susie Bioletti

SIS epidemics on Networks

Epidemics on networks

Quantum Mechanics. An essential theory to understand properties of matter and light. Chemical Electronic Magnetic Thermal Optical Etc.

Bayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2016

Rotational Motion. Chapter 10 of Essential University Physics, Richard Wolfson, 3 rd Edition

Gravitation. Chapter 8 of Essential University Physics, Richard Wolfson, 3 rd Edition

Grover s algorithm. We want to find aa. Search in an unordered database. QC oracle (as usual) Usual trick

PHL424: Nuclear Shell Model. Indian Institute of Technology Ropar

Hybrid Extended Kalman Filtering and Noise Statistics Optimization for Produce Wash State Estimation

Control of Mobile Robots

Review of Last Class 1

Worksheets for GCSE Mathematics. Algebraic Expressions. Mr Black 's Maths Resources for Teachers GCSE 1-9. Algebra

Analysis and Design of Control Dynamics of Manipulator Robot s Joint Drive

Probabilistic Graphical Models

Radial Basis Function (RBF) Networks

Problem 3.1 (Verdeyen 5.13) First, I calculate the ABCD matrix for beam traveling through the lens and space.

(2) Orbital angular momentum

Course Business. Homework 3 Due Now. Homework 4 Released. Professor Blocki is travelling, but will be back next week

13 : Variational Inference: Loopy Belief Propagation and Mean Field

A L T O SOLO LOWCLL. MICHIGAN, THURSDAY. DECEMBER 10,1931. ritt. Mich., to T h e Heights. Bos" l u T H I S COMMl'NiTY IN Wilcox

A A A A A A A A A A A A. a a a a a a a a a a a a a a a. Apples taste amazingly good.

Statistical Learning with the Lasso, spring The Lasso

Atomic fluorescence. The intensity of a transition line can be described with a transition probability inversely

Entropy Enhanced Covariance Matrix Adaptation Evolution Strategy (EE_CMAES)

Organization. I MCMC discussion. I project talks. I Lecture.

Jasmin Smajic1, Christian Hafner2, Jürg Leuthold2, March 23, 2015

Probabilistic Graphical Models (I)

CHAPTER 4 Structure of the Atom

TSOKOS READING ACTIVITY Section 7-2: The Greenhouse Effect and Global Warming (8 points)

6.867 Machine learning, lecture 23 (Jaakkola)

ECS 289 F / MAE 298, Lecture 15 May 20, Diffusion, Cascades and Influence

Transcription:

Inferring the origin of an epidemic with a dynamic message-passing algorithm HARSH GUPTA (Based on the original work done by Andrey Y. Lokhov, Marc Mézard, Hiroki Ohta, and Lenka Zdeborová)

Paper Andrey Y. Lokhov, Marc Mézard, Hiroki Ohta, and Lenka Zdeborová, Inferring the origin of an epidemic with a dynamic message-passing algorithm, Physical Review E 90, 2014 APS. Also available on arxiv.

Outline Introduction Infection spreading mechanism/model Susceptible-infected-recovered (SIR) model DMP algorithm for inference Dynamic message-passing (DMP) equations Performance comparison via simulations Comparing with Jordan and Distance centrality based algorithms

Introduction Understanding and controlling the spread of epidemics on networks of contacts is an important and significant task. It has far-reaching applications in mitigating the results of epidemics caused by infectious diseases, computer viruses, rumor spreading in social media, and others. We aim to address the specific problem of estimation of the origin of the epidemic outbreak, i.e., the so-called patient zero or infection source.

Introduction Information about the origin of the epidemic could be extremely useful to reduce or prevent future outbreaks (for e.g., one may fix a contaminated water body in the case of a cholera epidemic). Challenge: The only information available to us is a contact network and a snapshot of epidemic spread at a certain time. We aim to determine the infection source using this information. In order to define the problem, we need to know the mechanism with which the epidemic spreads. We assume the spreading mechanism follows the popular SIR model used widely in literature.

SIR model Susceptible-infected-recovered (SIR) model is one of the most popular and studied epidemiological models along with the susceptible-infected-susceptible (SIS) model. SIR model is defined as follows: Let GG (VV, EE) be a connected undirected graph containing NN nodes defined by the set of vertices VV and the set of edges EE. Each node ii VV at discrete time tt can be in one of the three states (state variable - qq ii tt ): susceptible qq ii tt = SS, infected qq ii tt = II, or recovered qq ii tt = RR. The epidemic process on a graph can be interpreted as the propagation of infection signals from infected to susceptible nodes. The infection signal dd ii jj (tt) is defined as a random variable which is equal to one with probability δδ qqii tt 1,IIλλ iiii, and equal to zero otherwise. Here λλ iijj measures the efficiency of spread from the node ii to node jj,

SIR model At each time step, an infected node ii will recover with probability μμ ii, and a susceptible node ii will become infected with probability 1 kk (1 λλ kkkk δδ qqkk tt,ii), where is the set of neighbors of node ii. The recovered nodes never change their state. In the image on the right is an example of a single instance of the inference problem generated using the SIR model. FIGURE 1

Dynamic message-passing algorithm The core idea behind the DMP algorithm is to infer epidemic origins using time-dependent marginal probabilities for each node in the graph. These probabilities are computed using dynamic message-passing equations. Let PP SS ii tt, PP II ii tt, and PP RR ii tt be the marginal probabilities that qq ii tt = SS, qq ii tt = II, and qq ii tt = RR respectively. Let OO be the snapshot of the contact network we have. We use DMP message-passing equations to compute the above marginal probabilities for any initial condition and time from which the snapshot might have been obtained.

Dynamic message-passing algorithm From the marginals, we know the probabilities PP jj SS tt, ii 0 (respectively, PP jj II (tt, ii 0 ) and PP jj RR (tt, ii 0 )) that a node jj is in each of the states SS, II, RR at a given time tt, for a given patient zero ii 0. For now, let us assume that we know the time tt 0 at which the snapshot was obtained. Using Bayes rule, PP ii OO, tt 0 PP(OO ii, tt 0 ). Intuitively, we compute the above posterior probability for all nodes ii, and choose the node which maximizes this probability as patient zero.

Dynamic message-passing algorithm There is no tractable way to compute exactly the joint probability of the observed snapshot OO. We approximate it using a mean-field type approach: PP OO ii, tt 0 PP SS kk (tt 0, ii) PP II ll (tt 0, ii) PP RR mm (tt 0, ii) kk OO,qq kk tt 0 =SS ll OO,qq ll tt 0 =II mm OO,qq mm tt 0 =RR If the value of tt 0 is not known, we estimate it by choosing the value that maximizes the partition function ZZ tt = ii PP(OO ii, tt).

DMP equations- How to compute marginals? Since the aforementioned marginals sum to one, therefore: PP II ii tt + 1 = 1 PP SS ii tt + 1 PP RR ii tt + 1 As the recovery process from the state II to state RR is independent of neighbors, for the SIR model we have: PP RR ii tt + 1 = PP RR ii tt + μμ ii PP II ii (tt) We need one more update equation, i.e., for PP SS ii (tt + 1) in order to complete the above set of equations. We will define 3 separate messages under different auxiliary dynamics to obtain an update rule for PP SS ii (tt + 1).

Dynamic message-passing equations We consider two kinds of auxiliary dynamics: DD jj where node jj receives infection signals, but ignores them and thus is fixed to the S state at all times. DD iiii where neighboring nodes ii and jj receive infection signals, but ignore them and are fixed to the S state at all times. We note that the auxiliary dynamics DD jj is identical to the original dynamics for all times such that qq jj tt = SS (similar analogy for DD iiii ). Also note that in dynamics DD jj, since the infection cannot propagate through node jj, different graph branches rooted at node jj become independent if the underlying graph is a tree.

Dynamic message-passing equations To obtain a closed system of marginal probability update equations, we will write the update rules for three different kinds of messages. The message θθ kk ii (tt) is the probability that the infection signal has not been passed from node kk to node ii up to time tt in the dynamics DD ii : θθ kk ii tt = Prob DD ii[ tt tt =0 dd kk ii tt = 0] The message φφ kk ii (tt) is the probability that the infection signal has not been passed from node kk to node ii up to time tt in the dynamics DD ii and that node kk is in the state II at time tt: φφ kk ii tt = Prob DD ii[ tt tt =0 dd kk ii tt = 0, qq kk tt = II]

Dynamic message-passing equations Finally, the message PP SS kk ii (tt) is the probability that node kk is in the state SS at time tt in the dynamics DD ii : PP SS kk ii tt = Prob DD ii[qq kk tt = SS] From the definition of the messages, we have: where ii means the set of neighbors of ii. PP ii SS tt = PP ii SS 0 θθ kk ii (tt + 1) kk It can be shown easily that: PP SS ii jj tt = PP SS ii 0 kk \j θθ kk ii (tt + 1)

Dynamic message-passing equations For the remaining two messages, the update rules are as follows: θθ kk ii tt + 1 θθ kk ii tt = λλ kkkk φφ kk ii tt, θθ ii jj 0 = 1. φφ kk ii tt + 1 = 1 λλ kkkk 1 μμ kk φφ kk ii tt PP SS kk ii tt + 1 PP SS kk ii tt, φφ kk ii 0 = δδ qqii 0,II. Combining the message-passing update equations, we get an iterative update rule for PP SS ii (tt + 1). This completes the set of equations required to compute all the marginal probabilities.

Experimental Results

Experimental Results

Thank you