Dynamic network sampling

Similar documents
Markov Model. Model representing the different resident states of a system, and the transitions between the different states

A&S 320: Mathematical Modeling in Biology

How to Use This Presentation

Markov Chains. X(t) is a Markov Process if, for arbitrary times t 1 < t 2 <... < t k < t k+1. If X(t) is discrete-valued. If X(t) is continuous-valued

All living organisms are limited by factors in the environment

The decoupling assumption in large stochastic system analysis Talk at ECLT

Project 1 Modeling of Epidemics

Social Influence in Online Social Networks. Epidemiological Models. Epidemic Process

Markov Chains, Random Walks on Graphs, and the Laplacian

Markov Processes Hamid R. Rabiee

AARMS Homework Exercises

CS 798: Homework Assignment 3 (Queueing Theory)

2. Transience and Recurrence

UNIVERSITY OF YORK. MSc Examinations 2004 MATHEMATICS Networks. Time Allowed: 3 hours.

Performance Evaluation of Queuing Systems

Markov Chains and MCMC

Statistics Canada International Symposium Series - Proceedings Symposium 2004: Innovative Methods for Surveying Difficult-to-reach Populations

Any live cell with less than 2 live neighbours dies. Any live cell with 2 or 3 live neighbours lives on to the next step.

1.3 Convergence of Regular Markov Chains

reversed chain is ergodic and has the same equilibrium probabilities (check that π j =

LIMITING PROBABILITY TRANSITION MATRIX OF A CONDENSED FIBONACCI TREE

4452 Mathematical Modeling Lecture 16: Markov Processes

Analytically tractable processes on networks

Markov Chains and Web Ranking: a Multilevel Adaptive Aggregation Method

Module 6:Random walks and related areas Lecture 24:Random woalk and other areas. The Lecture Contains: Random Walk.

Example: physical systems. If the state space. Example: speech recognition. Context can be. Example: epidemics. Suppose each infected

TEMPORAL EXPONENTIAL- FAMILY RANDOM GRAPH MODELING (TERGMS) WITH STATNET

Birth and Death Processes. Birth and Death Processes. Linear Growth with Immigration. Limiting Behaviour for Birth and Death Processes

COPYRIGHTED MATERIAL CONTENTS. Preface Preface to the First Edition

Lecture Notes: Markov chains Tuesday, September 16 Dannie Durand

Markov-Chain Monte Carlo

Recap. Probability, stochastic processes, Markov chains. ELEC-C7210 Modeling and analysis of communication networks

MARKOV PROCESSES. Valerio Di Valerio

Using Markov Chains To Model Human Migration in a Network Equilibrium Framework

Figure The Threshold Theorem of epidemiology

The Living World Continued: Populations and Communities

STA 4273H: Statistical Machine Learning

Powerful tool for sampling from complicated distributions. Many use Markov chains to model events that arise in nature.

Chapter 5 Lecture. Metapopulation Ecology. Spring 2013

An Introduction to Tuberculosis Disease Modeling

ESTIMATING LIKELIHOODS FOR SPATIO-TEMPORAL MODELS USING IMPORTANCE SAMPLING

Inference in Bayesian Networks

Markov Chain Monte Carlo The Metropolis-Hastings Algorithm

Use of Eigen values and eigen vectors to calculate higher transition probabilities

Link Analysis. Stony Brook University CSE545, Fall 2016

Detailed Balance and Branching Processes Monday, October 20, :04 PM

Stochastic Processes

IEOR 6711: Stochastic Models I, Fall 2003, Professor Whitt. Solutions to Final Exam: Thursday, December 18.

Math 142-2, Homework 2

Fundamentals of Applied Probability and Random Processes

PROBABILITY AND STOCHASTIC PROCESSES A Friendly Introduction for Electrical and Computer Engineers

On Influential Node Discovery in Dynamic Social Networks

CS242: Probabilistic Graphical Models Lecture 7B: Markov Chain Monte Carlo & Gibbs Sampling

Artificial Intelligence

STA 624 Practice Exam 2 Applied Stochastic Processes Spring, 2008

Adaptive Web Sampling

Horizontal versus vertical transmission of parasites in a stochastic spatial model

Slide source: Mining of Massive Datasets Jure Leskovec, Anand Rajaraman, Jeff Ullman Stanford University.

Introduction to Machine Learning CMU-10701

Machine Learning for Data Science (CS4786) Lecture 24

Lecture 15: MCMC Sanjeev Arora Elad Hazan. COS 402 Machine Learning and Artificial Intelligence Fall 2016

Lecture 20: Reversible Processes and Queues

Markov Chains Handout for Stat 110

Bayesian model selection in graphs by using BDgraph package

Adaptive Web Sampling

Thursday. Threshold and Sensitivity Analysis

Examples of Countable State Markov Chains Thursday, October 16, :12 PM

Phylogenetics: Bayesian Phylogenetic Analysis. COMP Spring 2015 Luay Nakhleh, Rice University

0.1 Naive formulation of PageRank

Challenges in Privacy-Preserving Analysis of Structured Data Kamalika Chaudhuri

Learning in Bayesian Networks

STAT 536: Migration. Karin S. Dorman. October 3, Department of Statistics Iowa State University

LAW OF LARGE NUMBERS FOR THE SIRS EPIDEMIC

FINAL REPORT. Project Title: Extreme Values of Queues, Point Processes and Stochastic Networks

De los ejercicios de abajo (sacados del libro de Georgii, Stochastics) se proponen los siguientes:

Multistate Modelling Vertical Transmission and Determination of R 0 Using Transition Intensities

DISCRETE STOCHASTIC PROCESSES Draft of 2nd Edition

Die-out Probability in SIS Epidemic Processes on Networks

10.2 A Stochastic Spatiotemporal Analysis of the Contribution of Primary versus Secondary Spread of HLB.

JUNIOR SECONDARY EXTERNAL EXAMINATION

Homework 2 will be posted by tomorrow morning, due Friday, October 16 at 5 PM.

Markov Chains and Stochastic Sampling

Math Homework 5 Solutions

At the boundary states, we take the same rules except we forbid leaving the state space, so,.

Ch.5 Evolution and Community Ecology How do organisms become so well suited to their environment? Evolution and Natural Selection

Research supported by NSF grant DMS

Stochastic modelling of epidemic spread

Ch. 4 - Population Ecology

Convex Optimization CMU-10725

The cost/reward formula has two specific widely used applications:

Unit 1 Lesson 3 Population Dynamics. Copyright Houghton Mifflin Harcourt Publishing Company

Mathematical and Algorithmic Models of Refugee Crises

Stochastic inference in Bayesian networks, Markov chain Monte Carlo methods

The parallel replica method for Markov chains

Copyright is owned by the Author of the thesis. Permission is given for a copy to be downloaded by an individual for the purpose of research and

An Introduction to Reversible Jump MCMC for Bayesian Networks, with Application

Lab 3. Adding Forces with a Force Table

Gozo College Boys Secondary Victoria Gozo, Malta Ninu Cremona

Bayesian Methods for Machine Learning

Temporal Point Processes the Conditional Intensity Function

Transcription:

Dynamic network sampling Steve Thompson Simon Fraser University thompson@sfu.ca Graybill Conference Colorado State University June 10, 2013

Dynamic network sampling The population of interest has spatial structure, often has network structure and moves or changes over time Steve Thompson () Dynamic network sampling 2 / 24

Dynamic network sampling The population of interest has spatial structure, often has network structure and moves or changes over time Designs for selecting a sample units use spatial and network relationships and progress dynamically Steve Thompson () Dynamic network sampling 2 / 24

Population and sample processes Population: A stochastic process {Y t }. Sample: A stochastic process {S t }. Time t such as day, with t = 0, 1, 2,.... Values Y t of units such as locations, states, and relationships between units at time t. Sample S t the set of units in the sample at time t. Steve Thompson () Dynamic network sampling 3 / 24

Spatial-temporal population model Purpose of dynamic population model is to evaluate the effectiveness of different sampling designs. We want it to be simple but to incorporate those characteristics that affect the effectiveness of sampling strategies. Steve Thompson () Dynamic network sampling 4 / 24

Spatial-temporal population model Purpose of dynamic population model is to evaluate the effectiveness of different sampling designs. We want it to be simple but to incorporate those characteristics that affect the effectiveness of sampling strategies. clustering, mixing, migration point process Steve Thompson () Dynamic network sampling 4 / 24

Spatial-temporal population model Purpose of dynamic population model is to evaluate the effectiveness of different sampling designs. We want it to be simple but to incorporate those characteristics that affect the effectiveness of sampling strategies. clustering, mixing, migration point process movements within and among groups small random displacements, MCMC selections, autoregressive processes Steve Thompson () Dynamic network sampling 4 / 24

Spatial-temporal population model Purpose of dynamic population model is to evaluate the effectiveness of different sampling designs. We want it to be simple but to incorporate those characteristics that affect the effectiveness of sampling strategies. clustering, mixing, migration point process movements within and among groups small random displacements, MCMC selections, autoregressive processes insertions and deletions of objects birth and death process, immigration and emigration. Steve Thompson () Dynamic network sampling 4 / 24

Dynamic network model builds on spatial temporal point process link probabilities dependent on distance between nodes, node characteristics, and current degree or target degree distribution renewal process for link formation, persistence, and dissolution Steve Thompson () Dynamic network sampling 5 / 24

Sampling process A sampling design in the static situation is a procedure for selecting units to include in the sample. For the dynamic situation we use a sampling process that includes Steve Thompson () Dynamic network sampling 6 / 24

Sampling process A sampling design in the static situation is a procedure for selecting units to include in the sample. For the dynamic situation we use a sampling process that includes an acquisition process by which units are added to the sample Steve Thompson () Dynamic network sampling 6 / 24

Sampling process A sampling design in the static situation is a procedure for selecting units to include in the sample. For the dynamic situation we use a sampling process that includes an acquisition process by which units are added to the sample an attrition process by which units are removed from the sample Steve Thompson () Dynamic network sampling 6 / 24

Equilibrium distributions Many properties of the population and sample process are ergodic and have stationary and limiting distributions. Steve Thompson () Dynamic network sampling 7 / 24

Equilibrium distributions Many properties of the population and sample process are ergodic and have stationary and limiting distributions. Sample size tends to increase when acquisition rate exceeds attrition rate and decrease when attrition rate exceeds attrition rate Steve Thompson () Dynamic network sampling 7 / 24

Uses of sampling designs Inference about population characteristics Experiments on sample units Interventions on sample units Steve Thompson () Dynamic network sampling 8 / 24

Intervention strategy Select a sample of units from the population, make interventions on those units changing their values. Steve Thompson () Dynamic network sampling 9 / 24

Intervention strategy Select a sample of units from the population, make interventions on those units changing their values. Objective is to change the population, not just sample units, in a desired way. Steve Thompson () Dynamic network sampling 9 / 24

Intervention strategy Select a sample of units from the population, make interventions on those units changing their values. Objective is to change the population, not just sample units, in a desired way. One strategy interacts with another Steve Thompson () Dynamic network sampling 9 / 24

Effect of an intervention An intervention strategy consists of a sampling design for finding units in the population on which to make interventions, and a plan for the types of interventions to be made, which may depend on sample unit characteristics. A simple way to view the effect of a strategy is the difference in the resulting equilibrium distribution, compared with the equilibrium distribution without the strategy, or with a different strategy. Steve Thompson () Dynamic network sampling 10 / 24

Natural sampling strategies virus selects a sample of people with a link-tracing design insects select a sample of plants with a temporal spatial distance design. Steve Thompson () Dynamic network sampling 11 / 24

Dynamic network sampling Steve Thompson () Dynamic network sampling 12 / 24

Dynamic network sampling initially and at ongoing rate units are selected using conventional or spatial design Steve Thompson () Dynamic network sampling 12 / 24

Dynamic network sampling initially and at ongoing rate units are selected using conventional or spatial design new units are added through link tracing tracing rate may depend on unit and link values Steve Thompson () Dynamic network sampling 12 / 24

Dynamic network sampling initially and at ongoing rate units are selected using conventional or spatial design new units are added through link tracing tracing rate may depend on unit and link values units are removed from sample through removal probability or deletion from population. Steve Thompson () Dynamic network sampling 12 / 24

Random walk in static network The classic random walk in a graph starts with an arbitrary node and at each step selects at random one of the links out from the current node to reach the next node. Steve Thompson () Dynamic network sampling 13 / 24

Random walk in static network The classic random walk in a graph starts with an arbitrary node and at each step selects at random one of the links out from the current node to reach the next node. The current sample S t consists of that one node. Steve Thompson () Dynamic network sampling 13 / 24

Random walk in static network The classic random walk in a graph starts with an arbitrary node and at each step selects at random one of the links out from the current node to reach the next node. The current sample S t consists of that one node. When sampling is with replacement, the sequence S 0, S 1, S 2 is a Markov chain with constant transition matrix. Steve Thompson () Dynamic network sampling 13 / 24

Random walk in static network The classic random walk in a graph starts with an arbitrary node and at each step selects at random one of the links out from the current node to reach the next node. The current sample S t consists of that one node. When sampling is with replacement, the sequence S 0, S 1, S 2 is a Markov chain with constant transition matrix. Connected components of networks form closed classes. Steve Thompson () Dynamic network sampling 13 / 24

Random walk in dynamic network links between nodes change over time component structure changes a random walk temporarily stuck in one component can eventually reach nodes of other components Steve Thompson () Dynamic network sampling 14 / 24

Random walk in dynamic network links between nodes change over time component structure changes a random walk temporarily stuck in one component can eventually reach nodes of other components deletions and insertions of nodes can interrupt the walk, requiring reseeding Steve Thompson () Dynamic network sampling 14 / 24

Random set designs current sample S t is a set of nodes acquisition of nodes includes tracing links out from sample attrition through removal of nodes from sample or deletion from population Steve Thompson () Dynamic network sampling 15 / 24

Simple random set design with-replacement selection independent link tracing independent removals Steve Thompson () Dynamic network sampling 16 / 24

Design options selection and removal probabilities depend on node and link values with or without replacent links followed from active sample units only Steve Thompson () Dynamic network sampling 17 / 24

Design options selection and removal probabilities depend on node and link values with or without replacent links followed from active sample units only replacement, activeness values between 0 and 1 Steve Thompson () Dynamic network sampling 17 / 24

Desired features trace rapidly at first; reseeding. have a target sample size distribution. find units with high degree or interesting values. Steve Thompson () Dynamic network sampling 18 / 24

Epidemic example HIV virus spreads with dynamic network sampling design seek and treat designs for interventions to reduce incidence. combination of interventions and counter-responses leads to new equilibrium distribution. Steve Thompson () Dynamic network sampling 19 / 24

What influences equilibrium distribution sample volume = number of nodes sample surface = number of links out Steve Thompson () Dynamic network sampling 20 / 24

What influences equilibrium distribution sample volume = number of nodes sample surface = number of links out may be weighted by tracing probabilities Steve Thompson () Dynamic network sampling 20 / 24

What influences equilibrium distribution sample volume = number of nodes sample surface = number of links out may be weighted by tracing probabilities surface to volume ratio tends to decrease as sample size increases Steve Thompson () Dynamic network sampling 20 / 24

Also affecting equilibrium level new node entering sample (HIV) tends to have higher degree than average for population and higher proportion of links-out than average for sample especially in early stages of epidemic Steve Thompson () Dynamic network sampling 21 / 24

0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.2 0.4 0.6 0.8 1.0 Steve Thompson () Dynamic network sampling 22 / 24

Proportion infected 0.0 0.2 0.4 0.6 0.8 1.0 0 50000 100000 150000 day Steve Thompson () Dynamic network sampling 23 / 24

Proportion infected 0.0 0.2 0.4 0.6 0.8 1.0 0 50000 100000 150000 day Steve Thompson () Dynamic network sampling 24 / 24