Machine learning for Dynamic Social Network Analysis

Similar documents
Learning with Temporal Point Processes

Introduc)on to Ar)ficial Intelligence

Graphical Models. Lecture 1: Mo4va4on and Founda4ons. Andrew McCallum

Least Squares Parameter Es.ma.on

Networks. Can (John) Bruce Keck Founda7on Biotechnology Lab Bioinforma7cs Resource

IS4200/CS6200 Informa0on Retrieval. PageRank Con+nued. with slides from Hinrich Schütze and Chris6na Lioma

DART Tutorial Sec'on 1: Filtering For a One Variable System

STA 4273H: Sta-s-cal Machine Learning

Computer Vision. Pa0ern Recogni4on Concepts Part I. Luis F. Teixeira MAP- i 2012/13

CSE 473: Ar+ficial Intelligence

STAD68: Machine Learning

An Introduc+on to Sta+s+cs and Machine Learning for Quan+ta+ve Biology. Anirvan Sengupta Dept. of Physics and Astronomy Rutgers University

CSE 473: Ar+ficial Intelligence. Probability Recap. Markov Models - II. Condi+onal probability. Product rule. Chain rule.

Sta$s$cal sequence recogni$on

Metropolis Sampler and Markov Chains

Introduc)on to Ar)ficial Intelligence

Outline. What is Machine Learning? Why Machine Learning? 9/29/08. Machine Learning Approaches to Biological Research: Bioimage Informa>cs and Beyond

Quan&fying Uncertainty. Sai Ravela Massachuse7s Ins&tute of Technology

Announcements. Topics: Homework: - sec0ons 1.2, 1.3, and 2.1 * Read these sec0ons and study solved examples in your textbook!

Reduced Models for Process Simula2on and Op2miza2on

CS 6140: Machine Learning Spring 2017

Gene Regulatory Networks II Computa.onal Genomics Seyoung Kim

Matrix products in integrable probability

Polynomials and Gröbner Bases

CSCI 360 Introduc/on to Ar/ficial Intelligence Week 2: Problem Solving and Op/miza/on

Founda'ons of Large- Scale Mul'media Informa'on Management and Retrieval. Lecture #3 Machine Learning. Edward Chang

CS 6140: Machine Learning Spring What We Learned Last Week 2/26/16

Par$cle Filters Part I: Theory. Peter Jan van Leeuwen Data- Assimila$on Research Centre DARC University of Reading

Two common difficul:es with HW 2: Problem 1c: v = r n ˆr

Experimental Designs for Planning Efficient Accelerated Life Tests

Quantum mechanics with indefinite causal order

Class Notes. Examining Repeated Measures Data on Individuals

Graphical Models. Lecture 3: Local Condi6onal Probability Distribu6ons. Andrew McCallum

Par$$oned Elias- Fano indexes. Giuseppe O)aviano Rossano Venturini

The BCS Model. Sara Changizi. This presenta5on closely follows parts of chapter 6 in Ring & Schuck The nuclear many body problem.

7. Quantum Monte Carlo (QMC)

Stat Lecture 20. Last class we introduced the covariance and correlation between two jointly distributed random variables.

CSCI 360 Introduc/on to Ar/ficial Intelligence Week 2: Problem Solving and Op/miza/on. Professor Wei-Min Shen Week 8.1 and 8.2

Examining the Evidence: Using data to pose and answer ques8ons about Earth s changing climate! Free Polar Explorer: Sea Level app and ac8vi8es

Computer Vision. Pa0ern Recogni4on Concepts. Luis F. Teixeira MAP- i 2014/15

DS504/CS586: Big Data Analytics Graph Mining II

Par$$oned Elias- Fano Indexes

Matrix models for the black hole informa4on paradox

Order- Revealing Encryp2on and the Hardness of Private Learning

Ensemble of Climate Models

UVA CS / Introduc8on to Machine Learning and Data Mining

DART Tutorial Part IV: Other Updates for an Observed Variable

CSE 473: Ar+ficial Intelligence. Hidden Markov Models. Bayes Nets. Two random variable at each +me step Hidden state, X i Observa+on, E i

Announcements. Topics: Work On: - sec0ons 1.2 and 1.3 * Read these sec0ons and study solved examples in your textbook!

Chapter 4. Fourier Analysis for Con5nuous-Time Signals and Systems Chapter Objec5ves

Systems Biology and Neuroengineering

Tsybakov noise adap/ve margin- based ac/ve learning

A6523 Modeling, Inference, and Mining Jim Cordes, Cornell University. False Positives in Fourier Spectra. For N = DFT length: Lecture 5 Reading

A6523 Modeling, Inference, and Mining Jim Cordes, Cornell University. Motivations: Detection & Characterization. Lecture 2.

Detec%ng and Analyzing Urban Regions with High Impact of Weather Change on Transport

Electricity & Magnetism Lecture 2: Electric Fields

CS 6140: Machine Learning Spring What We Learned Last Week. Survey 2/26/16. VS. Model

CS 6140: Machine Learning Spring 2016

MA/CS 109 Lecture 7. Back To Exponen:al Growth Popula:on Models

Bayesian networks Lecture 18. David Sontag New York University

CSCI1950 Z Computa3onal Methods for Biology Lecture 24. Ben Raphael April 29, hgp://cs.brown.edu/courses/csci1950 z/ Network Mo3fs

Cellular automata, entropy and box- coun4ng dimension

Quantum Energy Inequali1es

R Process Nucleosynthesis And Its Site. Mario A. Riquelme Theore<cal Seminar Fall 2009

Exponen'al func'ons and exponen'al growth. UBC Math 102

Mul$- model ensemble challenge ini$al/model uncertain$es

Least Mean Squares Regression. Machine Learning Fall 2017

Lecture 13: Tracking mo3on features op3cal flow

Some thoughts on linearity, nonlinearity, and partial separability

Machine Learning & Data Mining CS/CNS/EE 155. Lecture 11: Hidden Markov Models

CSE 473: Ar+ficial Intelligence. Example. Par+cle Filters for HMMs. An HMM is defined by: Ini+al distribu+on: Transi+ons: Emissions:

Latent Dirichlet Alloca/on

Probability and Structure in Natural Language Processing

Density Estimation: ML, MAP, Bayesian estimation

DS504/CS586: Big Data Analytics Graph Mining II

COMP 562: Introduction to Machine Learning

REGRESSION AND CORRELATION ANALYSIS

Rela%ons and Their Proper%es. Slides by A. Bloomfield

A6523 Modeling, Inference, and Mining Jim Cordes, Cornell University

PHYS1121 and MECHANICS

LBT for Procedural and Reac1ve Systems Part 2: Reac1ve Systems Basic Theory

Predicate abstrac,on and interpola,on. Many pictures and examples are borrowed from The So'ware Model Checker BLAST presenta,on.

SKA machine learning perspec1ves for imaging, processing and analysis

Aggrega?on of Epistemic Uncertainty

Sample Complexity of Learning Independent of Set Theory

Last Lecture Recap UVA CS / Introduc8on to Machine Learning and Data Mining. Lecture 3: Linear Regression

PSAAP Project Stanford

D. Dubois, H. Prade, F. Touazi (coopéra7on avec A. Hadjali, S. Kaci)

ODEs + Singulari0es + Monodromies + Boundary condi0ons. Kerr BH ScaRering: a systema0c study. Schwarzschild BH ScaRering: Quasi- normal modes

Part 2. Representation Learning Algorithms

Molecular Programming Models. Based on notes by Dave Doty

STA414/2104 Statistical Methods for Machine Learning II

Social and Economic Networks Matthew O. Jackson

Greedy Maximization Framework for Graph-based Influence Functions

Learning Deep Genera,ve Models

Gaussian Processes. Le Song. Machine Learning II: Advanced Topics CSE 8803ML, Spring 2012

Informa(on Retrieval

Parallel Repetition of entangled games on the uniform distribution

Last week. The diaba)c circula)on. Associated isentropic mass budget in the Middleworld. Possible implica7ons for poten7al vor7city

Statistical Models for sequencing data: from Experimental Design to Generalized Linear Models

Transcription:

Machine learning for Dynamic Social Network Analysis Manuel Gomez Rodriguez Max Planck Ins7tute for So;ware Systems UC3M, MAY 2017

Interconnected World SOCIAL NETWORKS TRANSPORTATION NETWORKS WORLD WIDE WEB PROTEIN INTERACTIONS INFORMATION NETWORKS INTERNET OF THINGS 2

Many discrete events in con7nuous 7me Qmee, 2013 3

Variety of processes behind these events Events are (noisy) observa7ons of a variety of complex dynamic processes FAST News spread in TwiXer Video becomes viral in Youtube Product reviews and sales in Amazon Ar7cle crea7on in Wikipedia A user gains recogni7on in Quora SLOW in a wide range of temporal scales. 4

Example I: Idea adop7on/viral marke7ng S D means D follows S 3.25pm Bob Chris7ne 3.00pm Beth 3.27pm Joe David 4.15pm t Friggeri et al., 2014 They can have an impact in the off- line world 5

Example II: Informa7on crea7on & cura7on Addi7on Refuta7on t Ques7on Answer Upvote t t

Example III: Learning trajectories 1st year computer science student Introduc9on to programming Discrete math Project presenta9on For/do- while loops Define Set theory func9ons Powerpoint Graph Theory Class vs. Keynote inheritance Export Geometry pptx to pdf t If else How to write Logic switch Private func9ons PP templates Class destructor Plot library 7

Detailed event traces DETAILED TRACES OF ACTIVITY The availability of event traces boosts a new generation of data- driven models and algorithms t 8

Previously: discrete- 7me models & algorithms Epoch 1 Epoch 2 Epoch 3 Epoch 4 Discrete- 7me models ar7ficially introduce epochs: 1. How long is each epoch? Data is very heterogeneous. 2. How to aggregate events within an epoch? 3. What if no event within an epoch? 4. Time is treated as index or condi7oning variable, not easy to deal with 7me- related queries. 9

Outline of the Seminar REPRESENTATION: TEMPORAL POINT PROCESSES 1. Intensity func7on 2. Basic building blocks 3. Superposi7on 4. Marks and SDEs with jumps This lecture APPLICATIONS: MODELS 1. Informa7on propaga7on 2. Opinion dynamics 3. Informa7on reliability 4. Knowledge acquisi7on APPLICATIONS: CONTROL 1. Influence maximiza7on 2. Ac7vity shaping 3. When- to- post Slides/references: learning.mpi-sws.org/uc3m-seminar 10

Representa7on: Temporal Point Processes 1. Intensity func7on 2. Basic building blocks 3. Superposi7on 4. Marks and SDEs with jumps 11

Temporal point processes Temporal point process: A random process whose realiza7on consists of discrete events localized in 7me Discrete events History, Dirac delta func7on Formally: 12

Model 7me as a random variable Prob. between [t, t+dt) density History, Prob. not before t Likelihood of a 7meline: 13

Problems of density parametriza7on (I) It is difficult for model design and interpretability: 1. Densi7es need to integrate to 1 (i.e., par77on func7on) 2. Difficult to combine 7melines 14

Problems of density parametriza7on (II) Difficult to combine 7melines: + = Sum of random processes 15

Intensity func7on density Prob. between [t, t+dt) History, Prob. not before t Intensity: Probability between [t, t+dt) but not before t Observa7on: It is a rate = # of events / unit of 7me 16

Advantages of intensity parametriza7on (I) Suitable for model design and interpretable: 1. Intensi7es only need to be nonnega7ve 2. Easy to combine 7melines 17

Advantages of intensity parametriza7on (II) Easy to combine 7meline: + = Sum of random processes 18

Rela7on between f*, F*, S*, λ* Central quan7ty we will use! 19

Representa7on: Temporal Point Processes 1. Intensity func7on 2. Basic building blocks 3. Superposi7on 4. Marks and SDEs with jumps 20

Poisson process Intensity of a Poisson process Observa7ons: 1. Intensity independent of history 2. Uniformly random occurrence 3. Time interval follows exponen7al distribu7on 21

Fisng a Poisson from (historical) 7meline Maximum likelihood 22

Sampling from a Poisson process We would like to sample: We sample using inversion sampling: 23

Inhomogeneous Poisson process Intensity of an inhomogeneous Poisson process Observa7ons: 1. Intensity independent of history 24

Fisng an inhomogeneous Poisson Maximum likelihood Design such that max. likelihood is convex (and use CVX)

Nonparametric inhomogeneous Poisson process Posi7ve combina7on of (Gaussian) RFB kernels: 26

Sampling from an inhomogeneous Poisson Thinning procedure (similar to rejec7on sampling): 1. Sample from Poisson process with intensity Inversion sampling 2. Generate 3. Keep the sample if Keep sample with prob.

Termina7ng (or survival) process Intensity of a termina7ng (or survival) process Observa7ons: 1. Limited number of occurrences 28

Self- exci7ng (or Hawkes) process History, Intensity of self- exci7ng (or Hawkes) process: Triggering kernel Observa7ons: 1. Clustered (or bursty) occurrence of events 2. Intensity is stochas7c and history dependent 29

Fisng a Hawkes process from a recorded 7meline Maximum likelihood The max. likelihood is jointly convex in and

Sampling from a Hawkes process Thinning procedure (similar to rejec7on sampling): 1. Sample from Poisson process with intensity Inversion sampling 2. Generate 3. Keep the sample if Keep sample with prob. 31

Summary Building blocks to represent different dynamic processes: Poisson processes: Inhomogeneous Poisson processes: We know how to fit them and how to sample from them Termina9ng point processes: Self- exci9ng point processes: 32

Representa7on: Temporal Point Processes 1. Intensity func7on 2. Basic building blocks 3. Superposi7on 4. Marks and SDEs with jumps 33

Superposi7on of processes Sample each intensity + take minimum = Addi7ve intensity 34

Mutually exci7ng process Bob History Chris7ne History Clustered occurrence affected by neighbors 35

Mutually exci7ng termina7ng process Bob Chris7ne History Clustered occurrence affected by neighbors 36

Representa7on: Temporal Point Processes 1. Intensity func7on 2. Basic building blocks 3. Superposi7on 4. Marks and SDEs with jumps 37

Marked temporal point processes Marked temporal point process: A random process whose realiza7on consists of discrete marked events localized in 7me History, 38

Independent iden7cally distributed marks Distribu7on for the marks: Observa7ons: 1. Marks independent of the temporal dynamics 2. Independent iden7cally distributed (I.I.D.) 39

Dependent marks: SDEs with jumps History, Marks given by stochas7c differen7al equa7on with jumps: Observa7ons: Dri; Event influence 1. Marks dependent of the temporal dynamics 2. Defined for all values of t 40

Dependent marks: distribu7on + SDE with jumps History, Distribu7on for the marks: Observa7ons: Dri; Event influence 1. Marks dependent on the temporal dynamics 2. Distribu7on represents addi7onal source of uncertainty 41

Mutually exci7ng + marks Bob Chris7ne Marks affected by neighbors Dri; Neighbor influence 42

REPRESENTATION: TEMPORAL POINT PROCESSES 1. Intensity func7on 2. Basic building blocks 3. Superposi7on 4. Marks and SDEs with jumps APPLICATIONS: MODELS 1. Informa7on propaga7on 2. Opinion dynamics 3. Informa7on reliability 4. Knowledge acquisi7on This lecture Next lecture APPLICATIONS: CONTROL 1. Influence maximiza7on 2. Ac7vity shaping 3. When- to- post Slides/references: learning.mpi-sws.org/sydney-seminar 43