CS224W: Social and Information Network Analysis Jure Leskovec, Stanford University
|
|
- Evelyn Thomas
- 5 years ago
- Views:
Transcription
1 CS224W: Social and Information Network Analysis Jure Leskovec, Stanford University
2 Find most influential set S of size k: largest expected cascade size f(s) if set S is activated a c b Want to solve: d e Network, each edge activates with prob. p max S = k 1 f ( S) = fi( S) I Consider S={a,d} then: f 1 (S)=5, f 2 (S)=4, f 3 (S)=3 and f(s) = 4 f i I Activate edges by coin flipping Multiple realizations i: 10/26/2011 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, 2 a a a influence set c c b b c b d d d e e f f e f
3 10/13/2009 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, 3 Algorithm: Hill Climbing At step i pick node u that: max u f(s i 1 u ) Thm: Hill climbing produces a solution S where: f(s) (1-1/e)*OPT Claim holds for functions f() with 2 properties: f is monotone: S T: f(s) f(t) and f({})=0 f is submodular: S T : f(s {u}) f(s) f(t {u}) f(t) Our f(s) is submodular! Why?
4 10/20/2010 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, 4 Submodularity: f(s {u}) f(s) f(t {u}) f(t) Basic fact 1: If f 1 (x),,f k (x) are submodular, and c 1,,c k 0 then F x = i c i f i x is also submodular Basic fact 2: A simple submodular function Sets A 1,, A m f S = S T T S A i i S (size of the union of sets A i, i S) u The more sets you already chose the less new area a new set u will cover
5 10/26/2011 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, 5 a c b d e f Activate edges by coin flipping Fix outcome i of coin flips f i (a) f i (u) = influence set: set of nodes a reachable from u on live-edge paths f i (S) = size of cascades from S on coin flips i f i S = u S f i (u) f i (S) is submodular! f i (v) are sets and f i (S) is the size of the union f i (a) Expected influence set size: f S = 1 i I f i (S) f(s) is submodular! I f(s) is linear combination of submodular functions a a f i (a) b c b c b c f i (d) e d f e d f i (d) f e d f f i (d)
6
7 10/13/2009 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, 7 Claim: If f(s) is monotone and submodular. Hill climbing produces a solution S where: f(s) (1-1/e)*OPT (f(s)>0.63*opt) Setting: Keep adding nodes that give the largest gain Start with S 0 ={}, produce sets S 1, S 2,,S k Add elements one by one Marginal gain: δ i = f(s i ) - f(s i-1 ) Let T={t 1 t k } be the optimal set of size k We need to show: f(s) (1-1/e) f(t)
8 10/13/2009 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, 8 k j=1 f(a B) f(a) [f(a {b j }) f(a)] where: B = {b 1,,b k } and f is submodular, Proof: Let B i = {b 1, b i }, so we have B 1, B 2,, B k =B f A B f A = k f A B i f A B i 1 k i=1 k i=1 i=1 = f A B i 1 b i f A B i 1 By submodularity since A X {b} A {b} f A {b i } f A Work out the sum. Everything but 1 st and last term cancels out: f A B 1 f A B 0 + f A B 2 f A B 1 + f A B 3 + f A B k f(a B k 1 )
9 10/13/2009 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, 9 f T f S i T Remember: δ i = f(s i ) - f(s i-1 ) (by monotonicity) = f S i T f S i + f S i k j=1 f S i {t j } f S i + f(s i ) k j=1 δ i+1 + f S i = f S i + k δ i+1 Thus: f T f S i + k δ i+1 δ i+1 1 k [f T f(s i)] (by prev. slide) T = {t 1, t k } t j is one choice of a next element. We greedily choose the best one, for a gain of δ i+1. This is the hill-climbing assumption.
10 10/13/2009 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, 10 We just showed: δ i+1 1 k [f T f(s i)] What is f(s i+1 )? f S i+1 = f S i + δ i+1 f S i + 1 k f T f S i = 1 1 k f S i + 1 k f(t) What is f(s k )?
11 10/13/2009 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, 11 Claim: f Proof by induction: i = 0: i 1 ( Si ) 1 1 f ( T ) k f S 0 = f({}) = k 0 f T = 0
12 10/13/2009 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, 12 Claim: f Proof by induction: At i + 1: i 1 ( Si ) 1 1 f ( T ) k f S i k f S i + 1 k f T 1 1 k k i f T + 1 k f T = k i+1 f(t)
13 10/26/2011 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, 13 Thus: f S = f S k k k f T Then: 1 e f S k 1 1 e f(t) qed.
14 10/26/2011 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, 14 We just proved: Hill climbing finds solution S which f(s) (1-1/e)*OPT i.e., f(s) 0.63*OPT This is a data independent bound This is a worst case bound No matter what is the input data (influence sets) we know that Hill Climbing won t do worse than 0.63*OPT Data dependent bound: Value of the bound depends on the input data On easy data, hill climbing may do better than 63%
15 Suppose S is some solution to f(s) s.t. S k f(s) is monotone & submodular Let T = {t 1,,t k } be the OPT solution For each u S let δ u = f(s {u})-f(s) Order δ u so that δ 1 δ 2 δ n Then: f S f T Note: k i=1 This is a data dependent bound (δ u depend on input data) Bound holds for any algorithm δ i Makes no assumption about how S is computed For some inputs the bound can be loose (worse than 63%) 10/20/2010 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, 15
16 For each u S let δ u = f(s {u})-f(s) Order δ u so that δ 1 δ 2 δ n k Then: f S f T i=1 δ i Proof: f T f T S = f S + k f S t 1 t i f S t 1 t i 1 i=1 k i=1 k i=1 f S + f S t i f S = f S + k δ ti k i=1 f S + i=1 δ i f T f S + δ i Instead of adding t i T (of benefit δ ti ), we add the best possible element (δ i ) 10/20/2010 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, 16
17 10/26/2011 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, 17
18 10/20/2010 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, 18 Hill-climbing a b c d e reward Add node with highest marginal gain What do we know about optimizing submodular functions? A hill-climbing is near optimal (1-1/e (~63%) of OPT) But Hill-climbing algorithm is slow At each iteration we need to reevaluate marginal gains It scales as O(n k)
19 [Leskovec et al., KDD 07] In round i+1: So far we picked S i = {s 1,,s i } Now pick s i+1 = argmax u f(s i {u}) - f(s i ) maximize the marginal benefit δ u (S i ) = f(s i {u}) - f(s i ) By submodularity property: f S i u f S i f S j u f S j for i<j Observation: By submodularity : For some u δ u (S i ) δ u (S j ) for i j since S i S j Marginal benefits δ x only shrink! δ u (S i ) δ u (S j ) Activating node u in step i helps more than activating it at step j (j>i) 10/20/2010 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, 19 u
20 [Leskovec et al., KDD 07] 10/20/2010 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, 20 Idea: Use δ i as upper-bound on δ j (j>i) Lazy hill-climbing: Keep an ordered list of marginal benefits δ i from previous iteration Re-evaluate δ i only for top node Re-sort and prune Marginal gain a b c d e S 1 ={a} f(s {u}) f(s) f(t {u}) f(t) S T
21 [Leskovec et al., KDD 07] 10/20/2010 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, 21 Idea: Use δ i as upper-bound on δ j (j>i) Lazy hill-climbing: Keep an ordered list of marginal benefits δ i from previous iteration Re-evaluate δ i only for top node Re-sort and prune Marginal gain a b c d e S 1 ={a} f(s {u}) f(s) f(t {u}) f(t) S T
22 [Leskovec et al., KDD 07] 10/20/2010 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, 22 Idea: Use δ i as upper-bound on δ j (j>i) Lazy hill-climbing: Keep an ordered list of marginal benefits δ i from previous iteration Re-evaluate δ i only for top node Re-sort and prune Marginal gain a d b e c S 1 ={a} S 2 ={a,b} f(s {u}) f(s) f(t {u}) f(t) S T
23
24 [Leskovec et al., KDD 07] 10/20/2010 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, 24 Given a real city water distribution network And data on how contaminants spread in the network Detect the contaminant as quickly as possible Problem posed by the US Environmental Protection Agency S
25 10/20/2010 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, 25 Utility of placing sensors Water flow dynamics, demands of households, For each subset A V compute utility F(A) High impact Contamination outbreak S3 S1 S2 Low impact outbreak Medium impact outbreak S3 S1 S4 Set V of all network junctions Sensor reduces impact through early detection! S1 S4 S2 High sensing quality F(A) = 0.9 Low sensing quality F(A)=0.01
26 [Leskovec et al., KDD 07] Given a graph G(V,E) And the data on how outbreaks spread over the network: For each outbreak i we know the time T(i,u) when outbreak i contaminated node u Goal: Select a subset of nodes A that maximize the expected reward: max S V f S = P i f i S Reward: Raise the alarm and save the most people i Expected reward for detecting outbreak i outbreak i Monitoring blue node saves more people than monitoring the green node 10/20/2010 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, 26
27 [Leskovec et al., KDD 07] 10/20/2010 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, 27 Observation: Diminishing returns S 1 New sensor: S s S 1 S 3 S 2 S 2 S 4 Placement S={s 1, s 2 } Adding s helps a lot Placement S ={s 1, s 2, s 3, s 4 } Adding s helps very little
28 [Leskovec et al., KDD 07] 10/20/2010 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, 28 Claim: The reward function is submodular Consider outbreak i: f i (u k ) = set of nodes saved from u k f i (S) = size of union f i (u k ), u k S f i is submodular Global optimization: f(s) = i P(i) f i (S) f(s) is submodular outbreak i u 2 R i (u 2 ) u 1 R i (u 1 )
29 [Leskovec et al., KDD 07] 10/20/2010 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, 29 Real metropolitan area network V = 21,000 nodes E = 25,000 pipes water Use a cluster of 50 machines for a month Simulate 3.6 million epidemic scenarios (152 GB of epidemic data) By exploiting sparsity we fit it into main memory (16GB)
30 Solution quality F(A) Higher is better Offline the (1-1/e) bound Data-dependent bound Hill Climbing Number of sensors placed Data-dependent bound is much tighter (gives more accurate estimate of alg. performance) 10/20/2010 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, 30
31 [Leskovec et al., KDD 07] 10/20/2010 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, 31 Placement heuristics perform much worse
32 10/20/2010 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, 32 = I have 10 minutes. Which blogs should I read to be most up to date?? = Who are the most influential bloggers?
33 Want to read things before others do. Detect blue & yellow soon but miss red. Detect all stories but late. 10/20/2010 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, 33
34 [Leskovec et al., KDD 07] 10/20/2010 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, 34 Online bound is much tighter: 87% instead of 63% (1-1/e) bound Data dependent bound Hill Climbing
35 [Leskovec et al., KDD 07] 10/20/2010 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, 35 Heuristics perform much worse
36 [Leskovec et al., KDD 07] 10/20/2010 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, 36 Naïve hill climbing Lazy hill climbing Lazy evaluation runs 700 times faster than naïve Hill Climbing algorithm
CS224W: Social and Information Network Analysis Jure Leskovec, Stanford University
CS224W: Social and Information Network Analysis Jure Leskovec, Stanford University http://cs224w.stanford.edu 10/24/2012 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, http://cs224w.stanford.edu
More informationCS 322: (Social and Information) Network Analysis Jure Leskovec Stanford University
CS 322: (Social and Inormation) Network Analysis Jure Leskovec Stanord University Initially some nodes S are active Each edge (a,b) has probability (weight) p ab b 0.4 0.4 0.2 a 0.4 0.2 0.4 g 0.2 Node
More information9. Submodular function optimization
Submodular function maximization 9-9. Submodular function optimization Submodular function maximization Greedy algorithm for monotone case Influence maximization Greedy algorithm for non-monotone case
More informationActive Learning and Optimized Information Gathering
Active Learning and Optimized Information Gathering Lecture 13 Submodularity (cont d) CS 101.2 Andreas Krause Announcements Homework 2: Due Thursday Feb 19 Project milestone due: Feb 24 4 Pages, NIPS format:
More informationCSCI 3210: Computational Game Theory. Cascading Behavior in Networks Ref: [AGT] Ch 24
CSCI 3210: Computational Game Theory Cascading Behavior in Networks Ref: [AGT] Ch 24 Mohammad T. Irfan Email: mirfan@bowdoin.edu Web: www.bowdoin.edu/~mirfan Course Website: www.bowdoin.edu/~mirfan/csci-3210.html
More informationMaximizing the Spread of Influence through a Social Network. David Kempe, Jon Kleinberg, Éva Tardos SIGKDD 03
Maximizing the Spread of Influence through a Social Network David Kempe, Jon Kleinberg, Éva Tardos SIGKDD 03 Influence and Social Networks Economics, sociology, political science, etc. all have studied
More informationMore Approximation Algorithms
CS 473: Algorithms, Spring 2018 More Approximation Algorithms Lecture 25 April 26, 2018 Most slides are courtesy Prof. Chekuri Ruta (UIUC) CS473 1 Spring 2018 1 / 28 Formal definition of approximation
More informationCS224W: Social and Information Network Analysis Jure Leskovec, Stanford University
CS224W: Social and Information Network Analysis Jure Leskovec Stanford University Jure Leskovec, Stanford University http://cs224w.stanford.edu Task: Find coalitions in signed networks Incentives: European
More informationCS224W: Analysis of Networks Jure Leskovec, Stanford University
Announcements: Please fill HW Survey Weekend Office Hours starting this weekend (Hangout only) Proposal: Can use 1 late period CS224W: Analysis of Networks Jure Leskovec, Stanford University http://cs224w.stanford.edu
More informationCSE541 Class 22. Jeremy Buhler. November 22, Today: how to generalize some well-known approximation results
CSE541 Class 22 Jeremy Buhler November 22, 2016 Today: how to generalize some well-known approximation results 1 Intuition: Behavior of Functions Consider a real-valued function gz) on integers or reals).
More informationSubmodular Functions Properties Algorithms Machine Learning
Submodular Functions Properties Algorithms Machine Learning Rémi Gilleron Inria Lille - Nord Europe & LIFL & Univ Lille Jan. 12 revised Aug. 14 Rémi Gilleron (Mostrare) Submodular Functions Jan. 12 revised
More informationCS224W: Analysis of Networks Jure Leskovec, Stanford University
CS224W: Analysis of Networks Jure Leskovec, Stanford University http://cs224w.stanford.edu 10/30/17 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, http://cs224w.stanford.edu 2
More informationInfluence Maximization in Continuous Time Diffusion Networks
Manuel Gomez-Rodriguez 1,2 Bernhard Schölkopf 1 1 MPI for Intelligent Systems and 2 Stanford University MANUELGR@STANFORD.EDU BS@TUEBINGEN.MPG.DE Abstract The problem of finding the optimal set of source
More informationMulti-Round Influence Maximization
Multi-Round Influence Maximization Lichao Sun 1, Weiran Huang 2, Philip S. Yu 1,2, Wei Chen 3, 1 University of Illinois at Chicago, 2 Tsinghua University, 3 Microsoft Research lsun29@uic.edu, huang.inbox@outlook.com,
More informationDiffusion of Innovation and Influence Maximization
Diffusion of Innovation and Influence Maximization Leonid E. Zhukov School of Data Analysis and Artificial Intelligence Department of Computer Science National Research University Higher School of Economics
More informationGreedy Maximization Framework for Graph-based Influence Functions
Greedy Maximization Framework for Graph-based Influence Functions Edith Cohen Google Research Tel Aviv University HotWeb '16 1 Large Graphs Model relations/interactions (edges) between entities (nodes)
More informationCS224W: Social and Information Network Analysis Jure Leskovec, Stanford University
CS224W: Social and Information Network Analysis Jure Leskovec, Stanford University http://cs224w.stanford.edu 12/3/2013 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, http://cs224w.stanford.edu
More information1 Submodular functions
CS 369P: Polyhedral techniques in combinatorial optimization Instructor: Jan Vondrák Lecture date: November 16, 2010 1 Submodular functions We have already encountered submodular functions. Let s recall
More informationUsing first-order logic, formalize the following knowledge:
Probabilistic Artificial Intelligence Final Exam Feb 2, 2016 Time limit: 120 minutes Number of pages: 19 Total points: 100 You can use the back of the pages if you run out of space. Collaboration on the
More informationJure Leskovec Stanford University
Jure Leskovec Stanford University 2 Part 1: Models for networks Part 2: Information flows in networks Information spread in social media 3 Information flows through networks Analyzing underlying mechanisms
More informationAdaptive Submodularity: A New Approach to Active Learning and Stochastic Optimization
Adaptive Submodularity: A New Approach to Active Learning and Stochastic Optimization Daniel Golovin California Institute of Technology Pasadena, CA 91125 dgolovin@caltech.edu Andreas Krause California
More informationCascading Behavior in Networks: Algorithmic and Economic Issues
CHAPTER 24 Cascading Behavior in Networks: Algorithmic and Economic Issues Jon Kleinberg Abstract The flow of information or influence through a large social network can be thought of as unfolding with
More informationMachine Learning and Bayesian Inference. Unsupervised learning. Can we find regularity in data without the aid of labels?
Machine Learning and Bayesian Inference Dr Sean Holden Computer Laboratory, Room FC6 Telephone extension 6372 Email: sbh11@cl.cam.ac.uk www.cl.cam.ac.uk/ sbh11/ Unsupervised learning Can we find regularity
More informationCSE446: Clustering and EM Spring 2017
CSE446: Clustering and EM Spring 2017 Ali Farhadi Slides adapted from Carlos Guestrin, Dan Klein, and Luke Zettlemoyer Clustering systems: Unsupervised learning Clustering Detect patterns in unlabeled
More informationOn-Line Social Systems with Long-Range Goals
On-Line Social Systems with Long-Range Goals Jon Kleinberg Cornell University Including joint work with Ashton Anderson, Dan Huttenlocher, Jure Leskovec, and Sigal Oren. Long-Range Planning Growth in on-line
More informationECS 253 / MAE 253, Lecture 15 May 17, I. Probability generating function recap
ECS 253 / MAE 253, Lecture 15 May 17, 2016 I. Probability generating function recap Part I. Ensemble approaches A. Master equations (Random graph evolution, cluster aggregation) B. Network configuration
More informationCS224W: Social and Information Network Analysis Jure Leskovec, Stanford University
CS224W: Social and Information Network Analysis Jure Leskovec, Stanford University http://cs224w.stanford.edu Non-overlapping vs. overlapping communities 11/10/2010 Jure Leskovec, Stanford CS224W: Social
More informationCS 188: Artificial Intelligence. Bayes Nets
CS 188: Artificial Intelligence Probabilistic Inference: Enumeration, Variable Elimination, Sampling Pieter Abbeel UC Berkeley Many slides over this course adapted from Dan Klein, Stuart Russell, Andrew
More informationCS599: Convex and Combinatorial Optimization Fall 2013 Lecture 24: Introduction to Submodular Functions. Instructor: Shaddin Dughmi
CS599: Convex and Combinatorial Optimization Fall 2013 Lecture 24: Introduction to Submodular Functions Instructor: Shaddin Dughmi Announcements Introduction We saw how matroids form a class of feasible
More information27 : Distributed Monte Carlo Markov Chain. 1 Recap of MCMC and Naive Parallel Gibbs Sampling
10-708: Probabilistic Graphical Models 10-708, Spring 2014 27 : Distributed Monte Carlo Markov Chain Lecturer: Eric P. Xing Scribes: Pengtao Xie, Khoa Luu In this scribe, we are going to review the Parallel
More informationCS224W: Social and Information Network Analysis Jure Leskovec, Stanford University
CS224W: Social and Information Network Analysis Jure Leskovec, Stanford University http://cs224w.stanford.edu Intro sessions to SNAP C++ and SNAP.PY: SNAP.PY: Friday 9/27, 4:5 5:30pm in Gates B03 SNAP
More informationThe Ties that Bind Characterizing Classes by Attributes and Social Ties
The Ties that Bind WWW April, 2017, Bryan Perozzi*, Leman Akoglu Stony Brook University *Now at Google. Introduction Outline Our problem: Characterizing Community Differences Proposed Method Experimental
More informationStochastic Submodular Cover with Limited Adaptivity
Stochastic Submodular Cover with Limited Adaptivity Arpit Agarwal Sepehr Assadi Sanjeev Khanna Abstract In the submodular cover problem, we are given a non-negative monotone submodular function f over
More informationDetermining the Diameter of Small World Networks
Determining the Diameter of Small World Networks Frank W. Takes & Walter A. Kosters Leiden University, The Netherlands CIKM 2011 October 2, 2011 Glasgow, UK NWO COMPASS project (grant #12.0.92) 1 / 30
More informationCS 188: Artificial Intelligence Spring Announcements
CS 188: Artificial Intelligence Spring 2011 Lecture 12: Probability 3/2/2011 Pieter Abbeel UC Berkeley Many slides adapted from Dan Klein. 1 Announcements P3 due on Monday (3/7) at 4:59pm W3 going out
More informationTime-Critical Influence Maximization in Social Networks with Time-Delayed Diffusion Process
Time-Critical Influence Maximization in Social Networks with Time-Delayed Diffusion Process Wei Chen Microsoft Research Asia Beijing, China weic@microsoft.com Wei Lu University of British Columbia Vancouver,
More informationCopyright 2000, Kevin Wayne 1
Algorithm runtime analysis and computational tractability Time Complexity of an Algorithm How do we measure the complexity (time, space requirements) of an algorithm. 1 microsecond? Units of time As soon
More informationCS 6901 (Applied Algorithms) Lecture 3
CS 6901 (Applied Algorithms) Lecture 3 Antonina Kolokolova September 16, 2014 1 Representative problems: brief overview In this lecture we will look at several problems which, although look somewhat similar
More informationAnnouncements. Inference. Mid-term. Inference by Enumeration. Reminder: Alarm Network. Introduction to Artificial Intelligence. V22.
Introduction to Artificial Intelligence V22.0472-001 Fall 2009 Lecture 15: Bayes Nets 3 Midterms graded Assignment 2 graded Announcements Rob Fergus Dept of Computer Science, Courant Institute, NYU Slides
More informationFig. 1. Securely distinguish three vehicle flows /17/$31.00 c 2017 IEEE
On the RSU-based Secure Distinguishability Among Vehicular Flows Wei Chang, Huanyang Zheng, and Jie Wu Department of Computer Science, Saint Joseph s University, Philadelphia, PA 19131 Department of Computer
More informationApproximate Inference
Approximate Inference Simulation has a name: sampling Sampling is a hot topic in machine learning, and it s really simple Basic idea: Draw N samples from a sampling distribution S Compute an approximate
More informationCS 781 Lecture 9 March 10, 2011 Topics: Local Search and Optimization Metropolis Algorithm Greedy Optimization Hopfield Networks Max Cut Problem Nash
CS 781 Lecture 9 March 10, 2011 Topics: Local Search and Optimization Metropolis Algorithm Greedy Optimization Hopfield Networks Max Cut Problem Nash Equilibrium Price of Stability Coping With NP-Hardness
More informationA Fast Approximation for Influence Maximization in Large Social Networks
A Fast Approximation for Influence Maximization in Large Social Networks ABSTRACT Jong-Ryul Lee Dept. of Computer Science, KAIST 29 Daehak-ro, Yuseong-gu, Daejeon, Korea jrlee@islab.kaist.ac.kr This paper
More informationECS 253 / MAE 253, Lecture 13 May 15, Diffusion, Cascades and Influence Mathematical models & generating functions
ECS 253 / MAE 253, Lecture 13 May 15, 2018 Diffusion, Cascades and Influence Mathematical models & generating functions Last week: spatial flows and game theory on networks Optimal location of facilities
More information2.6 Complexity Theory for Map-Reduce. Star Joins 2.6. COMPLEXITY THEORY FOR MAP-REDUCE 51
2.6. COMPLEXITY THEORY FOR MAP-REDUCE 51 Star Joins A common structure for data mining of commercial data is the star join. For example, a chain store like Walmart keeps a fact table whose tuples each
More informationAdaptive Submodularity: Theory and Applications in Active Learning and Stochastic Optimization
Journal of Artificial Intelligence Research 42 (2011) 427-486 Submitted 1/11; published 11/11 Adaptive Submodularity: Theory and Applications in Active Learning and Stochastic Optimization Daniel Golovin
More informationCS 188: Artificial Intelligence
CS 188: Artificial Intelligence Adversarial Search II Instructor: Anca Dragan University of California, Berkeley [These slides adapted from Dan Klein and Pieter Abbeel] Minimax Example 3 12 8 2 4 6 14
More informationCS340 Machine learning Lecture 5 Learning theory cont'd. Some slides are borrowed from Stuart Russell and Thorsten Joachims
CS340 Machine learning Lecture 5 Learning theory cont'd Some slides are borrowed from Stuart Russell and Thorsten Joachims Inductive learning Simplest form: learn a function from examples f is the target
More informationFaster Kriging on Graphs
Faster Kriging on Graphs Omkar Muralidharan Abstract [Xu et al. 2009] introduce a graph prediction method that is accurate but slow. My project investigates faster methods based on theirs that are nearly
More informationMaximizing the Spread of Influence through a Social Network
Maximizing the Spread of Influence through a Social Network David Kempe Dept. of Computer Science Cornell University, Ithaca NY kempe@cs.cornell.edu Jon Kleinberg Dept. of Computer Science Cornell University,
More informationCS246: Mining Massive Datasets Jure Leskovec, Stanford University
CS246: Mining Massive Datasets Jure Leskovec, Stanford University http://cs246.stanford.edu 2/26/2013 Jure Leskovec, Stanford CS246: Mining Massive Datasets, http://cs246.stanford.edu 2 More algorithms
More informationOverlapping Communities
Overlapping Communities Davide Mottin HassoPlattner Institute Graph Mining course Winter Semester 2017 Acknowledgements Most of this lecture is taken from: http://web.stanford.edu/class/cs224w/slides GRAPH
More informationBayes Nets III: Inference
1 Hal Daumé III (me@hal3.name) Bayes Nets III: Inference Hal Daumé III Computer Science University of Maryland me@hal3.name CS 421: Introduction to Artificial Intelligence 10 Apr 2012 Many slides courtesy
More informationAditya Bhaskara CS 5968/6968, Lecture 1: Introduction and Review 12 January 2016
Lecture 1: Introduction and Review We begin with a short introduction to the course, and logistics. We then survey some basics about approximation algorithms and probability. We also introduce some of
More informationCSE 421 Greedy Algorithms / Interval Scheduling
CSE 421 Greedy Algorithms / Interval Scheduling Yin Tat Lee 1 Interval Scheduling Job j starts at s(j) and finishes at f(j). Two jobs compatible if they don t overlap. Goal: find maximum subset of mutually
More informationOptimization of Submodular Functions Tutorial - lecture I
Optimization of Submodular Functions Tutorial - lecture I Jan Vondrák 1 1 IBM Almaden Research Center San Jose, CA Jan Vondrák (IBM Almaden) Submodular Optimization Tutorial 1 / 1 Lecture I: outline 1
More informationAn Analytic Solution to Discrete Bayesian Reinforcement Learning
An Analytic Solution to Discrete Bayesian Reinforcement Learning Pascal Poupart (U of Waterloo) Nikos Vlassis (U of Amsterdam) Jesse Hoey (U of Toronto) Kevin Regan (U of Waterloo) 1 Motivation Automated
More informationModels of Communication Dynamics for Simulation of Information Diffusion
Models of Communication Dynamics for Simulation of Information Diffusion Konstantin Mertsalov, Malik Magdon-Ismail, Mark Goldberg Rensselaer Polytechnic Institute Department of Computer Science 11 8th
More informationEdge-Weighted Personalized PageRank: Breaking a Decade-Old Performance Barrier
Edge-Weighted Personalized PageRank: Breaking a Decade-Old Performance Barrier W. Xie D. Bindel A. Demers J. Gehrke 12 Aug 2015 W. Xie, D. Bindel, A. Demers, J. Gehrke KDD2015 12 Aug 2015 1 / 1 PageRank
More informationLecture 11 October 11, Information Dissemination through Social Networks
CS 284r: Incentives and Information in Networks Fall 2013 Prof. Yaron Singer Lecture 11 October 11, 2013 Scribe: Michael Tingley, K. Nathaniel Tucker 1 Overview In today s lecture we will start the second
More informationProbability Models of Information Exchange on Networks Lecture 6
Probability Models of Information Exchange on Networks Lecture 6 UC Berkeley Many Other Models There are many models of information exchange on networks. Q: Which model to chose? My answer good features
More informationMaximization of Submodular Set Functions
Northeastern University Department of Electrical and Computer Engineering Maximization of Submodular Set Functions Biomedical Signal Processing, Imaging, Reasoning, and Learning BSPIRAL) Group Author:
More informationMotivation, Basic Concepts, Basic Methods, Travelling Salesperson Problem (TSP), Algorithms
Motivation, Basic Concepts, Basic Methods, Travelling Salesperson Problem (TSP), Algorithms 1 What is Combinatorial Optimization? Combinatorial Optimization deals with problems where we have to search
More informationPresentation for CSG399 By Jian Wen
Presentation for CSG399 By Jian Wen Outline Skyline Definition and properties Related problems Progressive Algorithms Analysis SFS Analysis Algorithm Cost Estimate Epsilon Skyline(overview and idea) Skyline
More informationDistributed Systems Gossip Algorithms
Distributed Systems Gossip Algorithms He Sun School of Informatics University of Edinburgh What is Gossip? Gossip algorithms In a gossip algorithm, each node in the network periodically exchanges information
More informationInferring the origin of an epidemic with a dynamic message-passing algorithm
Inferring the origin of an epidemic with a dynamic message-passing algorithm HARSH GUPTA (Based on the original work done by Andrey Y. Lokhov, Marc Mézard, Hiroki Ohta, and Lenka Zdeborová) Paper Andrey
More informationDistributed Optimization. Song Chong EE, KAIST
Distributed Optimization Song Chong EE, KAIST songchong@kaist.edu Dynamic Programming for Path Planning A path-planning problem consists of a weighted directed graph with a set of n nodes N, directed links
More informationReview: Probability. BM1: Advanced Natural Language Processing. University of Potsdam. Tatjana Scheffler
Review: Probability BM1: Advanced Natural Language Processing University of Potsdam Tatjana Scheffler tatjana.scheffler@uni-potsdam.de October 21, 2016 Today probability random variables Bayes rule expectation
More informationConsistency Under Sampling of Exponential Random Graph Models
Consistency Under Sampling of Exponential Random Graph Models Cosma Shalizi and Alessandro Rinaldo Summary by: Elly Kaizar Remember ERGMs (Exponential Random Graph Models) Exponential family models Sufficient
More informationECS 289 F / MAE 298, Lecture 15 May 20, Diffusion, Cascades and Influence
ECS 289 F / MAE 298, Lecture 15 May 20, 2014 Diffusion, Cascades and Influence Diffusion and cascades in networks (Nodes in one of two states) Viruses (human and computer) contact processes epidemic thresholds
More informationStochastic Submodular Cover with Limited Adaptivity
Stochastic Submodular Cover with Limited Adaptivity Arpit Agarwal Sepehr Assadi Sanjeev Khanna University of Pennsylvania {aarpit,sassadi,sanjeev}@cis.upenn.edu arxiv:1810.13351v1 [cs.ds] 31 Oct 2018 Abstract
More informationApplications of Submodular Functions in Speech and NLP
Applications of Submodular Functions in Speech and NLP Jeff Bilmes Department of Electrical Engineering University of Washington, Seattle http://ssli.ee.washington.edu/~bilmes June 27, 2011 J. Bilmes Applications
More informationMobiHoc 2014 MINIMUM-SIZED INFLUENTIAL NODE SET SELECTION FOR SOCIAL NETWORKS UNDER THE INDEPENDENT CASCADE MODEL
MobiHoc 2014 MINIMUM-SIZED INFLUENTIAL NODE SET SELECTION FOR SOCIAL NETWORKS UNDER THE INDEPENDENT CASCADE MODEL Jing (Selena) He Department of Computer Science, Kennesaw State University Shouling Ji,
More informationACM 116: Lecture 1. Agenda. Philosophy of the Course. Definition of probabilities. Equally likely outcomes. Elements of combinatorics
1 ACM 116: Lecture 1 Agenda Philosophy of the Course Definition of probabilities Equally likely outcomes Elements of combinatorics Conditional probabilities 2 Philosophy of the Course Probability is the
More informationApproximability of Adaptive Seeding under Knapsack Constraints
Approximability of Adaptive Seeding under Knapsack Constraints Aviad Rubinstein UC Berkeley aviad@cs.berkeley.edu Lior Seeman Cornell Univeristy lseeman@cs.cornell.edu May 6, 2015 Yaron Singer Harvard
More informationCS 188 Introduction to Fall 2007 Artificial Intelligence Midterm
NAME: SID#: Login: Sec: 1 CS 188 Introduction to Fall 2007 Artificial Intelligence Midterm You have 80 minutes. The exam is closed book, closed notes except a one-page crib sheet, basic calculators only.
More informationCS224W: Social and Information Network Analysis Jure Leskovec, Stanford University
CS224W: Social and Information Network Analysis Jure Leskovec, Stanford University http://cs224w.stanford.edu How to organize/navigate it? First try: Human curated Web directories Yahoo, DMOZ, LookSmart
More information1 Maximizing a Submodular Function
6.883 Learning with Combinatorial Structure Notes for Lecture 16 Author: Arpit Agarwal 1 Maximizing a Submodular Function In the last lecture we looked at maximization of a monotone submodular function,
More informationN/4 + N/2 + N = 2N 2.
CS61B Summer 2006 Instructor: Erin Korber Lecture 24, 7 Aug. 1 Amortized Analysis For some of the data structures we ve discussed (namely hash tables and splay trees), it was claimed that the average time
More informationNETS 412: Algorithmic Game Theory March 28 and 30, Lecture Approximation in Mechanism Design. X(v) = arg max v i (a)
NETS 412: Algorithmic Game Theory March 28 and 30, 2017 Lecture 16+17 Lecturer: Aaron Roth Scribe: Aaron Roth Approximation in Mechanism Design In the last lecture, we asked how far we can go beyond the
More information12. LOCAL SEARCH. gradient descent Metropolis algorithm Hopfield neural networks maximum cut Nash equilibria
12. LOCAL SEARCH gradient descent Metropolis algorithm Hopfield neural networks maximum cut Nash equilibria Lecture slides by Kevin Wayne Copyright 2005 Pearson-Addison Wesley h ttp://www.cs.princeton.edu/~wayne/kleinberg-tardos
More informationDimensional Analysis and Exponential Models Review Solutions
Dimensional Analysis and Exponential Models Review Solutions 1. Given 13 morks is 1 gloe, 5 gloes is 1 flit, 7 kits is one lonk, and 10 lonks is 1 gall. Convert 90 morks per kit to flits per gall. To solve
More informationDiscrete Optimization in Machine Learning. Colorado Reed
Discrete Optimization in Machine Learning Colorado Reed [ML-RCC] 31 Jan 2013 1 Acknowledgements Some slides/animations based on: Krause et al. tutorials: http://www.submodularity.org Pushmeet Kohli tutorial:
More informationCS341 info session is on Thu 3/1 5pm in Gates415. CS246: Mining Massive Datasets Jure Leskovec, Stanford University
CS341 info session is on Thu 3/1 5pm in Gates415 CS246: Mining Massive Datasets Jure Leskovec, Stanford University http://cs246.stanford.edu 2/28/18 Jure Leskovec, Stanford CS246: Mining Massive Datasets,
More informationAdministration. CSCI567 Machine Learning (Fall 2018) Outline. Outline. HW5 is available, due on 11/18. Practice final will also be available soon.
Administration CSCI567 Machine Learning Fall 2018 Prof. Haipeng Luo U of Southern California Nov 7, 2018 HW5 is available, due on 11/18. Practice final will also be available soon. Remaining weeks: 11/14,
More informationBubble Clustering: Set Covering via Union of Ellipses
Bubble Clustering: Set Covering via Union of Ellipses Matt Kraning, Arezou Keshavarz, Lei Zhao CS229, Stanford University Email: {mkraning,arezou,leiz}@stanford.edu Abstract We develop an algorithm called
More informationPrinciples of AI Planning
Principles of 5. Planning as search: progression and regression Malte Helmert and Bernhard Nebel Albert-Ludwigs-Universität Freiburg May 4th, 2010 Planning as (classical) search Introduction Classification
More informationTractable Models for Information Diffusion in Social Networks
Tractable Models for Information Diffusion in Social Networks Masahiro Kimura 1 and Kazumi Saito 2 1 Department of Electronics and Informatics, Ryukoku University Otsu, Shiga 520-2194, Japan 2 NTT Communication
More informationAnnouncements. CS 188: Artificial Intelligence Fall Causality? Example: Traffic. Topology Limits Distributions. Example: Reverse Traffic
CS 188: Artificial Intelligence Fall 2008 Lecture 16: Bayes Nets III 10/23/2008 Announcements Midterms graded, up on glookup, back Tuesday W4 also graded, back in sections / box Past homeworks in return
More informationLearning in State-Space Reinforcement Learning CIS 32
Learning in State-Space Reinforcement Learning CIS 32 Functionalia Syllabus Updated: MIDTERM and REVIEW moved up one day. MIDTERM: Everything through Evolutionary Agents. HW 2 Out - DUE Sunday before the
More informationOn Influential Node Discovery in Dynamic Social Networks
On Influential Node Discovery in Dynamic Social Networks Charu Aggarwal Shuyang Lin Philip S. Yu Abstract The problem of maximizing influence spread has been widely studied in social networks, because
More informationthe tree till a class assignment is reached
Decision Trees Decision Tree for Playing Tennis Prediction is done by sending the example down Prediction is done by sending the example down the tree till a class assignment is reached Definitions Internal
More informationMachine Learning
Machine Learning 10-601 Tom M. Mitchell Machine Learning Department Carnegie Mellon University August 30, 2017 Today: Decision trees Overfitting The Big Picture Coming soon Probabilistic learning MLE,
More informationThe Traveling Salesman Problem: An Overview. David P. Williamson, Cornell University Ebay Research January 21, 2014
The Traveling Salesman Problem: An Overview David P. Williamson, Cornell University Ebay Research January 21, 2014 (Cook 2012) A highly readable introduction Some terminology (imprecise) Problem Traditional
More informationDecision Trees. CSC411/2515: Machine Learning and Data Mining, Winter 2018 Luke Zettlemoyer, Carlos Guestrin, and Andrew Moore
Decision Trees Claude Monet, The Mulberry Tree Slides from Pedro Domingos, CSC411/2515: Machine Learning and Data Mining, Winter 2018 Luke Zettlemoyer, Carlos Guestrin, and Andrew Moore Michael Guerzhoy
More informationLocally Adaptive Optimization: Adaptive Seeding for Monotone Submodular Functions
Locally Adaptive Optimization: Adaptive Seeding for Monotone Submodular Functions Ashwinumar Badanidiyuru Google ashwinumarbv@gmail.com Aviad Rubinstein UC Bereley aviad@cs.bereley.edu Lior Seeman Cornell
More informationOnline Learning with Feedback Graphs
Online Learning with Feedback Graphs Claudio Gentile INRIA and Google NY clagentile@gmailcom NYC March 6th, 2018 1 Content of this lecture Regret analysis of sequential prediction problems lying between
More informationEffective placement of sensors for efficient early warning system in water distribution network MASHREKI ISLAM SAMI
Effective placement of sensors for efficient early warning system in water distribution network MASHREKI ISLAM SAMI AGENDA Introduction and Problem Statement Methodology Experimentation and Analysis Limitations
More informationPerformance Evaluation. Analyzing competitive influence maximization problems with partial information: An approximation algorithmic framework
Performance Evaluation ( ) Contents lists available at ScienceDirect Performance Evaluation journal homepage: www.elsevier.com/locate/peva Analyzing competitive influence maximization problems with partial
More informationCost and Preference in Recommender Systems Junhua Chen LESS IS MORE
Cost and Preference in Recommender Systems Junhua Chen, Big Data Research Center, UESTC Email:junmshao@uestc.edu.cn http://staff.uestc.edu.cn/shaojunming Abstract In many recommender systems (RS), user
More information