CS224W: Social and Information Network Analysis Jure Leskovec, Stanford University

Size: px
Start display at page:

Download "CS224W: Social and Information Network Analysis Jure Leskovec, Stanford University"

Transcription

1 CS224W: Social and Information Network Analysis Jure Leskovec, Stanford University

2 Find most influential set S of size k: largest expected cascade size f(s) if set S is activated a c b Want to solve: d e Network, each edge activates with prob. p max S = k 1 f ( S) = fi( S) I Consider S={a,d} then: f 1 (S)=5, f 2 (S)=4, f 3 (S)=3 and f(s) = 4 f i I Activate edges by coin flipping Multiple realizations i: 10/26/2011 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, 2 a a a influence set c c b b c b d d d e e f f e f

3 10/13/2009 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, 3 Algorithm: Hill Climbing At step i pick node u that: max u f(s i 1 u ) Thm: Hill climbing produces a solution S where: f(s) (1-1/e)*OPT Claim holds for functions f() with 2 properties: f is monotone: S T: f(s) f(t) and f({})=0 f is submodular: S T : f(s {u}) f(s) f(t {u}) f(t) Our f(s) is submodular! Why?

4 10/20/2010 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, 4 Submodularity: f(s {u}) f(s) f(t {u}) f(t) Basic fact 1: If f 1 (x),,f k (x) are submodular, and c 1,,c k 0 then F x = i c i f i x is also submodular Basic fact 2: A simple submodular function Sets A 1,, A m f S = S T T S A i i S (size of the union of sets A i, i S) u The more sets you already chose the less new area a new set u will cover

5 10/26/2011 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, 5 a c b d e f Activate edges by coin flipping Fix outcome i of coin flips f i (a) f i (u) = influence set: set of nodes a reachable from u on live-edge paths f i (S) = size of cascades from S on coin flips i f i S = u S f i (u) f i (S) is submodular! f i (v) are sets and f i (S) is the size of the union f i (a) Expected influence set size: f S = 1 i I f i (S) f(s) is submodular! I f(s) is linear combination of submodular functions a a f i (a) b c b c b c f i (d) e d f e d f i (d) f e d f f i (d)

6

7 10/13/2009 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, 7 Claim: If f(s) is monotone and submodular. Hill climbing produces a solution S where: f(s) (1-1/e)*OPT (f(s)>0.63*opt) Setting: Keep adding nodes that give the largest gain Start with S 0 ={}, produce sets S 1, S 2,,S k Add elements one by one Marginal gain: δ i = f(s i ) - f(s i-1 ) Let T={t 1 t k } be the optimal set of size k We need to show: f(s) (1-1/e) f(t)

8 10/13/2009 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, 8 k j=1 f(a B) f(a) [f(a {b j }) f(a)] where: B = {b 1,,b k } and f is submodular, Proof: Let B i = {b 1, b i }, so we have B 1, B 2,, B k =B f A B f A = k f A B i f A B i 1 k i=1 k i=1 i=1 = f A B i 1 b i f A B i 1 By submodularity since A X {b} A {b} f A {b i } f A Work out the sum. Everything but 1 st and last term cancels out: f A B 1 f A B 0 + f A B 2 f A B 1 + f A B 3 + f A B k f(a B k 1 )

9 10/13/2009 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, 9 f T f S i T Remember: δ i = f(s i ) - f(s i-1 ) (by monotonicity) = f S i T f S i + f S i k j=1 f S i {t j } f S i + f(s i ) k j=1 δ i+1 + f S i = f S i + k δ i+1 Thus: f T f S i + k δ i+1 δ i+1 1 k [f T f(s i)] (by prev. slide) T = {t 1, t k } t j is one choice of a next element. We greedily choose the best one, for a gain of δ i+1. This is the hill-climbing assumption.

10 10/13/2009 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, 10 We just showed: δ i+1 1 k [f T f(s i)] What is f(s i+1 )? f S i+1 = f S i + δ i+1 f S i + 1 k f T f S i = 1 1 k f S i + 1 k f(t) What is f(s k )?

11 10/13/2009 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, 11 Claim: f Proof by induction: i = 0: i 1 ( Si ) 1 1 f ( T ) k f S 0 = f({}) = k 0 f T = 0

12 10/13/2009 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, 12 Claim: f Proof by induction: At i + 1: i 1 ( Si ) 1 1 f ( T ) k f S i k f S i + 1 k f T 1 1 k k i f T + 1 k f T = k i+1 f(t)

13 10/26/2011 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, 13 Thus: f S = f S k k k f T Then: 1 e f S k 1 1 e f(t) qed.

14 10/26/2011 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, 14 We just proved: Hill climbing finds solution S which f(s) (1-1/e)*OPT i.e., f(s) 0.63*OPT This is a data independent bound This is a worst case bound No matter what is the input data (influence sets) we know that Hill Climbing won t do worse than 0.63*OPT Data dependent bound: Value of the bound depends on the input data On easy data, hill climbing may do better than 63%

15 Suppose S is some solution to f(s) s.t. S k f(s) is monotone & submodular Let T = {t 1,,t k } be the OPT solution For each u S let δ u = f(s {u})-f(s) Order δ u so that δ 1 δ 2 δ n Then: f S f T Note: k i=1 This is a data dependent bound (δ u depend on input data) Bound holds for any algorithm δ i Makes no assumption about how S is computed For some inputs the bound can be loose (worse than 63%) 10/20/2010 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, 15

16 For each u S let δ u = f(s {u})-f(s) Order δ u so that δ 1 δ 2 δ n k Then: f S f T i=1 δ i Proof: f T f T S = f S + k f S t 1 t i f S t 1 t i 1 i=1 k i=1 k i=1 f S + f S t i f S = f S + k δ ti k i=1 f S + i=1 δ i f T f S + δ i Instead of adding t i T (of benefit δ ti ), we add the best possible element (δ i ) 10/20/2010 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, 16

17 10/26/2011 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, 17

18 10/20/2010 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, 18 Hill-climbing a b c d e reward Add node with highest marginal gain What do we know about optimizing submodular functions? A hill-climbing is near optimal (1-1/e (~63%) of OPT) But Hill-climbing algorithm is slow At each iteration we need to reevaluate marginal gains It scales as O(n k)

19 [Leskovec et al., KDD 07] In round i+1: So far we picked S i = {s 1,,s i } Now pick s i+1 = argmax u f(s i {u}) - f(s i ) maximize the marginal benefit δ u (S i ) = f(s i {u}) - f(s i ) By submodularity property: f S i u f S i f S j u f S j for i<j Observation: By submodularity : For some u δ u (S i ) δ u (S j ) for i j since S i S j Marginal benefits δ x only shrink! δ u (S i ) δ u (S j ) Activating node u in step i helps more than activating it at step j (j>i) 10/20/2010 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, 19 u

20 [Leskovec et al., KDD 07] 10/20/2010 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, 20 Idea: Use δ i as upper-bound on δ j (j>i) Lazy hill-climbing: Keep an ordered list of marginal benefits δ i from previous iteration Re-evaluate δ i only for top node Re-sort and prune Marginal gain a b c d e S 1 ={a} f(s {u}) f(s) f(t {u}) f(t) S T

21 [Leskovec et al., KDD 07] 10/20/2010 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, 21 Idea: Use δ i as upper-bound on δ j (j>i) Lazy hill-climbing: Keep an ordered list of marginal benefits δ i from previous iteration Re-evaluate δ i only for top node Re-sort and prune Marginal gain a b c d e S 1 ={a} f(s {u}) f(s) f(t {u}) f(t) S T

22 [Leskovec et al., KDD 07] 10/20/2010 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, 22 Idea: Use δ i as upper-bound on δ j (j>i) Lazy hill-climbing: Keep an ordered list of marginal benefits δ i from previous iteration Re-evaluate δ i only for top node Re-sort and prune Marginal gain a d b e c S 1 ={a} S 2 ={a,b} f(s {u}) f(s) f(t {u}) f(t) S T

23

24 [Leskovec et al., KDD 07] 10/20/2010 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, 24 Given a real city water distribution network And data on how contaminants spread in the network Detect the contaminant as quickly as possible Problem posed by the US Environmental Protection Agency S

25 10/20/2010 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, 25 Utility of placing sensors Water flow dynamics, demands of households, For each subset A V compute utility F(A) High impact Contamination outbreak S3 S1 S2 Low impact outbreak Medium impact outbreak S3 S1 S4 Set V of all network junctions Sensor reduces impact through early detection! S1 S4 S2 High sensing quality F(A) = 0.9 Low sensing quality F(A)=0.01

26 [Leskovec et al., KDD 07] Given a graph G(V,E) And the data on how outbreaks spread over the network: For each outbreak i we know the time T(i,u) when outbreak i contaminated node u Goal: Select a subset of nodes A that maximize the expected reward: max S V f S = P i f i S Reward: Raise the alarm and save the most people i Expected reward for detecting outbreak i outbreak i Monitoring blue node saves more people than monitoring the green node 10/20/2010 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, 26

27 [Leskovec et al., KDD 07] 10/20/2010 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, 27 Observation: Diminishing returns S 1 New sensor: S s S 1 S 3 S 2 S 2 S 4 Placement S={s 1, s 2 } Adding s helps a lot Placement S ={s 1, s 2, s 3, s 4 } Adding s helps very little

28 [Leskovec et al., KDD 07] 10/20/2010 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, 28 Claim: The reward function is submodular Consider outbreak i: f i (u k ) = set of nodes saved from u k f i (S) = size of union f i (u k ), u k S f i is submodular Global optimization: f(s) = i P(i) f i (S) f(s) is submodular outbreak i u 2 R i (u 2 ) u 1 R i (u 1 )

29 [Leskovec et al., KDD 07] 10/20/2010 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, 29 Real metropolitan area network V = 21,000 nodes E = 25,000 pipes water Use a cluster of 50 machines for a month Simulate 3.6 million epidemic scenarios (152 GB of epidemic data) By exploiting sparsity we fit it into main memory (16GB)

30 Solution quality F(A) Higher is better Offline the (1-1/e) bound Data-dependent bound Hill Climbing Number of sensors placed Data-dependent bound is much tighter (gives more accurate estimate of alg. performance) 10/20/2010 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, 30

31 [Leskovec et al., KDD 07] 10/20/2010 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, 31 Placement heuristics perform much worse

32 10/20/2010 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, 32 = I have 10 minutes. Which blogs should I read to be most up to date?? = Who are the most influential bloggers?

33 Want to read things before others do. Detect blue & yellow soon but miss red. Detect all stories but late. 10/20/2010 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, 33

34 [Leskovec et al., KDD 07] 10/20/2010 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, 34 Online bound is much tighter: 87% instead of 63% (1-1/e) bound Data dependent bound Hill Climbing

35 [Leskovec et al., KDD 07] 10/20/2010 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, 35 Heuristics perform much worse

36 [Leskovec et al., KDD 07] 10/20/2010 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, 36 Naïve hill climbing Lazy hill climbing Lazy evaluation runs 700 times faster than naïve Hill Climbing algorithm

CS224W: Social and Information Network Analysis Jure Leskovec, Stanford University

CS224W: Social and Information Network Analysis Jure Leskovec, Stanford University CS224W: Social and Information Network Analysis Jure Leskovec, Stanford University http://cs224w.stanford.edu 10/24/2012 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, http://cs224w.stanford.edu

More information

CS 322: (Social and Information) Network Analysis Jure Leskovec Stanford University

CS 322: (Social and Information) Network Analysis Jure Leskovec Stanford University CS 322: (Social and Inormation) Network Analysis Jure Leskovec Stanord University Initially some nodes S are active Each edge (a,b) has probability (weight) p ab b 0.4 0.4 0.2 a 0.4 0.2 0.4 g 0.2 Node

More information

9. Submodular function optimization

9. Submodular function optimization Submodular function maximization 9-9. Submodular function optimization Submodular function maximization Greedy algorithm for monotone case Influence maximization Greedy algorithm for non-monotone case

More information

Active Learning and Optimized Information Gathering

Active Learning and Optimized Information Gathering Active Learning and Optimized Information Gathering Lecture 13 Submodularity (cont d) CS 101.2 Andreas Krause Announcements Homework 2: Due Thursday Feb 19 Project milestone due: Feb 24 4 Pages, NIPS format:

More information

CSCI 3210: Computational Game Theory. Cascading Behavior in Networks Ref: [AGT] Ch 24

CSCI 3210: Computational Game Theory. Cascading Behavior in Networks Ref: [AGT] Ch 24 CSCI 3210: Computational Game Theory Cascading Behavior in Networks Ref: [AGT] Ch 24 Mohammad T. Irfan Email: mirfan@bowdoin.edu Web: www.bowdoin.edu/~mirfan Course Website: www.bowdoin.edu/~mirfan/csci-3210.html

More information

Maximizing the Spread of Influence through a Social Network. David Kempe, Jon Kleinberg, Éva Tardos SIGKDD 03

Maximizing the Spread of Influence through a Social Network. David Kempe, Jon Kleinberg, Éva Tardos SIGKDD 03 Maximizing the Spread of Influence through a Social Network David Kempe, Jon Kleinberg, Éva Tardos SIGKDD 03 Influence and Social Networks Economics, sociology, political science, etc. all have studied

More information

More Approximation Algorithms

More Approximation Algorithms CS 473: Algorithms, Spring 2018 More Approximation Algorithms Lecture 25 April 26, 2018 Most slides are courtesy Prof. Chekuri Ruta (UIUC) CS473 1 Spring 2018 1 / 28 Formal definition of approximation

More information

CS224W: Social and Information Network Analysis Jure Leskovec, Stanford University

CS224W: Social and Information Network Analysis Jure Leskovec, Stanford University CS224W: Social and Information Network Analysis Jure Leskovec Stanford University Jure Leskovec, Stanford University http://cs224w.stanford.edu Task: Find coalitions in signed networks Incentives: European

More information

CS224W: Analysis of Networks Jure Leskovec, Stanford University

CS224W: Analysis of Networks Jure Leskovec, Stanford University Announcements: Please fill HW Survey Weekend Office Hours starting this weekend (Hangout only) Proposal: Can use 1 late period CS224W: Analysis of Networks Jure Leskovec, Stanford University http://cs224w.stanford.edu

More information

CSE541 Class 22. Jeremy Buhler. November 22, Today: how to generalize some well-known approximation results

CSE541 Class 22. Jeremy Buhler. November 22, Today: how to generalize some well-known approximation results CSE541 Class 22 Jeremy Buhler November 22, 2016 Today: how to generalize some well-known approximation results 1 Intuition: Behavior of Functions Consider a real-valued function gz) on integers or reals).

More information

Submodular Functions Properties Algorithms Machine Learning

Submodular Functions Properties Algorithms Machine Learning Submodular Functions Properties Algorithms Machine Learning Rémi Gilleron Inria Lille - Nord Europe & LIFL & Univ Lille Jan. 12 revised Aug. 14 Rémi Gilleron (Mostrare) Submodular Functions Jan. 12 revised

More information

CS224W: Analysis of Networks Jure Leskovec, Stanford University

CS224W: Analysis of Networks Jure Leskovec, Stanford University CS224W: Analysis of Networks Jure Leskovec, Stanford University http://cs224w.stanford.edu 10/30/17 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, http://cs224w.stanford.edu 2

More information

Influence Maximization in Continuous Time Diffusion Networks

Influence Maximization in Continuous Time Diffusion Networks Manuel Gomez-Rodriguez 1,2 Bernhard Schölkopf 1 1 MPI for Intelligent Systems and 2 Stanford University MANUELGR@STANFORD.EDU BS@TUEBINGEN.MPG.DE Abstract The problem of finding the optimal set of source

More information

Multi-Round Influence Maximization

Multi-Round Influence Maximization Multi-Round Influence Maximization Lichao Sun 1, Weiran Huang 2, Philip S. Yu 1,2, Wei Chen 3, 1 University of Illinois at Chicago, 2 Tsinghua University, 3 Microsoft Research lsun29@uic.edu, huang.inbox@outlook.com,

More information

Diffusion of Innovation and Influence Maximization

Diffusion of Innovation and Influence Maximization Diffusion of Innovation and Influence Maximization Leonid E. Zhukov School of Data Analysis and Artificial Intelligence Department of Computer Science National Research University Higher School of Economics

More information

Greedy Maximization Framework for Graph-based Influence Functions

Greedy Maximization Framework for Graph-based Influence Functions Greedy Maximization Framework for Graph-based Influence Functions Edith Cohen Google Research Tel Aviv University HotWeb '16 1 Large Graphs Model relations/interactions (edges) between entities (nodes)

More information

CS224W: Social and Information Network Analysis Jure Leskovec, Stanford University

CS224W: Social and Information Network Analysis Jure Leskovec, Stanford University CS224W: Social and Information Network Analysis Jure Leskovec, Stanford University http://cs224w.stanford.edu 12/3/2013 Jure Leskovec, Stanford CS224W: Social and Information Network Analysis, http://cs224w.stanford.edu

More information

1 Submodular functions

1 Submodular functions CS 369P: Polyhedral techniques in combinatorial optimization Instructor: Jan Vondrák Lecture date: November 16, 2010 1 Submodular functions We have already encountered submodular functions. Let s recall

More information

Using first-order logic, formalize the following knowledge:

Using first-order logic, formalize the following knowledge: Probabilistic Artificial Intelligence Final Exam Feb 2, 2016 Time limit: 120 minutes Number of pages: 19 Total points: 100 You can use the back of the pages if you run out of space. Collaboration on the

More information

Jure Leskovec Stanford University

Jure Leskovec Stanford University Jure Leskovec Stanford University 2 Part 1: Models for networks Part 2: Information flows in networks Information spread in social media 3 Information flows through networks Analyzing underlying mechanisms

More information

Adaptive Submodularity: A New Approach to Active Learning and Stochastic Optimization

Adaptive Submodularity: A New Approach to Active Learning and Stochastic Optimization Adaptive Submodularity: A New Approach to Active Learning and Stochastic Optimization Daniel Golovin California Institute of Technology Pasadena, CA 91125 dgolovin@caltech.edu Andreas Krause California

More information

Cascading Behavior in Networks: Algorithmic and Economic Issues

Cascading Behavior in Networks: Algorithmic and Economic Issues CHAPTER 24 Cascading Behavior in Networks: Algorithmic and Economic Issues Jon Kleinberg Abstract The flow of information or influence through a large social network can be thought of as unfolding with

More information

Machine Learning and Bayesian Inference. Unsupervised learning. Can we find regularity in data without the aid of labels?

Machine Learning and Bayesian Inference. Unsupervised learning. Can we find regularity in data without the aid of labels? Machine Learning and Bayesian Inference Dr Sean Holden Computer Laboratory, Room FC6 Telephone extension 6372 Email: sbh11@cl.cam.ac.uk www.cl.cam.ac.uk/ sbh11/ Unsupervised learning Can we find regularity

More information

CSE446: Clustering and EM Spring 2017

CSE446: Clustering and EM Spring 2017 CSE446: Clustering and EM Spring 2017 Ali Farhadi Slides adapted from Carlos Guestrin, Dan Klein, and Luke Zettlemoyer Clustering systems: Unsupervised learning Clustering Detect patterns in unlabeled

More information

On-Line Social Systems with Long-Range Goals

On-Line Social Systems with Long-Range Goals On-Line Social Systems with Long-Range Goals Jon Kleinberg Cornell University Including joint work with Ashton Anderson, Dan Huttenlocher, Jure Leskovec, and Sigal Oren. Long-Range Planning Growth in on-line

More information

ECS 253 / MAE 253, Lecture 15 May 17, I. Probability generating function recap

ECS 253 / MAE 253, Lecture 15 May 17, I. Probability generating function recap ECS 253 / MAE 253, Lecture 15 May 17, 2016 I. Probability generating function recap Part I. Ensemble approaches A. Master equations (Random graph evolution, cluster aggregation) B. Network configuration

More information

CS224W: Social and Information Network Analysis Jure Leskovec, Stanford University

CS224W: Social and Information Network Analysis Jure Leskovec, Stanford University CS224W: Social and Information Network Analysis Jure Leskovec, Stanford University http://cs224w.stanford.edu Non-overlapping vs. overlapping communities 11/10/2010 Jure Leskovec, Stanford CS224W: Social

More information

CS 188: Artificial Intelligence. Bayes Nets

CS 188: Artificial Intelligence. Bayes Nets CS 188: Artificial Intelligence Probabilistic Inference: Enumeration, Variable Elimination, Sampling Pieter Abbeel UC Berkeley Many slides over this course adapted from Dan Klein, Stuart Russell, Andrew

More information

CS599: Convex and Combinatorial Optimization Fall 2013 Lecture 24: Introduction to Submodular Functions. Instructor: Shaddin Dughmi

CS599: Convex and Combinatorial Optimization Fall 2013 Lecture 24: Introduction to Submodular Functions. Instructor: Shaddin Dughmi CS599: Convex and Combinatorial Optimization Fall 2013 Lecture 24: Introduction to Submodular Functions Instructor: Shaddin Dughmi Announcements Introduction We saw how matroids form a class of feasible

More information

27 : Distributed Monte Carlo Markov Chain. 1 Recap of MCMC and Naive Parallel Gibbs Sampling

27 : Distributed Monte Carlo Markov Chain. 1 Recap of MCMC and Naive Parallel Gibbs Sampling 10-708: Probabilistic Graphical Models 10-708, Spring 2014 27 : Distributed Monte Carlo Markov Chain Lecturer: Eric P. Xing Scribes: Pengtao Xie, Khoa Luu In this scribe, we are going to review the Parallel

More information

CS224W: Social and Information Network Analysis Jure Leskovec, Stanford University

CS224W: Social and Information Network Analysis Jure Leskovec, Stanford University CS224W: Social and Information Network Analysis Jure Leskovec, Stanford University http://cs224w.stanford.edu Intro sessions to SNAP C++ and SNAP.PY: SNAP.PY: Friday 9/27, 4:5 5:30pm in Gates B03 SNAP

More information

The Ties that Bind Characterizing Classes by Attributes and Social Ties

The Ties that Bind Characterizing Classes by Attributes and Social Ties The Ties that Bind WWW April, 2017, Bryan Perozzi*, Leman Akoglu Stony Brook University *Now at Google. Introduction Outline Our problem: Characterizing Community Differences Proposed Method Experimental

More information

Stochastic Submodular Cover with Limited Adaptivity

Stochastic Submodular Cover with Limited Adaptivity Stochastic Submodular Cover with Limited Adaptivity Arpit Agarwal Sepehr Assadi Sanjeev Khanna Abstract In the submodular cover problem, we are given a non-negative monotone submodular function f over

More information

Determining the Diameter of Small World Networks

Determining the Diameter of Small World Networks Determining the Diameter of Small World Networks Frank W. Takes & Walter A. Kosters Leiden University, The Netherlands CIKM 2011 October 2, 2011 Glasgow, UK NWO COMPASS project (grant #12.0.92) 1 / 30

More information

CS 188: Artificial Intelligence Spring Announcements

CS 188: Artificial Intelligence Spring Announcements CS 188: Artificial Intelligence Spring 2011 Lecture 12: Probability 3/2/2011 Pieter Abbeel UC Berkeley Many slides adapted from Dan Klein. 1 Announcements P3 due on Monday (3/7) at 4:59pm W3 going out

More information

Time-Critical Influence Maximization in Social Networks with Time-Delayed Diffusion Process

Time-Critical Influence Maximization in Social Networks with Time-Delayed Diffusion Process Time-Critical Influence Maximization in Social Networks with Time-Delayed Diffusion Process Wei Chen Microsoft Research Asia Beijing, China weic@microsoft.com Wei Lu University of British Columbia Vancouver,

More information

Copyright 2000, Kevin Wayne 1

Copyright 2000, Kevin Wayne 1 Algorithm runtime analysis and computational tractability Time Complexity of an Algorithm How do we measure the complexity (time, space requirements) of an algorithm. 1 microsecond? Units of time As soon

More information

CS 6901 (Applied Algorithms) Lecture 3

CS 6901 (Applied Algorithms) Lecture 3 CS 6901 (Applied Algorithms) Lecture 3 Antonina Kolokolova September 16, 2014 1 Representative problems: brief overview In this lecture we will look at several problems which, although look somewhat similar

More information

Announcements. Inference. Mid-term. Inference by Enumeration. Reminder: Alarm Network. Introduction to Artificial Intelligence. V22.

Announcements. Inference. Mid-term. Inference by Enumeration. Reminder: Alarm Network. Introduction to Artificial Intelligence. V22. Introduction to Artificial Intelligence V22.0472-001 Fall 2009 Lecture 15: Bayes Nets 3 Midterms graded Assignment 2 graded Announcements Rob Fergus Dept of Computer Science, Courant Institute, NYU Slides

More information

Fig. 1. Securely distinguish three vehicle flows /17/$31.00 c 2017 IEEE

Fig. 1. Securely distinguish three vehicle flows /17/$31.00 c 2017 IEEE On the RSU-based Secure Distinguishability Among Vehicular Flows Wei Chang, Huanyang Zheng, and Jie Wu Department of Computer Science, Saint Joseph s University, Philadelphia, PA 19131 Department of Computer

More information

Approximate Inference

Approximate Inference Approximate Inference Simulation has a name: sampling Sampling is a hot topic in machine learning, and it s really simple Basic idea: Draw N samples from a sampling distribution S Compute an approximate

More information

CS 781 Lecture 9 March 10, 2011 Topics: Local Search and Optimization Metropolis Algorithm Greedy Optimization Hopfield Networks Max Cut Problem Nash

CS 781 Lecture 9 March 10, 2011 Topics: Local Search and Optimization Metropolis Algorithm Greedy Optimization Hopfield Networks Max Cut Problem Nash CS 781 Lecture 9 March 10, 2011 Topics: Local Search and Optimization Metropolis Algorithm Greedy Optimization Hopfield Networks Max Cut Problem Nash Equilibrium Price of Stability Coping With NP-Hardness

More information

A Fast Approximation for Influence Maximization in Large Social Networks

A Fast Approximation for Influence Maximization in Large Social Networks A Fast Approximation for Influence Maximization in Large Social Networks ABSTRACT Jong-Ryul Lee Dept. of Computer Science, KAIST 29 Daehak-ro, Yuseong-gu, Daejeon, Korea jrlee@islab.kaist.ac.kr This paper

More information

ECS 253 / MAE 253, Lecture 13 May 15, Diffusion, Cascades and Influence Mathematical models & generating functions

ECS 253 / MAE 253, Lecture 13 May 15, Diffusion, Cascades and Influence Mathematical models & generating functions ECS 253 / MAE 253, Lecture 13 May 15, 2018 Diffusion, Cascades and Influence Mathematical models & generating functions Last week: spatial flows and game theory on networks Optimal location of facilities

More information

2.6 Complexity Theory for Map-Reduce. Star Joins 2.6. COMPLEXITY THEORY FOR MAP-REDUCE 51

2.6 Complexity Theory for Map-Reduce. Star Joins 2.6. COMPLEXITY THEORY FOR MAP-REDUCE 51 2.6. COMPLEXITY THEORY FOR MAP-REDUCE 51 Star Joins A common structure for data mining of commercial data is the star join. For example, a chain store like Walmart keeps a fact table whose tuples each

More information

Adaptive Submodularity: Theory and Applications in Active Learning and Stochastic Optimization

Adaptive Submodularity: Theory and Applications in Active Learning and Stochastic Optimization Journal of Artificial Intelligence Research 42 (2011) 427-486 Submitted 1/11; published 11/11 Adaptive Submodularity: Theory and Applications in Active Learning and Stochastic Optimization Daniel Golovin

More information

CS 188: Artificial Intelligence

CS 188: Artificial Intelligence CS 188: Artificial Intelligence Adversarial Search II Instructor: Anca Dragan University of California, Berkeley [These slides adapted from Dan Klein and Pieter Abbeel] Minimax Example 3 12 8 2 4 6 14

More information

CS340 Machine learning Lecture 5 Learning theory cont'd. Some slides are borrowed from Stuart Russell and Thorsten Joachims

CS340 Machine learning Lecture 5 Learning theory cont'd. Some slides are borrowed from Stuart Russell and Thorsten Joachims CS340 Machine learning Lecture 5 Learning theory cont'd Some slides are borrowed from Stuart Russell and Thorsten Joachims Inductive learning Simplest form: learn a function from examples f is the target

More information

Faster Kriging on Graphs

Faster Kriging on Graphs Faster Kriging on Graphs Omkar Muralidharan Abstract [Xu et al. 2009] introduce a graph prediction method that is accurate but slow. My project investigates faster methods based on theirs that are nearly

More information

Maximizing the Spread of Influence through a Social Network

Maximizing the Spread of Influence through a Social Network Maximizing the Spread of Influence through a Social Network David Kempe Dept. of Computer Science Cornell University, Ithaca NY kempe@cs.cornell.edu Jon Kleinberg Dept. of Computer Science Cornell University,

More information

CS246: Mining Massive Datasets Jure Leskovec, Stanford University

CS246: Mining Massive Datasets Jure Leskovec, Stanford University CS246: Mining Massive Datasets Jure Leskovec, Stanford University http://cs246.stanford.edu 2/26/2013 Jure Leskovec, Stanford CS246: Mining Massive Datasets, http://cs246.stanford.edu 2 More algorithms

More information

Overlapping Communities

Overlapping Communities Overlapping Communities Davide Mottin HassoPlattner Institute Graph Mining course Winter Semester 2017 Acknowledgements Most of this lecture is taken from: http://web.stanford.edu/class/cs224w/slides GRAPH

More information

Bayes Nets III: Inference

Bayes Nets III: Inference 1 Hal Daumé III (me@hal3.name) Bayes Nets III: Inference Hal Daumé III Computer Science University of Maryland me@hal3.name CS 421: Introduction to Artificial Intelligence 10 Apr 2012 Many slides courtesy

More information

Aditya Bhaskara CS 5968/6968, Lecture 1: Introduction and Review 12 January 2016

Aditya Bhaskara CS 5968/6968, Lecture 1: Introduction and Review 12 January 2016 Lecture 1: Introduction and Review We begin with a short introduction to the course, and logistics. We then survey some basics about approximation algorithms and probability. We also introduce some of

More information

CSE 421 Greedy Algorithms / Interval Scheduling

CSE 421 Greedy Algorithms / Interval Scheduling CSE 421 Greedy Algorithms / Interval Scheduling Yin Tat Lee 1 Interval Scheduling Job j starts at s(j) and finishes at f(j). Two jobs compatible if they don t overlap. Goal: find maximum subset of mutually

More information

Optimization of Submodular Functions Tutorial - lecture I

Optimization of Submodular Functions Tutorial - lecture I Optimization of Submodular Functions Tutorial - lecture I Jan Vondrák 1 1 IBM Almaden Research Center San Jose, CA Jan Vondrák (IBM Almaden) Submodular Optimization Tutorial 1 / 1 Lecture I: outline 1

More information

An Analytic Solution to Discrete Bayesian Reinforcement Learning

An Analytic Solution to Discrete Bayesian Reinforcement Learning An Analytic Solution to Discrete Bayesian Reinforcement Learning Pascal Poupart (U of Waterloo) Nikos Vlassis (U of Amsterdam) Jesse Hoey (U of Toronto) Kevin Regan (U of Waterloo) 1 Motivation Automated

More information

Models of Communication Dynamics for Simulation of Information Diffusion

Models of Communication Dynamics for Simulation of Information Diffusion Models of Communication Dynamics for Simulation of Information Diffusion Konstantin Mertsalov, Malik Magdon-Ismail, Mark Goldberg Rensselaer Polytechnic Institute Department of Computer Science 11 8th

More information

Edge-Weighted Personalized PageRank: Breaking a Decade-Old Performance Barrier

Edge-Weighted Personalized PageRank: Breaking a Decade-Old Performance Barrier Edge-Weighted Personalized PageRank: Breaking a Decade-Old Performance Barrier W. Xie D. Bindel A. Demers J. Gehrke 12 Aug 2015 W. Xie, D. Bindel, A. Demers, J. Gehrke KDD2015 12 Aug 2015 1 / 1 PageRank

More information

Lecture 11 October 11, Information Dissemination through Social Networks

Lecture 11 October 11, Information Dissemination through Social Networks CS 284r: Incentives and Information in Networks Fall 2013 Prof. Yaron Singer Lecture 11 October 11, 2013 Scribe: Michael Tingley, K. Nathaniel Tucker 1 Overview In today s lecture we will start the second

More information

Probability Models of Information Exchange on Networks Lecture 6

Probability Models of Information Exchange on Networks Lecture 6 Probability Models of Information Exchange on Networks Lecture 6 UC Berkeley Many Other Models There are many models of information exchange on networks. Q: Which model to chose? My answer good features

More information

Maximization of Submodular Set Functions

Maximization of Submodular Set Functions Northeastern University Department of Electrical and Computer Engineering Maximization of Submodular Set Functions Biomedical Signal Processing, Imaging, Reasoning, and Learning BSPIRAL) Group Author:

More information

Motivation, Basic Concepts, Basic Methods, Travelling Salesperson Problem (TSP), Algorithms

Motivation, Basic Concepts, Basic Methods, Travelling Salesperson Problem (TSP), Algorithms Motivation, Basic Concepts, Basic Methods, Travelling Salesperson Problem (TSP), Algorithms 1 What is Combinatorial Optimization? Combinatorial Optimization deals with problems where we have to search

More information

Presentation for CSG399 By Jian Wen

Presentation for CSG399 By Jian Wen Presentation for CSG399 By Jian Wen Outline Skyline Definition and properties Related problems Progressive Algorithms Analysis SFS Analysis Algorithm Cost Estimate Epsilon Skyline(overview and idea) Skyline

More information

Distributed Systems Gossip Algorithms

Distributed Systems Gossip Algorithms Distributed Systems Gossip Algorithms He Sun School of Informatics University of Edinburgh What is Gossip? Gossip algorithms In a gossip algorithm, each node in the network periodically exchanges information

More information

Inferring the origin of an epidemic with a dynamic message-passing algorithm

Inferring the origin of an epidemic with a dynamic message-passing algorithm Inferring the origin of an epidemic with a dynamic message-passing algorithm HARSH GUPTA (Based on the original work done by Andrey Y. Lokhov, Marc Mézard, Hiroki Ohta, and Lenka Zdeborová) Paper Andrey

More information

Distributed Optimization. Song Chong EE, KAIST

Distributed Optimization. Song Chong EE, KAIST Distributed Optimization Song Chong EE, KAIST songchong@kaist.edu Dynamic Programming for Path Planning A path-planning problem consists of a weighted directed graph with a set of n nodes N, directed links

More information

Review: Probability. BM1: Advanced Natural Language Processing. University of Potsdam. Tatjana Scheffler

Review: Probability. BM1: Advanced Natural Language Processing. University of Potsdam. Tatjana Scheffler Review: Probability BM1: Advanced Natural Language Processing University of Potsdam Tatjana Scheffler tatjana.scheffler@uni-potsdam.de October 21, 2016 Today probability random variables Bayes rule expectation

More information

Consistency Under Sampling of Exponential Random Graph Models

Consistency Under Sampling of Exponential Random Graph Models Consistency Under Sampling of Exponential Random Graph Models Cosma Shalizi and Alessandro Rinaldo Summary by: Elly Kaizar Remember ERGMs (Exponential Random Graph Models) Exponential family models Sufficient

More information

ECS 289 F / MAE 298, Lecture 15 May 20, Diffusion, Cascades and Influence

ECS 289 F / MAE 298, Lecture 15 May 20, Diffusion, Cascades and Influence ECS 289 F / MAE 298, Lecture 15 May 20, 2014 Diffusion, Cascades and Influence Diffusion and cascades in networks (Nodes in one of two states) Viruses (human and computer) contact processes epidemic thresholds

More information

Stochastic Submodular Cover with Limited Adaptivity

Stochastic Submodular Cover with Limited Adaptivity Stochastic Submodular Cover with Limited Adaptivity Arpit Agarwal Sepehr Assadi Sanjeev Khanna University of Pennsylvania {aarpit,sassadi,sanjeev}@cis.upenn.edu arxiv:1810.13351v1 [cs.ds] 31 Oct 2018 Abstract

More information

Applications of Submodular Functions in Speech and NLP

Applications of Submodular Functions in Speech and NLP Applications of Submodular Functions in Speech and NLP Jeff Bilmes Department of Electrical Engineering University of Washington, Seattle http://ssli.ee.washington.edu/~bilmes June 27, 2011 J. Bilmes Applications

More information

MobiHoc 2014 MINIMUM-SIZED INFLUENTIAL NODE SET SELECTION FOR SOCIAL NETWORKS UNDER THE INDEPENDENT CASCADE MODEL

MobiHoc 2014 MINIMUM-SIZED INFLUENTIAL NODE SET SELECTION FOR SOCIAL NETWORKS UNDER THE INDEPENDENT CASCADE MODEL MobiHoc 2014 MINIMUM-SIZED INFLUENTIAL NODE SET SELECTION FOR SOCIAL NETWORKS UNDER THE INDEPENDENT CASCADE MODEL Jing (Selena) He Department of Computer Science, Kennesaw State University Shouling Ji,

More information

ACM 116: Lecture 1. Agenda. Philosophy of the Course. Definition of probabilities. Equally likely outcomes. Elements of combinatorics

ACM 116: Lecture 1. Agenda. Philosophy of the Course. Definition of probabilities. Equally likely outcomes. Elements of combinatorics 1 ACM 116: Lecture 1 Agenda Philosophy of the Course Definition of probabilities Equally likely outcomes Elements of combinatorics Conditional probabilities 2 Philosophy of the Course Probability is the

More information

Approximability of Adaptive Seeding under Knapsack Constraints

Approximability of Adaptive Seeding under Knapsack Constraints Approximability of Adaptive Seeding under Knapsack Constraints Aviad Rubinstein UC Berkeley aviad@cs.berkeley.edu Lior Seeman Cornell Univeristy lseeman@cs.cornell.edu May 6, 2015 Yaron Singer Harvard

More information

CS 188 Introduction to Fall 2007 Artificial Intelligence Midterm

CS 188 Introduction to Fall 2007 Artificial Intelligence Midterm NAME: SID#: Login: Sec: 1 CS 188 Introduction to Fall 2007 Artificial Intelligence Midterm You have 80 minutes. The exam is closed book, closed notes except a one-page crib sheet, basic calculators only.

More information

CS224W: Social and Information Network Analysis Jure Leskovec, Stanford University

CS224W: Social and Information Network Analysis Jure Leskovec, Stanford University CS224W: Social and Information Network Analysis Jure Leskovec, Stanford University http://cs224w.stanford.edu How to organize/navigate it? First try: Human curated Web directories Yahoo, DMOZ, LookSmart

More information

1 Maximizing a Submodular Function

1 Maximizing a Submodular Function 6.883 Learning with Combinatorial Structure Notes for Lecture 16 Author: Arpit Agarwal 1 Maximizing a Submodular Function In the last lecture we looked at maximization of a monotone submodular function,

More information

N/4 + N/2 + N = 2N 2.

N/4 + N/2 + N = 2N 2. CS61B Summer 2006 Instructor: Erin Korber Lecture 24, 7 Aug. 1 Amortized Analysis For some of the data structures we ve discussed (namely hash tables and splay trees), it was claimed that the average time

More information

NETS 412: Algorithmic Game Theory March 28 and 30, Lecture Approximation in Mechanism Design. X(v) = arg max v i (a)

NETS 412: Algorithmic Game Theory March 28 and 30, Lecture Approximation in Mechanism Design. X(v) = arg max v i (a) NETS 412: Algorithmic Game Theory March 28 and 30, 2017 Lecture 16+17 Lecturer: Aaron Roth Scribe: Aaron Roth Approximation in Mechanism Design In the last lecture, we asked how far we can go beyond the

More information

12. LOCAL SEARCH. gradient descent Metropolis algorithm Hopfield neural networks maximum cut Nash equilibria

12. LOCAL SEARCH. gradient descent Metropolis algorithm Hopfield neural networks maximum cut Nash equilibria 12. LOCAL SEARCH gradient descent Metropolis algorithm Hopfield neural networks maximum cut Nash equilibria Lecture slides by Kevin Wayne Copyright 2005 Pearson-Addison Wesley h ttp://www.cs.princeton.edu/~wayne/kleinberg-tardos

More information

Dimensional Analysis and Exponential Models Review Solutions

Dimensional Analysis and Exponential Models Review Solutions Dimensional Analysis and Exponential Models Review Solutions 1. Given 13 morks is 1 gloe, 5 gloes is 1 flit, 7 kits is one lonk, and 10 lonks is 1 gall. Convert 90 morks per kit to flits per gall. To solve

More information

Discrete Optimization in Machine Learning. Colorado Reed

Discrete Optimization in Machine Learning. Colorado Reed Discrete Optimization in Machine Learning Colorado Reed [ML-RCC] 31 Jan 2013 1 Acknowledgements Some slides/animations based on: Krause et al. tutorials: http://www.submodularity.org Pushmeet Kohli tutorial:

More information

CS341 info session is on Thu 3/1 5pm in Gates415. CS246: Mining Massive Datasets Jure Leskovec, Stanford University

CS341 info session is on Thu 3/1 5pm in Gates415. CS246: Mining Massive Datasets Jure Leskovec, Stanford University CS341 info session is on Thu 3/1 5pm in Gates415 CS246: Mining Massive Datasets Jure Leskovec, Stanford University http://cs246.stanford.edu 2/28/18 Jure Leskovec, Stanford CS246: Mining Massive Datasets,

More information

Administration. CSCI567 Machine Learning (Fall 2018) Outline. Outline. HW5 is available, due on 11/18. Practice final will also be available soon.

Administration. CSCI567 Machine Learning (Fall 2018) Outline. Outline. HW5 is available, due on 11/18. Practice final will also be available soon. Administration CSCI567 Machine Learning Fall 2018 Prof. Haipeng Luo U of Southern California Nov 7, 2018 HW5 is available, due on 11/18. Practice final will also be available soon. Remaining weeks: 11/14,

More information

Bubble Clustering: Set Covering via Union of Ellipses

Bubble Clustering: Set Covering via Union of Ellipses Bubble Clustering: Set Covering via Union of Ellipses Matt Kraning, Arezou Keshavarz, Lei Zhao CS229, Stanford University Email: {mkraning,arezou,leiz}@stanford.edu Abstract We develop an algorithm called

More information

Principles of AI Planning

Principles of AI Planning Principles of 5. Planning as search: progression and regression Malte Helmert and Bernhard Nebel Albert-Ludwigs-Universität Freiburg May 4th, 2010 Planning as (classical) search Introduction Classification

More information

Tractable Models for Information Diffusion in Social Networks

Tractable Models for Information Diffusion in Social Networks Tractable Models for Information Diffusion in Social Networks Masahiro Kimura 1 and Kazumi Saito 2 1 Department of Electronics and Informatics, Ryukoku University Otsu, Shiga 520-2194, Japan 2 NTT Communication

More information

Announcements. CS 188: Artificial Intelligence Fall Causality? Example: Traffic. Topology Limits Distributions. Example: Reverse Traffic

Announcements. CS 188: Artificial Intelligence Fall Causality? Example: Traffic. Topology Limits Distributions. Example: Reverse Traffic CS 188: Artificial Intelligence Fall 2008 Lecture 16: Bayes Nets III 10/23/2008 Announcements Midterms graded, up on glookup, back Tuesday W4 also graded, back in sections / box Past homeworks in return

More information

Learning in State-Space Reinforcement Learning CIS 32

Learning in State-Space Reinforcement Learning CIS 32 Learning in State-Space Reinforcement Learning CIS 32 Functionalia Syllabus Updated: MIDTERM and REVIEW moved up one day. MIDTERM: Everything through Evolutionary Agents. HW 2 Out - DUE Sunday before the

More information

On Influential Node Discovery in Dynamic Social Networks

On Influential Node Discovery in Dynamic Social Networks On Influential Node Discovery in Dynamic Social Networks Charu Aggarwal Shuyang Lin Philip S. Yu Abstract The problem of maximizing influence spread has been widely studied in social networks, because

More information

the tree till a class assignment is reached

the tree till a class assignment is reached Decision Trees Decision Tree for Playing Tennis Prediction is done by sending the example down Prediction is done by sending the example down the tree till a class assignment is reached Definitions Internal

More information

Machine Learning

Machine Learning Machine Learning 10-601 Tom M. Mitchell Machine Learning Department Carnegie Mellon University August 30, 2017 Today: Decision trees Overfitting The Big Picture Coming soon Probabilistic learning MLE,

More information

The Traveling Salesman Problem: An Overview. David P. Williamson, Cornell University Ebay Research January 21, 2014

The Traveling Salesman Problem: An Overview. David P. Williamson, Cornell University Ebay Research January 21, 2014 The Traveling Salesman Problem: An Overview David P. Williamson, Cornell University Ebay Research January 21, 2014 (Cook 2012) A highly readable introduction Some terminology (imprecise) Problem Traditional

More information

Decision Trees. CSC411/2515: Machine Learning and Data Mining, Winter 2018 Luke Zettlemoyer, Carlos Guestrin, and Andrew Moore

Decision Trees. CSC411/2515: Machine Learning and Data Mining, Winter 2018 Luke Zettlemoyer, Carlos Guestrin, and Andrew Moore Decision Trees Claude Monet, The Mulberry Tree Slides from Pedro Domingos, CSC411/2515: Machine Learning and Data Mining, Winter 2018 Luke Zettlemoyer, Carlos Guestrin, and Andrew Moore Michael Guerzhoy

More information

Locally Adaptive Optimization: Adaptive Seeding for Monotone Submodular Functions

Locally Adaptive Optimization: Adaptive Seeding for Monotone Submodular Functions Locally Adaptive Optimization: Adaptive Seeding for Monotone Submodular Functions Ashwinumar Badanidiyuru Google ashwinumarbv@gmail.com Aviad Rubinstein UC Bereley aviad@cs.bereley.edu Lior Seeman Cornell

More information

Online Learning with Feedback Graphs

Online Learning with Feedback Graphs Online Learning with Feedback Graphs Claudio Gentile INRIA and Google NY clagentile@gmailcom NYC March 6th, 2018 1 Content of this lecture Regret analysis of sequential prediction problems lying between

More information

Effective placement of sensors for efficient early warning system in water distribution network MASHREKI ISLAM SAMI

Effective placement of sensors for efficient early warning system in water distribution network MASHREKI ISLAM SAMI Effective placement of sensors for efficient early warning system in water distribution network MASHREKI ISLAM SAMI AGENDA Introduction and Problem Statement Methodology Experimentation and Analysis Limitations

More information

Performance Evaluation. Analyzing competitive influence maximization problems with partial information: An approximation algorithmic framework

Performance Evaluation. Analyzing competitive influence maximization problems with partial information: An approximation algorithmic framework Performance Evaluation ( ) Contents lists available at ScienceDirect Performance Evaluation journal homepage: www.elsevier.com/locate/peva Analyzing competitive influence maximization problems with partial

More information

Cost and Preference in Recommender Systems Junhua Chen LESS IS MORE

Cost and Preference in Recommender Systems Junhua Chen LESS IS MORE Cost and Preference in Recommender Systems Junhua Chen, Big Data Research Center, UESTC Email:junmshao@uestc.edu.cn http://staff.uestc.edu.cn/shaojunming Abstract In many recommender systems (RS), user

More information